Sciweavers

KDD
2012
ACM

GigaTensor: scaling tensor analysis up by 100 times - algorithms and discoveries

12 years 2 months ago
GigaTensor: scaling tensor analysis up by 100 times - algorithms and discoveries
Many data are modeled as tensors, or multi dimensional arrays. Examples include the predicates (subject, verb, object) in knowledge bases, hyperlinks and anchor texts in the Web graphs, sensor streams (time, location, and type), social networks over time, and DBLP conference-author-keyword relations. Tensor decomposition is an important data mining tool with various applications including clustering, trend detection, and anomaly detection. However, current tensor decomposition algorithms are not scalable for large tensors with billions of sizes and hundreds millions of nonzeros: the largest tensor in the literature remains thousands of sizes and hundreds thousands of nonzeros. Consider a knowledge base tensor consisting of about 26 million noun-phrases. The intermediate data explosion problem, associated with naive implementations of tensor decomposition algorithms, would require the materialization and the storage of a matrix whose largest dimension would be ≈ 7 · 1014 ; this amou...
U. Kang, Evangelos E. Papalexakis, Abhay Harpale,
Added 28 Sep 2012
Updated 28 Sep 2012
Type Journal
Year 2012
Where KDD
Authors U. Kang, Evangelos E. Papalexakis, Abhay Harpale, Christos Faloutsos
Comments (0)