skip to main content
10.1145/3369583.3392692acmconferencesArticle/Chapter ViewAbstractPublication PageshpdcConference Proceedingsconference-collections
research-article

Space-Efficient k-d Tree-Based Storage Format for Sparse Tensors

Authors Info & Claims
Published:23 June 2020Publication History

ABSTRACT

Computations with tensors are widespread in many scientific areas. Usually, the used tensors are very large but sparse, i.e., the vast majority of their elements are zero. The space complexity of sparse tensor storage formats varies significantly. For overall efficiency, it is important to reduce the execution time and additional space requirements of the initial preprocessing (i.e., converting tensors from common storage formats to the given internal format).

The main contributions of this paper are threefold. Firstly, we present a new storage format for sparse tensors, which we call the succinct k-d tree-based tensor (SKTB) format. We compare the space complexity of common tensor storage formats and of the SKTB format and demonstrate the viability of using a tree as a data structurefor sparse tensors. Secondly, we present a parallel space-efficient algorithm for converting tensors to the SKTB format. Thirdly, we demonstrate the computational efficiency of the proposed format in sparse tensor-vector multiplication.

Skip Supplemental Material Section

Supplemental Material

3369583.3392692.mp4

mp4

14.8 MB

References

  1. Mart'in Abadi, Ashish Agarwal, Paul Barham, Eugene Brevdo, Zhifeng Chen, Craig Citro, Greg S. Corrado, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Ian Goodfellow, Andrew Harp, Geoffrey Irving, Michael Isard, Yangqing Jia, Rafal Jozefowicz, Lukasz Kaiser, Manjunath Kudlur, Josh Levenberg, Dandelion Mané, Rajat Monga, Sherry Moore, Derek Murray, Chris Olah, Mike Schuster, Jonathon Shlens, Benoit Steiner, Ilya Sutskever, Kunal Talwar, Paul Tucker, Vincent Vanhoucke, Vijay Vasudevan, Fernanda Viégas, Oriol Vinyals, Pete Warden, Martin Wattenberg, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng. 2015. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. https://www.tensorflow.org/ Software available from tensorflow.org.Google ScholarGoogle Scholar
  2. Evrim Acar, Seyit Ahmet Çamtepe, and Bülent Yener. 2006. Collective Sampling and Analysis of High Order Tensors for Chatroom Communications. In ISI (Lecture Notes in Computer Science), Sharad Mehrotra, Daniel Dajun Zeng, Hsinchun Chen, Bhavani M. Thuraisingham, and Fei-Yue Wang (Eds.), Vol. 3975. Springer, 213--224. http://dblp.uni-trier.de/db/conf/isi/isi2006.html#AcarCY06Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. D. Michal Aristotle. 2008. Matrix and Tensor Calculus: With Applications to Mechanics, Elasticity and Aeronautics .Dover Publications.Google ScholarGoogle Scholar
  4. Brett W. Bader and Tamara G. Kolda. 2007. Efficient MATLAB Computations with Sparse and Factored Tensors. SIAM Journal on Scientific Computing, Vol. 30, 1 (December 2007), 205--231. https://doi.org/10.1137/060676489Google ScholarGoogle Scholar
  5. R. Barrett, M. Berry, T. F. Chan, J. Demmel, J. Donato, J. Dongarra, V. Eijkhout, R. Pozo, C. Romine, and H. Van der Vorst. 1994. Templates for the Solution of Linear Systems: Building Blocks for Iterative Methods 2nd ed.). SIAM, Philadelphia, PA.Google ScholarGoogle Scholar
  6. JL Bentley. 1975. Multidimensional binary search trees used for associative searching. Commun. ACM, Vol. 18, 9 (1975), 517.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Iain Duff and John Reid. 1979. Some Design Features of a Sparse Matrix Code. ACM Trans. Math. Software, Vol. 5 (03 1979). https://doi.org/10.1145/355815.355817Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Fred G. Gustavson. 1972. Some Basic Techniques for Solving Sparse Systems of Linear Equations .Springer US, Boston, MA, 41--52. https://doi.org/10.1007/978--1--4615--8675--3_4Google ScholarGoogle Scholar
  9. G. Jacobson. 1989. Space-Efficient Static Trees and Graphs. In Proceedings of the 30th Annual Symposium on Foundations of Computer Science (SFCS '89). IEEE Computer Society, USA, 549--554. https://doi.org/10.1109/SFCS.1989.63533Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. T. Kolda and B. Bader. 2009. Tensor Decompositions and Applications. SIAM Rev., Vol. 51, 3 (2009), 455--500. https://doi.org/10.1137/07070111XGoogle ScholarGoogle ScholarDigital LibraryDigital Library
  11. Jiajia Li, Yuchen Ma, Chenggang Yan, and Richard Vuduc. 2016. Optimizing Sparse Tensor Times Matrix on Multi-core and Many-core Architectures. In Proceedings of the Sixth Workshop on Irregular Applications: Architectures and Algorithms (IA$^3$ '16). IEEE Press, Piscataway, NJ, USA, 26--33.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Jiajia Li, Jimeng Sun, and Richard Vuduc. 2018. HiCOO: Hierarchical Storage of Sparse Tensors. In Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis (SC '18). IEEE Press, Piscataway, NJ, USA, Article 19, 15 pages. http://dl.acm.org/citation.cfm?id=3291656.3291682Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Chun-Yuan Lin, Yeh-Ching Chung, and Jen-Shiuh Liu. 2003. Efficient Data Compression Methods for Multidimensional Sparse Array Operations Based on the EKMR Scheme. IEEE Trans. Comput., Vol. 52, 12 (Dec. 2003), 1640--1646. https://doi.org/10.1109/TC.2003.1252859Google ScholarGoogle Scholar
  14. Bangtian Liu, Chengyao Wen, Anand D. Sarwate, and Maryam Mehri Dehnavi. 2017. A Unified Optimization Approach for Sparse Tensor Operations on GPUs. 2017 IEEE International Conference on Cluster Computing (CLUSTER) (May 2017), 47--57.Google ScholarGoogle Scholar
  15. G. M. Morton. 1966. A computer Oriented Geodetic Data Base; and a New Technique in File Sequencing. IBM Ltd.Google ScholarGoogle Scholar
  16. Shaden Smith, Jee W. Choi, Jiajia Li, Richard Vuduc, Jongsoo Park, Xing Liu, and George Karypis. 2017. FROSTT: The Formidable Repository of Open Sparse Tensors and Tools. http://frostt.io/Google ScholarGoogle Scholar
  17. Shaden Smith, Niranjay Ravindran, Nicholas D. Sidiropoulos, and George Karypis. 2015. SPLATT: Efficient and Parallel Sparse Tensor-Matrix Multiplication. In Proceedings of the 2015 IEEE International Parallel and Distributed Processing Symposium (IPDPS '15). IEEE Computer Society, Washington, DC, USA, 61--70. https://doi.org/10.1109/IPDPS.2015.27Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Jimeng Sun, Dacheng Tao, and Christos Faloutsos. 2006. Beyond Streams and Graphs: Dynamic Tensor Analysis. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '06). Association for Computing Machinery, New York, NY, USA, 374--383. https://doi.org/10.1145/1150402.1150445Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Parker Allen Tew. 2016. An Investigation of Sparse Tensor Formats for Tensor Libraries. http://groups.csail.mit.edu/commit/papers/2016/parker-thesis.pdfGoogle ScholarGoogle Scholar
  20. Ivan Simevcek. 2009. Sparse Matrix Computations Using the Quadtree Storage Format. In Proceedings of the 2009 11th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC '09). IEEE Computer Society, Washington, DC, USA, 168--173. http://dx.doi.org/10.1109/SYNASC.2009.55Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. I. Simevcek, D. Langr, and J. Trdlivcka. 2014. Efficient Converting of Large Sparse Matrices to Quadtree Format. In Symbolic and Numeric Algorithms for Scientific Computing (SYNASC), 2014 16th International Symposium on. 122--129. https://doi.org/10.1109/SYNASC.2014.25Google ScholarGoogle ScholarCross RefCross Ref
  22. I. Simevcek, D. Langr, and P. Tvrdik. 2012. Minimal Quadtree Format for Compression of Sparse Matrices Storage. In 14th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC'2012) (SYNASC'2012). Timisoara, Romania, 359--364. https://doi.org/10.1109/SYNASC.2012.30Google ScholarGoogle Scholar

Index Terms

  1. Space-Efficient k-d Tree-Based Storage Format for Sparse Tensors

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            HPDC '20: Proceedings of the 29th International Symposium on High-Performance Parallel and Distributed Computing
            June 2020
            246 pages
            ISBN:9781450370523
            DOI:10.1145/3369583

            Copyright © 2020 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 23 June 2020

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            Overall Acceptance Rate166of966submissions,17%

            Upcoming Conference

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader