skip to main content
10.1145/3588195.3592985acmconferencesArticle/Chapter ViewAbstractPublication PageshpdcConference Proceedingsconference-collections
research-article

Auto-HPCnet: An Automatic Framework to Build Neural Network-based Surrogate for High-Performance Computing Applications

Published:07 August 2023Publication History

ABSTRACT

High-performance computing communities are increasingly adopting Neural Networks (NN) as surrogate models in their applications to generate scientific insights. Replacing an execution phase in the application with NN models can bring significant performance improvement. However, there is a lack of tools that can help domain scientists automatically apply NN-based surrogate models to HPC applications. We introduce a framework, named AutoHPC-net, to democratize the usage of NN-based surrogates. AutoHPC-net is the first end-to-end framework that makes past proposals for the NN-based surrogate model practical and disciplined. AutoHPC-net introduces a workflow to address unique challenges when applying the approximation, such as feature acquisition and meeting the application-specific constraint on the quality of final computation outcome. We show that AutoHPC-net can leverage NN for a set of HPC applications and achieve 5.50× speedup on average (up to 16.8× speedup and with data preparation cost included) while meeting the application-specific constraint on the final computation quality.

References

  1. Markus Reichstein, Gustau Camps-Valls, Bjorn Stevens, Martin Jung, Joachim Denzler, Nuno Carvalhais, et al. Deep learning and process understanding for data-driven earth system science. Nature, 566(7743):195--204, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  2. VladimirMKrasnopolsky and Michael S Fox-Rabinovitz. Complex hybrid models combining deterministic and machine learning components for numerical climate modeling and weather prediction. Neural Networks, 19(2):122--134, 2006.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Paul A O'Gorman and John G Dwyer. Using machine learning to parameterize moist convection: Potential for modeling of climate, climate change, and extreme events. Journal of Advances in Modeling Earth Systems, 10(10):2548--2563, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  4. Arvind T Mohan and Datta V Gaitonde. A deep learning based approach to reduced order modeling for turbulent flow control using lstm neural networks. arXiv preprint arXiv:1804.09269, 2018.Google ScholarGoogle Scholar
  5. Dunhui Xiao, CE Heaney, L Mottet, F Fang, W Lin, IM Navon, Y Guo, OK Matar, AG Robins, and CC Pain. A reduced order model for turbulent flows in the urban environment using machine learning. Building and Environment, 148:323--337, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  6. Mathis Bode, Michael Gauding, Zeyu Lian, Dominik Denker, Marco Davidovic, Konstantin Kleinheinz, Jenia Jitsev, and Heinz Pitsch. Using physics-informed enhanced super-resolution generative adversarial networks for subfilter modeling in turbulent reactive flows. Proceedings of the Combustion Institute, 38(2):2617-- 2625, 2021.Google ScholarGoogle ScholarCross RefCross Ref
  7. Akinori Tanaka, Akio Tomiya, and K¯oji Hashimoto. Deep Learning and Physics. Springer, 2021.Google ScholarGoogle ScholarCross RefCross Ref
  8. Rahul Rai and Chandan K Sahu. Driven by data or derived through physics? a review of hybrid physics guided machine learning techniques with cyber-physical system (cps) focus. IEEE Access, 8:71050--71073, 2020.Google ScholarGoogle ScholarCross RefCross Ref
  9. Paul Raccuglia, Katherine C Elbert, Philip DF Adler, Casey Falk, Malia B Wenny, AurelioMollo, Matthias Zeller, Sorelle A Friedler, Joshua Schrier, and Alexander J Norquist. Machine-learning-assisted materials discovery using failed experiments. Nature, 533(7601):73--76, 2016.Google ScholarGoogle ScholarCross RefCross Ref
  10. Gabriel R Schleder, Antonio CM Padilha, Carlos Mera Acosta, Marcio Costa, and Adalberto Fazzio. From dft to machine learning: recent approaches to materials science--a review. Journal of Physics: Materials, 2(3):032001, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  11. Ruijin Cang, Hechao Li, Hope Yao, Yang Jiao, and Yi Ren. Improving direct physical properties prediction of heterogeneous materials from imaging data via convolutional neural network and a morphology-aware generative model. Computational Materials Science, 150:212--221, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  12. Peter Sadowski, David Fooshee, Niranjan Subrahmanya, and Pierre Baldi. Synergies between quantum mechanics and machine learning in reaction prediction. Journal of chemical information and modeling, 56(11):2125--2128, 2016.Google ScholarGoogle Scholar
  13. Kristof Schütt, Pieter-Jan Kindermans, Huziel Enoc Sauceda Felix, Stefan Chmiela, Alexandre Tkatchenko, and Klaus-Robert Müller. Schnet: A continuousfilter convolutional neural network for modeling quantum interactions. Advances in neural information processing systems, 30, 2017.Google ScholarGoogle Scholar
  14. Suraj Pawar, Omer San, Burak Aksoylu, Adil Rasheed, and Trond Kvamsdal. Physics guided machine learning using simplified theories. Physics of Fluids, 33(1):011701, 2021.Google ScholarGoogle ScholarCross RefCross Ref
  15. Alireza Yazdani, Lu Lu, Maziar Raissi, and George Em Karniadakis. Systems biology informed deep learning for inferring parameters and hidden dynamics. PLoS computational biology, 16(11):e1007575, 2020.Google ScholarGoogle Scholar
  16. Mark Alber, Adrian Buganza Tepole, William R Cannon, Suvranu De, Salvador Dura-Bernal, Krishna Garikipati, George Karniadakis, William W Lytton, Paris Perdikaris, Linda Petzold, et al. Integrating machine learning and multiscale modeling- perspectives, challenges, and opportunities in the biological, biomedical, and behavioral sciences. NPJ digital medicine, 2(1):1--11, 2019.Google ScholarGoogle Scholar
  17. Charuleka Varadharajan, Vipin Kumar, Jared Willard, Jacob Zwart, Jeff Sadler, Helen Weierbach, Talita Perciano, Juliane Mueller, Valerie Hendrix, and Danielle Christianson. Using machine learning to develop a predictive understanding of the impacts of extreme water cycle perturbations on river water quality. Technical report, Lawrence Berkeley National Lab.(LBNL), Berkeley, CA (United States); Univ . . . , 2021.Google ScholarGoogle ScholarCross RefCross Ref
  18. Tianfang Xu and Albert J Valocchi. Data-driven methods to improve baseflow prediction of a regional groundwater model. Computers & Geosciences, 85:124-- 136, 2015.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Wenqian Dong, Jie Liu, Zhen Xie, and Dong Li. Adaptive neural network-based approximation to accelerate eulerian fluid simulation. In Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, pages 1--22, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Zongyi Li, Nikola Kovachki, Kamyar Azizzadenesheli, Burigede Liu, Kaushik Bhattacharya, Andrew Stuart, and Anima Anandkumar. Fourier neural operator for parametric partial differential equations. arXiv preprint arXiv:2010.08895, 2020.Google ScholarGoogle Scholar
  21. Denghui Lu, Han Wang, Mohan Chen, Lin Lin, Roberto Car, E Weinan, Weile Jia, and Linfeng Zhang. 86 pflops deep potential molecular dynamics simulation of 100 million atoms with ab initio accuracy. Computer Physics Communications, 259:107624, 2021.Google ScholarGoogle ScholarCross RefCross Ref
  22. Wenqian Dong, Zhen Xie, Gokcen Kestor, and Dong Li. Smart-pgsim: Using neural network to accelerate ac-opf power grid simulation. In Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, SC '20. IEEE Press, 2020.Google ScholarGoogle ScholarCross RefCross Ref
  23. Dalvan Griebler, Junior Loff, Gabriele Mencagli, Marco Danelutto, and Luiz Gustavo Fernandes. Efficient nas benchmark kernels with c parallel programming. In 2018 26th Euromicro International Conference on Parallel, Distributed and Network-based Processing (PDP), pages 733--740. IEEE, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  24. Haifeng Jin, Qingquan Song, and Xia Hu. Auto-keras: An efficient neural architecture search system. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 1946--1956, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Ekaba Bisong. Google automl: Cloud vision. In Building Machine Learning and Deep Learning Models on Google Cloud Platform, pages 581--598. Springer, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  26. Tom O'Malley, Elie Bursztein, James Long, François Chollet, Haifeng Jin, Luca Invernizzi, et al. Keras tuner. Retrieved May, 21:2020, 2019.Google ScholarGoogle Scholar
  27. Rudi Helfenstein and Jonas Koko. Parallel preconditioned conjugate gradient algorithm on gpu. Journal of Computational and Applied Mathematics, 236(15):3584-- 3590, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Aiichiro Nakano. Parallel multilevel preconditioned conjugate-gradient approach to variable-charge molecular dynamics. Computer Physics Communications, 104(1--3):59--69, 1997.Google ScholarGoogle Scholar
  29. Mehrzad Samadi, Davoud Anoushe Jamshidi, Janghaeng Lee, and Scott Mahlke. Paraprox: Pattern-Based Approximation for Data Parallel Applications. In Architectural Support for Programming Languages and Operating Systems (ASPLOS), 2014.Google ScholarGoogle Scholar
  30. Woongki Baek and Trishul M Chilimbi. Green: a framework for supporting energy-conscious programming using controlled approximation. In ACM Sigplan Notices, volume 45, pages 198--209. ACM, 2010.Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Marc De Kruijf, Shuou Nomura, and Karthikeyan Sankaralingam. Relax: An architectural framework for software recovery of hardware faults. ACM SIGARCH Computer Architecture News, 38(3):497--508, 2010.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Hadi Esmaeilzadeh, Adrian Sampson, Luis Ceze, and Doug Burger. Architecture support for disciplined approximate programming. In Proceedings of the seventeenth international conference on Architectural Support for Programming Languages and Operating Systems, pages 301--312, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. Stefano Cherubina Giovanni Agostaa Imane Lasrib, Erven Rohoub, and Olivier Sentieysb. Implications of reduced-precision computations in hpc: Performance, energy and error. Parallel Computing is Everywhere, 32:297, 2018.Google ScholarGoogle Scholar
  34. Jonathan Tompson, Kristofer Schlachter, Pablo Sprechmann, and Ken Perlin. Accelerating eulerian fluid simulation with convolutional networks. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 3424-- 3433. JMLR. org, 2017.Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. Kurt Binder, Jürgen Horbach, Walter Kob, Wolfgang Paul, and Fathollah Varnik. Molecular dynamics simulations. Journal of Physics: Condensed Matter, 16(5):S429, 2004.Google ScholarGoogle ScholarCross RefCross Ref
  36. Andrew C Lorenc. Analysis methods for numerical weather prediction. Quarterly Journal of the Royal Meteorological Society, 112(474):1177--1194, 1986.Google ScholarGoogle ScholarCross RefCross Ref
  37. Stelios Sidiroglou-Douskos, Sasa Misailovic, Henry Hoffmann, and Martin Rinard. Managing performance vs. accuracy trade-offs with loop perforation. In Proceedings of the 19th ACM SIGSOFT symposium and the 13th European conference on Foundations of software engineering, pages 124--134, 2011.Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. Henry Hoffmann, Sasa Misailovic, Stelios Sidiroglou, Anant Agarwal, and Martin Rinard. Using code perforation to improve performance, reduce energy consumption, and respond to failures. 2009.Google ScholarGoogle Scholar
  39. Branko Grünbaum and Geoffrey Colin Shephard. Tilings and patterns. Courier Dover Publications, 1987.Google ScholarGoogle Scholar
  40. Philip J Davis. Interpolation and approximation. Courier Corporation, 1975.Google ScholarGoogle Scholar
  41. Frank Noé, Gianni De Fabritiis, and Cecilia Clementi. Machine learning for protein folding and dynamics. Current opinion in structural biology, 60:77--84, 2020.Google ScholarGoogle Scholar
  42. Chiyu Max Jiang, Soheil Esmaeilzadeh, Kamyar Azizzadenesheli, Karthik Kashinath, Mustafa Mustafa, Hamdi A Tchelepi, Philip Marcus, Anima Anandkumar, et al. Meshfreeflownet: A physics-constrained deep continuous space-time super-resolution framework. arXiv preprint arXiv:2005.01463, 2020.Google ScholarGoogle Scholar
  43. Simone Campanoni, Giovanni Agosta, Stefano Crespi Reghizzi, and Andrea Di Biagio. A highly flexible, parallel virtual machine: design and experience of ildjit. Software: Practice and Experience, 40(2):177--207, 2010.Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. L. Guo, D. Li, I. Laguna, and M. Schulz. FlipTracker: Understanding Natural Error Resilience in HPC Applications. In Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis, 2018.Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. Mark Probst, Andreas Krall, and Bernhard Scholz. Register liveness analysis for optimizing dynamic binary translation. In Ninth Working Conference on Reverse Engineering, 2002. Proceedings., pages 35--44. IEEE, 2002.Google ScholarGoogle ScholarCross RefCross Ref
  46. Mary Jean Harrold and Mary Lou Soffa. Efficient computation of interprocedural definition-use chains. ACM Transactions on Programming Languages and Systems (TOPLAS), 16(2):175--204, 1994.Google ScholarGoogle Scholar
  47. Ian Goodfellow, Yoshua Bengio, Aaron Courville, and Yoshua Bengio. Deep learning, volume 1. MIT press Cambridge, 2016.Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. Tianqi Chen, Bing Xu, Chiyuan Zhang, and Carlos Guestrin. Training deep nets with sublinear memory cost. arXiv preprint arXiv:1604.06174, 2016.Google ScholarGoogle Scholar
  49. Martin Pelikan, David E Goldberg, Erick Cantú-Paz, et al. Boa: The bayesian optimization algorithm. In Proceedings of the genetic and evolutionary computation conference GECCO-99, volume 1, pages 525--532. Citeseer, 1999.Google ScholarGoogle Scholar
  50. Ian Dewancker, Michael McCourt, and Scott Clark. Bayesian optimization for machine learning: A practical guidebook. arXiv preprint arXiv:1612.04858, 2016.Google ScholarGoogle Scholar
  51. Sam Partee, Matthew Ellis, Alessandro Rigazzi, Scott Bachman, Gustavo Marques, Andrew Shao, and Benjamin Robbins. Using machine learning at scale in hpc simulations with smartsim: An application to ocean climate modeling. arXiv preprint arXiv:2104.09355, 2021.Google ScholarGoogle Scholar
  52. Raj Patel. Data education. redis is a cache or more? Technical report, EasyChair, 2021.Google ScholarGoogle Scholar
  53. Sharan Chetlur, Cliff Woolley, Philippe Vandermersch, Jonathan Cohen, John Tran, Bryan Catanzaro, and Evan Shelhamer. cudnn: Efficient primitives for deep learning. arXiv preprint arXiv:1410.0759, 2014.Google ScholarGoogle Scholar
  54. Martín Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, et al. Tensorflow: A system for large-scale machine learning. In 12th {USENIX} symposium on operating systems design and implementation ({OSDI} 16), pages 265--283, 2016.Google ScholarGoogle ScholarDigital LibraryDigital Library
  55. Christian Bienia. Benchmarking modern multiprocessors. Princeton University, 2011.Google ScholarGoogle ScholarDigital LibraryDigital Library
  56. Adrian Sampson, André Baixo, Benjamin Ransford, Thierry Moreau, Joshua Yip, Luis Ceze, and Mark Oskin. Accept: A programmer-guided compiler framework for practical approximate computing. University of Washington Technical Report UW-CSE-15-01, 1(2), 2015.Google ScholarGoogle Scholar
  57. Hadi Esmaeilzadeh, Adrian Sampson, Luis Ceze, and Doug Burger. Neural acceleration for general-purpose approximate programs. In 2012 45th Annual IEEE/ACM International Symposium on Microarchitecture, pages 449--460. IEEE, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  58. Divya Mahajan, Amir Yazdanbakhsh, Jongse Park, Bradley Thwaites, and Hadi Esmaeilzadeh. Towards statistical guarantees in controlling quality tradeoffs for approximate acceleration. ACM SIGARCH Computer Architecture News, 44(3):66--77, 2016.Google ScholarGoogle ScholarDigital LibraryDigital Library
  59. Paul Messina. The exascale computing project. Computing in Science & Engineering, 19(3):63--67, 2017.Google ScholarGoogle ScholarCross RefCross Ref
  60. Thierry Moreau, Mark Wyse, Jacob Nelson, Adrian Sampson, Hadi Esmaeilzadeh, Luis Ceze, and Mark Oskin. Snnap: Approximate computing on programmable socs via neural acceleration. In 2015 IEEE 21st International Symposium on High Performance Computer Architecture (HPCA), pages 603--614. IEEE, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  61. SIAM Journal on Scientific Computing. Amgx: A library for gpu accelerated algebraic multigrid and preconditioned iterative methods. Notices of the AMS, 37.5:S602--S626, 2015.Google ScholarGoogle Scholar
  62. Konstantinos Parasyris, Giorgis Georgakoudis, Harshitha Menon, James Diffenderfer, Ignacio Laguna, Daniel Osei-Kuffuor, and Markus Schordan. Hpac: evaluating approximate computing techniques on hpc openmp applications. In Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, pages 1--14, 2021.Google ScholarGoogle ScholarDigital LibraryDigital Library
  63. Anh Truong, Austin Walters, Jeremy Goodsitt, Keegan Hines, C Bayan Bruss, and Reza Farivar. Towards automated machine learning: Evaluation and comparison of automl approaches and tools. In 2019 IEEE 31st international conference on tools with artificial intelligence (ICTAI), pages 1471--1479. IEEE, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  64. Uri Shaham, Alexander Cloninger, and Ronald R Coifman. Provable approximation properties for deep neural networks. Applied and Computational Harmonic Analysis, 44(3):537--557, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  65. Sungmoon Jung and Jamshid Ghaboussi. Neural network constitutive model for rate-dependent materials. Computers & Structures, 84(15--16):955--963, 2006.Google ScholarGoogle Scholar
  66. Baris Sen and Suresh Menon. Representation of chemical kinetics by artificial neural networks for large eddy simulations. In 43rd Aiaa/Asme/Sae/Asee Joint Propulsion Conference & Exhibit, page 5635, 2007.Google ScholarGoogle ScholarCross RefCross Ref
  67. Erik Marchi, Fabio Vesperini, Florian Eyben, Stefano Squartini, and Björn Schuller. A novel approach for automatic acoustic novelty detection using a denoising autoencoder with bidirectional lstm neural networks. In Proceedings 40th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2015, pages 5--pages, 2015.Google ScholarGoogle Scholar
  68. Julia Ling, Andrew Kurzawski, and Jeremy Templeton. Reynolds averaged turbulence modelling using deep neural networks with embedded invariance. Journal of Fluid Mechanics, 807:155--166, 2016.Google ScholarGoogle ScholarCross RefCross Ref
  69. Giuseppe Carleo and Matthias Troyer. Solving the quantum many-body problem with artificial neural networks. Science, 355(6325):602--606, 2017.Google ScholarGoogle ScholarCross RefCross Ref
  70. Yunjie Liu, Evan Racah, Joaquin Correa, Amir Khosrowshahi, David Lavers, Kenneth Kunkel, Michael Wehner, William Collins, et al. Application of deep convolutional neural networks for detecting extreme weather in climate datasets. arXiv preprint arXiv:1605.01156, 2016.Google ScholarGoogle Scholar
  71. Guanghui Liang and K Chandrashekhara. Neural network based constitutive model for elastomeric foams. Engineering structures, 30(7):2002--2011, 2008.Google ScholarGoogle ScholarCross RefCross Ref
  72. Yuelin Shen, K Chandrashekhara, WF Breig, and LR Oliver. Finite element analysis of v-ribbed belts using neural network based hyperelastic material model. International Journal of Non-Linear Mechanics, 40(6):875--890, 2005.Google ScholarGoogle ScholarCross RefCross Ref
  73. Zhen Xie, Wenqian Dong, Jiawen Liu, Hang Liu, and Dong Li. Tahoe: tree structure-aware high performance inference engine for decision tree ensemble on gpu. In Proceedings of the Sixteenth European Conference on Computer Systems, pages 426--440, 2021.Google ScholarGoogle ScholarDigital LibraryDigital Library
  74. J. S. Smith, O. Isayev, and A. E. Roitberg. Ani-1: an extensible neural network potential with dft accuracy at force field computational cost. 2017.Google ScholarGoogle Scholar
  75. Michael Gastegger, Jörg Behlerb, and Philipp Marquetand. Machine Learning Molecular Dynamics for the Simulation of Infrared Spectra . Chemical Science, pages 6695--7270, 2017.Google ScholarGoogle ScholarCross RefCross Ref
  76. Wenqian Dong, Letian Kang, Zhe Quan, Kenli Li, Keqin Li, Ziyu Hao, and Xiang- Hui Xie. Implementing molecular dynamics simulation on sunway taihulight system. In 2016 IEEE 18th international conference on high performance computing and communications; IEEE 14th international conference on smart city; IEEE 2nd international conference on data science and systems (HPCC/SmartCity/DSS), pages 443--450. IEEE, 2016.Google ScholarGoogle Scholar
  77. Wenqian Dong, Kenli Li, Letian Kang, Zhe Quan, and Keqin Li. Implementing molecular dynamics simulation on the sunway taihulight system with heterogeneous many-core processors. Concurrency and Computation: Practice and Experience, 30(16):e4468, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  78. Hadi Esmaeilzadeh, Adrian Sampson, Luis Ceze, and Doug Burger. Architecture support for disciplined approximate programming. In ASPLOS, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  79. J. Han and M. Orshansky. Approximate computing: An emerging paradigm for energy-efficient design. In ETS, 2013.Google ScholarGoogle ScholarCross RefCross Ref
  80. Adrian Sampson, Jacob Nelson, Karin Strauss, and Luis Ceze. Approximate storage in solid-state memories. In ACM TOCS, 2014.Google ScholarGoogle ScholarDigital LibraryDigital Library
  81. Zeshi Liu, Zhen Xie, Wenqian Dong, Mengting Yuan, Haihang You, and Dong Li. A heterogeneous processing-in-memory approach to accelerate quantum chemistry simulation. Parallel Computing, 116:103017, 2023.Google ScholarGoogle ScholarDigital LibraryDigital Library
  82. Sasa Misailovic, Stelios Sidiroglou, Henry Hoffmann, and Martin Rinard. Quality of service profiling. In ICSE, 2010.Google ScholarGoogle ScholarDigital LibraryDigital Library
  83. Adrian Sampson, Andre Baixo, Benjamin Ransford, Thierry Moreau, Joshua Yip, Luis Ceze, and Mark Oskin. Accept: A programmer-guided compiler framework for practical approximate computing. University of Washington Technical Report UW-CSE-15-01, 2015.Google ScholarGoogle Scholar
  84. V. K. Chippa, D. Mohapatra, A. Raghunathan, K. Roy, and S. T. Chakradhar. Scalable effort hardware design: Exploiting algorithmic resilience for energy efficiency. In Design Automation Conference, 2010.Google ScholarGoogle ScholarDigital LibraryDigital Library
  85. Martin Rinard. Probabilistic accuracy bounds for fault-tolerant computations that discard tasks. In PInternational Conference on Supercomputing, 2006.Google ScholarGoogle ScholarDigital LibraryDigital Library
  86. Mehrzad Samadi, Janghaeng Lee, D Anoushe Jamshidi, Amir Hormati, and Scott Mahlke. Sage: Self-tuning approximation for graphics engines. In Microarchitecture (MICRO), 2013 46th Annual IEEE/ACM International Symposium on, 2013.Google ScholarGoogle Scholar
  87. Adrian Sampson, Werner Dietl, Emily Fortuna, Danushen Gnanapragasam, Luis Ceze, and Dan Grossman. Enerj: Approximate data types for safe and general low-power computation. In PLDI, 2011.Google ScholarGoogle ScholarDigital LibraryDigital Library
  88. Simone Campanoni, Glenn Holloway, Gu-Yeon Wei, and David Brooks. Helix-up: Relaxing program semantics to unleash parallelization. In IEEE/ACM International Symposium on Code Generation and Optimization, 2015.Google ScholarGoogle Scholar
  89. Inigo Goiri, Ricardo Bianchini, Santosh Nagarakatte, and Thu D. Nguyen. Approxhadoop: Bringing approximations to mapreduce frameworks. In ASPLOS, 2015.Google ScholarGoogle ScholarDigital LibraryDigital Library
  90. Shuangyan Yang, Minjia Zhang, Wenqian Dong, and Dong Li. Betty: Enabling large-scale gnn training with batch-level graph partitioning. 2023.Google ScholarGoogle Scholar
  91. Henry Hoffmann, Stelios Sidiroglou, Michael Carbin, Sasa Misailovic, Anant Agarwal, and Martin Rinard. Dynamic knobs for responsive power-aware computing. In ASPLOS, 2011.Google ScholarGoogle ScholarDigital LibraryDigital Library
  92. Jie Liu, Wenqian Dong, Qingqing Zhou, and Dong Li. Fauce: Fast and accurate deep ensembles with uncertainty for cardinality estimation. Proceedings of the VLDB Endowment, 14(11):1950--1963, 2021.Google ScholarGoogle ScholarDigital LibraryDigital Library
  93. Zhen Xie, Wenqian Dong, Jie Liu, Ivy Peng, Yanbao Ma, and Dong Li. Md-hm: memoization-based molecular dynamics simulations on big memory system. In Proceedings of the ACM International Conference on Supercomputing, pages 215--226, 2021.Google ScholarGoogle ScholarDigital LibraryDigital Library
  94. Daya S Khudia, Babak Zamirai, Mehrzad Samadi, and Scott Mahlke. Rumba: An online quality management system for approximate computing. In Computer Architecture (ISCA), 2015 ACM/IEEE 42nd Annual International Symposium on, 2015.Google ScholarGoogle Scholar
  95. Dhanya R. Krishnan, Do Le Quoc, Pramod Bhatotia, Christof Fetzer, and Rodrigo Rodrigues. Incapprox: A data analytics system for incremental approximate computing. In WWW, 2016.Google ScholarGoogle Scholar
  96. Xin Sui, Andrew Lenharth, Donald S Fussell, and Keshav Pingali. Proactive control of approximate programs. ACM SIGPLAN Notices, 51(4):607--621, 2016.Google ScholarGoogle ScholarDigital LibraryDigital Library
  97. Michael A Laurenzano, Parker Hill, Mehrzad Samadi, Scott Mahlke, Jason Mars, and Lingjia Tang. Input responsiveness: using canary inputs to dynamically steer approximation. In Proceedings of the 37th ACM SIGPLAN Conference on Programming Language Design and Implementation, pages 161--176, 2016.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Auto-HPCnet: An Automatic Framework to Build Neural Network-based Surrogate for High-Performance Computing Applications

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          HPDC '23: Proceedings of the 32nd International Symposium on High-Performance Parallel and Distributed Computing
          August 2023
          350 pages
          ISBN:9798400701559
          DOI:10.1145/3588195
          • General Chair:
          • Ali R. Butt,
          • Program Chairs:
          • Ningfang Mi,
          • Kyle Chard

          Copyright © 2023 ACM

          Publication rights licensed to ACM. ACM acknowledges that this contribution was authored or co-authored by an employee, contractor or affiliate of the United States government. As such, the Government retains a nonexclusive, royalty-free right to publish or reproduce this article, or to allow others to do so, for Government purposes only.

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 7 August 2023

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

          Acceptance Rates

          Overall Acceptance Rate166of966submissions,17%

          Upcoming Conference

        • Article Metrics

          • Downloads (Last 12 months)244
          • Downloads (Last 6 weeks)22

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader