CINXE.COM

Search | arXiv e-print repository

<!DOCTYPE html> <html lang="en"> <head> <meta charset="utf-8"/> <meta name="viewport" content="width=device-width, initial-scale=1"/> <!-- new favicon config and versions by realfavicongenerator.net --> <link rel="apple-touch-icon" sizes="180x180" href="https://static.arxiv.org/static/base/1.0.0a5/images/icons/apple-touch-icon.png"> <link rel="icon" type="image/png" sizes="32x32" href="https://static.arxiv.org/static/base/1.0.0a5/images/icons/favicon-32x32.png"> <link rel="icon" type="image/png" sizes="16x16" href="https://static.arxiv.org/static/base/1.0.0a5/images/icons/favicon-16x16.png"> <link rel="manifest" href="https://static.arxiv.org/static/base/1.0.0a5/images/icons/site.webmanifest"> <link rel="mask-icon" href="https://static.arxiv.org/static/base/1.0.0a5/images/icons/safari-pinned-tab.svg" color="#b31b1b"> <link rel="shortcut icon" href="https://static.arxiv.org/static/base/1.0.0a5/images/icons/favicon.ico"> <meta name="msapplication-TileColor" content="#b31b1b"> <meta name="msapplication-config" content="images/icons/browserconfig.xml"> <meta name="theme-color" content="#b31b1b"> <!-- end favicon config --> <title>Search | arXiv e-print repository</title> <script defer src="https://static.arxiv.org/static/base/1.0.0a5/fontawesome-free-5.11.2-web/js/all.js"></script> <link rel="stylesheet" href="https://static.arxiv.org/static/base/1.0.0a5/css/arxivstyle.css" /> <script type="text/x-mathjax-config"> MathJax.Hub.Config({ messageStyle: "none", extensions: ["tex2jax.js"], jax: ["input/TeX", "output/HTML-CSS"], tex2jax: { inlineMath: [ ['$','$'], ["\\(","\\)"] ], displayMath: [ ['$$','$$'], ["\\[","\\]"] ], processEscapes: true, ignoreClass: '.*', processClass: 'mathjax.*' }, TeX: { extensions: ["AMSmath.js", "AMSsymbols.js", "noErrors.js"], noErrors: { inlineDelimiters: ["$","$"], multiLine: false, style: { "font-size": "normal", "border": "" } } }, "HTML-CSS": { availableFonts: ["TeX"] } }); </script> <script src='//static.arxiv.org/MathJax-2.7.3/MathJax.js'></script> <script src="https://static.arxiv.org/static/base/1.0.0a5/js/notification.js"></script> <link rel="stylesheet" href="https://static.arxiv.org/static/search/0.5.6/css/bulma-tooltip.min.css" /> <link rel="stylesheet" href="https://static.arxiv.org/static/search/0.5.6/css/search.css" /> <script src="https://code.jquery.com/jquery-3.2.1.slim.min.js" integrity="sha256-k2WSCIexGzOj3Euiig+TlR8gA0EmPjuc79OEeY5L45g=" crossorigin="anonymous"></script> <script src="https://static.arxiv.org/static/search/0.5.6/js/fieldset.js"></script> <style> radio#cf-customfield_11400 { display: none; } </style> </head> <body> <header><a href="#main-container" class="is-sr-only">Skip to main content</a> <!-- contains Cornell logo and sponsor statement --> <div class="attribution level is-marginless" role="banner"> <div class="level-left"> <a class="level-item" href="https://cornell.edu/"><img src="https://static.arxiv.org/static/base/1.0.0a5/images/cornell-reduced-white-SMALL.svg" alt="Cornell University" width="200" aria-label="logo" /></a> </div> <div class="level-right is-marginless"><p class="sponsors level-item is-marginless"><span id="support-ack-url">We gratefully acknowledge support from<br /> the Simons Foundation, <a href="https://info.arxiv.org/about/ourmembers.html">member institutions</a>, and all contributors. <a href="https://info.arxiv.org/about/donate.html">Donate</a></span></p></div> </div> <!-- contains arXiv identity and search bar --> <div class="identity level is-marginless"> <div class="level-left"> <div class="level-item"> <a class="arxiv" href="https://arxiv.org/" aria-label="arxiv-logo"> <img src="https://static.arxiv.org/static/base/1.0.0a5/images/arxiv-logo-one-color-white.svg" aria-label="logo" alt="arxiv logo" width="85" style="width:85px;"/> </a> </div> </div> <div class="search-block level-right"> <form class="level-item mini-search" method="GET" action="https://arxiv.org/search"> <div class="field has-addons"> <div class="control"> <input class="input is-small" type="text" name="query" placeholder="Search..." aria-label="Search term or terms" /> <p class="help"><a href="https://info.arxiv.org/help">Help</a> | <a href="https://arxiv.org/search/advanced">Advanced Search</a></p> </div> <div class="control"> <div class="select is-small"> <select name="searchtype" aria-label="Field to search"> <option value="all" selected="selected">All fields</option> <option value="title">Title</option> <option value="author">Author</option> <option value="abstract">Abstract</option> <option value="comments">Comments</option> <option value="journal_ref">Journal reference</option> <option value="acm_class">ACM classification</option> <option value="msc_class">MSC classification</option> <option value="report_num">Report number</option> <option value="paper_id">arXiv identifier</option> <option value="doi">DOI</option> <option value="orcid">ORCID</option> <option value="author_id">arXiv author ID</option> <option value="help">Help pages</option> <option value="full_text">Full text</option> </select> </div> </div> <input type="hidden" name="source" value="header"> <button class="button is-small is-cul-darker">Search</button> </div> </form> </div> </div> <!-- closes identity --> <div class="container"> <div class="user-tools is-size-7 has-text-right has-text-weight-bold" role="navigation" aria-label="User menu"> <a href="https://arxiv.org/login">Login</a> </div> </div> </header> <main class="container" id="main-container"> <div class="level is-marginless"> <div class="level-left"> <h1 class="title is-clearfix"> Showing 1&ndash;50 of 72 results for author: <span class="mathjax">Chard, K</span> </h1> </div> <div class="level-right is-hidden-mobile"> <!-- feedback for mobile is moved to footer --> <span class="help" style="display: inline-block;"><a href="https://github.com/arXiv/arxiv-search/releases">Search v0.5.6 released 2020-02-24</a>&nbsp;&nbsp;</span> </div> </div> <div class="content"> <form method="GET" action="/search/cs" aria-role="search"> Searching in archive <strong>cs</strong>. <a href="/search/?searchtype=author&amp;query=Chard%2C+K">Search in all archives.</a> <div class="field has-addons-tablet"> <div class="control is-expanded"> <label for="query" class="hidden-label">Search term or terms</label> <input class="input is-medium" id="query" name="query" placeholder="Search term..." type="text" value="Chard, K"> </div> <div class="select control is-medium"> <label class="is-hidden" for="searchtype">Field</label> <select class="is-medium" id="searchtype" name="searchtype"><option value="all">All fields</option><option value="title">Title</option><option selected value="author">Author(s)</option><option value="abstract">Abstract</option><option value="comments">Comments</option><option value="journal_ref">Journal reference</option><option value="acm_class">ACM classification</option><option value="msc_class">MSC classification</option><option value="report_num">Report number</option><option value="paper_id">arXiv identifier</option><option value="doi">DOI</option><option value="orcid">ORCID</option><option value="license">License (URI)</option><option value="author_id">arXiv author ID</option><option value="help">Help pages</option><option value="full_text">Full text</option></select> </div> <div class="control"> <button class="button is-link is-medium">Search</button> </div> </div> <div class="field"> <div class="control is-size-7"> <label class="radio"> <input checked id="abstracts-0" name="abstracts" type="radio" value="show"> Show abstracts </label> <label class="radio"> <input id="abstracts-1" name="abstracts" type="radio" value="hide"> Hide abstracts </label> </div> </div> <div class="is-clearfix" style="height: 2.5em"> <div class="is-pulled-right"> <a href="/search/advanced?terms-0-term=Chard%2C+K&amp;terms-0-field=author&amp;size=50&amp;order=-announced_date_first">Advanced Search</a> </div> </div> <input type="hidden" name="order" value="-announced_date_first"> <input type="hidden" name="size" value="50"> </form> <div class="level breathe-horizontal"> <div class="level-left"> <form method="GET" action="/search/"> <div style="display: none;"> <select id="searchtype" name="searchtype"><option value="all">All fields</option><option value="title">Title</option><option selected value="author">Author(s)</option><option value="abstract">Abstract</option><option value="comments">Comments</option><option value="journal_ref">Journal reference</option><option value="acm_class">ACM classification</option><option value="msc_class">MSC classification</option><option value="report_num">Report number</option><option value="paper_id">arXiv identifier</option><option value="doi">DOI</option><option value="orcid">ORCID</option><option value="license">License (URI)</option><option value="author_id">arXiv author ID</option><option value="help">Help pages</option><option value="full_text">Full text</option></select> <input id="query" name="query" type="text" value="Chard, K"> <ul id="abstracts"><li><input checked id="abstracts-0" name="abstracts" type="radio" value="show"> <label for="abstracts-0">Show abstracts</label></li><li><input id="abstracts-1" name="abstracts" type="radio" value="hide"> <label for="abstracts-1">Hide abstracts</label></li></ul> </div> <div class="box field is-grouped is-grouped-multiline level-item"> <div class="control"> <span class="select is-small"> <select id="size" name="size"><option value="25">25</option><option selected value="50">50</option><option value="100">100</option><option value="200">200</option></select> </span> <label for="size">results per page</label>. </div> <div class="control"> <label for="order">Sort results by</label> <span class="select is-small"> <select id="order" name="order"><option selected value="-announced_date_first">Announcement date (newest first)</option><option value="announced_date_first">Announcement date (oldest first)</option><option value="-submitted_date">Submission date (newest first)</option><option value="submitted_date">Submission date (oldest first)</option><option value="">Relevance</option></select> </span> </div> <div class="control"> <button class="button is-small is-link">Go</button> </div> </div> </form> </div> </div> <nav class="pagination is-small is-centered breathe-horizontal" role="navigation" aria-label="pagination"> <a href="" class="pagination-previous is-invisible">Previous </a> <a href="/search/?searchtype=author&amp;query=Chard%2C+K&amp;start=50" class="pagination-next" >Next </a> <ul class="pagination-list"> <li> <a href="/search/?searchtype=author&amp;query=Chard%2C+K&amp;start=0" class="pagination-link is-current" aria-label="Goto page 1">1 </a> </li> <li> <a href="/search/?searchtype=author&amp;query=Chard%2C+K&amp;start=50" class="pagination-link " aria-label="Page 2" aria-current="page">2 </a> </li> </ul> </nav> <ol class="breathe-horizontal" start="1"> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2411.10637">arXiv:2411.10637</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2411.10637">pdf</a>, <a href="https://arxiv.org/format/2411.10637">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Software Engineering">cs.SE</span> </div> </div> <p class="title is-5 mathjax"> Exascale Workflow Applications and Middleware: An ExaWorks Retrospective </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Alsaadi%2C+A">Aymen Alsaadi</a>, <a href="/search/cs?searchtype=author&amp;query=Hategan-Marandiuc%2C+M">Mihael Hategan-Marandiuc</a>, <a href="/search/cs?searchtype=author&amp;query=Maheshwari%2C+K">Ketan Maheshwari</a>, <a href="/search/cs?searchtype=author&amp;query=Merzky%2C+A">Andre Merzky</a>, <a href="/search/cs?searchtype=author&amp;query=Titov%2C+M">Mikhail Titov</a>, <a href="/search/cs?searchtype=author&amp;query=Turilli%2C+M">Matteo Turilli</a>, <a href="/search/cs?searchtype=author&amp;query=Wilke%2C+A">Andreas Wilke</a>, <a href="/search/cs?searchtype=author&amp;query=Wozniak%2C+J+M">Justin M. Wozniak</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=da+Silva%2C+R+F">Rafael Ferreira da Silva</a>, <a href="/search/cs?searchtype=author&amp;query=Jha%2C+S">Shantenu Jha</a>, <a href="/search/cs?searchtype=author&amp;query=Laney%2C+D">Daniel Laney</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2411.10637v1-abstract-short" style="display: inline;"> Exascale computers offer transformative capabilities to combine data-driven and learning-based approaches with traditional simulation applications to accelerate scientific discovery and insight. However, these software combinations and integrations are difficult to achieve due to the challenges of coordinating and deploying heterogeneous software components on diverse and massive platforms. We pre&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2411.10637v1-abstract-full').style.display = 'inline'; document.getElementById('2411.10637v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2411.10637v1-abstract-full" style="display: none;"> Exascale computers offer transformative capabilities to combine data-driven and learning-based approaches with traditional simulation applications to accelerate scientific discovery and insight. However, these software combinations and integrations are difficult to achieve due to the challenges of coordinating and deploying heterogeneous software components on diverse and massive platforms. We present the ExaWorks project, which addresses many of these challenges. We developed a workflow Software Development Toolkit (SDK), a curated collection of workflow technologies that can be composed and interoperated through a common interface, engineered following current best practices, and specifically designed to work on HPC platforms. ExaWorks also developed PSI/J, a job management abstraction API, to simplify the construction of portable software components and applications that can be used over various HPC schedulers. The PSI/J API is a minimal interface for submitting and monitoring jobs and their execution state across multiple and commonly used HPC schedulers. We also describe several leading and innovative workflow examples of ExaWorks tools used on DOE leadership platforms. Furthermore, we discuss how our project is working with the workflow community, large computing facilities, and HPC platform vendors to address the requirements of workflows sustainably at the exascale. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2411.10637v1-abstract-full').style.display = 'none'; document.getElementById('2411.10637v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 15 November, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> November 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2411.04257">arXiv:2411.04257</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2411.04257">pdf</a>, <a href="https://arxiv.org/format/2411.04257">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> </div> </div> <p class="title is-5 mathjax"> LSHBloom: Memory-efficient, Extreme-scale Document Deduplication </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Khan%2C+A">Arham Khan</a>, <a href="/search/cs?searchtype=author&amp;query=Underwood%2C+R">Robert Underwood</a>, <a href="/search/cs?searchtype=author&amp;query=Siebenschuh%2C+C">Carlo Siebenschuh</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Ajith%2C+A">Aswathy Ajith</a>, <a href="/search/cs?searchtype=author&amp;query=Hippe%2C+K">Kyle Hippe</a>, <a href="/search/cs?searchtype=author&amp;query=Gokdemir%2C+O">Ozan Gokdemir</a>, <a href="/search/cs?searchtype=author&amp;query=Brace%2C+A">Alexander Brace</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2411.04257v1-abstract-short" style="display: inline;"> Deduplication is a major focus for assembling and curating training datasets for large language models (LLM) -- detecting and eliminating additional instances of the same content -- in large collections of technical documents. Unrestrained, duplicates in the training dataset increase training costs and lead to undesirable properties such as memorization in trained models or cheating on evaluation.&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2411.04257v1-abstract-full').style.display = 'inline'; document.getElementById('2411.04257v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2411.04257v1-abstract-full" style="display: none;"> Deduplication is a major focus for assembling and curating training datasets for large language models (LLM) -- detecting and eliminating additional instances of the same content -- in large collections of technical documents. Unrestrained, duplicates in the training dataset increase training costs and lead to undesirable properties such as memorization in trained models or cheating on evaluation. Contemporary approaches to document-level deduplication are often extremely expensive in both runtime and memory. We propose LSHBloom, an extension to MinhashLSH, which replaces the expensive LSHIndex with lightweight Bloom filters. LSHBloom demonstrates the same deduplication performance as MinhashLSH with only a marginal increase in false positives (as low as 1e-5 in our experiments); demonstrates competitive runtime (270\% faster than MinhashLSH on peS2o); and, crucially, uses just 0.6\% of the disk space required by MinhashLSH to deduplicate peS2o. We demonstrate that this space advantage scales with increased dataset size -- at the extreme scale of several billion documents, LSHBloom promises a 250\% speedup and a 54$\times$ space advantage over traditional MinHashLSH scaling deduplication of text datasets to many billions of documents. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2411.04257v1-abstract-full').style.display = 'none'; document.getElementById('2411.04257v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 6 November, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> November 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2410.14943">arXiv:2410.14943</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2410.14943">pdf</a>, <a href="https://arxiv.org/format/2410.14943">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.5281/zenodo.13844758">10.5281/zenodo.13844758 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> Workflows Community Summit 2024: Future Trends and Challenges in Scientific Workflows </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=da+Silva%2C+R+F">Rafael Ferreira da Silva</a>, <a href="/search/cs?searchtype=author&amp;query=Bard%2C+D">Deborah Bard</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=de+Witt%2C+S">Shaun de Witt</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I+T">Ian T. Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Gibbs%2C+T">Tom Gibbs</a>, <a href="/search/cs?searchtype=author&amp;query=Goble%2C+C">Carole Goble</a>, <a href="/search/cs?searchtype=author&amp;query=Godoy%2C+W">William Godoy</a>, <a href="/search/cs?searchtype=author&amp;query=Gustafsson%2C+J">Johan Gustafsson</a>, <a href="/search/cs?searchtype=author&amp;query=Haus%2C+U">Utz-Uwe Haus</a>, <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+S">Stephen Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Jha%2C+S">Shantenu Jha</a>, <a href="/search/cs?searchtype=author&amp;query=Los%2C+L">Laila Los</a>, <a href="/search/cs?searchtype=author&amp;query=Paine%2C+D">Drew Paine</a>, <a href="/search/cs?searchtype=author&amp;query=Suter%2C+F">Fr茅d茅ric Suter</a>, <a href="/search/cs?searchtype=author&amp;query=Ward%2C+L">Logan Ward</a>, <a href="/search/cs?searchtype=author&amp;query=Wilkinson%2C+S">Sean Wilkinson</a>, <a href="/search/cs?searchtype=author&amp;query=Amaris%2C+M">Marcos Amaris</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Bader%2C+J">Jonathan Bader</a>, <a href="/search/cs?searchtype=author&amp;query=Balin%2C+R">Riccardo Balin</a>, <a href="/search/cs?searchtype=author&amp;query=Balouek%2C+D">Daniel Balouek</a>, <a href="/search/cs?searchtype=author&amp;query=Beecroft%2C+S">Sarah Beecroft</a>, <a href="/search/cs?searchtype=author&amp;query=Belhajjame%2C+K">Khalid Belhajjame</a>, <a href="/search/cs?searchtype=author&amp;query=Bhattarai%2C+R">Rajat Bhattarai</a> , et al. (86 additional authors not shown) </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2410.14943v1-abstract-short" style="display: inline;"> The Workflows Community Summit gathered 111 participants from 18 countries to discuss emerging trends and challenges in scientific workflows, focusing on six key areas: time-sensitive workflows, AI-HPC convergence, multi-facility workflows, heterogeneous HPC environments, user experience, and FAIR computational workflows. The integration of AI and exascale computing has revolutionized scientific w&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2410.14943v1-abstract-full').style.display = 'inline'; document.getElementById('2410.14943v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2410.14943v1-abstract-full" style="display: none;"> The Workflows Community Summit gathered 111 participants from 18 countries to discuss emerging trends and challenges in scientific workflows, focusing on six key areas: time-sensitive workflows, AI-HPC convergence, multi-facility workflows, heterogeneous HPC environments, user experience, and FAIR computational workflows. The integration of AI and exascale computing has revolutionized scientific workflows, enabling higher-fidelity models and complex, time-sensitive processes, while introducing challenges in managing heterogeneous environments and multi-facility data dependencies. The rise of large language models is driving computational demands to zettaflop scales, necessitating modular, adaptable systems and cloud-service models to optimize resource utilization and ensure reproducibility. Multi-facility workflows present challenges in data movement, curation, and overcoming institutional silos, while diverse hardware architectures require integrating workflow considerations into early system design and developing standardized resource management tools. The summit emphasized improving user experience in workflow systems and ensuring FAIR workflows to enhance collaboration and accelerate scientific discovery. Key recommendations include developing standardized metrics for time-sensitive workflows, creating frameworks for cloud-HPC integration, implementing distributed-by-design workflow modeling, establishing multi-facility authentication protocols, and accelerating AI integration in HPC workflow management. The summit also called for comprehensive workflow benchmarks, workflow-specific UX principles, and a FAIR workflow maturity model, highlighting the need for continued collaboration in addressing the complex challenges posed by the convergence of AI, HPC, and multi-facility research environments. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2410.14943v1-abstract-full').style.display = 'none'; document.getElementById('2410.14943v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 18 October, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> October 2024. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Report number:</span> ORNL/TM-2024/3573 </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2410.12927">arXiv:2410.12927</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2410.12927">pdf</a>, <a href="https://arxiv.org/format/2410.12927">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Artificial Intelligence">cs.AI</span> </div> </div> <p class="title is-5 mathjax"> SoK: On Finding Common Ground in Loss Landscapes Using Deep Model Merging Techniques </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Khan%2C+A">Arham Khan</a>, <a href="/search/cs?searchtype=author&amp;query=Nief%2C+T">Todd Nief</a>, <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Sakarvadia%2C+M">Mansi Sakarvadia</a>, <a href="/search/cs?searchtype=author&amp;query=Grzenda%2C+D">Daniel Grzenda</a>, <a href="/search/cs?searchtype=author&amp;query=Ajith%2C+A">Aswathy Ajith</a>, <a href="/search/cs?searchtype=author&amp;query=Pettyjohn%2C+J">Jordan Pettyjohn</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2410.12927v1-abstract-short" style="display: inline;"> Understanding neural networks is crucial to creating reliable and trustworthy deep learning models. Most contemporary research in interpretability analyzes just one model at a time via causal intervention or activation analysis. Yet despite successes, these methods leave significant gaps in our understanding of the training behaviors of neural networks, how their inner representations emerge, and&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2410.12927v1-abstract-full').style.display = 'inline'; document.getElementById('2410.12927v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2410.12927v1-abstract-full" style="display: none;"> Understanding neural networks is crucial to creating reliable and trustworthy deep learning models. Most contemporary research in interpretability analyzes just one model at a time via causal intervention or activation analysis. Yet despite successes, these methods leave significant gaps in our understanding of the training behaviors of neural networks, how their inner representations emerge, and how we can predictably associate model components with task-specific behaviors. Seeking new insights from work in related fields, here we survey literature in the field of model merging, a field that aims to combine the abilities of various neural networks by merging their parameters and identifying task-specific model components in the process. We analyze the model merging literature through the lens of loss landscape geometry, an approach that enables us to connect observations from empirical studies on interpretability, security, model merging, and loss landscape analysis to phenomena that govern neural network training and the emergence of their inner representations. To systematize knowledge in this area, we present a novel taxonomy of model merging techniques organized by their core algorithmic principles. Additionally, we distill repeated empirical observations from the literature in these fields into characterizations of four major aspects of loss landscape geometry: mode convexity, determinism, directedness, and connectivity. We argue that by improving our understanding of the principles underlying model merging and loss landscape geometry, this work contributes to the goal of ensuring secure and trustworthy machine learning in practice. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2410.12927v1-abstract-full').style.display = 'none'; document.getElementById('2410.12927v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 16 October, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> October 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2410.12092">arXiv:2410.12092</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2410.12092">pdf</a>, <a href="https://arxiv.org/format/2410.12092">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Accelerating Python Applications with Dask and ProxyStore </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Rydzy%2C+K">Klaudiusz Rydzy</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Valerie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2410.12092v2-abstract-short" style="display: inline;"> Applications are increasingly written as dynamic workflows underpinned by an execution framework that manages asynchronous computations across distributed hardware. However, execution frameworks typically offer one-size-fits-all solutions for data flow management, which can restrict performance and scalability. ProxyStore, a middleware layer that optimizes data flow via an advanced pass-by-referen&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2410.12092v2-abstract-full').style.display = 'inline'; document.getElementById('2410.12092v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2410.12092v2-abstract-full" style="display: none;"> Applications are increasingly written as dynamic workflows underpinned by an execution framework that manages asynchronous computations across distributed hardware. However, execution frameworks typically offer one-size-fits-all solutions for data flow management, which can restrict performance and scalability. ProxyStore, a middleware layer that optimizes data flow via an advanced pass-by-reference paradigm, has shown to be an effective mechanism for addressing these limitations. Here, we investigate integrating ProxyStore with Dask Distributed, one of the most popular libraries for distributed computing in Python, with the goal of supporting scalable and portable scientific workflows. Dask provides an easy-to-use and flexible framework, but is less optimized for scaling certain data-intensive workflows. We investigate these limitations and detail the technical contributions necessary to develop a robust solution for distributed applications and demonstrate improved performance on synthetic benchmarks and real applications. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2410.12092v2-abstract-full').style.display = 'none'; document.getElementById('2410.12092v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 17 October, 2024; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 15 October, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> October 2024. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">To be presented as a demo at the SC24 Workshop on High Performance Python for Science at Scale (HPPSS)</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2410.02159">arXiv:2410.02159</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2410.02159">pdf</a>, <a href="https://arxiv.org/format/2410.02159">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Artificial Intelligence">cs.AI</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Computation and Language">cs.CL</span> </div> </div> <p class="title is-5 mathjax"> Mitigating Memorization In Language Models </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Sakarvadia%2C+M">Mansi Sakarvadia</a>, <a href="/search/cs?searchtype=author&amp;query=Ajith%2C+A">Aswathy Ajith</a>, <a href="/search/cs?searchtype=author&amp;query=Khan%2C+A">Arham Khan</a>, <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Geniesse%2C+C">Caleb Geniesse</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Yang%2C+Y">Yaoqing Yang</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Mahoney%2C+M+W">Michael W. Mahoney</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2410.02159v1-abstract-short" style="display: inline;"> Language models (LMs) can &#34;memorize&#34; information, i.e., encode training data in their weights in such a way that inference-time queries can lead to verbatim regurgitation of that data. This ability to extract training data can be problematic, for example, when data are private or sensitive. In this work, we investigate methods to mitigate memorization: three regularizer-based, three finetuning-bas&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2410.02159v1-abstract-full').style.display = 'inline'; document.getElementById('2410.02159v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2410.02159v1-abstract-full" style="display: none;"> Language models (LMs) can &#34;memorize&#34; information, i.e., encode training data in their weights in such a way that inference-time queries can lead to verbatim regurgitation of that data. This ability to extract training data can be problematic, for example, when data are private or sensitive. In this work, we investigate methods to mitigate memorization: three regularizer-based, three finetuning-based, and eleven machine unlearning-based methods, with five of the latter being new methods that we introduce. We also introduce TinyMem, a suite of small, computationally-efficient LMs for the rapid development and evaluation of memorization-mitigation methods. We demonstrate that the mitigation methods that we develop using TinyMem can successfully be applied to production-grade LMs, and we determine via experiment that: regularizer-based mitigation methods are slow and ineffective at curbing memorization; fine-tuning-based methods are effective at curbing memorization, but overly expensive, especially for retaining higher accuracies; and unlearning-based methods are faster and more effective, allowing for the precise localization and removal of memorized information from LM weights prior to inference. We show, in particular, that our proposed unlearning method BalancedSubnet outperforms other mitigation methods at removing memorized information while preserving performance on target tasks. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2410.02159v1-abstract-full').style.display = 'none'; document.getElementById('2410.02159v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 2 October, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> October 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2409.16495">arXiv:2409.16495</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2409.16495">pdf</a>, <a href="https://arxiv.org/format/2409.16495">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Flight: A FaaS-Based Framework for Complex and Hierarchical Federated Learning </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Valerie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Baughman%2C+M">Matt Baughman</a>, <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2409.16495v1-abstract-short" style="display: inline;"> Federated Learning (FL) is a decentralized machine learning paradigm where models are trained on distributed devices and are aggregated at a central server. Existing FL frameworks assume simple two-tier network topologies where end devices are directly connected to the aggregation server. While this is a practical mental model, it does not exploit the inherent topology of real-world distributed sy&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2409.16495v1-abstract-full').style.display = 'inline'; document.getElementById('2409.16495v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2409.16495v1-abstract-full" style="display: none;"> Federated Learning (FL) is a decentralized machine learning paradigm where models are trained on distributed devices and are aggregated at a central server. Existing FL frameworks assume simple two-tier network topologies where end devices are directly connected to the aggregation server. While this is a practical mental model, it does not exploit the inherent topology of real-world distributed systems like the Internet-of-Things. We present Flight, a novel FL framework that supports complex hierarchical multi-tier topologies, asynchronous aggregation, and decouples the control plane from the data plane. We compare the performance of Flight against Flower, a state-of-the-art FL framework. Our results show that Flight scales beyond Flower, supporting up to 2048 simultaneous devices, and reduces FL makespan across several models. Finally, we show that Flight&#39;s hierarchical FL model can reduce communication overheads by more than 60%. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2409.16495v1-abstract-full').style.display = 'none'; document.getElementById('2409.16495v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 24 September, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> September 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2408.14434">arXiv:2408.14434</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2408.14434">pdf</a>, <a href="https://arxiv.org/format/2408.14434">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> </div> </div> <p class="title is-5 mathjax"> Employing Artificial Intelligence to Steer Exascale Workflows with Colmena </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Ward%2C+L">Logan Ward</a>, <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Valerie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Brace%2C+A">Alexander Brace</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Thakur%2C+R">Rajeev Thakur</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2408.14434v1-abstract-short" style="display: inline;"> Computational workflows are a common class of application on supercomputers, yet the loosely coupled and heterogeneous nature of workflows often fails to take full advantage of their capabilities. We created Colmena to leverage the massive parallelism of a supercomputer by using Artificial Intelligence (AI) to learn from and adapt a workflow as it executes. Colmena allows scientists to define how&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2408.14434v1-abstract-full').style.display = 'inline'; document.getElementById('2408.14434v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2408.14434v1-abstract-full" style="display: none;"> Computational workflows are a common class of application on supercomputers, yet the loosely coupled and heterogeneous nature of workflows often fails to take full advantage of their capabilities. We created Colmena to leverage the massive parallelism of a supercomputer by using Artificial Intelligence (AI) to learn from and adapt a workflow as it executes. Colmena allows scientists to define how their application should respond to events (e.g., task completion) as a series of cooperative agents. In this paper, we describe the design of Colmena, the challenges we overcame while deploying applications on exascale systems, and the science workflows we have enhanced through interweaving AI. The scaling challenges we discuss include developing steering strategies that maximize node utilization, introducing data fabrics that reduce communication overhead of data-intensive tasks, and implementing workflow tasks that cache costly operations between invocations. These innovations coupled with a variety of application patterns accessible through our agent-based steering model have enabled science advances in chemistry, biophysics, and materials science using different types of AI. Our vision is that Colmena will spur creative solutions that harness AI across many domains of scientific computing. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2408.14434v1-abstract-full').style.display = 'none'; document.getElementById('2408.14434v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 26 August, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> August 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2408.07236">arXiv:2408.07236</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2408.07236">pdf</a>, <a href="https://arxiv.org/format/2408.07236">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> TaPS: A Performance Evaluation Suite for Task-based Execution Frameworks </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Valerie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Gonthier%2C+M">Maxime Gonthier</a>, <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Pan%2C+H">Haochen Pan</a>, <a href="/search/cs?searchtype=author&amp;query=Zhou%2C+S">Sicheng Zhou</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2408.07236v1-abstract-short" style="display: inline;"> Task-based execution frameworks, such as parallel programming libraries, computational workflow systems, and function-as-a-service platforms, enable the composition of distinct tasks into a single, unified application designed to achieve a computational goal. Task-based execution frameworks abstract the parallel execution of an application&#39;s tasks on arbitrary hardware. Research into these task ex&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2408.07236v1-abstract-full').style.display = 'inline'; document.getElementById('2408.07236v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2408.07236v1-abstract-full" style="display: none;"> Task-based execution frameworks, such as parallel programming libraries, computational workflow systems, and function-as-a-service platforms, enable the composition of distinct tasks into a single, unified application designed to achieve a computational goal. Task-based execution frameworks abstract the parallel execution of an application&#39;s tasks on arbitrary hardware. Research into these task executors has accelerated as computational sciences increasingly need to take advantage of parallel compute and/or heterogeneous hardware. However, the lack of evaluation standards makes it challenging to compare and contrast novel systems against existing implementations. Here, we introduce TaPS, the Task Performance Suite, to support continued research in parallel task executor frameworks. TaPS provides (1) a unified, modular interface for writing and evaluating applications using arbitrary execution frameworks and data management systems and (2) an initial set of reference synthetic and real-world science applications. We discuss how the design of TaPS supports the reliable evaluation of frameworks and demonstrate TaPS through a survey of benchmarks using the provided reference applications. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2408.07236v1-abstract-full').style.display = 'none'; document.getElementById('2408.07236v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 13 August, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> August 2024. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">To appear in the Proceedings of 20th IEEE International Conference on e-Science</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2407.16646">arXiv:2407.16646</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2407.16646">pdf</a>, <a href="https://arxiv.org/format/2407.16646">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Software Engineering">cs.SE</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> ExaWorks Software Development Kit: A Robust and Scalable Collection of Interoperable Workflow Technologies </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Turilli%2C+M">Matteo Turilli</a>, <a href="/search/cs?searchtype=author&amp;query=Hategan-Marandiuc%2C+M">Mihael Hategan-Marandiuc</a>, <a href="/search/cs?searchtype=author&amp;query=Titov%2C+M">Mikhail Titov</a>, <a href="/search/cs?searchtype=author&amp;query=Maheshwari%2C+K">Ketan Maheshwari</a>, <a href="/search/cs?searchtype=author&amp;query=Alsaadi%2C+A">Aymen Alsaadi</a>, <a href="/search/cs?searchtype=author&amp;query=Merzky%2C+A">Andre Merzky</a>, <a href="/search/cs?searchtype=author&amp;query=Arambula%2C+R">Ramon Arambula</a>, <a href="/search/cs?searchtype=author&amp;query=Zakharchanka%2C+M">Mikhail Zakharchanka</a>, <a href="/search/cs?searchtype=author&amp;query=Cowan%2C+M">Matt Cowan</a>, <a href="/search/cs?searchtype=author&amp;query=Wozniak%2C+J+M">Justin M. Wozniak</a>, <a href="/search/cs?searchtype=author&amp;query=Wilke%2C+A">Andreas Wilke</a>, <a href="/search/cs?searchtype=author&amp;query=Kilic%2C+O+O">Ozgur Ozan Kilic</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=da+Silva%2C+R+F">Rafael Ferreira da Silva</a>, <a href="/search/cs?searchtype=author&amp;query=Jha%2C+S">Shantenu Jha</a>, <a href="/search/cs?searchtype=author&amp;query=Laney%2C+D">Daniel Laney</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2407.16646v1-abstract-short" style="display: inline;"> Scientific discovery increasingly requires executing heterogeneous scientific workflows on high-performance computing (HPC) platforms. Heterogeneous workflows contain different types of tasks (e.g., simulation, analysis, and learning) that need to be mapped, scheduled, and launched on different computing. That requires a software stack that enables users to code their workflows and automate resour&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2407.16646v1-abstract-full').style.display = 'inline'; document.getElementById('2407.16646v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2407.16646v1-abstract-full" style="display: none;"> Scientific discovery increasingly requires executing heterogeneous scientific workflows on high-performance computing (HPC) platforms. Heterogeneous workflows contain different types of tasks (e.g., simulation, analysis, and learning) that need to be mapped, scheduled, and launched on different computing. That requires a software stack that enables users to code their workflows and automate resource management and workflow execution. Currently, there are many workflow technologies with diverse levels of robustness and capabilities, and users face difficult choices of software that can effectively and efficiently support their use cases on HPC machines, especially when considering the latest exascale platforms. We contributed to addressing this issue by developing the ExaWorks Software Development Kit (SDK). The SDK is a curated collection of workflow technologies engineered following current best practices and specifically designed to work on HPC platforms. We present our experience with (1) curating those technologies, (2) integrating them to provide users with new capabilities, (3) developing a continuous integration platform to test the SDK on DOE HPC platforms, (4) designing a dashboard to publish the results of those tests, and (5) devising an innovative documentation platform to help users to use those technologies. Our experience details the requirements and the best practices needed to curate workflow technologies, and it also serves as a blueprint for the capabilities and services that DOE will have to offer to support a variety of scientific heterogeneous workflows on the newly available exascale HPC platforms. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2407.16646v1-abstract-full').style.display = 'none'; document.getElementById('2407.16646v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 23 July, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2407.11432">arXiv:2407.11432</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2407.11432">pdf</a>, <a href="https://arxiv.org/format/2407.11432">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Octopus: Experiences with a Hybrid Event-Driven Architecture for Distributed Scientific Computing </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Pan%2C+H">Haochen Pan</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Zhou%2C+S">Sicheng Zhou</a>, <a href="/search/cs?searchtype=author&amp;query=Kamatar%2C+A">Alok Kamatar</a>, <a href="/search/cs?searchtype=author&amp;query=Vescovi%2C+R">Rafael Vescovi</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Val茅rie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Bauer%2C+A">Andr茅 Bauer</a>, <a href="/search/cs?searchtype=author&amp;query=Gonthier%2C+M">Maxime Gonthier</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2407.11432v2-abstract-short" style="display: inline;"> Scientific research increasingly relies on distributed computational resources, storage systems, networks, and instruments, ranging from HPC and cloud systems to edge devices. Event-driven architecture (EDA) benefits applications targeting distributed research infrastructures by enabling the organization, communication, processing, reliability, and security of events generated from many sources. T&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2407.11432v2-abstract-full').style.display = 'inline'; document.getElementById('2407.11432v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2407.11432v2-abstract-full" style="display: none;"> Scientific research increasingly relies on distributed computational resources, storage systems, networks, and instruments, ranging from HPC and cloud systems to edge devices. Event-driven architecture (EDA) benefits applications targeting distributed research infrastructures by enabling the organization, communication, processing, reliability, and security of events generated from many sources. To support the development of scientific EDA, we introduce Octopus, a hybrid, cloud-to-edge event fabric designed to link many local event producers and consumers with cloud-hosted brokers. Octopus can be scaled to meet demand, permits the deployment of highly available Triggers for automatic event processing, and enforces fine-grained access control. We identify requirements in self-driving laboratories, scientific data automation, online task scheduling, epidemic modeling, and dynamic workflow management use cases, and present results demonstrating Octopus&#39; ability to meet those requirements. Octopus supports producing and consuming events at a rate of over 4.2 M and 9.6 M events per second, respectively, from distributed clients. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2407.11432v2-abstract-full').style.display = 'none'; document.getElementById('2407.11432v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 28 September, 2024; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 16 July, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2024. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">12 pages and 8 figures. Camera-ready version for FTXS&#39;24 (https://sites.google.com/view/ftxs2024)</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2407.01764">arXiv:2407.01764</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2407.01764">pdf</a>, <a href="https://arxiv.org/format/2407.01764">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Object Proxy Patterns for Accelerating Distributed Applications </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Valerie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Ward%2C+L">Logan Ward</a>, <a href="/search/cs?searchtype=author&amp;query=Brace%2C+A">Alexander Brace</a>, <a href="/search/cs?searchtype=author&amp;query=Bauer%2C+A">Andr茅 Bauer</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2407.01764v1-abstract-short" style="display: inline;"> Workflow and serverless frameworks have empowered new approaches to distributed application design by abstracting compute resources. However, their typically limited or one-size-fits-all support for advanced data flow patterns leaves optimization to the application programmer -- optimization that becomes more difficult as data become larger. The transparent object proxy, which provides wide-area r&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2407.01764v1-abstract-full').style.display = 'inline'; document.getElementById('2407.01764v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2407.01764v1-abstract-full" style="display: none;"> Workflow and serverless frameworks have empowered new approaches to distributed application design by abstracting compute resources. However, their typically limited or one-size-fits-all support for advanced data flow patterns leaves optimization to the application programmer -- optimization that becomes more difficult as data become larger. The transparent object proxy, which provides wide-area references that can resolve to data regardless of location, has been demonstrated as an effective low-level building block in such situations. Here we propose three high-level proxy-based programming patterns -- distributed futures, streaming, and ownership -- that make the power of the proxy pattern usable for more complex and dynamic distributed program structures. We motivate these patterns via careful review of application requirements and describe implementations of each pattern. We evaluate our implementations through a suite of benchmarks and by applying them in three substantial scientific applications, in which we demonstrate substantial improvements in runtime, throughput, and memory usage. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2407.01764v1-abstract-full').style.display = 'none'; document.getElementById('2407.01764v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 1 July, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2406.17710">arXiv:2406.17710</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2406.17710">pdf</a>, <a href="https://arxiv.org/format/2406.17710">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> GreenFaaS: Maximizing Energy Efficiency of HPC Workloads with FaaS </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Kamatar%2C+A">Alok Kamatar</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Valerie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Bauer%2C+A">Andre Bauer</a>, <a href="/search/cs?searchtype=author&amp;query=Rattihalli%2C+G">Gourav Rattihalli</a>, <a href="/search/cs?searchtype=author&amp;query=Hogade%2C+N">Ninad Hogade</a>, <a href="/search/cs?searchtype=author&amp;query=Milojicic%2C+D">Dejan Milojicic</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2406.17710v1-abstract-short" style="display: inline;"> Application energy efficiency can be improved by executing each application component on the compute element that consumes the least energy while also satisfying time constraints. In principle, the function as a service (FaaS) paradigm should simplify such optimizations by abstracting away compute location, but existing FaaS systems do not provide for user transparency over application energy cons&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2406.17710v1-abstract-full').style.display = 'inline'; document.getElementById('2406.17710v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2406.17710v1-abstract-full" style="display: none;"> Application energy efficiency can be improved by executing each application component on the compute element that consumes the least energy while also satisfying time constraints. In principle, the function as a service (FaaS) paradigm should simplify such optimizations by abstracting away compute location, but existing FaaS systems do not provide for user transparency over application energy consumption or task placement. Here we present GreenFaaS, a novel open source framework that bridges this gap between energy-efficient applications and FaaS platforms. GreenFaaS can be deployed by end users or providers across systems to monitor energy use, provide task-specific feedback, and schedule tasks in an energy-aware manner. We demonstrate that intelligent placement of tasks can both reduce energy consumption and improve performance. For a synthetic workload, GreenFaaS reduces the energy-delay product by 45% compared to alternatives. Furthermore, running a molecular design application through GreenFaaS can reduce energy consumption by 21% and runtime by 63% by better matching tasks with machines. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2406.17710v1-abstract-full').style.display = 'none'; document.getElementById('2406.17710v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 25 June, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> June 2024. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">11 pages, 10 figures</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2404.19717">arXiv:2404.19717</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2404.19717">pdf</a>, <a href="https://arxiv.org/format/2404.19717">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Automated, Reliable, and Efficient Continental-Scale Replication of 7.3 Petabytes of Climate Simulation Data: A Case Study </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Lacinski%2C+L">Lukasz Lacinski</a>, <a href="/search/cs?searchtype=author&amp;query=Liming%2C+L">Lee Liming</a>, <a href="/search/cs?searchtype=author&amp;query=Turoscy%2C+S">Steven Turoscy</a>, <a href="/search/cs?searchtype=author&amp;query=Harr%2C+C">Cameron Harr</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Dart%2C+E">Eli Dart</a>, <a href="/search/cs?searchtype=author&amp;query=Durack%2C+P">Paul Durack</a>, <a href="/search/cs?searchtype=author&amp;query=Ames%2C+S">Sasha Ames</a>, <a href="/search/cs?searchtype=author&amp;query=Hoffman%2C+F+M">Forrest M. Hoffman</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I+T">Ian T. Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2404.19717v1-abstract-short" style="display: inline;"> We report on our experiences replicating 7.3 petabytes (PB) of Earth System Grid Federation (ESGF) climate simulation data from Lawrence Livermore National Laboratory (LLNL) in California to Argonne National Laboratory (ANL) in Illinois and Oak Ridge National Laboratory (ORNL) in Tennessee. This movement of some 29 million files, twice, undertaken in order to establish new ESGF nodes at ANL and OR&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2404.19717v1-abstract-full').style.display = 'inline'; document.getElementById('2404.19717v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2404.19717v1-abstract-full" style="display: none;"> We report on our experiences replicating 7.3 petabytes (PB) of Earth System Grid Federation (ESGF) climate simulation data from Lawrence Livermore National Laboratory (LLNL) in California to Argonne National Laboratory (ANL) in Illinois and Oak Ridge National Laboratory (ORNL) in Tennessee. This movement of some 29 million files, twice, undertaken in order to establish new ESGF nodes at ANL and ORNL, was performed largely automatically by a simple replication tool, a script that invoked Globus to transfer large bundles of files while tracking progress in a database. Under the covers, Globus organized transfers to make efficient use of the high-speed Energy Sciences network (ESnet) and the data transfer nodes deployed at participating sites, and also addressed security, integrity checking, and recovery from a variety of transient failures. This success demonstrates the considerable benefits that can accrue from the adoption of performant data replication infrastructure. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2404.19717v1-abstract-full').style.display = 'none'; document.getElementById('2404.19717v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 30 April, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> April 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2404.02163">arXiv:2404.02163</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2404.02163">pdf</a>, <a href="https://arxiv.org/format/2404.02163">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Information Theory">cs.IT</span> </div> </div> <p class="title is-5 mathjax"> FastqZip: An Improved Reference-Based Genome Sequence Lossy Compression Framework </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Liu%2C+Y">Yuanjian Liu</a>, <a href="/search/cs?searchtype=author&amp;query=Luo%2C+H">Huihao Luo</a>, <a href="/search/cs?searchtype=author&amp;query=Han%2C+Z">Zhijun Han</a>, <a href="/search/cs?searchtype=author&amp;query=Hu%2C+Y">Yao Hu</a>, <a href="/search/cs?searchtype=author&amp;query=Yang%2C+Y">Yehui Yang</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Di%2C+S">Sheng Di</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Wu%2C+J">Jiesheng Wu</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2404.02163v1-abstract-short" style="display: inline;"> Storing and archiving data produced by next-generation sequencing (NGS) is a huge burden for research institutions. Reference-based compression algorithms are effective in dealing with these data. Our work focuses on compressing FASTQ format files with an improved reference-based compression algorithm to achieve a higher compression ratio than other state-of-the-art algorithms. We propose FastqZip&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2404.02163v1-abstract-full').style.display = 'inline'; document.getElementById('2404.02163v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2404.02163v1-abstract-full" style="display: none;"> Storing and archiving data produced by next-generation sequencing (NGS) is a huge burden for research institutions. Reference-based compression algorithms are effective in dealing with these data. Our work focuses on compressing FASTQ format files with an improved reference-based compression algorithm to achieve a higher compression ratio than other state-of-the-art algorithms. We propose FastqZip, which uses a new method mapping the sequence to reference for compression, allows reads-reordering and lossy quality scores, and the BSC or ZPAQ algorithm to perform final lossless compression for a higher compression ratio and relatively fast speed. Our method ensures the sequence can be losslessly reconstructed while allowing lossless or lossy compression for the quality scores. We reordered the reads to get a higher compression ratio. We evaluate our algorithms on five datasets and show that FastqZip can outperform the SOTA algorithm Genozip by around 10% in terms of compression ratio while having an acceptable slowdown. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2404.02163v1-abstract-full').style.display = 'none'; document.getElementById('2404.02163v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 22 February, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> April 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2403.19257">arXiv:2403.19257</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2403.19257">pdf</a>, <a href="https://arxiv.org/format/2403.19257">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1109/IPDPS57955.2024.00027">10.1109/IPDPS57955.2024.00027 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> UniFaaS: Programming across Distributed Cyberinfrastructure with Federated Function Serving </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Li%2C+Y">Yifei Li</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Li%2C+Z">Zhuozhao Li</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2403.19257v1-abstract-short" style="display: inline;"> Modern scientific applications are increasingly decomposable into individual functions that may be deployed across distributed and diverse cyberinfrastructure such as supercomputers, clouds, and accelerators. Such applications call for new approaches to programming, distributed execution, and function-level management. We present UniFaaS, a parallel programming framework that relies on a federated&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2403.19257v1-abstract-full').style.display = 'inline'; document.getElementById('2403.19257v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2403.19257v1-abstract-full" style="display: none;"> Modern scientific applications are increasingly decomposable into individual functions that may be deployed across distributed and diverse cyberinfrastructure such as supercomputers, clouds, and accelerators. Such applications call for new approaches to programming, distributed execution, and function-level management. We present UniFaaS, a parallel programming framework that relies on a federated function-as-a-service (FaaS) model to enable composition of distributed, scalable, and high-performance scientific workflows, and to support fine-grained function-level management. UniFaaS provides a unified programming interface to compose dynamic task graphs with transparent wide-area data management. UniFaaS exploits an observe-predict-decide approach to efficiently map workflow tasks to target heterogeneous and dynamic resources. We propose a dynamic heterogeneity-aware scheduling algorithm that employs a delay mechanism and a re-scheduling mechanism to accommodate dynamic resource capacity. Our experiments show that UniFaaS can efficiently execute workflows across computing resources with minimal scheduling overhead. We show that UniFaaS can improve the performance of a real-world drug screening workflow by as much as 22.99% when employing an additional 19.48% of resources and a montage workflow by 54.41% when employing an additional 47.83% of resources across multiple distributed clusters, in contrast to using a single cluster <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2403.19257v1-abstract-full').style.display = 'none'; document.getElementById('2403.19257v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 28 March, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> March 2024. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">13 pages, 13 figures, IPDPS2024</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2403.06077">arXiv:2403.06077</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2403.06077">pdf</a>, <a href="https://arxiv.org/format/2403.06077">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Steering a Fleet: Adaptation for Large-Scale, Workflow-Based Experiments </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Pruyne%2C+J">Jim Pruyne</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Valerie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Zheng%2C+W">Weijian Zheng</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Wozniak%2C+J+M">Justin M. Wozniak</a>, <a href="/search/cs?searchtype=author&amp;query=Bicer%2C+T">Tekin Bicer</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I+T">Ian T. Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2403.06077v1-abstract-short" style="display: inline;"> Experimental science is increasingly driven by instruments that produce vast volumes of data and thus a need to manage, compute, describe, and index this data. High performance and distributed computing provide the means of addressing the computing needs; however, in practice, the variety of actions required and the distributed set of resources involved, requires sophisticated &#34;flows&#34; defining the&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2403.06077v1-abstract-full').style.display = 'inline'; document.getElementById('2403.06077v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2403.06077v1-abstract-full" style="display: none;"> Experimental science is increasingly driven by instruments that produce vast volumes of data and thus a need to manage, compute, describe, and index this data. High performance and distributed computing provide the means of addressing the computing needs; however, in practice, the variety of actions required and the distributed set of resources involved, requires sophisticated &#34;flows&#34; defining the steps to be performed on data. As each scan or measurement is performed by an instrument, a new instance of the flow is initiated resulting in a &#34;fleet&#34; of concurrently running flows, with the overall goal to process all the data collected during a potentially long-running experiment. During the course of the experiment, each flow may need to adapt its execution due to changes in the environment, such as computational or storage resource availability, or based on the progress of the fleet as a whole such as completion or discovery of an intermediate result leading to a change in subsequent flow&#39;s behavior. We introduce a cloud-based decision engine, Braid, which flows consult during execution to query their run-time environment and coordinate with other flows within their fleet. Braid accepts streams of measurements taken from the run-time environment or from within flow runs which can then be statistically aggregated and compared to other streams to determine a strategy to guide flow execution. For example, queue lengths in execution environments can be used to direct a flow to run computations in one environment or another, or experiment progress as measured by individual flows can be aggregated to determine the progress and subsequent direction of the flows within a fleet. We describe Braid, its interface, implementation and performance characteristics. We further show through examples and experience modifying an existing scientific flow how Braid is used to make adaptable flows. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2403.06077v1-abstract-full').style.display = 'none'; document.getElementById('2403.06077v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 9 March, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> March 2024. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2402.14129">arXiv:2402.14129</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2402.14129">pdf</a>, <a href="https://arxiv.org/format/2402.14129">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Information Retrieval">cs.IR</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Computation and Language">cs.CL</span> </div> </div> <p class="title is-5 mathjax"> Combining Language and Graph Models for Semi-structured Information Extraction on the Web </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Hong%2C+Z">Zhi Hong</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2402.14129v1-abstract-short" style="display: inline;"> Relation extraction is an efficient way of mining the extraordinary wealth of human knowledge on the Web. Existing methods rely on domain-specific training data or produce noisy outputs. We focus here on extracting targeted relations from semi-structured web pages given only a short description of the relation. We present GraphScholarBERT, an open-domain information extraction method based on a jo&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2402.14129v1-abstract-full').style.display = 'inline'; document.getElementById('2402.14129v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2402.14129v1-abstract-full" style="display: none;"> Relation extraction is an efficient way of mining the extraordinary wealth of human knowledge on the Web. Existing methods rely on domain-specific training data or produce noisy outputs. We focus here on extracting targeted relations from semi-structured web pages given only a short description of the relation. We present GraphScholarBERT, an open-domain information extraction method based on a joint graph and language model structure. GraphScholarBERT can generalize to previously unseen domains without additional data or training and produces only clean extraction results matched to the search keyword. Experiments show that GraphScholarBERT can improve extraction F1 scores by as much as 34.8\% compared to previous work in a zero-shot domain and zero-shot website setting. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2402.14129v1-abstract-full').style.display = 'none'; document.getElementById('2402.14129v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 21 February, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> February 2024. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">7 pages, 2 figures</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2402.03480">arXiv:2402.03480</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2402.03480">pdf</a>, <a href="https://arxiv.org/format/2402.03480">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Artificial Intelligence">cs.AI</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Trillion Parameter AI Serving Infrastructure for Scientific Discovery: A Survey and Vision </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Baughman%2C+M">Matt Baughman</a>, <a href="/search/cs?searchtype=author&amp;query=Kamatar%2C+A">Alok Kamatar</a>, <a href="/search/cs?searchtype=author&amp;query=Sakarvadia%2C+M">Mansi Sakarvadia</a>, <a href="/search/cs?searchtype=author&amp;query=Ward%2C+L">Logan Ward</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Bauer%2C+A">Andr茅 Bauer</a>, <a href="/search/cs?searchtype=author&amp;query=Levental%2C+M">Maksim Levental</a>, <a href="/search/cs?searchtype=author&amp;query=Wang%2C+W">Wenyi Wang</a>, <a href="/search/cs?searchtype=author&amp;query=Engler%2C+W">Will Engler</a>, <a href="/search/cs?searchtype=author&amp;query=Skelly%2C+O+P">Owen Price Skelly</a>, <a href="/search/cs?searchtype=author&amp;query=Blaiszik%2C+B">Ben Blaiszik</a>, <a href="/search/cs?searchtype=author&amp;query=Stevens%2C+R">Rick Stevens</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2402.03480v1-abstract-short" style="display: inline;"> Deep learning methods are transforming research, enabling new techniques, and ultimately leading to new discoveries. As the demand for more capable AI models continues to grow, we are now entering an era of Trillion Parameter Models (TPM), or models with more than a trillion parameters -- such as Huawei&#39;s PanGu-$危$. We describe a vision for the ecosystem of TPM users and providers that caters to t&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2402.03480v1-abstract-full').style.display = 'inline'; document.getElementById('2402.03480v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2402.03480v1-abstract-full" style="display: none;"> Deep learning methods are transforming research, enabling new techniques, and ultimately leading to new discoveries. As the demand for more capable AI models continues to grow, we are now entering an era of Trillion Parameter Models (TPM), or models with more than a trillion parameters -- such as Huawei&#39;s PanGu-$危$. We describe a vision for the ecosystem of TPM users and providers that caters to the specific needs of the scientific community. We then outline the significant technical challenges and open problems in system design for serving TPMs to enable scientific research and discovery. Specifically, we describe the requirements of a comprehensive software stack and interfaces to support the diverse and flexible requirements of researchers. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2402.03480v1-abstract-full').style.display = 'none'; document.getElementById('2402.03480v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 5 February, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> February 2024. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">10 pages, 3 figures, accepted for publication in the proceedings of the 10th IEEE/ACM International Conference on Big Data Computing, Applications and Technologies (BDCAT2023)</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2401.02524">arXiv:2401.02524</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2401.02524">pdf</a>, <a href="https://arxiv.org/format/2401.02524">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Artificial Intelligence">cs.AI</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Computer Vision and Pattern Recognition">cs.CV</span> </div> </div> <p class="title is-5 mathjax"> Comprehensive Exploration of Synthetic Data Generation: A Survey </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Bauer%2C+A">Andr茅 Bauer</a>, <a href="/search/cs?searchtype=author&amp;query=Trapp%2C+S">Simon Trapp</a>, <a href="/search/cs?searchtype=author&amp;query=Stenger%2C+M">Michael Stenger</a>, <a href="/search/cs?searchtype=author&amp;query=Leppich%2C+R">Robert Leppich</a>, <a href="/search/cs?searchtype=author&amp;query=Kounev%2C+S">Samuel Kounev</a>, <a href="/search/cs?searchtype=author&amp;query=Leznik%2C+M">Mark Leznik</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2401.02524v2-abstract-short" style="display: inline;"> Recent years have witnessed a surge in the popularity of Machine Learning (ML), applied across diverse domains. However, progress is impeded by the scarcity of training data due to expensive acquisition and privacy legislation. Synthetic data emerges as a solution, but the abundance of released models and limited overview literature pose challenges for decision-making. This work surveys 417 Synthe&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2401.02524v2-abstract-full').style.display = 'inline'; document.getElementById('2401.02524v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2401.02524v2-abstract-full" style="display: none;"> Recent years have witnessed a surge in the popularity of Machine Learning (ML), applied across diverse domains. However, progress is impeded by the scarcity of training data due to expensive acquisition and privacy legislation. Synthetic data emerges as a solution, but the abundance of released models and limited overview literature pose challenges for decision-making. This work surveys 417 Synthetic Data Generation (SDG) models over the last decade, providing a comprehensive overview of model types, functionality, and improvements. Common attributes are identified, leading to a classification and trend analysis. The findings reveal increased model performance and complexity, with neural network-based approaches prevailing, except for privacy-preserving data generation. Computer vision dominates, with GANs as primary generative models, while diffusion models, transformers, and RNNs compete. Implications from our performance evaluation highlight the scarcity of common metrics and datasets, making comparisons challenging. Additionally, the neglect of training and computational costs in literature necessitates attention in future research. This work serves as a guide for SDG model selection and identifies crucial areas for future exploration. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2401.02524v2-abstract-full').style.display = 'none'; document.getElementById('2401.02524v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 1 February, 2024; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 4 January, 2024; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> January 2024. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">Fixed bug in Figure 44</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2310.16270">arXiv:2310.16270</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2310.16270">pdf</a>, <a href="https://arxiv.org/format/2310.16270">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Computation and Language">cs.CL</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Artificial Intelligence">cs.AI</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> </div> </div> <p class="title is-5 mathjax"> Attention Lens: A Tool for Mechanistically Interpreting the Attention Head Information Retrieval Mechanism </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Sakarvadia%2C+M">Mansi Sakarvadia</a>, <a href="/search/cs?searchtype=author&amp;query=Khan%2C+A">Arham Khan</a>, <a href="/search/cs?searchtype=author&amp;query=Ajith%2C+A">Aswathy Ajith</a>, <a href="/search/cs?searchtype=author&amp;query=Grzenda%2C+D">Daniel Grzenda</a>, <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Bauer%2C+A">Andr茅 Bauer</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2310.16270v1-abstract-short" style="display: inline;"> Transformer-based Large Language Models (LLMs) are the state-of-the-art for natural language tasks. Recent work has attempted to decode, by reverse engineering the role of linear layers, the internal mechanisms by which LLMs arrive at their final predictions for text completion tasks. Yet little is known about the specific role of attention heads in producing the final token prediction. We propose&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2310.16270v1-abstract-full').style.display = 'inline'; document.getElementById('2310.16270v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2310.16270v1-abstract-full" style="display: none;"> Transformer-based Large Language Models (LLMs) are the state-of-the-art for natural language tasks. Recent work has attempted to decode, by reverse engineering the role of linear layers, the internal mechanisms by which LLMs arrive at their final predictions for text completion tasks. Yet little is known about the specific role of attention heads in producing the final token prediction. We propose Attention Lens, a tool that enables researchers to translate the outputs of attention heads into vocabulary tokens via learned attention-head-specific transformations called lenses. Preliminary findings from our trained lenses indicate that attention heads play highly specialized roles in language models. The code for Attention Lens is available at github.com/msakarvadia/AttentionLens. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2310.16270v1-abstract-full').style.display = 'none'; document.getElementById('2310.16270v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 24 October, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> October 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2309.05605">arXiv:2309.05605</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2309.05605">pdf</a>, <a href="https://arxiv.org/format/2309.05605">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Computation and Language">cs.CL</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Artificial Intelligence">cs.AI</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> </div> </div> <p class="title is-5 mathjax"> Memory Injections: Correcting Multi-Hop Reasoning Failures during Inference in Transformer-Based Language Models </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Sakarvadia%2C+M">Mansi Sakarvadia</a>, <a href="/search/cs?searchtype=author&amp;query=Ajith%2C+A">Aswathy Ajith</a>, <a href="/search/cs?searchtype=author&amp;query=Khan%2C+A">Arham Khan</a>, <a href="/search/cs?searchtype=author&amp;query=Grzenda%2C+D">Daniel Grzenda</a>, <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Bauer%2C+A">Andr茅 Bauer</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2309.05605v3-abstract-short" style="display: inline;"> Answering multi-hop reasoning questions requires retrieving and synthesizing information from diverse sources. Large Language Models (LLMs) struggle to perform such reasoning consistently. Here we propose an approach to pinpoint and rectify multi-hop reasoning failures through targeted memory injections on LLM attention heads. First, we analyze the per-layer activations of GPT-2 models in response&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2309.05605v3-abstract-full').style.display = 'inline'; document.getElementById('2309.05605v3-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2309.05605v3-abstract-full" style="display: none;"> Answering multi-hop reasoning questions requires retrieving and synthesizing information from diverse sources. Large Language Models (LLMs) struggle to perform such reasoning consistently. Here we propose an approach to pinpoint and rectify multi-hop reasoning failures through targeted memory injections on LLM attention heads. First, we analyze the per-layer activations of GPT-2 models in response to single and multi-hop prompts. We then propose a mechanism that allows users to inject pertinent prompt-specific information, which we refer to as &#34;memories,&#34; at critical LLM locations during inference. By thus enabling the LLM to incorporate additional relevant information during inference, we enhance the quality of multi-hop prompt completions. We show empirically that a simple, efficient, and targeted memory injection into a key attention layer can often increase the probability of the desired next token in multi-hop tasks, by up to 424%. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2309.05605v3-abstract-full').style.display = 'none'; document.getElementById('2309.05605v3-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 28 February, 2024; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 11 September, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> September 2023. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">Oral Presentation at BlackboxNLP Workshop at EMNLP 2023</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2308.14658">arXiv:2308.14658</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2308.14658">pdf</a>, <a href="https://arxiv.org/format/2308.14658">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Adversarial Predictions of Data Distributions Across Federated Internet-of-Things Devices </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Rajani%2C+S">Samir Rajani</a>, <a href="/search/cs?searchtype=author&amp;query=Dematties%2C+D">Dario Dematties</a>, <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Ferrier%2C+N">Nicola Ferrier</a>, <a href="/search/cs?searchtype=author&amp;query=Sankaran%2C+R">Rajesh Sankaran</a>, <a href="/search/cs?searchtype=author&amp;query=Beckman%2C+P">Peter Beckman</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2308.14658v1-abstract-short" style="display: inline;"> Federated learning (FL) is increasingly becoming the default approach for training machine learning models across decentralized Internet-of-Things (IoT) devices. A key advantage of FL is that no raw data are communicated across the network, providing an immediate layer of privacy. Despite this, recent works have demonstrated that data reconstruction can be done with the locally trained model updat&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2308.14658v1-abstract-full').style.display = 'inline'; document.getElementById('2308.14658v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2308.14658v1-abstract-full" style="display: none;"> Federated learning (FL) is increasingly becoming the default approach for training machine learning models across decentralized Internet-of-Things (IoT) devices. A key advantage of FL is that no raw data are communicated across the network, providing an immediate layer of privacy. Despite this, recent works have demonstrated that data reconstruction can be done with the locally trained model updates which are communicated across the network. However, many of these works have limitations with regard to how the gradients are computed in backpropagation. In this work, we demonstrate that the model weights shared in FL can expose revealing information about the local data distributions of IoT devices. This leakage could expose sensitive information to malicious actors in a distributed system. We further discuss results which show that injecting noise into model weights is ineffective at preventing data leakage without seriously harming the global model accuracy. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2308.14658v1-abstract-full').style.display = 'none'; document.getElementById('2308.14658v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 28 August, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> August 2023. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">6 pages, 6 figures, accepted for publication through 2023 IEEE World Forum on Internet of Things</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2308.09793">arXiv:2308.09793</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2308.09793">pdf</a>, <a href="https://arxiv.org/format/2308.09793">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Robotics">cs.RO</span> </div> </div> <p class="title is-5 mathjax"> Towards a Modular Architecture for Science Factories </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Vescovi%2C+R">Rafael Vescovi</a>, <a href="/search/cs?searchtype=author&amp;query=Ginsburg%2C+T">Tobias Ginsburg</a>, <a href="/search/cs?searchtype=author&amp;query=Hippe%2C+K">Kyle Hippe</a>, <a href="/search/cs?searchtype=author&amp;query=Ozgulbas%2C+D">Doga Ozgulbas</a>, <a href="/search/cs?searchtype=author&amp;query=Stone%2C+C">Casey Stone</a>, <a href="/search/cs?searchtype=author&amp;query=Stroka%2C+A">Abraham Stroka</a>, <a href="/search/cs?searchtype=author&amp;query=Butler%2C+R">Rory Butler</a>, <a href="/search/cs?searchtype=author&amp;query=Blaiszik%2C+B">Ben Blaiszik</a>, <a href="/search/cs?searchtype=author&amp;query=Brettin%2C+T">Tom Brettin</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Hereld%2C+M">Mark Hereld</a>, <a href="/search/cs?searchtype=author&amp;query=Ramanathan%2C+A">Arvind Ramanathan</a>, <a href="/search/cs?searchtype=author&amp;query=Stevens%2C+R">Rick Stevens</a>, <a href="/search/cs?searchtype=author&amp;query=Vriza%2C+A">Aikaterini Vriza</a>, <a href="/search/cs?searchtype=author&amp;query=Xu%2C+J">Jie Xu</a>, <a href="/search/cs?searchtype=author&amp;query=Zhang%2C+Q">Qingteng Zhang</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2308.09793v2-abstract-short" style="display: inline;"> Advances in robotic automation, high-performance computing (HPC), and artificial intelligence (AI) encourage us to conceive of science factories: large, general-purpose computation- and AI-enabled self-driving laboratories (SDLs) with the generality and scale needed both to tackle large discovery problems and to support thousands of scientists. Science factories require modular hardware and softwa&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2308.09793v2-abstract-full').style.display = 'inline'; document.getElementById('2308.09793v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2308.09793v2-abstract-full" style="display: none;"> Advances in robotic automation, high-performance computing (HPC), and artificial intelligence (AI) encourage us to conceive of science factories: large, general-purpose computation- and AI-enabled self-driving laboratories (SDLs) with the generality and scale needed both to tackle large discovery problems and to support thousands of scientists. Science factories require modular hardware and software that can be replicated for scale and (re)configured to support many applications. To this end, we propose a prototype modular science factory architecture in which reconfigurable modules encapsulating scientific instruments are linked with manipulators to form workcells, that can themselves be combined to form larger assemblages, and linked with distributed computing for simulation, AI model training and inference, and related tasks. Workflows that perform sets of actions on modules can be specified, and various applications, comprising workflows plus associated computational and data manipulation steps, can be run concurrently. We report on our experiences prototyping this architecture and applying it in experiments involving 15 different robotic apparatus, five applications (one in education, two in biology, two in materials), and a variety of workflows, across four laboratories. We describe the reuse of modules, workcells, and workflows in different applications, the migration of applications between workcells, and the use of digital twins, and suggest directions for future work aimed at yet more generality and scalability. Code and data are available at https://ad-sdl.github.io/wei2023 and in the Supplementary Information <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2308.09793v2-abstract-full').style.display = 'none'; document.getElementById('2308.09793v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 17 October, 2023; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 18 August, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> August 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2308.04602">arXiv:2308.04602</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2308.04602">pdf</a>, <a href="https://arxiv.org/format/2308.04602">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Computers and Society">cs.CY</span> </div> </div> <p class="title is-5 mathjax"> NSF RESUME HPC Workshop: High-Performance Computing and Large-Scale Data Management in Service of Epidemiological Modeling </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Stevens%2C+A">Abby Stevens</a>, <a href="/search/cs?searchtype=author&amp;query=Ozik%2C+J">Jonathan Ozik</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Gerardin%2C+J">Jaline Gerardin</a>, <a href="/search/cs?searchtype=author&amp;query=Wozniak%2C+J+M">Justin M. Wozniak</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2308.04602v1-abstract-short" style="display: inline;"> The NSF-funded Robust Epidemic Surveillance and Modeling (RESUME) project successfully convened a workshop entitled &#34;High-performance computing and large-scale data management in service of epidemiological modeling&#34; at the University of Chicago on May 1-2, 2023. This was part of a series of workshops designed to foster sustainable and interdisciplinary co-design for predictive intelligence and pan&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2308.04602v1-abstract-full').style.display = 'inline'; document.getElementById('2308.04602v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2308.04602v1-abstract-full" style="display: none;"> The NSF-funded Robust Epidemic Surveillance and Modeling (RESUME) project successfully convened a workshop entitled &#34;High-performance computing and large-scale data management in service of epidemiological modeling&#34; at the University of Chicago on May 1-2, 2023. This was part of a series of workshops designed to foster sustainable and interdisciplinary co-design for predictive intelligence and pandemic prevention. The event brought together 31 experts in epidemiological modeling, high-performance computing (HPC), HPC workflows, and large-scale data management to develop a shared vision for capabilities needed for computational epidemiology to better support pandemic prevention. Through the workshop, participants identified key areas in which HPC capabilities could be used to improve epidemiological modeling, particularly in supporting public health decision-making, with an emphasis on HPC workflows, data integration, and HPC access. The workshop explored nascent HPC workflow and large-scale data management approaches currently in use for epidemiological modeling and sought to draw from approaches used in other domains to determine which practices could be best adapted for use in epidemiological modeling. This report documents the key findings and takeaways from the workshop. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2308.04602v1-abstract-full').style.display = 'none'; document.getElementById('2308.04602v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 8 August, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> August 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2307.16080">arXiv:2307.16080</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2307.16080">pdf</a>, <a href="https://arxiv.org/format/2307.16080">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Programming Languages">cs.PL</span> </div> </div> <p class="title is-5 mathjax"> nelli: a lightweight frontend for MLIR </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Levental%2C+M">Maksim Levental</a>, <a href="/search/cs?searchtype=author&amp;query=Kamatar%2C+A">Alok Kamatar</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2307.16080v2-abstract-short" style="display: inline;"> Multi-Level Intermediate Representation (MLIR) is a novel compiler infrastructure that aims to provide modular and extensible components to facilitate building domain specific compilers. However, since MLIR models programs at an intermediate level of abstraction, and most extant frontends are at a very high level of abstraction, the semantics and mechanics of the fundamental transformations availa&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2307.16080v2-abstract-full').style.display = 'inline'; document.getElementById('2307.16080v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2307.16080v2-abstract-full" style="display: none;"> Multi-Level Intermediate Representation (MLIR) is a novel compiler infrastructure that aims to provide modular and extensible components to facilitate building domain specific compilers. However, since MLIR models programs at an intermediate level of abstraction, and most extant frontends are at a very high level of abstraction, the semantics and mechanics of the fundamental transformations available in MLIR are difficult to investigate and employ in and of themselves. To address these challenges, we have developed \texttt{nelli}, a lightweight, Python-embedded, domain-specific, language for generating MLIR code. \texttt{nelli} leverages existing MLIR infrastructure to develop Pythonic syntax and semantics for various MLIR features. We describe \texttt{nelli}&#39;s design goals, discuss key details of our implementation, and demonstrate how \texttt{nelli} enables easily defining and lowering compute kernels to diverse hardware platforms. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2307.16080v2-abstract-full').style.display = 'none'; document.getElementById('2307.16080v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 14 August, 2023; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 29 July, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2307.11060">arXiv:2307.11060</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2307.11060">pdf</a>, <a href="https://arxiv.org/ps/2307.11060">ps</a>, <a href="https://arxiv.org/format/2307.11060">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Software Engineering">cs.SE</span> </div> </div> <p class="title is-5 mathjax"> The Changing Role of RSEs over the Lifetime of Parsl </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Katz%2C+D+S">Daniel S. Katz</a>, <a href="/search/cs?searchtype=author&amp;query=Clifford%2C+B">Ben Clifford</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Kesling%2C+K+H">Kevin Hunter Kesling</a>, <a href="/search/cs?searchtype=author&amp;query=Woodard%2C+A">Anna Woodard</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2307.11060v2-abstract-short" style="display: inline;"> This position paper describes the Parsl open source research software project and its various phases over seven years. It defines four types of research software engineers (RSEs) who have been important to the project in those phases; we believe this is also applicable to other research software projects. </span> <span class="abstract-full has-text-grey-dark mathjax" id="2307.11060v2-abstract-full" style="display: none;"> This position paper describes the Parsl open source research software project and its various phases over seven years. It defines four types of research software engineers (RSEs) who have been important to the project in those phases; we believe this is also applicable to other research software projects. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2307.11060v2-abstract-full').style.display = 'none'; document.getElementById('2307.11060v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 20 July, 2023; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 20 July, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2023. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">3 pages</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2307.07895">arXiv:2307.07895</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2307.07895">pdf</a>, <a href="https://arxiv.org/format/2307.07895">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1109/e-Science58273.2023.10254912">10.1109/e-Science58273.2023.10254912 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> PSI/J: A Portable Interface for Submitting, Monitoring, and Managing Jobs </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Hategan-Marandiuc%2C+M">Mihael Hategan-Marandiuc</a>, <a href="/search/cs?searchtype=author&amp;query=Merzky%2C+A">Andre Merzky</a>, <a href="/search/cs?searchtype=author&amp;query=Collier%2C+N">Nicholson Collier</a>, <a href="/search/cs?searchtype=author&amp;query=Maheshwari%2C+K">Ketan Maheshwari</a>, <a href="/search/cs?searchtype=author&amp;query=Ozik%2C+J">Jonathan Ozik</a>, <a href="/search/cs?searchtype=author&amp;query=Turilli%2C+M">Matteo Turilli</a>, <a href="/search/cs?searchtype=author&amp;query=Wilke%2C+A">Andreas Wilke</a>, <a href="/search/cs?searchtype=author&amp;query=Wozniak%2C+J+M">Justin M. Wozniak</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=da+Silva%2C+R+F">Rafael Ferreira da Silva</a>, <a href="/search/cs?searchtype=author&amp;query=Jha%2C+S">Shantenu Jha</a>, <a href="/search/cs?searchtype=author&amp;query=Laney%2C+D">Daniel Laney</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2307.07895v2-abstract-short" style="display: inline;"> It is generally desirable for high-performance computing (HPC) applications to be portable between HPC systems, for example to make use of more performant hardware, make effective use of allocations, and to co-locate compute jobs with large datasets. Unfortunately, moving scientific applications between HPC systems is challenging for various reasons, most notably that HPC systems have different HP&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2307.07895v2-abstract-full').style.display = 'inline'; document.getElementById('2307.07895v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2307.07895v2-abstract-full" style="display: none;"> It is generally desirable for high-performance computing (HPC) applications to be portable between HPC systems, for example to make use of more performant hardware, make effective use of allocations, and to co-locate compute jobs with large datasets. Unfortunately, moving scientific applications between HPC systems is challenging for various reasons, most notably that HPC systems have different HPC schedulers. We introduce PSI/J, a job management abstraction API intended to simplify the construction of software components and applications that are portable over various HPC scheduler implementations. We argue that such a system is both necessary and that no viable alternative currently exists. We analyze similar notable APIs and attempt to determine the factors that influenced their evolution and adoption by the HPC community. We base the design of PSI/J on that analysis. We describe how PSI/J has been integrated in three workflow systems and one application, and also show via experiments that PSI/J imposes minimal overhead. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2307.07895v2-abstract-full').style.display = 'none'; document.getElementById('2307.07895v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 20 September, 2023; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 15 July, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2307.05416">arXiv:2307.05416</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2307.05416">pdf</a>, <a href="https://arxiv.org/format/2307.05416">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Databases">cs.DB</span> </div> </div> <p class="title is-5 mathjax"> Optimizing Scientific Data Transfer on Globus with Error-bounded Lossy Compression </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Liu%2C+Y">Yuanjian Liu</a>, <a href="/search/cs?searchtype=author&amp;query=Di%2C+S">Sheng Di</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Cappello%2C+F">Franck Cappello</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2307.05416v1-abstract-short" style="display: inline;"> The increasing volume and velocity of science data necessitate the frequent movement of enormous data volumes as part of routine research activities. As a result, limited wide-area bandwidth often leads to bottlenecks in research progress. However, in many cases, consuming applications (e.g., for analysis, visualization, and machine learning) can achieve acceptable performance on reduced-precision&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2307.05416v1-abstract-full').style.display = 'inline'; document.getElementById('2307.05416v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2307.05416v1-abstract-full" style="display: none;"> The increasing volume and velocity of science data necessitate the frequent movement of enormous data volumes as part of routine research activities. As a result, limited wide-area bandwidth often leads to bottlenecks in research progress. However, in many cases, consuming applications (e.g., for analysis, visualization, and machine learning) can achieve acceptable performance on reduced-precision data, and thus researchers may wish to compromise on data precision to reduce transfer and storage costs. Error-bounded lossy compression presents a promising approach as it can significantly reduce data volumes while preserving data integrity based on user-specified error bounds. In this paper, we propose a novel data transfer framework called Ocelot that integrates error-bounded lossy compression into the Globus data transfer infrastructure. We note four key contributions: (1) Ocelot is the first integration of lossy compression in Globus to significantly improve scientific data transfer performance over wide area network (WAN). (2) We propose an effective machine-learning based lossy compression quality estimation model that can predict the quality of error-bounded lossy compressors, which is fundamental to ensure that transferred data are acceptable to users. (3) We develop optimized strategies to reduce the compression time overhead, counter the compute-node waiting time, and improve transfer speed for compressed files. (4) We perform evaluations using many real-world scientific applications across different domains and distributed Globus endpoints. Our experiments show that Ocelot can improve dataset transfer performance substantially, and the quality of lossy compression (time, ratio and data distortion) can be predicted accurately for the purpose of quality assurance. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2307.05416v1-abstract-full').style.display = 'none'; document.getElementById('2307.05416v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 11 July, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2305.09593">arXiv:2305.09593</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2305.09593">pdf</a>, <a href="https://arxiv.org/format/2305.09593">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Accelerating Communications in Federated Applications with Transparent Object Proxies </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Valerie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Ward%2C+L">Logan Ward</a>, <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Sabino%2C+C">Charlie Sabino</a>, <a href="/search/cs?searchtype=author&amp;query=Baughman%2C+M">Matt Baughman</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2305.09593v3-abstract-short" style="display: inline;"> Advances in networks, accelerators, and cloud services encourage programmers to reconsider where to compute -- such as when fast networks make it cost-effective to compute on remote accelerators despite added latency. Workflow and cloud-hosted serverless computing frameworks can manage multi-step computations spanning federated collections of cloud, high-performance computing (HPC), and edge syste&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2305.09593v3-abstract-full').style.display = 'inline'; document.getElementById('2305.09593v3-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2305.09593v3-abstract-full" style="display: none;"> Advances in networks, accelerators, and cloud services encourage programmers to reconsider where to compute -- such as when fast networks make it cost-effective to compute on remote accelerators despite added latency. Workflow and cloud-hosted serverless computing frameworks can manage multi-step computations spanning federated collections of cloud, high-performance computing (HPC), and edge systems, but passing data among computational steps via cloud storage can incur high costs. Here, we overcome this obstacle with a new programming paradigm that decouples control flow from data flow by extending the pass-by-reference model to distributed applications. We describe ProxyStore, a system that implements this paradigm by providing object proxies that act as wide-area object references with just-in-time resolution. This proxy model enables data producers to communicate data unilaterally, transparently, and efficiently to both local and remote consumers. We demonstrate the benefits of this model with synthetic benchmarks and real-world scientific applications, running across various computing platforms. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2305.09593v3-abstract-full').style.display = 'none'; document.getElementById('2305.09593v3-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 29 August, 2023; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 16 May, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> May 2023. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">Accepted for publication at the International Conference for High Performance Computing, Networking, Storage and Analysis (SC23)</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2305.03842">arXiv:2305.03842</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2305.03842">pdf</a>, <a href="https://arxiv.org/format/2305.03842">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Databases">cs.DB</span> </div> </div> <p class="title is-5 mathjax"> Data Station: Delegated, Trustworthy, and Auditable Computation to Enable Data-Sharing Consortia with a Data Escrow </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Xia%2C+S">Siyuan Xia</a>, <a href="/search/cs?searchtype=author&amp;query=Zhu%2C+Z">Zhiru Zhu</a>, <a href="/search/cs?searchtype=author&amp;query=Zhu%2C+C">Chris Zhu</a>, <a href="/search/cs?searchtype=author&amp;query=Zhao%2C+J">Jinjin Zhao</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Elmore%2C+A+J">Aaron J. Elmore</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Franklin%2C+M">Michael Franklin</a>, <a href="/search/cs?searchtype=author&amp;query=Krishnan%2C+S">Sanjay Krishnan</a>, <a href="/search/cs?searchtype=author&amp;query=Fernandez%2C+R+C">Raul Castro Fernandez</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2305.03842v1-abstract-short" style="display: inline;"> Pooling and sharing data increases and distributes its value. But since data cannot be revoked once shared, scenarios that require controlled release of data for regulatory, privacy, and legal reasons default to not sharing. Because selectively controlling what data to release is difficult, the few data-sharing consortia that exist are often built around data-sharing agreements resulting from long&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2305.03842v1-abstract-full').style.display = 'inline'; document.getElementById('2305.03842v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2305.03842v1-abstract-full" style="display: none;"> Pooling and sharing data increases and distributes its value. But since data cannot be revoked once shared, scenarios that require controlled release of data for regulatory, privacy, and legal reasons default to not sharing. Because selectively controlling what data to release is difficult, the few data-sharing consortia that exist are often built around data-sharing agreements resulting from long and tedious one-off negotiations. We introduce Data Station, a data escrow designed to enable the formation of data-sharing consortia. Data owners share data with the escrow knowing it will not be released without their consent. Data users delegate their computation to the escrow. The data escrow relies on delegated computation to execute queries without releasing the data first. Data Station leverages hardware enclaves to generate trust among participants, and exploits the centralization of data and computation to generate an audit log. We evaluate Data Station on machine learning and data-sharing applications while running on an untrusted intermediary. In addition to important qualitative advantages, we show that Data Station: i) outperforms federated learning baselines in accuracy and runtime for the machine learning application; ii) is orders of magnitude faster than alternative secure data-sharing frameworks; and iii) introduces small overhead on the critical path. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2305.03842v1-abstract-full').style.display = 'none'; document.getElementById('2305.03842v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 5 May, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> May 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2304.14982">arXiv:2304.14982</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2304.14982">pdf</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Hierarchical and Decentralised Federated Learning </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Rana%2C+O">Omer Rana</a>, <a href="/search/cs?searchtype=author&amp;query=Spyridopoulos%2C+T">Theodoros Spyridopoulos</a>, <a href="/search/cs?searchtype=author&amp;query=Hudson%2C+N">Nathaniel Hudson</a>, <a href="/search/cs?searchtype=author&amp;query=Baughman%2C+M">Matt Baughman</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Khan%2C+A">Aftab Khan</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2304.14982v1-abstract-short" style="display: inline;"> Federated learning has shown enormous promise as a way of training ML models in distributed environments while reducing communication costs and protecting data privacy. However, the rise of complex cyber-physical systems, such as the Internet-of-Things, presents new challenges that are not met with traditional FL methods. Hierarchical Federated Learning extends the traditional FL process to enable&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2304.14982v1-abstract-full').style.display = 'inline'; document.getElementById('2304.14982v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2304.14982v1-abstract-full" style="display: none;"> Federated learning has shown enormous promise as a way of training ML models in distributed environments while reducing communication costs and protecting data privacy. However, the rise of complex cyber-physical systems, such as the Internet-of-Things, presents new challenges that are not met with traditional FL methods. Hierarchical Federated Learning extends the traditional FL process to enable more efficient model aggregation based on application needs or characteristics of the deployment environment (e.g., resource capabilities and/or network connectivity). It illustrates the benefits of balancing processing across the cloud-edge continuum. Hierarchical Federated Learning is likely to be a key enabler for a wide range of applications, such as smart farming and smart energy management, as it can improve performance and reduce costs, whilst also enabling FL workflows to be deployed in environments that are not well-suited to traditional FL. Model aggregation algorithms, software frameworks, and infrastructures will need to be designed and implemented to make such solutions accessible to researchers and engineers across a growing set of domains. H-FL also introduces a number of new challenges. For instance, there are implicit infrastructural challenges. There is also a trade-off between having generalised models and personalised models. If there exist geographical patterns for data (e.g., soil conditions in a smart farm likely are related to the geography of the region itself), then it is crucial that models used locally can consider their own locality in addition to a globally-learned model. H-FL will be crucial to future FL solutions as it can aggregate and distribute models at multiple levels to optimally serve the trade-off between locality dependence and global anomaly robustness. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2304.14982v1-abstract-full').style.display = 'none'; document.getElementById('2304.14982v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 28 April, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> April 2023. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">11 pages, 6 figures, 25 references</span> </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">ACM Class:</span> C.2.4; I.2.11 </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2304.14244">arXiv:2304.14244</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2304.14244">pdf</a>, <a href="https://arxiv.org/format/2304.14244">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Developing Distributed High-performance Computing Capabilities of an Open Science Platform for Robust Epidemic Analysis </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Collier%2C+N">Nicholson Collier</a>, <a href="/search/cs?searchtype=author&amp;query=Wozniak%2C+J+M">Justin M. Wozniak</a>, <a href="/search/cs?searchtype=author&amp;query=Stevens%2C+A">Abby Stevens</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Binois%2C+M">Micka毛l Binois</a>, <a href="/search/cs?searchtype=author&amp;query=Fadikar%2C+A">Arindam Fadikar</a>, <a href="/search/cs?searchtype=author&amp;query=W%C3%BCrth%2C+A">Alexandra W眉rth</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Ozik%2C+J">Jonathan Ozik</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2304.14244v2-abstract-short" style="display: inline;"> COVID-19 had an unprecedented impact on scientific collaboration. The pandemic and its broad response from the scientific community has forged new relationships among domain experts, mathematical modelers, and scientific computing specialists. Computationally, however, it also revealed critical gaps in the ability of researchers to exploit advanced computing systems. These challenging areas includ&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2304.14244v2-abstract-full').style.display = 'inline'; document.getElementById('2304.14244v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2304.14244v2-abstract-full" style="display: none;"> COVID-19 had an unprecedented impact on scientific collaboration. The pandemic and its broad response from the scientific community has forged new relationships among domain experts, mathematical modelers, and scientific computing specialists. Computationally, however, it also revealed critical gaps in the ability of researchers to exploit advanced computing systems. These challenging areas include gaining access to scalable computing systems, porting models and workflows to new systems, sharing data of varying sizes, and producing results that can be reproduced and validated by others. Informed by our team&#39;s work in supporting public health decision makers during the COVID-19 pandemic and by the identified capability gaps in applying high-performance computing (HPC) to the modeling of complex social systems, we present the goals, requirements, and initial implementation of OSPREY, an open science platform for robust epidemic analysis. The prototype implementation demonstrates an integrated, algorithm-driven HPC workflow architecture, coordinating tasks across federated HPC resources, with robust, secure and automated access to each of the resources. We demonstrate scalable and fault-tolerant task execution, an asynchronous API to support fast time-to-solution algorithms, an inclusive, multi-language approach, and efficient wide-area data management. The example OSPREY code is made available on a public repository. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2304.14244v2-abstract-full').style.display = 'none'; document.getElementById('2304.14244v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 10 May, 2023; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 27 April, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> April 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2304.00019">arXiv:2304.00019</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2304.00019">pdf</a>, <a href="https://arxiv.org/format/2304.00019">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.5281/zenodo.7750670">10.5281/zenodo.7750670 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> Workflows Community Summit 2022: A Roadmap Revolution </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=da+Silva%2C+R+F">Rafael Ferreira da Silva</a>, <a href="/search/cs?searchtype=author&amp;query=Badia%2C+R+M">Rosa M. Badia</a>, <a href="/search/cs?searchtype=author&amp;query=Bala%2C+V">Venkat Bala</a>, <a href="/search/cs?searchtype=author&amp;query=Bard%2C+D">Debbie Bard</a>, <a href="/search/cs?searchtype=author&amp;query=Bremer%2C+P">Peer-Timo Bremer</a>, <a href="/search/cs?searchtype=author&amp;query=Buckley%2C+I">Ian Buckley</a>, <a href="/search/cs?searchtype=author&amp;query=Caino-Lores%2C+S">Silvina Caino-Lores</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Goble%2C+C">Carole Goble</a>, <a href="/search/cs?searchtype=author&amp;query=Jha%2C+S">Shantenu Jha</a>, <a href="/search/cs?searchtype=author&amp;query=Katz%2C+D+S">Daniel S. Katz</a>, <a href="/search/cs?searchtype=author&amp;query=Laney%2C+D">Daniel Laney</a>, <a href="/search/cs?searchtype=author&amp;query=Parashar%2C+M">Manish Parashar</a>, <a href="/search/cs?searchtype=author&amp;query=Suter%2C+F">Frederic Suter</a>, <a href="/search/cs?searchtype=author&amp;query=Tyler%2C+N">Nick Tyler</a>, <a href="/search/cs?searchtype=author&amp;query=Uram%2C+T">Thomas Uram</a>, <a href="/search/cs?searchtype=author&amp;query=Altintas%2C+I">Ilkay Altintas</a>, <a href="/search/cs?searchtype=author&amp;query=Andersson%2C+S">Stefan Andersson</a>, <a href="/search/cs?searchtype=author&amp;query=Arndt%2C+W">William Arndt</a>, <a href="/search/cs?searchtype=author&amp;query=Aznar%2C+J">Juan Aznar</a>, <a href="/search/cs?searchtype=author&amp;query=Bader%2C+J">Jonathan Bader</a>, <a href="/search/cs?searchtype=author&amp;query=Balis%2C+B">Bartosz Balis</a>, <a href="/search/cs?searchtype=author&amp;query=Blanton%2C+C">Chris Blanton</a>, <a href="/search/cs?searchtype=author&amp;query=Braghetto%2C+K+R">Kelly Rosa Braghetto</a>, <a href="/search/cs?searchtype=author&amp;query=Brodutch%2C+A">Aharon Brodutch</a> , et al. (80 additional authors not shown) </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2304.00019v1-abstract-short" style="display: inline;"> Scientific workflows have become integral tools in broad scientific computing use cases. Science discovery is increasingly dependent on workflows to orchestrate large and complex scientific experiments that range from execution of a cloud-based data preprocessing pipeline to multi-facility instrument-to-edge-to-HPC computational workflows. Given the changing landscape of scientific computing and t&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2304.00019v1-abstract-full').style.display = 'inline'; document.getElementById('2304.00019v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2304.00019v1-abstract-full" style="display: none;"> Scientific workflows have become integral tools in broad scientific computing use cases. Science discovery is increasingly dependent on workflows to orchestrate large and complex scientific experiments that range from execution of a cloud-based data preprocessing pipeline to multi-facility instrument-to-edge-to-HPC computational workflows. Given the changing landscape of scientific computing and the evolving needs of emerging scientific applications, it is paramount that the development of novel scientific workflows and system functionalities seek to increase the efficiency, resilience, and pervasiveness of existing systems and applications. Specifically, the proliferation of machine learning/artificial intelligence (ML/AI) workflows, need for processing large scale datasets produced by instruments at the edge, intensification of near real-time data processing, support for long-term experiment campaigns, and emergence of quantum computing as an adjunct to HPC, have significantly changed the functional and operational requirements of workflow systems. Workflow systems now need to, for example, support data streams from the edge-to-cloud-to-HPC enable the management of many small-sized files, allow data reduction while ensuring high accuracy, orchestrate distributed services (workflows, instruments, data movement, provenance, publication, etc.) across computing and user facilities, among others. Further, to accelerate science, it is also necessary that these systems implement specifications/standards and APIs for seamless (horizontal and vertical) integration between systems and applications, as well as enabling the publication of workflows and their associated products according to the FAIR principles. This document reports on discussions and findings from the 2022 international edition of the Workflows Community Summit that took place on November 29 and 30, 2022. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2304.00019v1-abstract-full').style.display = 'none'; document.getElementById('2304.00019v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 31 March, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> April 2023. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Report number:</span> ORNL/TM-2023/2885 </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2303.08803">arXiv:2303.08803</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2303.08803">pdf</a>, <a href="https://arxiv.org/format/2303.08803">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Artificial Intelligence">cs.AI</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1109/IPDPSW59300.2023.00018">10.1109/IPDPSW59300.2023.00018 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> Cloud Services Enable Efficient AI-Guided Simulation Workflows across Heterogeneous Resources </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Ward%2C+L">Logan Ward</a>, <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Hayot-Sasson%2C+V">Valerie Hayot-Sasson</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Sivaraman%2C+G">Ganesh Sivaraman</a>, <a href="/search/cs?searchtype=author&amp;query=Choudhury%2C+S">Sutanay Choudhury</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Thakur%2C+R">Rajeev Thakur</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2303.08803v1-abstract-short" style="display: inline;"> Applications that fuse machine learning and simulation can benefit from the use of multiple computing resources, with, for example, simulation codes running on highly parallel supercomputers and AI training and inference tasks on specialized accelerators. Here, we present our experiences deploying two AI-guided simulation workflows across such heterogeneous systems. A unique aspect of our approach&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2303.08803v1-abstract-full').style.display = 'inline'; document.getElementById('2303.08803v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2303.08803v1-abstract-full" style="display: none;"> Applications that fuse machine learning and simulation can benefit from the use of multiple computing resources, with, for example, simulation codes running on highly parallel supercomputers and AI training and inference tasks on specialized accelerators. Here, we present our experiences deploying two AI-guided simulation workflows across such heterogeneous systems. A unique aspect of our approach is our use of cloud-hosted management services to manage challenging aspects of cross-resource authentication and authorization, function-as-a-service (FaaS) function invocation, and data transfer. We show that these methods can achieve performance parity with systems that rely on direct connection between resources. We achieve parity by integrating the FaaS system and data transfer capabilities with a system that passes data by reference among managers and workers, and a user-configurable steering algorithm to hide data transfer latencies. We anticipate that this ease of use can enable routine use of heterogeneous resources in computational science. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2303.08803v1-abstract-full').style.display = 'none'; document.getElementById('2303.08803v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 15 March, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> March 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2302.06751">arXiv:2302.06751</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2302.06751">pdf</a>, <a href="https://arxiv.org/format/2302.06751">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Hardware Architecture">cs.AR</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> </div> </div> <p class="title is-5 mathjax"> OpenHLS: High-Level Synthesis for Low-Latency Deep Neural Networks for Experimental Science </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Levental%2C+M">Maksim Levental</a>, <a href="/search/cs?searchtype=author&amp;query=Khan%2C+A">Arham Khan</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Yoshii%2C+K">Kazutomo Yoshii</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2302.06751v4-abstract-short" style="display: inline;"> In many experiment-driven scientific domains, such as high-energy physics, material science, and cosmology, high data rate experiments impose hard constraints on data acquisition systems: collected data must either be indiscriminately stored for post-processing and analysis, thereby necessitating large storage capacity, or accurately filtered in real-time, thereby necessitating low-latency process&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2302.06751v4-abstract-full').style.display = 'inline'; document.getElementById('2302.06751v4-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2302.06751v4-abstract-full" style="display: none;"> In many experiment-driven scientific domains, such as high-energy physics, material science, and cosmology, high data rate experiments impose hard constraints on data acquisition systems: collected data must either be indiscriminately stored for post-processing and analysis, thereby necessitating large storage capacity, or accurately filtered in real-time, thereby necessitating low-latency processing. Deep neural networks, effective in other filtering tasks, have not been widely employed in such data acquisition systems, due to design and deployment difficulties. We present an open source, lightweight, compiler framework, without any proprietary dependencies, OpenHLS, based on high-level synthesis techniques, for translating high-level representations of deep neural networks to low-level representations, suitable for deployment to near-sensor devices such as field-programmable gate arrays. We evaluate OpenHLS on various workloads and present a case-study implementation of a deep neural network for Bragg peak detection in the context of high-energy diffraction microscopy. We show OpenHLS is able to produce an implementation of the network with a throughput 4.8 $渭$s/sample, which is approximately a 4$\times$ improvement over the existing implementation <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2302.06751v4-abstract-full').style.display = 'none'; document.getElementById('2302.06751v4-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 15 March, 2023; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 13 February, 2023; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> February 2023. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2209.11631">arXiv:2209.11631</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2209.11631">pdf</a>, <a href="https://arxiv.org/format/2209.11631">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1109/TPDS.2022.3208767">10.1109/TPDS.2022.3208767 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> funcX: Federated Function as a Service for Science </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Li%2C+Z">Zhuozhao Li</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Galewsky%2C+B">Ben Galewsky</a>, <a href="/search/cs?searchtype=author&amp;query=Skluzacek%2C+T">Tyler Skluzacek</a>, <a href="/search/cs?searchtype=author&amp;query=Nagaitsev%2C+K">Kirill Nagaitsev</a>, <a href="/search/cs?searchtype=author&amp;query=Woodard%2C+A">Anna Woodard</a>, <a href="/search/cs?searchtype=author&amp;query=Blaiszik%2C+B">Ben Blaiszik</a>, <a href="/search/cs?searchtype=author&amp;query=Bryan%2C+J">Josh Bryan</a>, <a href="/search/cs?searchtype=author&amp;query=Katz%2C+D+S">Daniel S. Katz</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2209.11631v1-abstract-short" style="display: inline;"> funcX is a distributed function as a service (FaaS) platform that enables flexible, scalable, and high performance remote function execution. Unlike centralized FaaS systems, funcX decouples the cloud-hosted management functionality from the edge-hosted execution functionality. funcX&#39;s endpoint software can be deployed, by users or administrators, on arbitrary laptops, clouds, clusters, and superc&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2209.11631v1-abstract-full').style.display = 'inline'; document.getElementById('2209.11631v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2209.11631v1-abstract-full" style="display: none;"> funcX is a distributed function as a service (FaaS) platform that enables flexible, scalable, and high performance remote function execution. Unlike centralized FaaS systems, funcX decouples the cloud-hosted management functionality from the edge-hosted execution functionality. funcX&#39;s endpoint software can be deployed, by users or administrators, on arbitrary laptops, clouds, clusters, and supercomputers, in effect turning them into function serving systems. funcX&#39;s cloud-hosted service provides a single location for registering, sharing, and managing both functions and endpoints. It allows for transparent, secure, and reliable function execution across the federated ecosystem of endpoints--enabling users to route functions to endpoints based on specific needs. funcX uses containers (e.g., Docker, Singularity, and Shifter) to provide common execution environments across endpoints. funcX implements various container management strategies to execute functions with high performance and efficiency on diverse funcX endpoints. funcX also integrates with an in-memory data store and Globus for managing data that may span endpoints. We motivate the need for funcX, present our prototype design and implementation, and demonstrate, via experiments on two supercomputers, that funcX can scale to more than 130 000 concurrent workers. We show that funcX&#39;s container warming-aware routing algorithm can reduce the completion time for 3000 functions by up to 61% compared to a randomized algorithm and the in-memory data store can speed up data transfers by up to 3x compared to a shared file system. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2209.11631v1-abstract-full').style.display = 'none'; document.getElementById('2209.11631v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 23 September, 2022; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> September 2022. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">arXiv admin note: substantial text overlap with arXiv:2005.04215</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2208.09513">arXiv:2208.09513</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2208.09513">pdf</a>, <a href="https://arxiv.org/format/2208.09513">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Artificial Intelligence">cs.AI</span> </div> </div> <p class="title is-5 mathjax"> Globus Automation Services: Research process automation across the space-time continuum </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Pruyne%2C+J">Jim Pruyne</a>, <a href="/search/cs?searchtype=author&amp;query=McKee%2C+K">Kurt McKee</a>, <a href="/search/cs?searchtype=author&amp;query=Bryan%2C+J">Josh Bryan</a>, <a href="/search/cs?searchtype=author&amp;query=Raumann%2C+B">Brigitte Raumann</a>, <a href="/search/cs?searchtype=author&amp;query=Ananthakrishnan%2C+R">Rachana Ananthakrishnan</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2208.09513v2-abstract-short" style="display: inline;"> Research process automation -- the reliable, efficient, and reproducible execution of linked sets of actions on scientific instruments, computers, data stores, and other resources -- has emerged as an essential element of modern science. We report here on new services within the Globus research data management platform that enable the specification of diverse research processes as reusable sets of&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2208.09513v2-abstract-full').style.display = 'inline'; document.getElementById('2208.09513v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2208.09513v2-abstract-full" style="display: none;"> Research process automation -- the reliable, efficient, and reproducible execution of linked sets of actions on scientific instruments, computers, data stores, and other resources -- has emerged as an essential element of modern science. We report here on new services within the Globus research data management platform that enable the specification of diverse research processes as reusable sets of actions, \emph{flows}, and the execution of such flows in heterogeneous research environments. To support flows with broad spatial extent (e.g., from scientific instrument to remote data center) and temporal extent (from seconds to weeks), these Globus automation services feature: 1) cloud hosting for reliable execution of even long-lived flows despite sporadic failures; 2) a simple specification and extensible asynchronous action provider API, for defining and executing a wide variety of actions and flows involving heterogeneous resources; 3) an event-driven execution model for automating execution of flows in response to arbitrary events; and 4) a rich security model enabling authorization delegation mechanisms for secure execution of long-running actions across distributed resources. These services permit researchers to outsource and automate the management of a broad range of research tasks to a reliable, scalable, and secure cloud platform. We present use cases for Globus automation services, describe their design and implementation, present microbenchmark studies, and review experiences applying the services in a range of applications. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2208.09513v2-abstract-full').style.display = 'none'; document.getElementById('2208.09513v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 6 December, 2022; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 19 August, 2022; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> August 2022. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2207.00611">arXiv:2207.00611</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2207.00611">pdf</a>, <a href="https://arxiv.org/format/2207.00611">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Artificial Intelligence">cs.AI</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Materials Science">cond-mat.mtrl-sci</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1038/s41597-022-01712-9">10.1038/s41597-022-01712-9 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> FAIR principles for AI models with a practical application for accelerated high energy diffraction microscopy </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Ravi%2C+N">Nikil Ravi</a>, <a href="/search/cs?searchtype=author&amp;query=Chaturvedi%2C+P">Pranshu Chaturvedi</a>, <a href="/search/cs?searchtype=author&amp;query=Huerta%2C+E+A">E. A. Huerta</a>, <a href="/search/cs?searchtype=author&amp;query=Liu%2C+Z">Zhengchun Liu</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Scourtas%2C+A">Aristana Scourtas</a>, <a href="/search/cs?searchtype=author&amp;query=Schmidt%2C+K+J">K. J. Schmidt</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Blaiszik%2C+B">Ben Blaiszik</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2207.00611v3-abstract-short" style="display: inline;"> A concise and measurable set of FAIR (Findable, Accessible, Interoperable and Reusable) principles for scientific data is transforming the state-of-practice for data management and stewardship, supporting and enabling discovery and innovation. Learning from this initiative, and acknowledging the impact of artificial intelligence (AI) in the practice of science and engineering, we introduce a set o&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2207.00611v3-abstract-full').style.display = 'inline'; document.getElementById('2207.00611v3-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2207.00611v3-abstract-full" style="display: none;"> A concise and measurable set of FAIR (Findable, Accessible, Interoperable and Reusable) principles for scientific data is transforming the state-of-practice for data management and stewardship, supporting and enabling discovery and innovation. Learning from this initiative, and acknowledging the impact of artificial intelligence (AI) in the practice of science and engineering, we introduce a set of practical, concise, and measurable FAIR principles for AI models. We showcase how to create and share FAIR data and AI models within a unified computational framework combining the following elements: the Advanced Photon Source at Argonne National Laboratory, the Materials Data Facility, the Data and Learning Hub for Science, and funcX, and the Argonne Leadership Computing Facility (ALCF), in particular the ThetaGPU supercomputer and the SambaNova DataScale system at the ALCF AI Testbed. We describe how this domain-agnostic computational framework may be harnessed to enable autonomous AI-driven discovery. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2207.00611v3-abstract-full').style.display = 'none'; document.getElementById('2207.00611v3-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 21 December, 2022; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 1 July, 2022; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2022. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">11 pages, 3 figures; Accepted to Scientific Data; for press release see https://www.anl.gov/article/argonne-scientists-promote-fair-standards-for-managing-artificial-intelligence-models and https://www.ncsa.illinois.edu/ncsa-student-researchers-lead-authors-on-award-winning-paper; Received 2022 HPCwire Readers&#39; Choice Award on Best Use of High Performance Data Analytics &amp; Artificial Intelligence</span> </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">MSC Class:</span> 68T01; 68T05 <span class="has-text-black-bis has-text-weight-semibold">ACM Class:</span> I.2; J.2 </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Journal ref:</span> Scientific Data 9, 657 (2022) </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2205.11342">arXiv:2205.11342</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2205.11342">pdf</a>, <a href="https://arxiv.org/format/2205.11342">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Computation and Language">cs.CL</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> </div> </div> <p class="title is-5 mathjax"> The Diminishing Returns of Masked Language Models to Science </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Hong%2C+Z">Zhi Hong</a>, <a href="/search/cs?searchtype=author&amp;query=Ajith%2C+A">Aswathy Ajith</a>, <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+G">Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Duede%2C+E">Eamon Duede</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2205.11342v2-abstract-short" style="display: inline;"> Transformer-based masked language models such as BERT, trained on general corpora, have shown impressive performance on downstream tasks. It has also been demonstrated that the downstream task performance of such models can be improved by pretraining larger models for longer on more data. In this work, we empirically evaluate the extent to which these results extend to tasks in science. We use 14&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2205.11342v2-abstract-full').style.display = 'inline'; document.getElementById('2205.11342v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2205.11342v2-abstract-full" style="display: none;"> Transformer-based masked language models such as BERT, trained on general corpora, have shown impressive performance on downstream tasks. It has also been demonstrated that the downstream task performance of such models can be improved by pretraining larger models for longer on more data. In this work, we empirically evaluate the extent to which these results extend to tasks in science. We use 14 domain-specific transformer-based models (including ScholarBERT, a new 770M-parameter science-focused masked language model pretrained on up to 225B tokens) to evaluate the impact of training data, model size, pretraining and finetuning time on 12 downstream scientific tasks. Interestingly, we find that increasing model sizes, training data, or compute time does not always lead to significant improvements (i.e., &gt;1% F1), if at all, in scientific information extraction tasks and offered possible explanations for the surprising performance differences. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2205.11342v2-abstract-full').style.display = 'none'; document.getElementById('2205.11342v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 3 May, 2023; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 23 May, 2022; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> May 2022. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">12 pages. 3 figures. 5 tables. Accepted to the Findings of ACL 2023</span> </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">ACM Class:</span> I.2.7 </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2205.01527">arXiv:2205.01527</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2205.01527">pdf</a>, <a href="https://arxiv.org/ps/2205.01527">ps</a>, <a href="https://arxiv.org/format/2205.01527">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1145/3463478.3463486">10.1145/3463478.3463486 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> Extended Abstract: Productive Parallel Programming with Parsl </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Woodard%2C+A">Anna Woodard</a>, <a href="/search/cs?searchtype=author&amp;query=Clifford%2C+B">Ben Clifford</a>, <a href="/search/cs?searchtype=author&amp;query=Li%2C+Z">Zhuozhao Li</a>, <a href="/search/cs?searchtype=author&amp;query=Hategan%2C+M">Mihael Hategan</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Wilde%2C+M">Mike Wilde</a>, <a href="/search/cs?searchtype=author&amp;query=Katz%2C+D+S">Daniel S. Katz</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2205.01527v2-abstract-short" style="display: inline;"> Parsl is a parallel programming library for Python that aims to make it easy to specify parallelism in programs and to realize that parallelism on arbitrary parallel and distributed computing systems. Parsl relies on developers annotating Python functions-wrapping either Python or external applications-to indicate that these functions may be executed concurrently. Developers can then link together&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2205.01527v2-abstract-full').style.display = 'inline'; document.getElementById('2205.01527v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2205.01527v2-abstract-full" style="display: none;"> Parsl is a parallel programming library for Python that aims to make it easy to specify parallelism in programs and to realize that parallelism on arbitrary parallel and distributed computing systems. Parsl relies on developers annotating Python functions-wrapping either Python or external applications-to indicate that these functions may be executed concurrently. Developers can then link together functions via the exchange of data. Parsl establishes a dynamic dependency graph and sends tasks for execution on connected resources when dependencies are resolved. Parsl&#39;s runtime system enables different compute resources to be used, from laptops to supercomputers, without modification to the Parsl program. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2205.01527v2-abstract-full').style.display = 'none'; document.getElementById('2205.01527v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 4 May, 2022; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 3 May, 2022; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> May 2022. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Journal ref:</span> ACM SIGAda Ada Letters 40 (2), 73-75, 2020 </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2204.05128">arXiv:2204.05128</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2204.05128">pdf</a>, <a href="https://arxiv.org/format/2204.05128">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> Linking Scientific Instruments and HPC: Patterns, Technologies, Experiences </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Vescovi%2C+R">Rafael Vescovi</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Saint%2C+N">Nickolaus Saint</a>, <a href="/search/cs?searchtype=author&amp;query=Blaiszik%2C+B">Ben Blaiszik</a>, <a href="/search/cs?searchtype=author&amp;query=Pruyne%2C+J">Jim Pruyne</a>, <a href="/search/cs?searchtype=author&amp;query=Bicer%2C+T">Tekin Bicer</a>, <a href="/search/cs?searchtype=author&amp;query=Lavens%2C+A">Alex Lavens</a>, <a href="/search/cs?searchtype=author&amp;query=Liu%2C+Z">Zhengchun Liu</a>, <a href="/search/cs?searchtype=author&amp;query=Papka%2C+M+E">Michael E. Papka</a>, <a href="/search/cs?searchtype=author&amp;query=Narayanan%2C+S">Suresh Narayanan</a>, <a href="/search/cs?searchtype=author&amp;query=Schwarz%2C+N">Nicholas Schwarz</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2204.05128v2-abstract-short" style="display: inline;"> Powerful detectors at modern experimental facilities routinely collect data at multiple GB/s. Online analysis methods are needed to enable the collection of only interesting subsets of such massive data streams, such as by explicitly discarding some data elements or by directing instruments to relevant areas of experimental space. Such online analyses require methods for configuring and running hi&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2204.05128v2-abstract-full').style.display = 'inline'; document.getElementById('2204.05128v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2204.05128v2-abstract-full" style="display: none;"> Powerful detectors at modern experimental facilities routinely collect data at multiple GB/s. Online analysis methods are needed to enable the collection of only interesting subsets of such massive data streams, such as by explicitly discarding some data elements or by directing instruments to relevant areas of experimental space. Such online analyses require methods for configuring and running high-performance distributed computing pipelines--what we call flows--linking instruments, HPC (e.g., for analysis, simulation, AI model training), edge computing (for analysis), data stores, metadata catalogs, and high-speed networks. In this article, we review common patterns associated with such flows and describe methods for instantiating those patterns. We also present experiences with the application of these methods to the processing of data from five different scientific instruments, each of which engages HPC resources for data inversion, machine learning model training, or other purposes. We also discuss implications of these new methods for operators and users of scientific facilities. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2204.05128v2-abstract-full').style.display = 'none'; document.getElementById('2204.05128v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 22 August, 2022; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 11 April, 2022; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> April 2022. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2201.07435">arXiv:2201.07435</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2201.07435">pdf</a>, <a href="https://arxiv.org/format/2201.07435">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.5281/zenodo.5815332">10.5281/zenodo.5815332 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> Workflows Community Summit: Tightening the Integration between Computing Facilities and Scientific Workflows </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=da+Silva%2C+R+F">Rafael Ferreira da Silva</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Casanova%2C+H">Henri Casanova</a>, <a href="/search/cs?searchtype=author&amp;query=Laney%2C+D">Dan Laney</a>, <a href="/search/cs?searchtype=author&amp;query=Ahn%2C+D">Dong Ahn</a>, <a href="/search/cs?searchtype=author&amp;query=Jha%2C+S">Shantenu Jha</a>, <a href="/search/cs?searchtype=author&amp;query=Allcock%2C+W+E">William E. Allcock</a>, <a href="/search/cs?searchtype=author&amp;query=Bauer%2C+G">Gregory Bauer</a>, <a href="/search/cs?searchtype=author&amp;query=Duplyakin%2C+D">Dmitry Duplyakin</a>, <a href="/search/cs?searchtype=author&amp;query=Enders%2C+B">Bjoern Enders</a>, <a href="/search/cs?searchtype=author&amp;query=Heer%2C+T+M">Todd M. Heer</a>, <a href="/search/cs?searchtype=author&amp;query=Lancon%2C+E">Eric Lancon</a>, <a href="/search/cs?searchtype=author&amp;query=Sanielevici%2C+S">Sergiu Sanielevici</a>, <a href="/search/cs?searchtype=author&amp;query=Sayers%2C+K">Kevin Sayers</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2201.07435v1-abstract-short" style="display: inline;"> The importance of workflows is highlighted by the fact that they have underpinned some of the most significant discoveries of the past decades. Many of these workflows have significant computational, storage, and communication demands, and thus must execute on a range of large-scale computer systems, from local clusters to public clouds and upcoming exascale HPC platforms. Historically, infrastruc&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2201.07435v1-abstract-full').style.display = 'inline'; document.getElementById('2201.07435v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2201.07435v1-abstract-full" style="display: none;"> The importance of workflows is highlighted by the fact that they have underpinned some of the most significant discoveries of the past decades. Many of these workflows have significant computational, storage, and communication demands, and thus must execute on a range of large-scale computer systems, from local clusters to public clouds and upcoming exascale HPC platforms. Historically, infrastructures for workflow execution consisted of complex, integrated systems, developed in-house by workflow practitioners with strong dependencies on a range of legacy technologies. Due to the increasing need to support workflows, dedicated workflow systems were developed to provide abstractions for creating, executing, and adapting workflows conveniently and efficiently while ensuring portability. While these efforts are all worthwhile individually, there are now hundreds of independent workflow systems. The resulting workflow system technology landscape is fragmented, which may present significant barriers for future workflow users due to many seemingly comparable, yet usually mutually incompatible, systems that exist. In order to tackle some of these challenges, the DOE-funded ExaWorks and NSF-funded WorkflowsRI projects have organized in 2021 a series of events entitled the &#34;Workflows Community Summit&#34;. The third edition of the ``Workflows Community Summit&#34; explored workflows challenges and opportunities from the perspective of computing centers and facilities. The third summit brought together a small group of facilities representatives with the aim to understand how workflows are currently being used at each facility, how facilities would like to interact with workflow developers and users, how workflows fit with facility roadmaps, and what opportunities there are for tighter integration between facilities and workflows. More information at: https://workflowsri.org/summits/facilities/ <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2201.07435v1-abstract-full').style.display = 'none'; document.getElementById('2201.07435v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 19 January, 2022; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> January 2022. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">arXiv admin note: text overlap with arXiv:2110.02168</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2110.02827">arXiv:2110.02827</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2110.02827">pdf</a>, <a href="https://arxiv.org/format/2110.02827">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Materials Science">cond-mat.mtrl-sci</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1109/MLHPC54614.2021.00007">10.1109/MLHPC54614.2021.00007 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> Colmena: Scalable Machine-Learning-Based Steering of Ensemble Simulations for High Performance Computing </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Ward%2C+L">Logan Ward</a>, <a href="/search/cs?searchtype=author&amp;query=Sivaraman%2C+G">Ganesh Sivaraman</a>, <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Dandu%2C+N">Naveen Dandu</a>, <a href="/search/cs?searchtype=author&amp;query=Redfern%2C+P+C">Paul C. Redfern</a>, <a href="/search/cs?searchtype=author&amp;query=Assary%2C+R+S">Rajeev S. Assary</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Curtiss%2C+L+A">Larry A. Curtiss</a>, <a href="/search/cs?searchtype=author&amp;query=Thakur%2C+R">Rajeev Thakur</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2110.02827v1-abstract-short" style="display: inline;"> Scientific applications that involve simulation ensembles can be accelerated greatly by using experiment design methods to select the best simulations to perform. Methods that use machine learning (ML) to create proxy models of simulations show particular promise for guiding ensembles but are challenging to deploy because of the need to coordinate dynamic mixes of simulation and learning tasks. We&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2110.02827v1-abstract-full').style.display = 'inline'; document.getElementById('2110.02827v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2110.02827v1-abstract-full" style="display: none;"> Scientific applications that involve simulation ensembles can be accelerated greatly by using experiment design methods to select the best simulations to perform. Methods that use machine learning (ML) to create proxy models of simulations show particular promise for guiding ensembles but are challenging to deploy because of the need to coordinate dynamic mixes of simulation and learning tasks. We present Colmena, an open-source Python framework that allows users to steer campaigns by providing just the implementations of individual tasks plus the logic used to choose which tasks to execute when. Colmena handles task dispatch, results collation, ML model invocation, and ML model (re)training, using Parsl to execute tasks on HPC systems. We describe the design of Colmena and illustrate its capabilities by applying it to electrolyte design, where it both scales to 65536 CPUs and accelerates the discovery rate for high-performance molecules by a factor of 100 over unguided searches. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2110.02827v1-abstract-full').style.display = 'none'; document.getElementById('2110.02827v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 6 October, 2021; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> October 2021. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">camera-ready version for ML in HPC Environments 2021</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2110.02168">arXiv:2110.02168</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2110.02168">pdf</a>, <a href="https://arxiv.org/ps/2110.02168">ps</a>, <a href="https://arxiv.org/format/2110.02168">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1109/WORKS54523.2021.00016">10.1109/WORKS54523.2021.00016 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> A Community Roadmap for Scientific Workflows Research and Development </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=da+Silva%2C+R+F">Rafael Ferreira da Silva</a>, <a href="/search/cs?searchtype=author&amp;query=Casanova%2C+H">Henri Casanova</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Altintas%2C+I">Ilkay Altintas</a>, <a href="/search/cs?searchtype=author&amp;query=Badia%2C+R+M">Rosa M Badia</a>, <a href="/search/cs?searchtype=author&amp;query=Balis%2C+B">Bartosz Balis</a>, <a href="/search/cs?searchtype=author&amp;query=Coleman%2C+T">Tain茫 Coleman</a>, <a href="/search/cs?searchtype=author&amp;query=Coppens%2C+F">Frederik Coppens</a>, <a href="/search/cs?searchtype=author&amp;query=Di+Natale%2C+F">Frank Di Natale</a>, <a href="/search/cs?searchtype=author&amp;query=Enders%2C+B">Bjoern Enders</a>, <a href="/search/cs?searchtype=author&amp;query=Fahringer%2C+T">Thomas Fahringer</a>, <a href="/search/cs?searchtype=author&amp;query=Filgueira%2C+R">Rosa Filgueira</a>, <a href="/search/cs?searchtype=author&amp;query=Fursin%2C+G">Grigori Fursin</a>, <a href="/search/cs?searchtype=author&amp;query=Garijo%2C+D">Daniel Garijo</a>, <a href="/search/cs?searchtype=author&amp;query=Goble%2C+C">Carole Goble</a>, <a href="/search/cs?searchtype=author&amp;query=Howell%2C+D">Dorran Howell</a>, <a href="/search/cs?searchtype=author&amp;query=Jha%2C+S">Shantenu Jha</a>, <a href="/search/cs?searchtype=author&amp;query=Katz%2C+D+S">Daniel S. Katz</a>, <a href="/search/cs?searchtype=author&amp;query=Laney%2C+D">Daniel Laney</a>, <a href="/search/cs?searchtype=author&amp;query=Leser%2C+U">Ulf Leser</a>, <a href="/search/cs?searchtype=author&amp;query=Malawski%2C+M">Maciej Malawski</a>, <a href="/search/cs?searchtype=author&amp;query=Mehta%2C+K">Kshitij Mehta</a>, <a href="/search/cs?searchtype=author&amp;query=Pottier%2C+L">Lo茂c Pottier</a>, <a href="/search/cs?searchtype=author&amp;query=Ozik%2C+J">Jonathan Ozik</a>, <a href="/search/cs?searchtype=author&amp;query=Peterson%2C+J+L">J. Luc Peterson</a> , et al. (4 additional authors not shown) </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2110.02168v2-abstract-short" style="display: inline;"> The landscape of workflow systems for scientific applications is notoriously convoluted with hundreds of seemingly equivalent workflow systems, many isolated research claims, and a steep learning curve. To address some of these challenges and lay the groundwork for transforming workflows research and development, the WorkflowsRI and ExaWorks projects partnered to bring the international workflows&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2110.02168v2-abstract-full').style.display = 'inline'; document.getElementById('2110.02168v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2110.02168v2-abstract-full" style="display: none;"> The landscape of workflow systems for scientific applications is notoriously convoluted with hundreds of seemingly equivalent workflow systems, many isolated research claims, and a steep learning curve. To address some of these challenges and lay the groundwork for transforming workflows research and development, the WorkflowsRI and ExaWorks projects partnered to bring the international workflows community together. This paper reports on discussions and findings from two virtual &#34;Workflows Community Summits&#34; (January and April, 2021). The overarching goals of these workshops were to develop a view of the state of the art, identify crucial research challenges in the workflows community, articulate a vision for potential community efforts, and discuss technical approaches for realizing this vision. To this end, participants identified six broad themes: FAIR computational workflows; AI workflows; exascale challenges; APIs, interoperability, reuse, and standards; training and education; and building a workflows community. We summarize discussions and recommendations for each of these themes. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2110.02168v2-abstract-full').style.display = 'none'; document.getElementById('2110.02168v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 8 October, 2021; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 5 October, 2021; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> October 2021. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">arXiv admin note: substantial text overlap with arXiv:2103.09181</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2109.12060">arXiv:2109.12060</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2109.12060">pdf</a>, <a href="https://arxiv.org/format/2109.12060">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Instrumentation and Methods for Astrophysics">astro-ph.IM</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1109/eScience51609.2021.00031">10.1109/eScience51609.2021.00031 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> Extreme Scale Survey Simulation with Python Workflows </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Villarreal%2C+A+S">A. S. Villarreal</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Uram%2C+T">Tom Uram</a>, <a href="/search/cs?searchtype=author&amp;query=Katz%2C+D+S">Daniel S. Katz</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Heitmann%2C+K">Katrin Heitmann</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2109.12060v1-abstract-short" style="display: inline;"> The Vera C. Rubin Observatory Legacy Survey of Space and Time (LSST) will soon carry out an unprecedented wide, fast, and deep survey of the sky in multiple optical bands. The data from LSST will open up a new discovery space in astronomy and cosmology, simultaneously providing clues toward addressing burning issues of the day, such as the origin of dark energy and and the nature of dark matter, w&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2109.12060v1-abstract-full').style.display = 'inline'; document.getElementById('2109.12060v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2109.12060v1-abstract-full" style="display: none;"> The Vera C. Rubin Observatory Legacy Survey of Space and Time (LSST) will soon carry out an unprecedented wide, fast, and deep survey of the sky in multiple optical bands. The data from LSST will open up a new discovery space in astronomy and cosmology, simultaneously providing clues toward addressing burning issues of the day, such as the origin of dark energy and and the nature of dark matter, while at the same time yielding data that will, in turn, pose fresh new questions. To prepare for the imminent arrival of this remarkable data set, it is crucial that the associated scientific communities be able to develop the software needed to analyze it. Computational power now available allows us to generate synthetic data sets that can be used as a realistic training ground for such an effort. This effort raises its own challenges -- the need to generate very large simulations of the night sky, scaling up simulation campaigns to large numbers of compute nodes across multiple computing centers with different architectures, and optimizing the complex workload around memory requirements and widely varying wall clock times. We describe here a large-scale workflow that melds together Python code to steer the workflow, Parsl to manage the large-scale distributed execution of workflow components, and containers to carry out the image simulation campaign across multiple sites. Taking advantage of these tools, we developed an extreme-scale computational framework and used it to simulate five years of observations for 300 square degrees of sky area. We describe our experiences and lessons learned in developing this workflow capability, and highlight how the scalability and portability of our approach enabled us to efficiently execute it on up to 4000 compute nodes on two supercomputers. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2109.12060v1-abstract-full').style.display = 'none'; document.getElementById('2109.12060v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 24 September, 2021; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> September 2021. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">Proceeding for eScience 2021, 9 pages, 5 figures</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2108.13521">arXiv:2108.13521</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2108.13521">pdf</a>, <a href="https://arxiv.org/format/2108.13521">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> </div> <p class="title is-5 mathjax"> ExaWorks: Workflows for Exascale </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Al-Saadi%2C+A">Aymen Al-Saadi</a>, <a href="/search/cs?searchtype=author&amp;query=Ahn%2C+D+H">Dong H. Ahn</a>, <a href="/search/cs?searchtype=author&amp;query=Babuji%2C+Y">Yadu Babuji</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Corbett%2C+J">James Corbett</a>, <a href="/search/cs?searchtype=author&amp;query=Hategan%2C+M">Mihael Hategan</a>, <a href="/search/cs?searchtype=author&amp;query=Herbein%2C+S">Stephen Herbein</a>, <a href="/search/cs?searchtype=author&amp;query=Jha%2C+S">Shantenu Jha</a>, <a href="/search/cs?searchtype=author&amp;query=Laney%2C+D">Daniel Laney</a>, <a href="/search/cs?searchtype=author&amp;query=Merzky%2C+A">Andre Merzky</a>, <a href="/search/cs?searchtype=author&amp;query=Munson%2C+T">Todd Munson</a>, <a href="/search/cs?searchtype=author&amp;query=Salim%2C+M">Michael Salim</a>, <a href="/search/cs?searchtype=author&amp;query=Titov%2C+M">Mikhail Titov</a>, <a href="/search/cs?searchtype=author&amp;query=Turilli%2C+M">Matteo Turilli</a>, <a href="/search/cs?searchtype=author&amp;query=Wozniak%2C+J+M">Justin M. Wozniak</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2108.13521v1-abstract-short" style="display: inline;"> Exascale computers will offer transformative capabilities to combine data-driven and learning-based approaches with traditional simulation applications to accelerate scientific discovery and insight. These software combinations and integrations, however, are difficult to achieve due to challenges of coordination and deployment of heterogeneous software components on diverse and massive platforms.&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2108.13521v1-abstract-full').style.display = 'inline'; document.getElementById('2108.13521v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2108.13521v1-abstract-full" style="display: none;"> Exascale computers will offer transformative capabilities to combine data-driven and learning-based approaches with traditional simulation applications to accelerate scientific discovery and insight. These software combinations and integrations, however, are difficult to achieve due to challenges of coordination and deployment of heterogeneous software components on diverse and massive platforms. We present the ExaWorks project, which can address many of these challenges: ExaWorks is leading a co-design process to create a workflow software development Toolkit (SDK) consisting of a wide range of workflow management tools that can be composed and interoperate through common interfaces. We describe the initial set of tools and interfaces supported by the SDK, efforts to make them easier to apply to complex science challenges, and examples of their application to exemplar cases. Furthermore, we discuss how our project is working with the workflows community, large computing facilities as well as HPC platform vendors to sustainably address the requirements of workflows at the exascale. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2108.13521v1-abstract-full').style.display = 'none'; document.getElementById('2108.13521v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 30 August, 2021; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> August 2021. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2108.12050">arXiv:2108.12050</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2108.12050">pdf</a>, <a href="https://arxiv.org/format/2108.12050">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Image and Video Processing">eess.IV</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Computer Vision and Pattern Recognition">cs.CV</span> </div> </div> <p class="title is-5 mathjax"> Ultrafast Focus Detection for Automated Microscopy </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Levental%2C+M">Maksim Levental</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+R">Ryan Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Wildenberg%2C+G+A">Gregg A. Wildenberg</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2108.12050v3-abstract-short" style="display: inline;"> Technological advancements in modern scientific instruments, such as scanning electron microscopes (SEMs), have significantly increased data acquisition rates and image resolutions enabling new questions to be explored; however, the resulting data volumes and velocities, combined with automated experiments, are quickly overwhelming scientists as there remain crucial steps that require human interv&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2108.12050v3-abstract-full').style.display = 'inline'; document.getElementById('2108.12050v3-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2108.12050v3-abstract-full" style="display: none;"> Technological advancements in modern scientific instruments, such as scanning electron microscopes (SEMs), have significantly increased data acquisition rates and image resolutions enabling new questions to be explored; however, the resulting data volumes and velocities, combined with automated experiments, are quickly overwhelming scientists as there remain crucial steps that require human intervention, for example reviewing image focus. We present a fast out-of-focus detection algorithm for electron microscopy images collected serially and demonstrate that it can be used to provide near-real-time quality control for neuroscience workflows. Our technique, \textit{Multi-scale Histologic Feature Detection}, adapts classical computer vision techniques and is based on detecting various fine-grained histologic features. We exploit the inherent parallelism in the technique to employ GPU primitives in order to accelerate characterization. We show that our method can detect of out-of-focus conditions within just 20ms. To make these capabilities generally available, we deploy our feature detector as an on-demand service and show that it can be used to determine the degree of focus in approximately 230ms, enabling near-real-time use. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2108.12050v3-abstract-full').style.display = 'none'; document.getElementById('2108.12050v3-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 22 February, 2022; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 26 August, 2021; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> August 2021. </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2107.01739">arXiv:2107.01739</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2107.01739">pdf</a>, <a href="https://arxiv.org/format/2107.01739">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Machine Learning">cs.LG</span> <span class="tag is-small is-grey tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1145/3458817.3476152">10.1145/3458817.3476152 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> KAISA: An Adaptive Second-Order Optimizer Framework for Deep Neural Networks </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Pauloski%2C+J+G">J. Gregory Pauloski</a>, <a href="/search/cs?searchtype=author&amp;query=Huang%2C+Q">Qi Huang</a>, <a href="/search/cs?searchtype=author&amp;query=Huang%2C+L">Lei Huang</a>, <a href="/search/cs?searchtype=author&amp;query=Venkataraman%2C+S">Shivaram Venkataraman</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Foster%2C+I">Ian Foster</a>, <a href="/search/cs?searchtype=author&amp;query=Zhang%2C+Z">Zhao Zhang</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2107.01739v2-abstract-short" style="display: inline;"> Kronecker-factored Approximate Curvature (K-FAC) has recently been shown to converge faster in deep neural network (DNN) training than stochastic gradient descent (SGD); however, K-FAC&#39;s larger memory footprint hinders its applicability to large models. We present KAISA, a K-FAC-enabled, Adaptable, Improved, and ScAlable second-order optimizer framework that adapts the memory footprint, communicat&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2107.01739v2-abstract-full').style.display = 'inline'; document.getElementById('2107.01739v2-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2107.01739v2-abstract-full" style="display: none;"> Kronecker-factored Approximate Curvature (K-FAC) has recently been shown to converge faster in deep neural network (DNN) training than stochastic gradient descent (SGD); however, K-FAC&#39;s larger memory footprint hinders its applicability to large models. We present KAISA, a K-FAC-enabled, Adaptable, Improved, and ScAlable second-order optimizer framework that adapts the memory footprint, communication, and computation given specific models and hardware to improve performance and increase scalability. We quantify the tradeoffs between memory and communication cost and evaluate KAISA on large models, including ResNet-50, Mask R-CNN, U-Net, and BERT, on up to 128 NVIDIA A100 GPUs. Compared to the original optimizers, KAISA converges 18.1-36.3% faster across applications with the same global batch size. Under a fixed memory budget, KAISA converges 32.5% and 41.6% faster in ResNet-50 and BERT-Large, respectively. KAISA can balance memory and communication to achieve scaling efficiency equal to or better than the baseline optimizers. KAISA is open source and available at https://github.com/gpauloski/kfac_pytorch. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2107.01739v2-abstract-full').style.display = 'none'; document.getElementById('2107.01739v2-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 20 September, 2021; <span class="has-text-black-bis has-text-weight-semibold">v1</span> submitted 4 July, 2021; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2021. </p> <p class="comments is-size-7"> <span class="has-text-black-bis has-text-weight-semibold">Comments:</span> <span class="has-text-grey-dark mathjax">Accepted for publication at the International Conference for High Performance Computing, Networking, Storage and Analysis (SC21)</span> </p> </li> <li class="arxiv-result"> <div class="is-marginless"> <p class="list-title is-inline-block"><a href="https://arxiv.org/abs/2107.00417">arXiv:2107.00417</a> <span>&nbsp;[<a href="https://arxiv.org/pdf/2107.00417">pdf</a>, <a href="https://arxiv.org/ps/2107.00417">ps</a>, <a href="https://arxiv.org/format/2107.00417">other</a>]&nbsp;</span> </p> <div class="tags is-inline-block"> <span class="tag is-small is-link tooltip is-tooltip-top" data-tooltip="Distributed, Parallel, and Cluster Computing">cs.DC</span> </div> <div class="is-inline-block" style="margin-left: 0.5rem"> <div class="tags has-addons"> <span class="tag is-dark is-size-7">doi</span> <span class="tag is-light is-size-7"><a class="" href="https://doi.org/10.1145/3311790.3400848">10.1145/3311790.3400848 <i class="fa fa-external-link" aria-hidden="true"></i></a></span> </div> </div> </div> <p class="title is-5 mathjax"> Toward Interoperable Cyberinfrastructure: Common Descriptions for Computational Resources and Applications </p> <p class="authors"> <span class="search-hit">Authors:</span> <a href="/search/cs?searchtype=author&amp;query=Stubbs%2C+J">Joe Stubbs</a>, <a href="/search/cs?searchtype=author&amp;query=Marru%2C+S">Suresh Marru</a>, <a href="/search/cs?searchtype=author&amp;query=Mejia%2C+D">Daniel Mejia</a>, <a href="/search/cs?searchtype=author&amp;query=Katz%2C+D+S">Daniel S. Katz</a>, <a href="/search/cs?searchtype=author&amp;query=Chard%2C+K">Kyle Chard</a>, <a href="/search/cs?searchtype=author&amp;query=Dahan%2C+M">Maytal Dahan</a>, <a href="/search/cs?searchtype=author&amp;query=Pierce%2C+M">Marlon Pierce</a>, <a href="/search/cs?searchtype=author&amp;query=Zentner%2C+M">Michael Zentner</a> </p> <p class="abstract mathjax"> <span class="has-text-black-bis has-text-weight-semibold">Abstract</span>: <span class="abstract-short has-text-grey-dark mathjax" id="2107.00417v1-abstract-short" style="display: inline;"> The user-facing components of the Cyberinfrastructure (CI) ecosystem, science gateways and scientific workflow systems, share a common need of interfacing with physical resources (storage systems and execution environments) to manage data and execute codes (applications). However, there is no uniform, platform-independent way to describe either the resources or the applications. To address this, w&hellip; <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2107.00417v1-abstract-full').style.display = 'inline'; document.getElementById('2107.00417v1-abstract-short').style.display = 'none';">&#9661; More</a> </span> <span class="abstract-full has-text-grey-dark mathjax" id="2107.00417v1-abstract-full" style="display: none;"> The user-facing components of the Cyberinfrastructure (CI) ecosystem, science gateways and scientific workflow systems, share a common need of interfacing with physical resources (storage systems and execution environments) to manage data and execute codes (applications). However, there is no uniform, platform-independent way to describe either the resources or the applications. To address this, we propose uniform semantics for describing resources and applications that will be relevant to a diverse set of stakeholders. We sketch a solution to the problem of a common description and catalog of resources: we describe an approach to implementing a resource registry for use by the community and discuss potential approaches to some long-term challenges. We conclude by looking ahead to the application description language. <a class="is-size-7" style="white-space: nowrap;" onclick="document.getElementById('2107.00417v1-abstract-full').style.display = 'none'; document.getElementById('2107.00417v1-abstract-short').style.display = 'inline';">&#9651; Less</a> </span> </p> <p class="is-size-7"><span class="has-text-black-bis has-text-weight-semibold">Submitted</span> 1 July, 2021; <span class="has-text-black-bis has-text-weight-semibold">originally announced</span> July 2021. </p> </li> </ol> <nav class="pagination is-small is-centered breathe-horizontal" role="navigation" aria-label="pagination"> <a href="" class="pagination-previous is-invisible">Previous </a> <a href="/search/?searchtype=author&amp;query=Chard%2C+K&amp;start=50" class="pagination-next" >Next </a> <ul class="pagination-list"> <li> <a href="/search/?searchtype=author&amp;query=Chard%2C+K&amp;start=0" class="pagination-link is-current" aria-label="Goto page 1">1 </a> </li> <li> <a href="/search/?searchtype=author&amp;query=Chard%2C+K&amp;start=50" class="pagination-link " aria-label="Page 2" aria-current="page">2 </a> </li> </ul> </nav> <div class="is-hidden-tablet"> <!-- feedback for mobile only --> <span class="help" style="display: inline-block;"><a href="https://github.com/arXiv/arxiv-search/releases">Search v0.5.6 released 2020-02-24</a>&nbsp;&nbsp;</span> </div> </div> </main> <footer> <div class="columns is-desktop" role="navigation" aria-label="Secondary"> <!-- MetaColumn 1 --> <div class="column"> <div class="columns"> <div class="column"> <ul class="nav-spaced"> <li><a href="https://info.arxiv.org/about">About</a></li> <li><a href="https://info.arxiv.org/help">Help</a></li> </ul> </div> <div class="column"> <ul class="nav-spaced"> <li> <svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 512 512" class="icon filter-black" role="presentation"><title>contact arXiv</title><desc>Click here to contact arXiv</desc><path d="M502.3 190.8c3.9-3.1 9.7-.2 9.7 4.7V400c0 26.5-21.5 48-48 48H48c-26.5 0-48-21.5-48-48V195.6c0-5 5.7-7.8 9.7-4.7 22.4 17.4 52.1 39.5 154.1 113.6 21.1 15.4 56.7 47.8 92.2 47.6 35.7.3 72-32.8 92.3-47.6 102-74.1 131.6-96.3 154-113.7zM256 320c23.2.4 56.6-29.2 73.4-41.4 132.7-96.3 142.8-104.7 173.4-128.7 5.8-4.5 9.2-11.5 9.2-18.9v-19c0-26.5-21.5-48-48-48H48C21.5 64 0 85.5 0 112v19c0 7.4 3.4 14.3 9.2 18.9 30.6 23.9 40.7 32.4 173.4 128.7 16.8 12.2 50.2 41.8 73.4 41.4z"/></svg> <a href="https://info.arxiv.org/help/contact.html"> Contact</a> </li> <li> <svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 512 512" class="icon filter-black" role="presentation"><title>subscribe to arXiv mailings</title><desc>Click here to subscribe</desc><path d="M476 3.2L12.5 270.6c-18.1 10.4-15.8 35.6 2.2 43.2L121 358.4l287.3-253.2c5.5-4.9 13.3 2.6 8.6 8.3L176 407v80.5c0 23.6 28.5 32.9 42.5 15.8L282 426l124.6 52.2c14.2 6 30.4-2.9 33-18.2l72-432C515 7.8 493.3-6.8 476 3.2z"/></svg> <a href="https://info.arxiv.org/help/subscribe"> Subscribe</a> </li> </ul> </div> </div> </div> <!-- end MetaColumn 1 --> <!-- MetaColumn 2 --> <div class="column"> <div class="columns"> <div class="column"> <ul class="nav-spaced"> <li><a href="https://info.arxiv.org/help/license/index.html">Copyright</a></li> <li><a href="https://info.arxiv.org/help/policies/privacy_policy.html">Privacy Policy</a></li> </ul> </div> <div class="column sorry-app-links"> <ul class="nav-spaced"> <li><a href="https://info.arxiv.org/help/web_accessibility.html">Web Accessibility Assistance</a></li> <li> <p class="help"> <a class="a11y-main-link" href="https://status.arxiv.org" target="_blank">arXiv Operational Status <svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 256 512" class="icon filter-dark_grey" role="presentation"><path d="M224.3 273l-136 136c-9.4 9.4-24.6 9.4-33.9 0l-22.6-22.6c-9.4-9.4-9.4-24.6 0-33.9l96.4-96.4-96.4-96.4c-9.4-9.4-9.4-24.6 0-33.9L54.3 103c9.4-9.4 24.6-9.4 33.9 0l136 136c9.5 9.4 9.5 24.6.1 34z"/></svg></a><br> Get status notifications via <a class="is-link" href="https://subscribe.sorryapp.com/24846f03/email/new" target="_blank"><svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 512 512" class="icon filter-black" role="presentation"><path d="M502.3 190.8c3.9-3.1 9.7-.2 9.7 4.7V400c0 26.5-21.5 48-48 48H48c-26.5 0-48-21.5-48-48V195.6c0-5 5.7-7.8 9.7-4.7 22.4 17.4 52.1 39.5 154.1 113.6 21.1 15.4 56.7 47.8 92.2 47.6 35.7.3 72-32.8 92.3-47.6 102-74.1 131.6-96.3 154-113.7zM256 320c23.2.4 56.6-29.2 73.4-41.4 132.7-96.3 142.8-104.7 173.4-128.7 5.8-4.5 9.2-11.5 9.2-18.9v-19c0-26.5-21.5-48-48-48H48C21.5 64 0 85.5 0 112v19c0 7.4 3.4 14.3 9.2 18.9 30.6 23.9 40.7 32.4 173.4 128.7 16.8 12.2 50.2 41.8 73.4 41.4z"/></svg>email</a> or <a class="is-link" href="https://subscribe.sorryapp.com/24846f03/slack/new" target="_blank"><svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 448 512" class="icon filter-black" role="presentation"><path d="M94.12 315.1c0 25.9-21.16 47.06-47.06 47.06S0 341 0 315.1c0-25.9 21.16-47.06 47.06-47.06h47.06v47.06zm23.72 0c0-25.9 21.16-47.06 47.06-47.06s47.06 21.16 47.06 47.06v117.84c0 25.9-21.16 47.06-47.06 47.06s-47.06-21.16-47.06-47.06V315.1zm47.06-188.98c-25.9 0-47.06-21.16-47.06-47.06S139 32 164.9 32s47.06 21.16 47.06 47.06v47.06H164.9zm0 23.72c25.9 0 47.06 21.16 47.06 47.06s-21.16 47.06-47.06 47.06H47.06C21.16 243.96 0 222.8 0 196.9s21.16-47.06 47.06-47.06H164.9zm188.98 47.06c0-25.9 21.16-47.06 47.06-47.06 25.9 0 47.06 21.16 47.06 47.06s-21.16 47.06-47.06 47.06h-47.06V196.9zm-23.72 0c0 25.9-21.16 47.06-47.06 47.06-25.9 0-47.06-21.16-47.06-47.06V79.06c0-25.9 21.16-47.06 47.06-47.06 25.9 0 47.06 21.16 47.06 47.06V196.9zM283.1 385.88c25.9 0 47.06 21.16 47.06 47.06 0 25.9-21.16 47.06-47.06 47.06-25.9 0-47.06-21.16-47.06-47.06v-47.06h47.06zm0-23.72c-25.9 0-47.06-21.16-47.06-47.06 0-25.9 21.16-47.06 47.06-47.06h117.84c25.9 0 47.06 21.16 47.06 47.06 0 25.9-21.16 47.06-47.06 47.06H283.1z"/></svg>slack</a> </p> </li> </ul> </div> </div> </div> <!-- end MetaColumn 2 --> </div> </footer> <script src="https://static.arxiv.org/static/base/1.0.0a5/js/member_acknowledgement.js"></script> </body> </html>

Pages: 1 2 3 4 5 6 7 8 9 10