liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Parallel Empirical Evaluations: Resilience despite Concurrency
Linköping University, Department of Computer and Information Science, Software and Systems. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0002-8681-7470
KBS Group, Institute for Logic and Computation, TU Wien, Austria.
Massachusetts Institute of Technology, USA.
KBS Group, Institute for Logic and Computation, TU Wien, Austria.
2024 (English)In: Proceedings of the 38th AAAI Conference on Artificial Intelligence, ASSOC ADVANCEMENT ARTIFICIAL INTELLIGENCE , 2024Conference paper, Published paper (Refereed)
Abstract [en]

 Computational evaluations are crucial in modern problem-solving when we surpass theoretical algorithms or bounds. These experiments frequently take much work, and the sheer amount of needed resources makes it impossible to execute them on a single personal computer or laptop. Cluster schedulers allow for automatizing these tasks and scale to many computers. But, when we evaluate implementations of combinatorial algorithms, we depend on stable runtime results. Common approaches either limit parallelism or suffer from unstable runtime measurements due to interference among jobs on modern hardware. The former is inefficient and not sustainable. The latter results in unreplicable experiments. In this work, we address this issue and offer an acceptable balance between efficiency, software, hardware complexity, reliability, and replicability. We investigate effects towards replicability stability and illustrate how to efficiently use widely employed cluster resources for parallel evaluations. Furthermore, we present solutions which mitigate issues that emerge from the concurrent execution of benchmark jobs. Our experimental evaluation shows that – despite parallel execution – our approach reduces the runtime instability on the majority of instances to one second. 

Place, publisher, year, edition, pages
ASSOC ADVANCEMENT ARTIFICIAL INTELLIGENCE , 2024.
Series
AAAI Conference on Artificial Intelligence, ISSN 2159-5399
Keywords [en]
CSO: Solvers and Tools, SO: Algorithm Configuration, SO: Evaluation and Analysis
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:liu:diva-202590DOI: 10.1609/aaai.v38i8.28638ISI: 001239938200011OAI: oai:DiVA.org:liu-202590DiVA, id: diva2:1852212
Conference
AAAI'24, Vancouver, Canada, February 20–27, 2024
Funder
ELLIIT - The Linköping‐Lund Initiative on IT and Mobile Communications
Note

Funding Agencies|ELLIIT - Swedish government; Austrian Academy of Sciences (OAW), DOC Fellowship; Austrian Science Fund (FWF) [J4656]; Society for Research Funding in Lower Austria (GFF) [ExzF-0004]

Available from: 2024-04-17 Created: 2024-04-17 Last updated: 2024-09-06Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Authority records

Fichte, Johannes Klaus

Search in DiVA

By author/editor
Fichte, Johannes Klaus
By organisation
Software and SystemsFaculty of Science & Engineering
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 75 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf