TUHH Open Research
Help
  • Log In
    New user? Click here to register.Have you forgotten your password?
  • English
  • Deutsch
  • Communities & Collections
  • Publications
  • Research Data
  • People
  • Institutions
  • Projects
  • Statistics
  1. Home
  2. TUHH
  3. Research Data
  4. haRTStone - Collection of Existing ANSI-C Benchmarks
 
Options

haRTStone - Collection of Existing ANSI-C Benchmarks

Citation Link: https://doi.org/10.15480/336.4209
Type
Software
Version
1.0
Date Issued
2022-02-13
Author(s)
Falk, Heiko  orcid-logo
Eingebettete Systeme E-13  
Gandyra, Max 
Eingebettete Systeme E-13  
Language
English
Institute
Eingebettete Systeme E-13  
DOI
10.15480/336.4209
TORE-URI
http://hdl.handle.net/11420/11807
Is Supplement To
10.5281/zenodo.6062597
Abstract
Many embedded systems are safety-critical real-time systems that have to meet hard deadlines (e.g., airbag or flight control systems). When designing such real-time systems, it is of utmost importance to guarantee that all tasks of a system meet their given deadlines. For this purpose, dedicated timing analyses are required that examine the worst-case behavior of a system and are able to provide such guarantees. In the case that deadlines are not met, optimizations need to be applied in order to modify the code of the system such that timing constraints are nevertheless finally met.

Research on such analyses and optimizations for hard real-time systems is an extremely lively area where new results are presented regularly and at a very fast pace. Naturally, the evaluation of such analyses and optimizations plays a very important role. Nowadays, evaluation typically relies on benchmarking such that new analyses or optimizations are applied to existing collections of applications, tasks or program codes. The currently used benchmarks are, however, highly limited and not sufficient in order to perform a sound and scientific evaluation, especially if massively parallel multi-task systems are considered.

For a well-founded and reproducible evaluation of analyses and optimizations, there is a strong demand for universally applicable benchmark approaches that are freely available for the entire scientific community. Benchmarks should satisfy the needs and requirements of various branches of research (e.g., schedulability analysis, WCET analysis, compiler optimization) on the one hand, but should also, on the other hand, realistically represent different application domains like, e.g., control or signal processing applications.

This project aims at the realization of a flexible and parameterizable benchmark generator that produces benchmark programs in an automated, pseudo-randomized and reproducible fashion. This benchmark generator will in particular cover the system and the code level by producing both complete task sets and also actual program codes for the individual tasks. In order to enable a widespread use of the generator and a broad collaboration with arbitrary interested people and groups, this project will be inclusive and the developed software will be openly available right from the beginning. In the end, this project shall lead to a methodology for benchmarking-based evaluation that describes clearly and reproducibly for the different real-time communities, how to use the benchmark generator in order to obtain plausible, sound and scientifically accepted evaluation results.

In order to be able to generate realistic and useful benchmarks, it is necessary to characterize key features of real-life applications and benchmarks, and to classify such applications according to their respective application domains. For this purpose, this archive contains a large collection of already existing and freely available ANSI-C benchmark collections which will serve as templates for a future classification and as blueprints for a future, randomized code generation.
DDC Class
004: Informatik
Funding(s)
Automatisierte Generierung von Benchmark-Programmen zur Evaluation von Analyse- und Optimierungsverfahren Harter Echtzeitsysteme  
Funding Organisations
Deutsche Forschungsgemeinschaft (DFG)  
License
https://mit-license.org/
No Thumbnail Available
Name

haRTStone-Training_Benchmarks.tgz

Size

1.07 GB

Format

tgz

TUHH
Weiterführende Links
  • Contact
  • Send Feedback
  • Cookie settings
  • Privacy policy
  • Impress
DSpace Software

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science
Design by effective webwork GmbH

  • Deutsche NationalbibliothekDeutsche Nationalbibliothek
  • ORCiD Member OrganizationORCiD Member Organization
  • DataCiteDataCite
  • Re3DataRe3Data
  • OpenDOAROpenDOAR
  • OpenAireOpenAire
  • BASE Bielefeld Academic Search EngineBASE Bielefeld Academic Search Engine
Feedback