Automatic testing of OCE, a human-centered reinforcement learning system for automated software composition - Intelligence Collective et Interaction Access content directly
Conference Papers Year : 2022

Automatic testing of OCE, a human-centered reinforcement learning system for automated software composition

Abstract

More and more applications rely on Machine Learning (ML) techniques, e.g., to automate software engineering. Like other applications, they need to be tested and validated. Testing ML-based software differs from testing software which do not rely on AI and ML: nondeterminism, lack of oracle, high dependence on training and evaluation data are hard points. The problem is even more complicated when humans are involved in the process. In this paper, we analyze the problem of evaluating OCE, a human-centered intelligent system based on reinforcement learning that automatically builds user-tailored software. We present a test environment composed of two tools which are based on the notions of "scenario" and "ideal assembly": OCE Scenario Maker to edit scenarios and OCE Scenario Runner to automate and repeat their execution.
Fichier principal
Vignette du fichier
Version finale.pdf (245.64 Ko) Télécharger le fichier
Origin Files produced by the author(s)
licence

Dates and versions

hal-04048372 , version 1 (27-03-2023)

Licence

Identifiers

  • HAL Id : hal-04048372 , version 1

Cite

Maxence Demougeot, Kévin Delcourt, Jean-Paul Arcangeli, Sylvie Trouilhet, Françoise Adreit. Automatic testing of OCE, a human-centered reinforcement learning system for automated software composition. 1st International Workshop on Intelligent Software Engineering (ISE 2022), Dec 2022, Virtual Event, Japan. ⟨hal-04048372⟩
98 View
43 Download

Share

Gmail Mastodon Facebook X LinkedIn More