On Reporting Performance and Accuracy Bugs for Deep Learning Frameworks: An Exploratory Study from GitHub

Guoming Long, Tao Chen*

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

The tremendous success of Deep Learning (DL) has significantly boosted the number of open-sourced DL frameworks hosted on GitHub. Among others, performance and accuracy bugs are critical factors that affect the reputation of these DL frameworks, therefore understanding the practice of discovering and investigating them for DL is important. In this paper, we conduct an exploratory study on the nature of reporting performance and accuracy bugs for DL frameworks, aiming to improve our knowledge on this topic. Our study covers 10 most popular open-sourced DL frameworks on GitHub (e.g., TensorFlow, Keras, and PyTorch), based on which we sample 664 representative performance and accuracy bug reports out of a total population of 22,522. Through systematic analysis, we found that: (1) low speed is the primary reason that a performance bug related report is submitted but we see no consistent pattern for accuracy related ones; (2) most of the reports are about issues encountered in the training stage; (3) only a small proportion of the reports provide insufficient information to investigate; (4) the majority of the performance and accuracy bug reports (from 69% to 100%) are not related to the actual bug or regarded as unclassified; (5) around 50% of the performance and accuracy bug reports, which indeed reveal bugs, are not resolved by direct patches. Deriving from the above, we discuss a set of actionable implications to the researchers, maintainers, and report submitters. To promote open science, the labeled dataset has been made publicly available at https://zenodo.org/record/6371676 .
Original languageEnglish
Title of host publicationEASE '22
Subtitle of host publicationProceedings of the 26th International Conference on Evaluation and Assessment in Software Engineering
EditorsMiroslaw Staron, Christian Berger, Jocelyn Simmonds, Rafael Prikladnicki
PublisherAssociation for Computing Machinery (ACM)
Pages90–99
Number of pages10
ISBN (Electronic)9781450396134
Publication statusPublished - 13 Jun 2022
EventThe International Conference on Evaluation and Assessment in Software Engineering 202 - Gothenburg , Sweden
Duration: 13 Jun 202215 Jun 2022

Publication series

NameEASE: Evaluation and Assessment in Software Engineering

Conference

ConferenceThe International Conference on Evaluation and Assessment in Software Engineering 202
Abbreviated titleEASE 2022
Country/TerritorySweden
CityGothenburg
Period13/06/2215/06/22

Keywords

  • Empirical software engineering
  • mining software repositories
  • artificial intelligence
  • performance engineering

Fingerprint

Dive into the research topics of 'On Reporting Performance and Accuracy Bugs for Deep Learning Frameworks: An Exploratory Study from GitHub'. Together they form a unique fingerprint.

Cite this