Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Large-scale Information Retrieval in Software Engineering - An Experience Report from Industrial Application
Blekinge Tekniska Högskola, Institutionen för programvaruteknik.
Blekinge Tekniska Högskola, Institutionen för programvaruteknik.
Blekinge Tekniska Högskola, Institutionen för programvaruteknik.
Blekinge Tekniska Högskola, Institutionen för datalogi och datorsystemteknik.ORCID iD: 0000-0002-0535-1761
2016 (English)In: Journal of Empirical Software Engineering, ISSN 1382-3256, E-ISSN 1573-7616, Vol. 21, no 6, p. 2324-2365Article in journal (Refereed) Published
Abstract [en]

Background: Software Engineering activities are information intensive. Research proposes Information Retrieval (IR) techniques to support engineers in their daily tasks, such as establishing and maintaining traceability links, fault identification, and software maintenance.

Objective: We describe an engineering task, test case selection, and illustrate our problem analysis and solution discovery process. The objective of the study is to gain an understanding of to what extent IR techniques (one potential solution) can be applied to test case selection and provide decision support in a large-scale, industrial setting.

Method: We analyze, in the context of the studied company, how test case selection is performed and design a series of experiments evaluating the performance of different IR techniques. Each experiment provides lessons learned from implementation, execution, and results, feeding to its successor.

Results: The three experiments led to the following observations: 1) there is a lack of research on scalable parameter optimization of IR techniques for software engineering problems; 2) scaling IR techniques to industry data is challenging, in particular for latent semantic analysis; 3) the IR context poses constraints on the empirical evaluation of IR techniques, requiring more research on developing valid statistical approaches.

Conclusions: We believe that our experiences in conducting a series of IR experiments with industry grade data are valuable for peer researchers so that they can avoid the pitfalls that we have encountered. Furthermore, we identified challenges that need to be addressed in order to bridge the gap between laboratory IR experiments and real applications of IR in the industry.

Place, publisher, year, edition, pages
Springer, 2016. Vol. 21, no 6, p. 2324-2365
Keyword [en]
Test Case Selection, Information Retrieval, Data Mining, Experiment
National Category
Software Engineering
Identifiers
URN: urn:nbn:se:hj:diva-37971DOI: 10.1007/s10664-015-9410-8ISI: 000389085500004OAI: oai:DiVA.org:hj-37971DiVA, id: diva2:1159709
Available from: 2015-11-20 Created: 2017-11-23 Last updated: 2018-01-13Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Authority records BETA

Lavesson, Niklas

Search in DiVA

By author/editor
Lavesson, Niklas
In the same journal
Journal of Empirical Software Engineering
Software Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 6 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf