Authors
Tim Menzies, Burak Turhan, Ayse Bener, Gregory Gay, Bojan Cukic, Yue Jiang
Publication date
2008/5/12
Book
Proceedings of the 4th international workshop on Predictor models in software engineering
Pages
47-54
Description
Context
There are many methods that input static code features and output a predictor for faulty code modules. These data mining methods have hit a "performance ceiling"; i.e., some inherent upper bound on the amount of information offered by, say, static code features when identifying modules which contain faults.
Objective
We seek an explanation for this ceiling effect. Perhaps static code features have "limited information content"; i.e. their information can be quickly and completely discovered by even simple learners.
Method
An initial literature review documents the ceiling effect in other work. Next, using three sub-sampling techniques (under-, over-, and micro-sampling), we look for the lower useful bound on the number of training instances.
Results
Using micro-sampling, we find that as few as 50 instances yield as much information as larger training sets.
Conclusions
We have found much evidence for the …
Total citations
200720082009201020112012201320142015201620172018201920202021202220232024191282019201116259171281711165
Scholar articles
T Menzies, B Turhan, A Bener, G Gay, B Cukic, Y Jiang - Proceedings of the 4th international workshop on …, 2008