The IRMA Community
Newsletters
Research IRM
Click a keyword to search titles using our InfoSci-OnDemand powered search:
|
Improving Credibility of Machine Learner Models in Software Engineering
Abstract
Given a choice, software project managers frequently prefer traditional methods of making decisions rather than relying on empirical software engineering (empirical/machine learning-based models). One reason for this choice is the perceived lack of credibility associated with these models. To promote better empirical software engineering, a series of experiments are conducted on various NASA datasets to demonstrate the importance of assessing the ease/difficulty of a modeling situation. Each dataset is divided into three groups, a training set, and “nice/nasty” neighbor test sets. Using a nearest neighbor approach, “nice neighbors” align closest to same class training instances. “Nasty neighbors” align to the opposite class training instances. The “nice”, “nasty” experiments average 94% and 20% accuracy, respectively. Another set of experiments show how a ten-fold cross-validation is not sufficient in characterizing a dataset. Finally, a set of metric equations is proposed for improving the credibility assessment of empirical/machine learning models.
Related Content
Babita Srivastava.
© 2024.
21 pages.
|
Sakuntala Rao, Shalini Chandra, Dhrupad Mathur.
© 2024.
27 pages.
|
Satya Sekhar Venkata Gudimetla, Naveen Tirumalaraju.
© 2024.
24 pages.
|
Neeta Baporikar.
© 2024.
23 pages.
|
Shankar Subramanian Subramanian, Amritha Subhayan Krishnan, Arumugam Seetharaman.
© 2024.
35 pages.
|
Charu Banga, Farhan Ujager.
© 2024.
24 pages.
|
Munir Ahmad.
© 2024.
27 pages.
|
|
|