Skip to main content

Showing 1–120 of 120 results for all: tim menzies

.
  1. arXiv:2401.09622  [pdf, other

    cs.SE cs.LG

    SMOOTHIE: A Theory of Hyper-parameter Optimization for Software Analytics

    Authors: Rahul Yedida, Tim Menzies

    Abstract: Hyper-parameter optimization is the black art of tuning a learner's control parameters. In software analytics, a repeated result is that such tuning can result in dramatic performance improvements. Despite this, hyper-parameter optimization is often applied rarely or poorly in software analytics--perhaps due to the CPU cost of exploring all those parameter options can be prohibitive. We theorize… ▽ More

    Submitted 17 January, 2024; originally announced January 2024.

    Comments: v1

  2. arXiv:2401.01883  [pdf, other

    cs.CR cs.IR cs.LG cs.SE

    Mining Temporal Attack Patterns from Cyberthreat Intelligence Reports

    Authors: Md Rayhanur Rahman, Brandon Wroblewski, Quinn Matthews, Brantley Morgan, Tim Menzies, Laurie Williams

    Abstract: Defending from cyberattacks requires practitioners to operate on high-level adversary behavior. Cyberthreat intelligence (CTI) reports on past cyberattack incidents describe the chain of malicious actions with respect to time. To avoid repeating cyberattack incidents, practitioners must proactively identify and defend against recurring chain of actions - which we refer to as temporal attack patter… ▽ More

    Submitted 3 January, 2024; originally announced January 2024.

    Comments: A modified version of this pre-print is submitted to IEEE Transactions on Software Engineering, and is under review

  3. arXiv:2312.05436  [pdf, other

    cs.SE

    Trading Off Scalability, Privacy, and Performance in Data Synthesis

    Authors: Xiao Ling, Tim Menzies, Christopher Hazard, Jack Shu, Jacob Beel

    Abstract: Synthetic data has been widely applied in the real world recently. One typical example is the creation of synthetic data for privacy concerned datasets. In this scenario, synthetic data substitute the real data which contains the privacy information, and is used to public testing for machine learning models. Another typical example is the unbalance data over-sampling which the synthetic data is ge… ▽ More

    Submitted 8 December, 2023; originally announced December 2023.

    Comments: 13 pages, 2 figures, 6 tables, submitted to IEEEAccess

  4. arXiv:2310.19125  [pdf, other

    cs.SE

    Partial Orderings as Heuristic for Multi-Objective Model-Based Reasoning

    Authors: Andre Lustosa, Tim Menzies

    Abstract: Model-based reasoning is becoming increasingly common in software engineering. The process of building and analyzing models helps stakeholders to understand the ramifications of their software decisions. But complex models can confuse and overwhelm stakeholders when these models have too many candidate solutions. We argue here that a technique based on partial orderings lets humans find acceptable… ▽ More

    Submitted 29 October, 2023; originally announced October 2023.

  5. arXiv:2310.07109  [pdf, other

    cs.SE

    SparseCoder: Advancing Source Code Analysis with Sparse Attention and Learned Token Pruning

    Authors: Xueqi Yang, Mariusz Jakubowski, Kelly Kang, Haojie Yu, Tim Menzies

    Abstract: As software projects rapidly evolve, software artifacts become more complex and defects behind get harder to identify. The emerging Transformer-based approaches, though achieving remarkable performance, struggle with long code sequences due to their self-attention mechanism, which scales quadratically with the sequence length. This paper introduces SparseCoder, an innovative approach incorporating… ▽ More

    Submitted 10 October, 2023; originally announced October 2023.

    Comments: 11 pages, 8 figures, pre-print

  6. Model Review: A PROMISEing Opportunity

    Authors: Tim Menzies

    Abstract: To make models more understandable and correctable, I propose that the PROMISE community pivots to the problem of model review. Over the years, there have been many reports that very simple models can perform exceptionally well. Yet, where are the researchers asking "say, does that mean that we could make software analytics simpler and more comprehensible?" This is an important question, since hum… ▽ More

    Submitted 6 September, 2023; v1 submitted 3 September, 2023; originally announced September 2023.

    Comments: 5 pages, 1 figure

  7. arXiv:2305.03714  [pdf, other

    cs.SE

    On the Benefits of Semi-Supervised Test Case Generation for Simulation Models

    Authors: Xiao Ling, Tim Menzies

    Abstract: Testing complex simulation models can be expensive and time consuming. Current state-of-the-art methods that explore this problem are fully-supervised; i.e. they require that all examples are labeled. On the other hand, the GenClu system (introduced in this paper) takes a semi-supervised approach; i.e. (a) only a small subset of information is actually labeled (via simulation) and (b) those labels… ▽ More

    Submitted 1 December, 2023; v1 submitted 5 May, 2023; originally announced May 2023.

    Comments: 14 pages, 4 figures, 6 tables, first round review in TSE

  8. arXiv:2302.01997  [pdf, other

    cs.SE cs.AI

    Less, but Stronger: On the Value of Strong Heuristics in Semi-supervised Learning for Software Analytics

    Authors: Huy Tu, Tim Menzies

    Abstract: In many domains, there are many examples and far fewer labels for those examples; e.g. we may have access to millions of lines of source code, but access to only a handful of warnings about that code. In those domains, semi-supervised learners (SSL) can extrapolate labels from a small number of examples to the rest of the data. Standard SSL algorithms use ``weak'' knowledge (i.e. those not based o… ▽ More

    Submitted 3 February, 2023; originally announced February 2023.

    Comments: Submitting to EMSE

  9. arXiv:2301.10407  [pdf, other

    cs.SE cs.AI cs.CR

    Don't Lie to Me: Avoiding Malicious Explanations with STEALTH

    Authors: Lauren Alvarez, Tim Menzies

    Abstract: STEALTH is a method for using some AI-generated model, without suffering from malicious attacks (i.e. lying) or associated unfairness issues. After recursively bi-clustering the data, STEALTH system asks the AI model a limited number of queries about class labels. STEALTH asks so few queries (1 per data cluster) that malicious algorithms (a) cannot detect its operation, nor (b) know when to lie.

    Submitted 25 January, 2023; originally announced January 2023.

    Comments: 6 pages, 6 Tables, 3 figures

  10. arXiv:2301.06577  [pdf, other

    cs.SE cs.LG

    Learning from Very Little Data: On the Value of Landscape Analysis for Predicting Software Project Health

    Authors: Andre Lustosa, Tim Menzies

    Abstract: When data is scarce, software analytics can make many mistakes. For example, consider learning predictors for open source project health (e.g. the number of closed pull requests in twelve months time). The training data for this task may be very small (e.g. five years of data, collected every month means just 60 rows of training data). The models generated from such tiny data sets can make many pr… ▽ More

    Submitted 11 October, 2023; v1 submitted 16 January, 2023; originally announced January 2023.

  11. A Tale of Two Cities: Data and Configuration Variances in Robust Deep Learning

    Authors: Guanqin Zhang, Jiankun Sun, Feng Xu, H. M. N. Dilum Bandara, Shiping Chen, Yulei Sui, Tim Menzies

    Abstract: Deep neural networks (DNNs), are widely used in many industries such as image recognition, supply chain, medical diagnosis, and autonomous driving. However, prior work has shown the high accuracy of a DNN model does not imply high robustness (i.e., consistent performances on new and future datasets) because the input data and external environment (e.g., software and model configurations) for a dep… ▽ More

    Submitted 25 November, 2022; v1 submitted 17 November, 2022; originally announced November 2022.

  12. arXiv:2211.05920  [pdf, other

    cs.SE cs.LG

    When Less is More: On the Value of "Co-training" for Semi-Supervised Software Defect Predictors

    Authors: Suvodeep Majumder, Joymallya Chakraborty, Tim Menzies

    Abstract: Labeling a module defective or non-defective is an expensive task. Hence, there are often limits on how much-labeled data is available for training. Semi-supervised classifiers use far fewer labels for training models. However, there are numerous semi-supervised methods, including self-labeling, co-training, maximal-margin, and graph-based methods, to name a few. Only a handful of these methods ha… ▽ More

    Submitted 15 February, 2024; v1 submitted 10 November, 2022; originally announced November 2022.

    Comments: 36 pages, 10 figures, 5 tables

  13. arXiv:2208.01595  [pdf, other

    cs.SE cs.CR

    Do I really need all this work to find vulnerabilities? An empirical case study comparing vulnerability detection techniques on a Java application

    Authors: Sarah Elder, Nusrat Zahan, Rui Shu, Monica Metro, Valeri Kozarev, Tim Menzies, Laurie Williams

    Abstract: CONTEXT: Applying vulnerability detection techniques is one of many tasks using the limited resources of a software project. OBJECTIVE: The goal of this research is to assist managers and other decision-makers in making informed choices about the use of software vulnerability detection techniques through an empirical study of the efficiency and effectiveness of four techniques on a Java-based we… ▽ More

    Submitted 2 August, 2022; originally announced August 2022.

    ACM Class: D.2.5

  14. arXiv:2205.10504  [pdf, other

    cs.SE cs.LG

    How to Find Actionable Static Analysis Warnings: A Case Study with FindBugs

    Authors: Rahul Yedida, Hong Jin Kang, Huy Tu, Xueqi Yang, David Lo, Tim Menzies

    Abstract: Automatically generated static code warnings suffer from a large number of false alarms. Hence, developers only take action on a small percent of those warnings. To better predict which static code warnings should not be ignored, we suggest that analysts need to look deeper into their algorithms to find choices that better improve the particulars of their specific problem. Specifically, we show he… ▽ More

    Submitted 23 December, 2022; v1 submitted 21 May, 2022; originally announced May 2022.

    Comments: Accepted to TSE

  15. arXiv:2205.00665  [pdf, other

    cs.CR cs.SE

    Reducing the Cost of Training Security Classifier (via Optimized Semi-Supervised Learning)

    Authors: Rui Shu, Tianpei Xia, Huy Tu, Laurie Williams, Tim Menzies

    Abstract: Background: Most of the existing machine learning models for security tasks, such as spam detection, malware detection, or network intrusion detection, are built on supervised machine learning algorithms. In such a paradigm, models need a large amount of labeled data to learn the useful relationships between selected features and the target class. However, such labeled data can be scarce and expen… ▽ More

    Submitted 2 May, 2022; originally announced May 2022.

  16. arXiv:2203.11410  [pdf, other

    cs.CR cs.LG cs.SE

    Dazzle: Using Optimized Generative Adversarial Networks to Address Security Data Class Imbalance Issue

    Authors: Rui Shu, Tianpei Xia, Laurie Williams, Tim Menzies

    Abstract: Background: Machine learning techniques have been widely used and demonstrate promising performance in many software security tasks such as software vulnerability prediction. However, the class ratio within software vulnerability datasets is often highly imbalanced (since the percentage of observed vulnerability is usually very low). Goal: To help security practitioners address software security d… ▽ More

    Submitted 2 May, 2022; v1 submitted 21 March, 2022; originally announced March 2022.

  17. arXiv:2202.01322  [pdf, other

    cs.SE

    How to Improve Deep Learning for Software Analytics (a case study with code smell detection)

    Authors: Rahul Yedida, Tim Menzies

    Abstract: To reduce technical debt and make code more maintainable, it is important to be able to warn programmers about code smells. State-of-the-art code small detectors use deep learners, without much exploration of alternatives within that technology. One promising alternative for software analytics and deep learning is GHOST (from TSE'21) that relies on a combination of hyper-parameter optimization o… ▽ More

    Submitted 27 March, 2022; v1 submitted 2 February, 2022; originally announced February 2022.

    Comments: Accepted to MSR 2022

  18. arXiv:2201.13296  [pdf, other

    astro-ph.SR astro-ph.GA astro-ph.HE

    An Isolated Stellar-Mass Black Hole Detected Through Astrometric Microlensing

    Authors: Kailash C. Sahu, Jay Anderson, Stefano Casertano, Howard E. Bond, Andrzej Udalski, Martin Dominik, Annalisa Calamida, Andrea Bellini, Thomas M. Brown, Marina Rejkuba, Varun Bajaj, Noe Kains, Henry C. Ferguson, Chris L. Fryer, Philip Yock, Przemek Mroz, Szymon Kozlowski, Pawel Pietrukowicz, Radek Poleski, Jan Skowron, Igor Soszynski, Michael K. Szymanski, Krzysztof Ulaczyk, Lukasz Wyrzykowski, Richard Barry , et al. (68 additional authors not shown)

    Abstract: We report the first unambiguous detection and mass measurement of an isolated stellar-mass black hole (BH). We used the Hubble Space Telescope (HST) to carry out precise astrometry of the source star of the long-duration (t_E~270 days), high-magnification microlensing event MOA-2011-BLG-191/OGLE-2011-BLG-0462 (hereafter designated as MOA-11-191/OGLE-11-462), in the direction of the Galactic bulge.… ▽ More

    Submitted 22 July, 2022; v1 submitted 31 January, 2022; originally announced January 2022.

    Comments: 37 pages, Published in ApJ

    Journal ref: ApJ, 933, 83 (2022)

  19. arXiv:2201.10592  [pdf, other

    cs.SE cs.AI

    DebtFree: Minimizing Labeling Cost in Self-Admitted Technical Debt Identification using Semi-Supervised Learning

    Authors: Huy Tu, Tim Menzies

    Abstract: Keeping track of and managing Self-Admitted Technical Debts (SATDs) is important for maintaining a healthy software project. Current active-learning SATD recognition tool involves manual inspection of 24% of the test comments on average to reach 90% of the recall. Among all the test comments, about 5% are SATDs. The human experts are then required to read almost a quintuple of the SATD comments wh… ▽ More

    Submitted 25 January, 2022; originally announced January 2022.

    Comments: Accepted at EMSE

  20. arXiv:2112.01598  [pdf, other

    cs.SE

    What Not to Test (for Cyber-Physical Systems)

    Authors: Xiao Ling, Tim Menzies

    Abstract: For simulation-based systems, finding a set of test cases with the least cost by exploring multiple goals is a complex task. Domain-specific optimization goals (e.g. maximize output variance) are useful for guiding the rapid selection of test cases via mutation. But evaluating the selected test cases via mutation (that can distinguish the current program from the mutated systems) is a different go… ▽ More

    Submitted 5 May, 2023; v1 submitted 2 December, 2021; originally announced December 2021.

    Comments: 17 pages, 5 figures, 7 tables. Accepted by TSE

  21. arXiv:2110.13029  [pdf, other

    cs.LG cs.CY cs.SE

    Fair Enough: Searching for Sufficient Measures of Fairness

    Authors: Suvodeep Majumder, Joymallya Chakraborty, Gina R. Bai, Kathryn T. Stolee, Tim Menzies

    Abstract: Testing machine learning software for ethical bias has become a pressing current concern. In response, recent research has proposed a plethora of new fairness metrics, for example, the dozens of fairness metrics in the IBM AIF360 toolkit. This raises the question: How can any fairness tool satisfy such a diverse range of goals? While we cannot completely simplify the task of fairness testing, we c… ▽ More

    Submitted 21 March, 2022; v1 submitted 25 October, 2021; originally announced October 2021.

    Comments: 8 tables and 1 figure

  22. arXiv:2110.02922   

    cs.SE

    SNEAK: Faster Interactive Search-based SE

    Authors: Andre Lustosa, Jaydeep Patel, Venkata Sai Teja Malapati, Tim Menzies

    Abstract: When AI tools can generate many solutions, some human preference must be applied to determine which solution is relevant to the current project. One way to find those preferences is interactive search-based software engineering (iSBSE) where humans can influence the search process. This paper argues that when optimizing a model using human-in-the-loop, data mining methods such as our SNEAK tool (t… ▽ More

    Submitted 16 January, 2023; v1 submitted 6 October, 2021; originally announced October 2021.

    Comments: removal for resubmission under different title and more information

  23. arXiv:2110.01710  [pdf, other

    cs.SE

    PyTorrent: A Python Library Corpus for Large-scale Language Models

    Authors: Mehdi Bahrami, N. C. Shrikanth, Shade Ruangwan, Lei Liu, Yuji Mizobuchi, Masahiro Fukuyori, Wei-Peng Chen, Kazuki Munakata, Tim Menzies

    Abstract: A large scale collection of both semantic and natural language resources is essential to leverage active Software Engineering research areas such as code reuse and code comprehensibility. Existing machine learning models ingest data from Open Source repositories (like GitHub projects) and forum discussions (like Stackoverflow.com), whereas, in this showcase, we took a step backward to orchestrate… ▽ More

    Submitted 4 October, 2021; originally announced October 2021.

    Comments: 10 pages, 2 figures, 5 tables

  24. arXiv:2110.01109  [pdf, other

    cs.LG cs.SE

    FairMask: Better Fairness via Model-based Rebalancing of Protected Attributes

    Authors: Kewen Peng, Joymallya Chakraborty, Tim Menzies

    Abstract: Context: Machine learning software can generate models that inappropriately discriminate against specific protected social groups (e.g., groups based on gender, ethnicity, etc). Motivated by those results, software engineering researchers have proposed many methods for mitigating those discriminatory effects. While those methods are effective in mitigating bias, few of them can provide explanation… ▽ More

    Submitted 27 October, 2022; v1 submitted 3 October, 2021; originally announced October 2021.

    Comments: 14 pages, 6 figures, 7 tables, accepted by TSE

    ACM Class: D.2

  25. arXiv:2109.14569  [pdf, other

    cs.LG cs.SE stat.ML

    An Expert System for Redesigning Software for Cloud Applications

    Authors: Rahul Yedida, Rahul Krishna, Anup Kalia, Tim Menzies, Jin Xiao, Maja Vukovic

    Abstract: Cloud-based software has many advantages. When services are divided into many independent components, they are easier to update. Also, during peak demand, it is easier to scale cloud services (just hire more CPUs). Hence, many organizations are partitioning their monolithic enterprise applications into cloud-based microservices. Recently there has been much work using machine learning to simplif… ▽ More

    Submitted 27 June, 2022; v1 submitted 29 September, 2021; originally announced September 2021.

    Comments: version 3

  26. arXiv:2108.09847  [pdf, other

    cs.SE cs.LG

    FRUGAL: Unlocking SSL for Software Analytics

    Authors: Huy Tu, Tim Menzies

    Abstract: Standard software analytics often involves having a large amount of data with labels in order to commission models with acceptable performance. However, prior work has shown that such requirements can be expensive, taking several weeks to label thousands of commits, and not always available when traversing new research problems and domains. Unsupervised Learning is a promising direction to learn h… ▽ More

    Submitted 22 August, 2021; originally announced August 2021.

    Comments: Accepted for ASE 2022

  27. Crowdsourcing the State of the Art(ifacts)

    Authors: Maria Teresa Baldassarre, Neil Ernst, Ben Hermann, Tim Menzies, Rahul Yedida

    Abstract: In any field, finding the "leading edge" of research is an on-going challenge. Researchers cannot appease reviewers and educators cannot teach to the leading edge of their field if no one agrees on what is the state-of-the-art. Using a novel crowdsourced "reuse graph" approach, we propose here a new method to learn this state-of-the-art. Our reuse graphs are less effort to build and verify than… ▽ More

    Submitted 15 August, 2021; originally announced August 2021.

    Comments: Submitted to Communications ACM

    Journal ref: CACM February 2023 (Vol. 66, No. 2)

  28. arXiv:2107.08310  [pdf, other

    cs.LG

    FairBalance: How to Achieve Equalized Odds With Data Pre-processing

    Authors: Zhe Yu, Joymallya Chakraborty, Tim Menzies

    Abstract: This research seeks to benefit the software engineering society by providing a simple yet effective pre-processing approach to achieve equalized odds fairness in machine learning software. Fairness issues have attracted increasing attention since machine learning software is increasingly used for high-stakes and high-risk decisions. Amongst all the existing fairness notions, this work specifically… ▽ More

    Submitted 26 April, 2023; v1 submitted 17 July, 2021; originally announced July 2021.

    Comments: 13 pages

  29. arXiv:2107.05088  [pdf, other

    cs.SE cs.AI

    Fairer Software Made Easier (using "Keys")

    Authors: Tim Menzies, Kewen Peng, Andre Lustosa

    Abstract: Can we simplify explanations for software analytics? Maybe. Recent results show that systems often exhibit a "keys effect"; i.e. a few key features control the rest. Just to say the obvious, for systems controlled by a few keys, explanation and control is just a matter of running a handful of "what-if" queries across the keys. By exploiting the keys effect, it should be possible to dramatically si… ▽ More

    Submitted 11 July, 2021; originally announced July 2021.

    Comments: Submitted to NIER ASE 2021 (new ideas, emerging research)

  30. arXiv:2106.06652  [pdf, ps, other

    cs.SE

    Lessons learned from hyper-parameter tuning for microservice candidate identification

    Authors: Rahul Yedida, Rahul Krishna, Anup Kalia, Tim Menzies, Jin Xiao, Maja Vukovic

    Abstract: When optimizing software for the cloud, monolithic applications need to be partitioned into many smaller *microservices*. While many tools have been proposed for this task, we warn that the evaluation of those approaches has been incomplete; e.g. minimal prior exploration of hyperparameter optimization. Using a set of open source Java EE applications, we show here that (a) such optimization can si… ▽ More

    Submitted 10 August, 2021; v1 submitted 11 June, 2021; originally announced June 2021.

    Comments: Accepted to ASE 2021 (industry track, short paper)

  31. arXiv:2106.03792   

    cs.SE

    Preference Discovery in Large Product Lines

    Authors: Andre Lustosa, Tim Menzies

    Abstract: When AI tools can generate many solutions, some human preference must be applied to determine which solution is relevant to the current project. One way to find those preferences is interactive search-based software engineering (iSBSE) where humans can influence the search process. Current iSBSE methods can lead to cognitive fatigue (when they overwhelm humans with too many overly elaborate questi… ▽ More

    Submitted 16 January, 2023; v1 submitted 7 June, 2021; originally announced June 2021.

    Comments: Reformatting and republishing of the paper under a different name

  32. arXiv:2106.02716  [pdf, other

    cs.SE

    VEER: Enhancing the Interpretability of Model-based Optimizations

    Authors: Kewen Peng, Christian Kaltenecker, Norbert Siegmund, Sven Apel, Tim Menzies

    Abstract: Many software systems can be tuned for multiple objectives (e.g., faster runtime, less required memory, less network traffic or energy consumption, etc.). Optimizers built for different objectives suffer from "model disagreement"; i.e., they have different (or even opposite) insights and tactics on how to optimize a system. Model disagreement is rampant (at least for configuration problems). Yet p… ▽ More

    Submitted 12 February, 2023; v1 submitted 4 June, 2021; originally announced June 2021.

    Comments: 27 pages, 7 figures, 4 tables, accepted by EMSE

    ACM Class: D.2; K.6.3

  33. Bias in Machine Learning Software: Why? How? What to do?

    Authors: Joymallya Chakraborty, Suvodeep Majumder, Tim Menzies

    Abstract: Increasingly, software is making autonomous decisions in case of criminal sentencing, approving credit cards, hiring employees, and so on. Some of these decisions show bias and adversely affect certain social groups (e.g. those defined by sex, race, age, marital status). Many prior works on bias mitigation take the following form: change the data or learners in multiple ways, then see if any of th… ▽ More

    Submitted 9 July, 2021; v1 submitted 25 May, 2021; originally announced May 2021.

    Journal ref: ESEC/FSE'2021: The 29th ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering (ESEC/FSE), Athens, Greece, August 23-28, 2021

  34. arXiv:2105.11082  [pdf, other

    cs.SE cs.AI cs.LG

    Assessing the Early Bird Heuristic (for Predicting Project Quality)

    Authors: N. C. Shrikanth, Tim Menzies

    Abstract: Before researchers rush to reason across all available data or try complex methods, perhaps it is prudent to first check for simpler alternatives. Specifically, if the historical data has the most information in some small region, perhaps a model learned from that region would suffice for the rest of the project. To support this claim, we offer a case study with 240 projects, where we find that… ▽ More

    Submitted 11 January, 2023; v1 submitted 23 May, 2021; originally announced May 2021.

    Comments: 38 pages (Accepted TOSEM Jan 2023)

  35. arXiv:2103.12221  [pdf, other

    cs.SE cs.NI

    Mining Scientific Workflows for Anomalous Data Transfers

    Authors: Huy Tu, George Papadimitriou, Mariam Kiran, Cong Wang, Anirban Mandal, Ewa Deelman, Tim Menzies

    Abstract: Modern scientific workflows are data-driven and are often executed on distributed, heterogeneous, high-performance computing infrastructures. Anomalies and failures in the workflow execution cause loss of scientific productivity and inefficient use of the infrastructure. Hence, detecting, diagnosing, and mitigating these anomalies are immensely important for reliable and performant scientific work… ▽ More

    Submitted 22 March, 2021; originally announced March 2021.

    Comments: Accepted for MSR 2021: Working Conference on Mining Software Repositories (https://2021.msrconf.org/details/msr-2021-technical-papers/1/Mining-Workflows-for-Anomalous-Data-Transfers)

  36. arXiv:2103.05088  [pdf, other

    cs.SE cs.CR

    Structuring a Comprehensive Software Security Course Around the OWASP Application Security Verification Standard

    Authors: Sarah Elder, Nusrat Zahan, Val Kozarev, Rui Shu, Tim Menzies, Laurie Williams

    Abstract: Lack of security expertise among software practitioners is a problem with many implications. First, there is a deficit of security professionals to meet current needs. Additionally, even practitioners who do not plan to work in security may benefit from increased understanding of security. The goal of this paper is to aid software engineering educators in designing a comprehensive software securit… ▽ More

    Submitted 8 March, 2021; originally announced March 2021.

    Comments: 10 pages, 5 figures, 1 table, submitted to International Conference on Software Engineering: Joint Track on Software Engineering Education and Training (ICSE-JSEET)

    ACM Class: K.3.0; D.2.0; K.6.5

  37. arXiv:2101.06319  [pdf, other

    cs.SE cs.AI

    Old but Gold: Reconsidering the value of feedforward learners for software analytics

    Authors: Rahul Yedida, Xueqi Yang, Tim Menzies

    Abstract: There has been an increased interest in the use of deep learning approaches for software analytics tasks. State-of-the-art techniques leverage modern deep learning techniques such as LSTMs, yielding competitive performance, albeit at the price of longer training times. Recently, Galke and Scherp [18] showed that at least for image recognition, a decades-old feedforward neural network can match t… ▽ More

    Submitted 5 February, 2022; v1 submitted 15 January, 2021; originally announced January 2021.

    Comments: v2

  38. arXiv:2101.02817  [pdf, other

    cs.SE

    Faster SAT Solving for Software with Repeated Structures (with Case Studies on Software Test Suite Minimization)

    Authors: Jianfeng Chen, Xipeng Shen, Tim Menzies

    Abstract: Theorem provers has been used extensively in software engineering for software testing or verification. However, software is now so large and complex that additional architecture is needed to guide theorem provers as they try to generate test suites. The SNAP test suite generator (introduced in this paper) combines the Z3 theorem prover with the following tactic: cluster some candidate tests, then… ▽ More

    Submitted 7 January, 2021; originally announced January 2021.

    Comments: Submitted to Journal Software and Systems. arXiv admin note: substantial text overlap with arXiv:1905.05358

  39. arXiv:2011.13071  [pdf, other

    cs.SE cs.LG

    Early Life Cycle Software Defect Prediction. Why? How?

    Authors: N. C. Shrikanth, Suvodeep Majumder, Tim Menzies

    Abstract: Many researchers assume that, for software analytics, "more data is better." We write to show that, at least for learning defect predictors, this may not be true. To demonstrate this, we analyzed hundreds of popular GitHub projects. These projects ran for 84 months and contained 3,728 commits (median values). Across these projects, most of the defects occur very early in their life cycle. Hence, d… ▽ More

    Submitted 8 February, 2021; v1 submitted 25 November, 2020; originally announced November 2020.

    Comments: 12 pages (To appear ICSE 2021)

  40. arXiv:2011.12720  [pdf, other

    cs.CR cs.LG

    Omni: Automated Ensemble with Unexpected Models against Adversarial Evasion Attack

    Authors: Rui Shu, Tianpei Xia, Laurie Williams, Tim Menzies

    Abstract: Background: Machine learning-based security detection models have become prevalent in modern malware and intrusion detection systems. However, previous studies show that such models are susceptible to adversarial evasion attacks. In this type of attack, inputs (i.e., adversarial examples) are specially crafted by intelligent malicious adversaries, with the aim of being misclassified by existing st… ▽ More

    Submitted 12 October, 2021; v1 submitted 23 November, 2020; originally announced November 2020.

    Comments: Submitted to EMSE

  41. arXiv:2010.03525  [pdf

    cs.SE cs.GL

    Empirical Standards for Software Engineering Research

    Authors: Paul Ralph, Nauman bin Ali, Sebastian Baltes, Domenico Bianculli, Jessica Diaz, Yvonne Dittrich, Neil Ernst, Michael Felderer, Robert Feldt, Antonio Filieri, Breno Bernard Nicolau de França, Carlo Alberto Furia, Greg Gay, Nicolas Gold, Daniel Graziotin, Pinjia He, Rashina Hoda, Natalia Juristo, Barbara Kitchenham, Valentina Lenarduzzi, Jorge Martínez, Jorge Melegati, Daniel Mendez, Tim Menzies, Jefferson Molleri , et al. (18 additional authors not shown)

    Abstract: Empirical Standards are natural-language models of a scientific community's expectations for a specific kind of study (e.g. a questionnaire survey). The ACM SIGSOFT Paper and Peer Review Quality Initiative generated empirical standards for research methods commonly used in software engineering. These living documents, which should be continuously revised to reflect evolving consensus around resear… ▽ More

    Submitted 4 March, 2021; v1 submitted 7 October, 2020; originally announced October 2020.

    Comments: For the complete standards, supplements and other resources, see https://github.com/acmsigsoft/EmpiricalStandards

  42. Revisiting Process versus Product Metrics: a Large Scale Analysis

    Authors: Suvodeep Majumder, Pranav Mody, Tim Menzies

    Abstract: Numerous methods can build predictive models from software data. However, what methods and conclusions should we endorse as we move from analytics in-the-small (dealing with a handful of projects) to analytics in-the-large (dealing with hundreds of projects)? To answer this question, we recheck prior small-scale results (about process versus product metrics for defect prediction and the granular… ▽ More

    Submitted 26 October, 2021; v1 submitted 21 August, 2020; originally announced August 2020.

    Comments: 36 pages, 12 figures and 5 tables

    Journal ref: Empirical Software Engineering, Volume 27, Issue 3, May 2022

  43. arXiv:2008.07334   

    cs.SE

    Simpler Hyperparameter Optimization for Software Analytics: Why, How, When?

    Authors: Amritanshu Agrawal, Xueqi Yang, Rishabh Agrawal, Xipeng Shen, Tim Menzies

    Abstract: How to make software analytics simpler and faster? One method is to match the complexity of analysis to the intrinsic complexity of the data being explored. For example, hyperparameter optimizers find the control settings for data miners that improve for improving the predictions generated via software analytics. Sometimes, very fast hyperparameter optimization can be achieved by just DODGE-ing aw… ▽ More

    Submitted 22 April, 2021; v1 submitted 13 August, 2020; originally announced August 2020.

    Comments: made a mistake with my co-author. the current version of this doc is their version arXiv:1912.04061

  44. arXiv:2008.03835  [pdf, other

    cs.SE

    On the Value of Oversampling for Deep Learning in Software Defect Prediction

    Authors: Rahul Yedida, Tim Menzies

    Abstract: One truism of deep learning is that the automatic feature engineering (seen in the first layers of those networks) excuses data scientists from performing tedious manual feature engineering prior to running DL. For the specific case of deep learning for defect prediction, we show that that truism is false. Specifically, when we preprocess data with a novel oversampling technique called fuzzy sampl… ▽ More

    Submitted 20 April, 2021; v1 submitted 9 August, 2020; originally announced August 2020.

    Comments: v3, revision 2 (minor revision); submitted to TSE

  45. arXiv:2008.00612  [pdf, other

    cs.SE

    How Different is Test Case Prioritization for Open and Closed Source Projects?

    Authors: Xiao Ling, Rishabh Agrawal, Tim Menzies

    Abstract: Improved test case prioritization means that software developers can detect and fix more software faults sooner than usual. But is there one "best" prioritization algorithm? Or do different kinds of projects deserve special kinds of prioritization? To answer these questions, this paper applies nine prioritization schemes to 31 projects that range from (a) highly rated open-source Github projects t… ▽ More

    Submitted 20 February, 2021; v1 submitted 2 August, 2020; originally announced August 2020.

    Comments: 15 pages, 4 figures, 16 tables, accepted to TSE

  46. Making Fair ML Software using Trustworthy Explanation

    Authors: Joymallya Chakraborty, Kewen Peng, Tim Menzies

    Abstract: Machine learning software is being used in many applications (finance, hiring, admissions, criminal justice) having a huge social impact. But sometimes the behavior of this software is biased and it shows discrimination based on some sensitive attributes such as sex, race, etc. Prior works concentrated on finding and mitigating bias in ML models. A recent trend is using instance-based model-agnost… ▽ More

    Submitted 18 August, 2020; v1 submitted 6 July, 2020; originally announced July 2020.

    Comments: New Ideas and Emerging Results (NIER) track; The 35th IEEE/ACM International Conference on Automated Software Engineering; Melbourne, Australia

    Journal ref: ASE 2020: The 35th IEEE/ACM International Conference on Automated Software Engineering, Melbourne, Australia, Mon 21 - Fri 25 September 2020

  47. arXiv:2006.07416  [pdf, other

    cs.SE

    Defect Reduction Planning (using TimeLIME)

    Authors: Kewen Peng, Tim Menzies

    Abstract: Software comes in releases. An implausible change to software is something that has never been changed in prior releases. When planning how to reduce defects, it is better to use plausible changes, i.e., changes with some precedence in the prior releases. To demonstrate these points, this paper compares several defect reduction planning tools. LIME is a local sensitivity analysis tool that can r… ▽ More

    Submitted 15 February, 2021; v1 submitted 12 June, 2020; originally announced June 2020.

    Comments: 15 pages, 5 figures, 12 tables, accepted by TSE. arXiv admin note: substantial text overlap with arXiv:2003.06887

  48. arXiv:2006.07240  [pdf, other

    cs.SE

    Predicting Health Indicators for Open Source Projects (using Hyperparameter Optimization)

    Authors: Tianpei Xia, Wei Fu, Rui Shu, Rishabh Agrawal, Tim Menzies

    Abstract: Software developed on public platform is a source of data that can be used to make predictions about those projects. While the individual developing activity may be random and hard to predict, the developing behavior on project level can be predicted with good accuracy when large groups of developers work together on software projects. To demonstrate this, we use 64,181 months of data from 1,159… ▽ More

    Submitted 17 March, 2022; v1 submitted 12 June, 2020; originally announced June 2020.

    Comments: Accepted to EMSE 2022

  49. Assessing Practitioner Beliefs about Software Engineering

    Authors: N. C. Shrikanth, William Nichols, Fahmid Morshed Fahid, Tim Menzies

    Abstract: Software engineering is a highly dynamic discipline. Hence, as times change, so too might our beliefs about core processes in this field. This paper checks some five beliefs that originated in the past decades that comment on the relationships between (i) developer productivity; (ii) software quality and (iii) years of developer experience. Using data collected from 1,356 developers in the period… ▽ More

    Submitted 24 May, 2021; v1 submitted 9 June, 2020; originally announced June 2020.

    Comments: 32 pages, published https://link.springer.com/article/10.1007/s10664-021-09957-5

  50. arXiv:2006.00444  [pdf, other

    cs.SE

    Learning to Recognize Actionable Static Code Warnings (is Intrinsically Easy)

    Authors: Xueqi Yang, Jianfeng Chen, Rahul Yedida, Zhe Yu, Tim Menzies

    Abstract: Static code warning tools often generate warnings that programmers ignore. Such tools can be made more useful via data mining algorithms that select the "actionable" warnings; i.e. the warnings that are usually not ignored. In this paper, we look for actionable warnings within a sample of 5,675 actionable warnings seen in 31,058 static code warnings from FindBugs. We find that data mining algori… ▽ More

    Submitted 10 January, 2021; v1 submitted 31 May, 2020; originally announced June 2020.

    Comments: 24 pages, 5 figures, 7 tables, accepted to Empirical Software Engineering and to appear

  51. Fairway: A Way to Build Fair ML Software

    Authors: Joymallya Chakraborty, Suvodeep Majumder, Zhe Yu, Tim Menzies

    Abstract: Machine learning software is increasingly being used to make decisions that affect people's lives. But sometimes, the core part of this software (the learned model), behaves in a biased manner that gives undue advantages to a specific group of people (where those groups are determined by sex, race, etc.). This "algorithmic discrimination" in the AI software systems has become a matter of serious c… ▽ More

    Submitted 6 October, 2020; v1 submitted 23 March, 2020; originally announced March 2020.

    Comments: ESEC/FSE'20: The 28th ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering

    Journal ref: ESEC/FSE'2020: The 28th ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering, Sacramento, California, United States, November 8-13, 2020

  52. arXiv:2003.06887  [pdf, other

    cs.SE

    How to Improve AI Tools (by Adding in SE Knowledge): Experiments with the TimeLIME Defect Reduction Tool

    Authors: Kewen Peng, Tim Menzies

    Abstract: AI algorithms are being used with increased frequency in SE research and practice. Such algorithms are usually commissioned and certified using data from outside the SE domain. Can we assume that such algorithms can be used ''off-the-shelf'' (i.e. with no modifications)? To say that another way, are there special features of SE problems that suggest a different and better way to use AI tools? To… ▽ More

    Submitted 15 March, 2020; originally announced March 2020.

    Comments: 11 pages, 7 figures, submitted to FSE

  53. arXiv:2003.05922  [pdf, other

    cs.SE

    The Changing Nature of Computational Science Software

    Authors: Huy Tu, Rishabh Agrawal, Tim Menzies

    Abstract: How should software engineering be adapted for Computational Science (CS)? If we understood that, then we could better support software sustainability, verifiability, reproducibility, comprehension, and usability for CS community. For example, improving the maintainability of the CS code could lead to: (a) faster adaptation of scientific project simulations to new and efficient hardware (multi-cor… ▽ More

    Submitted 12 March, 2020; originally announced March 2020.

    Comments: 10 pages, submitted for FSE

  54. Identifying Self-Admitted Technical Debts with Jitterbug: A Two-step Approach

    Authors: Zhe Yu, Fahmid Morshed Fahid, Huy Tu, Tim Menzies

    Abstract: Keeping track of and managing Self-Admitted Technical Debts (SATDs) are important to maintaining a healthy software project. This requires much time and effort from human experts to identify the SATDs manually. The current automated solutions do not have satisfactory precision and recall in identifying SATDs to fully automate the process. To solve the above problems, we propose a two-step framewor… ▽ More

    Submitted 16 October, 2020; v1 submitted 25 February, 2020; originally announced February 2020.

    Comments: 14 pages, 3 pages for appendix, 6+3 figures, 10 tables. Accepted by TSE journal

  55. Assessing Practitioner Beliefs about Software Defect Prediction

    Authors: N. C. Shrikanth, Tim Menzies

    Abstract: Just because software developers say they believe in "X", that does not necessarily mean that "X" is true. As shown here, there exist numerous beliefs listed in the recent Software Engineering literature which are only supported by small portions of the available data. Hence we ask what is the source of this disconnect between beliefs and evidence?. To answer this question we look for evidence for… ▽ More

    Submitted 8 April, 2020; v1 submitted 20 December, 2019; originally announced December 2019.

    Comments: 9 pages, 3 Figures, 4 Tables, ICSE SEIP 2020

  56. arXiv:1912.04189  [pdf, other

    cs.SE

    Sequential Model Optimization for Software Process Control

    Authors: Tianpei Xia, Rui Shu, Xipeng Shen, Tim Menzies

    Abstract: Many methods have been proposed to estimate how much effort is required to build and maintain software. Much of that research assumes a ``classic'' waterfall-based approach rather than contemporary projects (where the developing process may be more iterative than linear in nature). Also, much of that work tries to recommend a single method-- an approach that makes the dubious assumption that one m… ▽ More

    Submitted 17 February, 2020; v1 submitted 9 December, 2019; originally announced December 2019.

  57. arXiv:1912.04061  [pdf, other

    cs.SE cs.AI cs.LG

    Simpler Hyperparameter Optimization for Software Analytics: Why, How, When?

    Authors: Amritanshu Agrawal, Xueqi Yang, Rishabh Agrawal, Rahul Yedida, Xipeng Shen, Tim Menzies

    Abstract: How can we make software analytics simpler and faster? One method is to match the complexity of analysis to the intrinsic complexity of the data being explored. For example, hyperparameter optimizers find the control settings for data miners that improve the predictions generated via software analytics. Sometimes, very fast hyperparameter optimization can be achieved by "DODGE-ing"; i.e. simply st… ▽ More

    Submitted 22 April, 2021; v1 submitted 9 December, 2019; originally announced December 2019.

    Comments: 15 pages

    Journal ref: Transactions on Software Engineering, 2021

  58. arXiv:1911.04250  [pdf, other

    cs.SE cs.LG

    Methods for Stabilizing Models across Large Samples of Projects (with case studies on Predicting Defect and Project Health)

    Authors: Suvodeep Majumder, Tianpei Xia, Rahul Krishna, Tim Menzies

    Abstract: Despite decades of research, SE lacks widely accepted models (that offer precise quantitative stable predictions) about what factors most influence software quality. This paper provides a promising result showing such stable models can be generated using a new transfer learning framework called "STABILIZER". Given a tree of recursively clustered projects (using project meta-data), STABILIZER promo… ▽ More

    Submitted 21 March, 2022; v1 submitted 6 November, 2019; originally announced November 2019.

    Comments: 12 pages, 4 figures, 5 Tables

  59. arXiv:1911.02476  [pdf, other

    cs.SE

    How to Better Distinguish Security Bug Reports (using Dual Hyperparameter Optimization

    Authors: Rui Shu, Tianpei Xia, Jianfeng Chen, Laurie Williams, Tim Menzies

    Abstract: Background: In order that the general public is not vulnerable to hackers, security bug reports need to be handled by small groups of engineers before being widely discussed. But learning how to distinguish the security bug reports from other bug reports is challenging since they may occur rarely. Data mining methods that can find such scarce targets require extensive optimization effort. Goal:… ▽ More

    Submitted 17 March, 2021; v1 submitted 4 November, 2019; originally announced November 2019.

    Comments: arXiv admin note: substantial text overlap with arXiv:1905.06872

  60. arXiv:1911.01817  [pdf, other

    cs.SE

    Whence to Learn? Transferring Knowledge in Configurable Systems using BEETLE

    Authors: Rahul Krishna, Vivek Nair, Pooyan Jamshidi, Tim Menzies

    Abstract: As software systems grow in complexity and the space of possible configurations increases exponentially, finding the near-optimal configuration of a software system becomes challenging. Recent approaches address this challenge by learning performance models based on a sample set of configurations. However, collecting enough sample configurations can be very expensive since each such sample require… ▽ More

    Submitted 25 March, 2020; v1 submitted 1 November, 2019; originally announced November 2019.

    Comments: Accepted, to appear in IEEE TSE. arXiv admin note: text overlap with arXiv:1803.03900

  61. arXiv:1911.01387  [pdf, other

    cs.SE

    Understanding Static Code Warnings: an Incremental AI Approach

    Authors: Xueqi Yang, Zhe Yu, Junjie Wang, Tim Menzies

    Abstract: Knowledge-based systems reason over some knowledge base. Hence, an important issue for such systems is how to acquire the knowledge needed for their inference. This paper assesses active learning methods for acquiring knowledge for "static code warnings". Static code analysis is a widely-used method for detecting bugs and security vulnerabilities in software systems. As software becomes more com… ▽ More

    Submitted 22 October, 2020; v1 submitted 4 November, 2019; originally announced November 2019.

    Comments: Accepted to Expert Systems with Applications

  62. arXiv:1909.07249  [pdf, other

    cs.SE

    Assessing Expert System-Assisted Literature Reviews With a Case Study

    Authors: Zhe Yu, Jeffrey C. Carver, Gregg Rothermel, Tim Menzies

    Abstract: Given the large number of publications in software engineering, frequent literature reviews are required to keep current on work in specific areas. One tedious work in literature reviews is to find relevant studies amongst thousands of non-relevant search results. In theory, expert systems can assist in finding relevant work but those systems have primarily been tested in simulations rather than i… ▽ More

    Submitted 8 April, 2022; v1 submitted 16 September, 2019; originally announced September 2019.

    Comments: 23+8 pages, 9 figures, 3 tables. Accepted by Expert Systems with Applications

  63. arXiv:1905.08297  [pdf, other

    cs.SE cs.AI

    Better Technical Debt Detection via SURVEYing

    Authors: Fahmid M. Fahid, Zhe Yu, Tim Menzies

    Abstract: Software analytics can be improved by surveying; i.e. rechecking and (possibly) revising the labels offered by prior analysis. Surveying is a time-consuming task and effective surveyors must carefully manage their time. Specifically, they must balance the cost of further surveying against the additional benefits of that extra effort. This paper proposes SURVEY0, an incremental Logistic Regression… ▽ More

    Submitted 20 May, 2019; originally announced May 2019.

    Comments: 10 pages, 4 figures, 4 tables, conference

  64. TERMINATOR: Better Automated UI Test Case Prioritization

    Authors: Zhe Yu, Fahmid M. Fahid, Tim Menzies, Gregg Rothermel, Kyle Patrick, Snehit Cherian

    Abstract: Automated UI testing is an important component of the continuous integration process of software development. A modern web-based UI is an amalgam of reports from dozens of microservices written by multiple teams. Queries on a page that opens up another will fail if any of that page's microservices fails. As a result, the overall cost for automated UI testing is high since the UI elements cannot be… ▽ More

    Submitted 18 June, 2019; v1 submitted 16 May, 2019; originally announced May 2019.

    Comments: 10+2 pages, 4 figures, 3 tables, ESEC/FSE 2019 industry track

  65. arXiv:1905.06872  [pdf, other

    cs.SE

    Better Security Bug Report Classification via Hyperparameter Optimization

    Authors: Rui Shu, Tianpei Xia, Laurie Williams, Tim Menzies

    Abstract: When security bugs are detected, they should be (a)~discussed privately by security software engineers; and (b)~not mentioned to the general public until security patches are available. Software engineers usually report bugs to bug tracking system, and label them as security bug reports (SBRs) or not-security bug reports (NSBRs), while SBRs have a higher priority to be fixed before exploited by at… ▽ More

    Submitted 16 May, 2019; originally announced May 2019.

    Comments: 12 pages, 1 figure, submitted to 34th IEEE/ACM International Conference on Automated Software Engineering (ASE 2019)

  66. Predicting Breakdowns in Cloud Services (with SPIKE)

    Authors: Jianfeng Chen, Joymallya Chakraborty, Philip Clark, Kevin Haverlock, Snehit Cherian, Tim Menzies

    Abstract: Maintaining web-services is a mission-critical task where any down-time means loss of revenue and reputation (of being a reliable service provider). In the current competitive web services market, such a loss of reputation causes extensive loss of future revenue. To address this issue, we developed SPIKE, a data mining tool which can predict upcoming service breakdowns, half an hour into the futur… ▽ More

    Submitted 14 June, 2019; v1 submitted 15 May, 2019; originally announced May 2019.

    Comments: 9 pages, 6 figures, in Proceedings of The 27th ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering (ESEC/FSE'19), industry track

  67. arXiv:1905.05786  [pdf, ps, other

    cs.SE cs.LG

    Software Engineering for Fairness: A Case Study with Hyperparameter Optimization

    Authors: Joymallya Chakraborty, Tianpei Xia, Fahmid M. Fahid, Tim Menzies

    Abstract: We assert that it is the ethical duty of software engineers to strive to reduce software discrimination. This paper discusses how that might be done. This is an important topic since machine learning software is increasingly being used to make decisions that affect people's lives. Potentially, the application of that software will result in fairer decisions because (unlike humans) machine learning… ▽ More

    Submitted 30 October, 2019; v1 submitted 14 May, 2019; originally announced May 2019.

  68. arXiv:1905.05358  [pdf, other

    cs.SE

    Building Very Small Test Suites (with Snap)

    Authors: Jianfeng Chen, Xipeng Shen, Tim Menzies

    Abstract: Software is now so large and complex that additional architecture is needed to guide theorem provers as they try to generate test suites. For example, the SNAP test suite generator (introduced in this paper) combines the Z3 theorem prover with the following tactic: sample around the average values seen in a few randomly selected valid tests. This tactic is remarkably effective. For 27 real-world p… ▽ More

    Submitted 14 July, 2020; v1 submitted 13 May, 2019; originally announced May 2019.

    Comments: 13 pages, 9 figures, submitted to IEEE Transactions on Software Engineering (Journal First)

  69. arXiv:1905.01719  [pdf, other

    cs.SE

    Better Data Labelling with EMBLEM (and how that Impacts Defect Prediction)

    Authors: Huy Tu, Zhe Yu, Tim Menzies

    Abstract: Standard automatic methods for recognizing problematic development commits can be greatly improved via the incremental application of human+artificial expertise. In this approach, called EMBLEM, an AI tool first explore the software development process to label commits that are most problematic. Humans then apply their expertise to check those labels (perhaps resulting in the AI updating the suppo… ▽ More

    Submitted 7 April, 2020; v1 submitted 5 May, 2019; originally announced May 2019.

    Comments: 17 pages, 2 pages references, submitted for TSE journal

  70. arXiv:1904.09954  [pdf, other

    cs.SE

    Communication and Code Dependency Effects on Software Code Quality: An Empirical Analysis of Herbsleb Hypothesis

    Authors: Suvodeep Majumder, Joymallya Chakraborty, Amritanshu Agrawal, Tim Menzies

    Abstract: Prior literature has suggested that in many projects 80\% or more of the contributions are made by a small called group of around 20% of the development team. Most prior studies deprecate a reliance on such a small inner group of "heroes", arguing that it causes bottlenecks in development and communication. Despite this, such projects are very common in open source projects. So what exactly is the… ▽ More

    Submitted 21 March, 2022; v1 submitted 22 April, 2019; originally announced April 2019.

    Comments: 12 pages, 7 figures, 2 tables

  71. arXiv:1904.05794  [pdf, other

    cs.SE

    Assessing Developer Beliefs: A Reply to "Perceptions, Expectations, and Challenges in Defect Prediction"

    Authors: Shrikanth N. C., Tim Menzies

    Abstract: It can be insightful to extend qualitative studies with a secondary quantitative analysis (where the former suggests insightful questions that the latter can answer). Documenting developer beliefs should be the start, not the end, of Software Engineering research. Once prevalent beliefs are found, they should be checked against real-world data. For example, this paper finds several notable discrep… ▽ More

    Submitted 11 April, 2019; originally announced April 2019.

    Comments: 3 pages, 1 figure, 2 tables, Submitted to TSE Journal

  72. arXiv:1902.04060  [pdf, other

    cs.SE

    Replication Can Improve Prior Results: A GitHub Study of Pull Request Acceptance

    Authors: Di Chen, Kathyrn Stolee, Tim Menzies

    Abstract: Crowdsourcing and data mining can be used to effectively reduce the effort associated with the partial replication and enhancement of qualitative studies. For example, in a primary study, other researchers explored factors influencing the fate of GitHub pull requests using an extensive qualitative analysis of 20 pull requests. Guided by their findings, we mapped some of their qualitative insight… ▽ More

    Submitted 8 February, 2019; originally announced February 2019.

    Comments: 12 pages, submitted to ICPC 2019. arXiv admin note: substantial text overlap with arXiv:1702.08571

  73. arXiv:1902.01838  [pdf, other

    cs.SE cs.AI cs.LG cs.NE

    How to "DODGE" Complex Software Analytics?

    Authors: Amritanshu Agrawal, Wei Fu, Di Chen, Xipeng Shen, Tim Menzies

    Abstract: Machine learning techniques applied to software engineering tasks can be improved by hyperparameter optimization, i.e., automatic tools that find good settings for a learner's control parameters. We show that such hyperparameter optimization can be unnecessarily slow, particularly when the optimizers waste time exploring "redundant tunings"', i.e., pairs of tunings which lead to indistinguishabl… ▽ More

    Submitted 1 December, 2019; v1 submitted 5 February, 2019; originally announced February 2019.

    Comments: 13 Pages, Accepted to IEEE Transactions in Software Engineering, 2019

  74. Better Software Analytics via "DUO": Data Mining Algorithms Using/Used-by Optimizers

    Authors: Amritanshu Agrawal, Tim Menzies, Leandro L. Minku, Markus Wagner, Zhe Yu

    Abstract: This paper claims that a new field of empirical software engineering research and practice is emerging: data mining using/used-by optimizers for empirical studies or DUO. For example, data miners can generate models that are explored by optimizers. Also, optimizers can advise how to best adjust the control parameters of a data miner. This combined approach acts like an agent leaning over the shoul… ▽ More

    Submitted 13 December, 2019; v1 submitted 4 December, 2018; originally announced December 2018.

    Comments: 35 Pages, Accepted to EMSE Journal

  75. Total Recall, Language Processing, and Software Engineering

    Authors: Zhe Yu, Tim Menzies

    Abstract: A broad class of software engineering problems can be generalized as the "total recall problem". This short paper claims that identifying and exploring total recall language processing problems in software engineering is an important task with wide applicability. To make that case, we show that by applying and adapting the state of the art active learning and text mining, solutions of the total… ▽ More

    Submitted 31 August, 2018; originally announced September 2018.

    Comments: 4 pages, 2 figures. Submitted to NL4SE@ESEC/FSE 2018

  76. arXiv:1805.12124  [pdf, other

    cs.DL

    Better Metrics for Ranking SE Researchers

    Authors: George Mathew, Tim Menzies

    Abstract: This paper studies how SE researchers are ranked using a variety of metrics and data from 35,406 authors of 35,391 papers from 34 top SE venues in the period 1992-2016. Based on that analysis, we: deprecate the widely used "h-index", favoring instead an alternate Weighted Page Rank(PR_W) metric that is somewhat analogous to the PageRank(PR) metric developed at Google. Unlike the h-index, PR_W rewa… ▽ More

    Submitted 29 May, 2018; originally announced May 2018.

    Comments: 5 pages, 2 figures, 4 tables. Submitted to IEEE TSE 2019

  77. Crowdtesting : When is The Party Over?

    Authors: Junjie Wang, Ye Yang, Zhe Yu, Tim Menzies, Qing Wang

    Abstract: Trade-offs such as "how much testing is enough" are critical yet challenging project decisions in software engineering. Most existing approaches adopt risk-driven or value-based analysis to prioritize test cases and minimize test runs. However, none of these is applicable to the emerging crowd testing paradigm where task requesters typically have no control over online crowdworkers's dynamic behav… ▽ More

    Submitted 8 May, 2018; originally announced May 2018.

    Comments: 12 pages

    Journal ref: iSENSE: completion-aware crowdtesting management. ICSE 2019: 912-923

  78. Cutting Away the Confusion From Crowdtesting

    Authors: Junjie Wang, Mingyang Li, Song Wang, Tim Menzies, Qing Wang

    Abstract: Crowdtesting is effective especially when it comes to the feedback on GUI systems, or subjective opinions about features. Despite of this, we find crowdtesting reports are highly replicated, i.e., 82% of them are replicates of others. Hence automatically detecting replicate reports could help reduce triaging efforts. Most of the existing approaches mainly adopted textual information for replicate… ▽ More

    Submitted 13 November, 2018; v1 submitted 7 May, 2018; originally announced May 2018.

    Comments: 12 pages

    Journal ref: Information and Software Technology. 110 (2019). 139-155

  79. arXiv:1805.02744  [pdf, other

    cs.SE

    Effective Automated Decision Support for Managing Crowdtesting

    Authors: Junjie Wang, Ye Yang, Rahul Krishna, Tim Menzies, Qing Wang

    Abstract: Crowdtesting has grown to be an effective alter-native to traditional testing, especially in mobile apps. However,crowdtesting is hard to manage in nature. Given the complexity of mobile applications and unpredictability of distributed, parallel crowdtesting process, it is difficult to estimate (a) the remaining number of bugs as yet undetected or (b) the required cost to find those bugs. Experien… ▽ More

    Submitted 7 May, 2018; originally announced May 2018.

    Comments: 12 pages

  80. arXiv:1805.00336  [pdf, other

    cs.SE

    Hyperparameter Optimization for Effort Estimation

    Authors: Tianpei Xia, Rahul Krishna, Jianfeng Chen, George Mathew, Xipeng Shen, Tim Menzies

    Abstract: Software analytics has been widely used in software engineering for many tasks such as generating effort estimates for software projects. One of the "black arts" of software analytics is tuning the parameters controlling a data mining algorithm. Such hyperparameter optimization has been widely studied in other software analytics domains (e.g. defect prediction and text mining) but, so far, has not… ▽ More

    Submitted 31 January, 2019; v1 submitted 27 April, 2018; originally announced May 2018.

    Comments: Submitted to EMSE

  81. arXiv:1804.10657  [pdf, other

    cs.SE

    Can You Explain That, Better? Comprehensible Text Analytics for SE Applications

    Authors: Amritanshu Agrawal, Huy Tu, Tim Menzies

    Abstract: Text mining methods are used for a wide range of Software Engineering (SE) tasks. The biggest challenge of text mining is high dimensional data, i.e., a corpus of documents can contain $10^4$ to $10^6$ unique words. To address this complexity, some very convoluted text mining methods have been applied. Is that complexity necessary? Are there simpler ways to quickly generate models that perform as… ▽ More

    Submitted 27 April, 2018; originally announced April 2018.

    Comments: 10+2 pages, submitted to ASE 2018

  82. arXiv:1804.00626  [pdf, other

    cs.SE

    Why Software Effort Estimation Needs SBSE

    Authors: Tianpei Xia, Jianfeng Chen, George Mathew, Xipeng Shen, Tim Menzies

    Abstract: Industrial practitioners now face a bewildering array of possible configurations for effort estimation. How to select the best one for a particular dataset? This paper introduces OIL (short for optimized learning), a novel configuration tool for effort estimation based on differential evolution. When tested on 945 software projects, OIL significantly improved effort estimations, after exploring… ▽ More

    Submitted 2 April, 2018; originally announced April 2018.

    Comments: 15 pages, submitted to SSBSE'18

  83. Improving Vulnerability Inspection Efficiency Using Active Learning

    Authors: Zhe Yu, Christopher Theisen, Laurie Williams, Tim Menzies

    Abstract: Software engineers can find vulnerabilities with less effort if they are directed towards code that might contain more vulnerabilities. HARMLESS is an incremental support vector machine tool that builds a vulnerability prediction model from the sourcecode inspected to date, then suggests what source code files should be inspected next. In this way, HARMLESS can reduce the time and effort required… ▽ More

    Submitted 26 October, 2019; v1 submitted 17 March, 2018; originally announced March 2018.

    Comments: 17+1 pages, 4 figures, 7 tables. Accepted by IEEE Transactions on Software Engineering

  84. arXiv:1803.05587  [pdf, other

    cs.DC

    Micky: A Cheaper Alternative for Selecting Cloud Instances

    Authors: Chin-Jung Hsu, Vivek Nair, Tim Menzies, Vincent Freeh

    Abstract: Most cloud computing optimizers explore and improve one workload at a time. When optimizing many workloads, the single-optimizer approach can be prohibitively expensive. Accordingly, we examine "collective optimizer" that concurrently explore and improve a set of workloads significantly reducing the measurement costs. Our large-scale empirical study shows that there is often a single cloud configu… ▽ More

    Submitted 15 March, 2018; originally announced March 2018.

  85. arXiv:1803.05518  [pdf, other

    cs.SE

    Bad Smells in Software Analytics Papers

    Authors: Tim Menzies, Martin Shepperd

    Abstract: CONTEXT: There has been a rapid growth in the use of data analytics to underpin evidence-based software engineering. However the combination of complex techniques, diverse reporting standards and poorly understood underlying phenomena are causing some concern as to the reliability of studies. OBJECTIVE: Our goal is to provide guidance for producers and consumers of software analytics studies (co… ▽ More

    Submitted 12 April, 2019; v1 submitted 14 March, 2018; originally announced March 2018.

    Comments: Accepted April 2019. To appear

    Journal ref: Information Software Technology, 2019

  86. arXiv:1803.05067  [pdf, other

    cs.SE

    Applications of Psychological Science for Actionable Analytics

    Authors: Di Chen, Wei Fu, Rahul Krishna, Tim Menzies

    Abstract: Actionable analytics are those that humans can understand, and operationalize. What kind of data mining models generate such actionable analytics? According to psychological scientists, humans understand models that most match their own internal models, which they characterize as lists of "heuristic" (i.e., lists of very succinct rules). One such heuristic rule generator is the Fast-and-Frugal Tre… ▽ More

    Submitted 13 March, 2018; originally announced March 2018.

    Comments: 10 pages, 5 figures

  87. arXiv:1803.04608  [pdf, other

    cs.SE

    Building Better Quality Predictors Using "$ε$-Dominance"

    Authors: Wei Fu, Tim Menzies, Di Chen, Amritanshu Agrawal

    Abstract: Despite extensive research, many methods in software quality prediction still exhibit some degree of uncertainty in their results. Rather than treating this as a problem, this paper asks if this uncertainty is a resource that can simplify software quality prediction. For example, Deb's principle of $ε$-dominance states that if there exists some $ε$ value below which it is useless or impossible t… ▽ More

    Submitted 12 March, 2018; originally announced March 2018.

    Comments: 10 pages

  88. arXiv:1803.03900  [pdf, other

    cs.SE

    Transfer Learning with Bellwethers to find Good Configurations

    Authors: Vivek Nair, Rahul Krishna, Tim Menzies, Pooyan Jamshidi

    Abstract: As software systems grow in complexity, the space of possible configurations grows exponentially. Within this increasing complexity, developers, maintainers, and users cannot keep track of the interactions between all the various configuration options. Finding the optimally performing configuration of a software system for a given setting is challenging. Recent approaches address this challenge by… ▽ More

    Submitted 14 March, 2018; v1 submitted 10 March, 2018; originally announced March 2018.

  89. arXiv:1803.01296  [pdf, other

    cs.DC

    Scout: An Experienced Guide to Find the Best Cloud Configuration

    Authors: Chin-Jung Hsu, Vivek Nair, Tim Menzies, Vincent W. Freeh

    Abstract: Finding the right cloud configuration for workloads is an essential step to ensure good performance and contain running costs. A poor choice of cloud configuration decreases application performance and increases running cost significantly. While Bayesian Optimization is effective and applicable to any workloads, it is fragile because performance and workload are hard to model (to predict). In th… ▽ More

    Submitted 3 March, 2018; originally announced March 2018.

  90. arXiv:1802.05319  [pdf, other

    cs.SE cs.LG stat.ML

    500+ Times Faster Than Deep Learning (A Case Study Exploring Faster Methods for Text Mining StackOverflow)

    Authors: Suvodeep Majumder, Nikhila Balaji, Katie Brey, Wei Fu, Tim Menzies

    Abstract: Deep learning methods are useful for high-dimensional data and are becoming widely used in many areas of software engineering. Deep learners utilizes extensive computational power and can take a long time to train-- making it difficult to widely validate and repeat and improve their results. Further, they are not the best solution in all domains. For example, recent results show that for finding r… ▽ More

    Submitted 14 February, 2018; originally announced February 2018.

    Journal ref: MSR '18, Proceedings of the 15th International Conference on Mining Software Repositories, May 2018, Pages 554 to 563

  91. Data-Driven Search-based Software Engineering

    Authors: Vivek Nair, Amritanshu Agrawal, Jianfeng Chen, Wei Fu, George Mathew, Tim Menzies, Leandro Minku, Markus Wagner, Zhe Yu

    Abstract: This paper introduces Data-Driven Search-based Software Engineering (DSE), which combines insights from Mining Software Repositories (MSR) and Search-based Software Engineering (SBSE). While MSR formulates software engineering problems as data mining problems, SBSE reformulates SE problems as optimization problems and use meta-heuristic algorithms to solve them. Both MSR and SBSE share the common… ▽ More

    Submitted 16 March, 2018; v1 submitted 30 January, 2018; originally announced January 2018.

    Comments: 10 pages, 7 figures

  92. arXiv:1801.02175  [pdf, other

    cs.SE

    Finding Faster Configurations using FLASH

    Authors: Vivek Nair, Zhe Yu, Tim Menzies, Norbert Siegmund, Sven Apel

    Abstract: Finding good configurations for a software system is often challenging since the number of configuration options can be large. Software engineers often make poor choices about configuration or, even worse, they usually use a sub-optimal configuration in production, which leads to inadequate performance. To assist engineers in finding the (near) optimal configuration, this paper introduces FLASH, a… ▽ More

    Submitted 1 September, 2018; v1 submitted 7 January, 2018; originally announced January 2018.

  93. arXiv:1712.10081  [pdf, other

    cs.DC

    Low-Level Augmented Bayesian Optimization for Finding the Best Cloud VM

    Authors: Chin-Jung Hsu, Vivek Nair, Vincent W. Freeh, Tim Menzies

    Abstract: With the advent of big data applications, which tends to have longer execution time, choosing the right cloud VM to run these applications has significant performance as well as economic implications. For example, in our large-scale empirical study of 107 different workloads on three popular big data systems, we found that a wrong choice can lead to a 20 times slowdown or an increase in cost by 10… ▽ More

    Submitted 28 December, 2017; originally announced December 2017.

  94. We Don't Need Another Hero? The Impact of "Heroes" on Software Development

    Authors: Amritanshu Agrawal, Akond Rahman, Rahul Krishna, Alexander Sobran, Tim Menzies

    Abstract: A software project has "Hero Developers" when 80% of contributions are delivered by 20% of the developers. Are such heroes a good idea? Are too many heroes bad for software quality? Is it better to have more/less heroes for different kinds of projects? To answer these questions, we studied 661 open source projects from Public open source software (OSS) Github and 171 projects from an Enterprise Gi… ▽ More

    Submitted 20 February, 2018; v1 submitted 24 October, 2017; originally announced October 2017.

    Comments: 8 pages + 1 references, Accepted to International conference on Software Engineering - Software Engineering in Practice, 2018

    Journal ref: International conference on Software Engineering - Software Engineering in Practice, 2018

  95. What is the Connection Between Issues, Bugs, and Enhancements? (Lessons Learned from 800+ Software Projects)

    Authors: Rahul Krishna, Amritanshu Agrawal, Akond Rahman, Alexander Sobran, Tim Menzies

    Abstract: Agile teams juggle multiple tasks so professionals are often assigned to multiple projects, especially in service organizations that monitor and maintain a large suite of software for a large user base. If we could predict changes in project conditions changes, then managers could better adjust the staff allocated to those projects.This paper builds such a predictor using data from 832 open source… ▽ More

    Submitted 5 September, 2018; v1 submitted 24 October, 2017; originally announced October 2017.

    Comments: Accepted to 2018 International Conference on Software Engineering, at the software engineering in practice track. 10 pages, 10 figures

  96. arXiv:1708.08127  [pdf, other

    cs.SE cs.AI cs.NE

    RIOT: a Stochastic-based Method for Workflow Scheduling in the Cloud

    Authors: Jianfeng Chen, Tim Menzies

    Abstract: Cloud computing provides engineers or scientists a place to run complex computing tasks. Finding a workflow's deployment configuration in a cloud environment is not easy. Traditional workflow scheduling algorithms were based on some heuristics, e.g. reliability greedy, cost greedy, cost-time balancing, etc., or more recently, the meta-heuristic methods, such as genetic algorithms. These methods ar… ▽ More

    Submitted 22 April, 2018; v1 submitted 27 August, 2017; originally announced August 2017.

    Comments: 8 pages, 4 figures, 3 tables. In Proceedings of IEEE international conference on Cloud Computing'18

  97. arXiv:1708.05442  [pdf, other

    cs.SE

    Learning Actionable Analytics from Multiple Software Projects

    Authors: Rahul Krishna, Tim Menzies

    Abstract: The current generation of software analytics tools are mostly prediction algorithms (e.g. support vector machines, naive bayes, logistic regression, etc). While prediction is useful, after prediction comes planning about what actions to take in order to improve quality. This research seeks methods that generate demonstrably useful guidance on "what to do" within the context of a specific software… ▽ More

    Submitted 24 January, 2020; v1 submitted 17 August, 2017; originally announced August 2017.

    Comments: Accepted at Empirical Software Engineering Journal

  98. FAST$^2$: an Intelligent Assistant for Finding Relevant Papers

    Authors: Zhe Yu, Tim Menzies

    Abstract: Literature reviews are essential for any researcher trying to keep up to date with the burgeoning software engineering literature. FAST$^2$ is a novel tool for reducing the effort required for conducting literature reviews by assisting the researchers to find the next promising paper to read (among a set of unread papers). This paper describes FAST$^2$ and tests it on four large software engineeri… ▽ More

    Submitted 11 November, 2018; v1 submitted 15 May, 2017; originally announced May 2017.

    Comments: 20+3 pages, 6 figures, 5 tables, and 4 algorithms. Accepted by Journal of Expert Systems with Applications

    MSC Class: 68N01; 68T50 ACM Class: D.2.0; I.2.7

  99. arXiv:1705.05018  [pdf, other

    cs.SE

    FLASH: A Faster Optimizer for SBSE Tasks

    Authors: Vivek Nair, Zhe Yu, Tim Menzies

    Abstract: Most problems in search-based software engineering involve balancing conflicting objectives. Prior approaches to this task have required a large number of evaluations- making them very slow to execute and very hard to comprehend. To solve these problems, this paper introduces FLASH, a decision tree based optimizer that incrementally grows one decision tree per objective. These trees are then used… ▽ More

    Submitted 18 May, 2017; v1 submitted 14 May, 2017; originally announced May 2017.

    Comments: 11 pages, 11 figures

  100. Is "Better Data" Better than "Better Data Miners"? (On the Benefits of Tuning SMOTE for Defect Prediction)

    Authors: Amritanshu Agrawal, Tim Menzies

    Abstract: We report and fix an important systematic error in prior studies that ranked classifiers for software analytics. Those studies did not (a) assess classifiers on multiple criteria and they did not (b) study how variations in the data affect the results. Hence, this paper applies (a) multi-criteria tests while (b) fixing the weaker regions of the training data (using SMOTUNED, which is a self-tuning… ▽ More

    Submitted 20 February, 2018; v1 submitted 10 May, 2017; originally announced May 2017.

    Comments: 10 pages + 2 references. Accepted to International Conference of Software Engineering (ICSE), 2018

    Journal ref: International Conference of Software Engineering (ICSE), 2018

  101. arXiv:1703.06218  [pdf, ps, other

    cs.SE

    Bellwethers: A Baseline Method For Transfer Learning

    Authors: Rahul Krishna, Tim Menzies

    Abstract: Software analytics builds quality prediction models for software projects. Experience shows that (a) the more projects studied, the more varied are the conclusions; and (b) project managers lose faith in the results of software analytics if those results keep changing. To reduce this conclusion instability, we propose the use of "bellwethers": given N projects from a community the bellwether is th… ▽ More

    Submitted 21 January, 2018; v1 submitted 17 March, 2017; originally announced March 2017.

    Comments: 23 Pages

  102. Easy over Hard: A Case Study on Deep Learning

    Authors: Wei Fu, Tim Menzies

    Abstract: While deep learning is an exciting new technique, the benefits of this method need to be assessed with respect to its computational cost. This is particularly important for deep learning since these learners need hours (to weeks) to train the model. Such long training time limits the ability of (a)~a researcher to test the stability of their conclusion via repeated runs with different random seeds… ▽ More

    Submitted 24 June, 2017; v1 submitted 28 February, 2017; originally announced March 2017.

    Comments: 12 pages, 6 figures, accepted at FSE2017

  103. Revisiting Unsupervised Learning for Defect Prediction

    Authors: Wei Fu, Tim Menzies

    Abstract: Collecting quality data from software projects can be time-consuming and expensive. Hence, some researchers explore "unsupervised" approaches to quality prediction that does not require labelled data. An alternate technique is to use "supervised" approaches that learn models from project data labelled with, say, "defective" or "not-defective". Most researchers use these supervised models since, it… ▽ More

    Submitted 24 June, 2017; v1 submitted 28 February, 2017; originally announced March 2017.

    Comments: 11 pages, 5 figures. Accepted at FSE2017

  104. arXiv:1702.08571  [pdf, other

    cs.SE

    Replicating and Scaling up Qualitative Analysis using Crowdsourcing: A Github-based Case Study

    Authors: Di Chen, Kathryn T. Stolee, Tim Menzies

    Abstract: Due to the difficulties in replicating and scaling up qualitative studies, such studies are rarely verified. Accordingly, in this paper, we leverage the advantages of crowdsourcing (low costs, fast speed, scalable workforce) to replicate and scale-up one state-of-the-art qualitative study. That qualitative study explored 20 GitHub pull requests to learn factors that influence the fate of pull requ… ▽ More

    Submitted 1 March, 2017; v1 submitted 27 February, 2017; originally announced February 2017.

    Comments: Submitted to FSE'17, 12 pages

  105. arXiv:1702.07735  [pdf, other

    cs.SE

    Better Predictors for Issue Lifetime

    Authors: Mitch Rees-Jones, Matthew Martin, Tim Menzies

    Abstract: Predicting issue lifetime can help software developers, managers, and stakeholders effectively prioritize work, allocate development resources, and better understand project timelines. Progress had been made on this prediction problem, but prior work has reported low precision and high false alarms. The latest results also use complex models such as random forests that detract from their readabili… ▽ More

    Submitted 4 April, 2017; v1 submitted 24 February, 2017; originally announced February 2017.

    Comments: 9 pages, 3 figures, 5 tables

  106. Using Bad Learners to find Good Configurations

    Authors: Vivek Nair, Tim Menzies, Norbert Siegmund, Sven Apel

    Abstract: Finding the optimally performing configuration of a software system for a given setting is often challenging. Recent approaches address this challenge by learning performance models based on a sample set of configurations. However, building an accurate performance model can be very expensive (and is often infeasible in practice). The central insight of this paper is that exact performance values (… ▽ More

    Submitted 28 June, 2017; v1 submitted 19 February, 2017; originally announced February 2017.

    Comments: 11 pages, 11 figures

  107. arXiv:1702.05568  [pdf, other

    cs.SE

    "SHORT"er Reasoning About Larger Requirements Models

    Authors: George Mathew, Tim Menzies, Neil A. Ernst, John Klein

    Abstract: When Requirements Engineering(RE) models are unreasonably complex, they cannot support efficient decision making. SHORT is a tool to simplify that reasoning by exploiting the "key" decisions within RE models. These "keys" have the property that once values are assigned to them, it is very fast to reason over the remaining decisions. Using these "keys", reasoning about RE models can be greatly SHOR… ▽ More

    Submitted 22 August, 2017; v1 submitted 17 February, 2017; originally announced February 2017.

    Comments: 10 pages, 5 figures, IEEE Requirements Engineering 2017

  108. arXiv:1701.08106  [pdf, other

    cs.SE cs.LG

    Faster Discovery of Faster System Configurations with Spectral Learning

    Authors: Vivek Nair, Tim Menzies, Norbert Siegmund, Sven Apel

    Abstract: Despite the huge spread and economical importance of configurable software systems, there is unsatisfactory support in utilizing the full potential of these systems with respect to finding performance-optimal configurations. Prior work on predicting the performance of software configurations suffered from either (a) requiring far too many sample configurations or (b) large variances in their predi… ▽ More

    Submitted 3 August, 2017; v1 submitted 27 January, 2017; originally announced January 2017.

    Comments: 26 pages, 6 figures

  109. Beyond Evolutionary Algorithms for Search-based Software Engineering

    Authors: Jianfeng Chen, Vivek Nair, Tim Menzies

    Abstract: Context: Evolutionary algorithms typically require a large number of evaluations (of solutions) to converge - which can be very slow and expensive to evaluate.Objective: To solve search-based software engineering (SE) problems, using fewer evaluations than evolutionary methods.Method: Instead of mutating a small population, we build a very large initial population which is then culled using a recu… ▽ More

    Submitted 17 September, 2017; v1 submitted 27 January, 2017; originally announced January 2017.

    Comments: 17 pages, 10 figures, Information and Software Technology 2017

  110. arXiv:1612.03240  [pdf, other

    cs.SE

    Impacts of Bad ESP (Early Size Predictions) on Software Effort Estimation

    Authors: George Mathew, Tim Menzies, Jairus Hihn

    Abstract: Context: Early size predictions (ESP) can lead to errors in effort predictions for software projects. This problem is particular acute in parametric effort models that give extra weight to size factors (for example, the COCOMO model assumes that effort is exponentially proportional to project size). Objective: To test if effort estimates are crippled by bad ESP. Method: Document inaccuracies in ea… ▽ More

    Submitted 19 February, 2018; v1 submitted 9 December, 2016; originally announced December 2016.

    Comments: 17 pages. Submitted to EMSE journal, 5 figures, 2 tables

  111. Finding Better Active Learners for Faster Literature Reviews

    Authors: Zhe Yu, Nicholas A. Kraft, Tim Menzies

    Abstract: Literature reviews can be time-consuming and tedious to complete. By cataloging and refactoring three state-of-the-art active learning techniques from evidence-based medicine and legal electronic discovery, this paper finds and implements FASTREAD, a faster technique for studying a large corpus of documents. This paper assesses FASTREAD using datasets generated from existing SE literature reviews… ▽ More

    Submitted 2 February, 2018; v1 submitted 9 December, 2016; originally announced December 2016.

    Comments: 23 pages, 5 figures, 3 tables, accepted for publication in EMSE journal

    MSC Class: 68N01; 68T50 ACM Class: D.2.0; I.2.7

  112. arXiv:1609.05563  [pdf, other

    cs.SE

    Negative Results for Software Effort Estimation

    Authors: Tim Menzies, Ye Yang, George Mathew, Barry Boehm, Jairus Hihn

    Abstract: Context:More than half the literature on software effort estimation (SEE) focuses on comparisons of new estimation methods. Surprisingly, there are no studies comparing state of the art latest methods with decades-old approaches. Objective:To check if new SEE methods generated better estimates than older methods. Method: Firstly, collect effort estimation methods ranging from "classical" COCOMO (p… ▽ More

    Submitted 29 September, 2016; v1 submitted 18 September, 2016; originally announced September 2016.

    Comments: 22 pages, EMSE 2016 accepted submission. Affiliated to "Jet Propulsion Laboratory, California Institute of Technology"

  113. Are Delayed Issues Harder to Resolve? Revisiting Cost-to-Fix of Defects throughout the Lifecycle

    Authors: Tim Menzies, William Nichols, Forrest Shull, Lucas Layman

    Abstract: Many practitioners and academics believe in a delayed issue effect (DIE); i.e. the longer an issue lingers in the system, the more effort it requires to resolve. This belief is often used to justify major investments in new development processes that promise to retire more issues sooner. This paper tests for the delayed issue effect in 171 software projects conducted around the world in the peri… ▽ More

    Submitted 15 September, 2016; originally announced September 2016.

    Comments: 31 pages. Accepted with minor revisions to Journal of Empirical Software Engineering. Keywords: software economics, phase delay, cost to fix

    ACM Class: D.2.8

    Journal ref: Empirical Software Engineering, Volume 22, Issue 4, 1903--1935 (2017)

  114. Less is More: Minimizing Code Reorganization using XTREE

    Authors: Rahul Krishna, Tim Menzies, Lucas Layman

    Abstract: Context: Developers use bad code smells to guide code reorganization. Yet developers, text books, tools, and researchers disagree on which bad smells are important. Objective: To evaluate the likelihood that a code reorganization to address bad code smells will yield improvement in the defect-proneness of the code. Method: We introduce XTREE, a tool that analyzes a historical log of defects seen p… ▽ More

    Submitted 14 May, 2017; v1 submitted 12 September, 2016; originally announced September 2016.

    Comments: 16 pages, 9 figures (2 colour images), and 68 References

    Journal ref: Information and Software Technology, Volume 88, August 2017, Pages 53-66, ISSN 0950-5849

  115. arXiv:1609.02613  [pdf, other

    cs.SE cs.LG stat.ML

    Why is Differential Evolution Better than Grid Search for Tuning Defect Predictors?

    Authors: Wei Fu, Vivek Nair, Tim Menzies

    Abstract: Context: One of the black arts of data mining is learning the magic parameters which control the learners. In software analytics, at least for defect prediction, several methods, like grid search and differential evolution (DE), have been proposed to learn these parameters, which has been proved to be able to improve the performance scores of learners. Objective: We want to evaluate which method… ▽ More

    Submitted 10 March, 2017; v1 submitted 8 September, 2016; originally announced September 2016.

    Comments: 12 pages, 8 figures, submitted to Information and Software Technology

  116. Tuning for Software Analytics: is it Really Necessary?

    Authors: Wei Fu, Tim Menzies, Xipeng Shen

    Abstract: Context: Data miners have been widely used in software engineering to, say, generate defect predictors from static code measures. Such static code defect predictors perform well compared to manual methods, and they are easy to use and useful to use. But one of the "black art" of data mining is setting the tunings that control the miner. Objective:We seek simple, automatic, and very effective metho… ▽ More

    Submitted 6 September, 2016; originally announced September 2016.

    Journal ref: Information and Software Technology 76 (2016): 135-146

  117. arXiv:1609.00489  [pdf, other

    cs.SE cs.LG stat.ML

    A deep learning model for estimating story points

    Authors: Morakot Choetkiertikul, Hoa Khanh Dam, Truyen Tran, Trang Pham, Aditya Ghose, Tim Menzies

    Abstract: Although there has been substantial research in software analytics for effort estimation in traditional software projects, little work has been done for estimation in agile projects, especially estimating user stories or issues. Story points are the most common unit of measure used for estimating the effort involved in implementing a user story or resolving an issue. In this paper, we offer for th… ▽ More

    Submitted 6 September, 2016; v1 submitted 2 September, 2016; originally announced September 2016.

    Comments: Submitted to ICSE'17

  118. arXiv:1608.08176  [pdf, other

    cs.SE cs.AI cs.CL cs.IR

    What is Wrong with Topic Modeling? (and How to Fix it Using Search-based Software Engineering)

    Authors: Amritanshu Agrawal, Wei Fu, Tim Menzies

    Abstract: Context: Topic modeling finds human-readable structures in unstructured textual data. A widely used topic modeler is Latent Dirichlet allocation. When run on different datasets, LDA suffers from "order effects" i.e. different topics are generated if the order of training data is shuffled. Such order effects introduce a systematic error for any study. This error can relate to misleading results;spe… ▽ More

    Submitted 20 February, 2018; v1 submitted 29 August, 2016; originally announced August 2016.

    Comments: 15 pages + 2 page references. Accepted to IST

    Journal ref: Information and Software Technology Journal, 2018

  119. Finding Trends in Software Research

    Authors: George Mathew, Amritanshu Agrawal, Tim Menzies

    Abstract: This paper explores the structure of research papers in software engineering. Using text mining, we study 35,391 software engineering (SE) papers from 34 leading SE venues over the last 25 years. These venues were divided, nearly evenly, between conferences and journals. An important aspect of this analysis is that it is fully automated and repeatable. To achieve that automation, we used a stable… ▽ More

    Submitted 2 October, 2018; v1 submitted 29 August, 2016; originally announced August 2016.

    Comments: 12 pages, 4 tables, 11 figures, Accepted in IEEE Transactions on Software Engineering 2018

  120. arXiv:1608.07617  [pdf, other

    cs.SE

    "Sampling"' as a Baseline Optimizer for Search-based Software Engineering

    Authors: Jianfeng Chen, Vivek Nair, Rahul Krishna, Tim Menzies

    Abstract: Increasingly, Software Engineering (SE) researchers use search-based optimization techniques to solve SE problems with multiple conflicting objectives. These techniques often apply CPU-intensive evolutionary algorithms to explore generations of mutations to a population of candidate solutions. An alternative approach, proposed in this paper, is to start with a very large population and sample down… ▽ More

    Submitted 5 January, 2018; v1 submitted 26 August, 2016; originally announced August 2016.

    Comments: 15 pages, 11 figures, 4 tables. To appear, IEEE Trans Software Engineering 2018