SlideShare a Scribd company logo
Keynote
On the Effectiveness of SBSE
Techniques through Instance
Space Analysis
Aldeida Aleti
Monash University, Australia
@AldeidaAleti aldeida.aleti@monash.edu
Effectiveness of SBSE - Status Quo
A large focus of SBSE research is in introducing new SBSE approaches
As part of the evaluation process, usually a set of experiments are conducted
- A benchmark is selected, e..g., Defects4J
- The new approach is compared against the state of the art
- Averages/medians are reported
- Some statistical tests are conducted
Instance Space Analysis
1. to understand and visualise the strengths and weaknesses of different approaches
2. to help with the objective assessment of different approaches
a. Scrutinising how approaches perform under different conditions, and stress testing them
Motivation 1: Are the problem instances adequate?
Problem 1: How were the problem instances selected?
Common benchmark problems are important for fair comparison, but are they
- demonstrably diverse
- unbiased
- representative of a range of real world context,
- challenging
- discriminating
ICSE 2022 review criteria
Motivation 2: Reporting averages/medians obscures
important information
A. Perera, A. Aleti, M. Böhme and B. Turhan, "Defect Prediction Guided Search-Based
Software Testing," 2020 35th IEEE/ACM International Conference on Automated Software
Engineering (ASE), 2020, pp. 448-460.
Problem 2: Performance is often problem dependent
(NFT)
- What are the strengths and weaknesses of the approaches?
- Which are the problem instances where an approach performs really well and
why?
- Which are the problem instances where an approach struggles and why?
- How do features of the problem instances affect the performance of the
approaches?
- Which features give an algorithm competitive advantage?
- Given a problem instance with particular features, which approach should I use?
Which algorithm is suitable for future problems?
Example
Which approach is better? SF110
C. Oliveira, A. Aleti, L. Grunske and K. Smith-Miles, "Mapping the Effectiveness of Automated Test Suite Generation
Techniques," in IEEE Transactions on Reliability, vol. 67, no. 3, pp. 771-785, Sept. 2018, doi: 10.1109/TR.2018.2832072.
Instance Space Analysis for Search Based Software Engineering
Open Questions
● What impacts the effectiveness of SBSE techniques?
○ How can features of problem instances help us infer what are the strengths and weaknesses of
different SBSE approaches?
○ How can we objectively assess different SBSE techniques
● How easy or hard are existing benchmarks? How diverse are they? Are they biased
towards a particular technique?
● Can we select the most suitable SBSE technique given a problem with particular
features?
Empirical Review of Program Repair Tools: A Large-Scale Experiment on 2 141 Bugs and 23 551
Repair Attempts. T. Durieux, F. Madeiral, M. Martinez, R. Abreu. ESEC/FSE Foundations of Software
Engineering (2019) doi: 10.1145/ 3338906.3338911.
ISA
K. Smith-Miles et al. / Computers & Operations Research 45 (2014) 12–24
Steps of ISA
1. Create the metadata
a. Features
b. SBSE performances
2. Create instance space
3. Visualise footprints
4. Explain strengths/weaknesses
Instance Space Analysis for Search Based Software Engineering
Features (56)
What makes the problem easy or hard?
Problem instances SF110
Performance measure
● Branch coverage.
● An approach is considered superior if its branch coverage is at least 1% higher than
the other techniques; otherwise, we use the label “Equal.”
Approaches
● Whole Test Suite with Archive (WSA)
● Many Objective Sorting Algorithm (MOSA)
● Random Testing (RT)
Significant features
● coupling between object classes
○ the number of classes coupled to a given class (method calls, field accesses, inheritance,
arguments, return types, and exceptions)
● response for a class
○ number of different methods that can be executed when a method is invoked for that object
of a class
SBST Footprints
SBST selection
Instance Space Analysis for Search Based Software Engineering
Instance Space Analysis for Search Based Software Engineering
Instance Space Analysis for Search Based Software Engineering
E-APR
Metadata
Features (146)
Observation-based features (Yu et al. 2019)
Significant Features (9)
(F1) MOA: Measure of Aggregation.
(F2) CAM: Cohesion Among Methods
(F3) AMC: Average Method Complexity
(F4) PMC: Private Method Count
(F5) AECSL: Atomic Expression Comparison Same Left indicates the number of statements
with a binary expression that have more than an atomic expression (e.g., variable access).
(F6) SPTWNG: Similar Primitive Type With Normal Guard indicates the number of
statements that contain a variable (local or global) that is also used in another statement
contained inside a guard (i.e., an If condition).
(F7) CVNI: Compatible Variable Not Included is the number of local primitive type variables
within the scope of a statement that involves primitive variables that are not part of that
statement.
(F8) VCTC: Variable Compatible Type in Condition measures the number of variables within
an If condition that are compatible with another variable in the scope.
(F9) PUIA: Primitive Used In Assignment - the number of primitive variables in assignments.
Instance Space Analysis for Search Based Software Engineering
● Little overlap between
IntroClassJava/Defects4J and the other
datasets
● Bugs.jar has the most diverse bugs
APR selection
For ISA to reveal useful insights
● Diverse features
● Diverse instances
● Diverse approaches
● A good performance measure
So what
We have a responsibility to find the weaknesses of the approaches we develop
We need to make sure that the chosen problem instances are demonstrably diverse,
unbiased, representative of a range of real world context, challenging,
discriminating of approach performance
To understand which approach is suitable for future problems, we must understand
which features impact its performance

More Related Content

What's hot (20)

PDF
VST2022.pdf
Annibale Panichella
 
PPTX
A software fault localization technique based on program mutations
Tao He
 
DOC
Testing survey by_directions
Tao He
 
PPT
Experiments on Design Pattern Discovery
Tim Menzies
 
PDF
Controlled experiments, Hypothesis Testing, Test Selection, Threats to Validity
alessio_ferrari
 
PDF
Survey Research In Empirical Software Engineering
alessio_ferrari
 
PDF
Wcre13a.ppt
Ptidej Team
 
PDF
Programming with GUTs
catherinewall
 
PDF
[Tho Quan] Fault Localization - Where is the root cause of a bug?
Ho Chi Minh City Software Testing Club
 
PDF
Wcre13b.ppt
Ptidej Team
 
PPT
Using Developer Information as a Prediction Factor
Tim Menzies
 
PDF
Exploratory testing STEW 2016
Per Runeson
 
PDF
130411 francis palma - detection of process antipatterns -- a bpel perspective
Ptidej Team
 
PDF
Ssbse12b.ppt
Ptidej Team
 
PDF
Experimental design
Dan Toma
 
PDF
Sound Empirical Evidence in Software Testing
Jaguaraci Silva
 
PPTX
Software testing using genetic algorithms
Nurhussen Menza
 
PDF
Model-Driven Run-Time Enforcement of Complex Role-Based Access Control Policies
Lionel Briand
 
PDF
An Empirical Comparison of Model Validation Techniques for Defect Prediction ...
Chakkrit (Kla) Tantithamthavorn
 
PPT
Cause-Effect Graphing: Rigorous Test Case Design
TechWell
 
VST2022.pdf
Annibale Panichella
 
A software fault localization technique based on program mutations
Tao He
 
Testing survey by_directions
Tao He
 
Experiments on Design Pattern Discovery
Tim Menzies
 
Controlled experiments, Hypothesis Testing, Test Selection, Threats to Validity
alessio_ferrari
 
Survey Research In Empirical Software Engineering
alessio_ferrari
 
Wcre13a.ppt
Ptidej Team
 
Programming with GUTs
catherinewall
 
[Tho Quan] Fault Localization - Where is the root cause of a bug?
Ho Chi Minh City Software Testing Club
 
Wcre13b.ppt
Ptidej Team
 
Using Developer Information as a Prediction Factor
Tim Menzies
 
Exploratory testing STEW 2016
Per Runeson
 
130411 francis palma - detection of process antipatterns -- a bpel perspective
Ptidej Team
 
Ssbse12b.ppt
Ptidej Team
 
Experimental design
Dan Toma
 
Sound Empirical Evidence in Software Testing
Jaguaraci Silva
 
Software testing using genetic algorithms
Nurhussen Menza
 
Model-Driven Run-Time Enforcement of Complex Role-Based Access Control Policies
Lionel Briand
 
An Empirical Comparison of Model Validation Techniques for Defect Prediction ...
Chakkrit (Kla) Tantithamthavorn
 
Cause-Effect Graphing: Rigorous Test Case Design
TechWell
 

Similar to Instance Space Analysis for Search Based Software Engineering (20)

PDF
Guidelines to Understanding Design of Experiment and Reliability Prediction
ijsrd.com
 
PPTX
Specification based or black box techniques
Muhammad Ibnu Wardana
 
PPTX
Specification based or black box techniques
Irvan Febry
 
PPTX
Specification based or black box techniques
muhammad afif
 
PPTX
Specification based or black box techniques (andika m)
Andika Mardanu
 
PPTX
Specification based or black box techniques
Dinul
 
PPTX
Specification based or black box techniques
Aji Pamungkas Prasetio
 
PPTX
Specification based or black box techniques 3
alex swandi
 
PPTX
Specification based or black box techniques
M Branikno Ramadhan
 
PPTX
Specification based or black box techniques
M Branikno Ramadhan
 
PPTX
Specification Based or Black Box Techniques
Nadia Chairunissa
 
PDF
A Software Measurement Using Artificial Neural Network and Support Vector Mac...
ijseajournal
 
PPTX
Specification based or black box techniques
Yoga Setiawan
 
PPTX
Specification based or black box techniques
M HiDayat
 
PPTX
Specification Based or Black Box Techniques
RakhesLeoPutra
 
PDF
Comprehensive Testing Tool for Automatic Test Suite Generation, Prioritizatio...
CSCJournals
 
PPTX
ders 6 Panel data analysis.pptx
Ergin Akalpler
 
PPTX
Specification based or black box techniques
Yoga Pratama Putra
 
PDF
Software Cost Estimation Using Clustering and Ranking Scheme
Editor IJMTER
 
PDF
Investigate the Matrix: Leveraging Variability to Specialize Software and Tes...
paultemple20
 
Guidelines to Understanding Design of Experiment and Reliability Prediction
ijsrd.com
 
Specification based or black box techniques
Muhammad Ibnu Wardana
 
Specification based or black box techniques
Irvan Febry
 
Specification based or black box techniques
muhammad afif
 
Specification based or black box techniques (andika m)
Andika Mardanu
 
Specification based or black box techniques
Dinul
 
Specification based or black box techniques
Aji Pamungkas Prasetio
 
Specification based or black box techniques 3
alex swandi
 
Specification based or black box techniques
M Branikno Ramadhan
 
Specification based or black box techniques
M Branikno Ramadhan
 
Specification Based or Black Box Techniques
Nadia Chairunissa
 
A Software Measurement Using Artificial Neural Network and Support Vector Mac...
ijseajournal
 
Specification based or black box techniques
Yoga Setiawan
 
Specification based or black box techniques
M HiDayat
 
Specification Based or Black Box Techniques
RakhesLeoPutra
 
Comprehensive Testing Tool for Automatic Test Suite Generation, Prioritizatio...
CSCJournals
 
ders 6 Panel data analysis.pptx
Ergin Akalpler
 
Specification based or black box techniques
Yoga Pratama Putra
 
Software Cost Estimation Using Clustering and Ranking Scheme
Editor IJMTER
 
Investigate the Matrix: Leveraging Variability to Specialize Software and Tes...
paultemple20
 
Ad

Recently uploaded (20)

PDF
Meet in the Middle: Solving the Low-Latency Challenge for Agentic AI
Alluxio, Inc.
 
PPTX
Function & Procedure: Function Vs Procedure in PL/SQL
Shani Tiwari
 
PDF
Australian Enterprises Need Project Service Automation
Navision India
 
PDF
Understanding the EU Cyber Resilience Act
ICS
 
PDF
Show Which Projects Support Your Strategy and Deliver Results with OnePlan df
OnePlan Solutions
 
PPTX
Get Started with Maestro: Agent, Robot, and Human in Action – Session 5 of 5
klpathrudu
 
PPTX
TexSender Pro 8.9.1 Crack Full Version Download
cracked shares
 
PPTX
Odoo Migration Services by CandidRoot Solutions
CandidRoot Solutions Private Limited
 
PDF
How to get the licensing right for Microsoft Core Infrastructure Server Suite...
Q-Advise
 
PPTX
Transforming Lending with IntelliGrow – Advanced Loan Software Solutions
Intelli grow
 
PPTX
prodad heroglyph crack 2.0.214.2 Full Free Download
cracked shares
 
PPTX
Cutting Optimization Pro 5.18.2 Crack With Free Download
cracked shares
 
PDF
Salesforce Experience Cloud Consultant.pdf
VALiNTRY360
 
PDF
ERP Consulting Services and Solutions by Contetra Pvt Ltd
jayjani123
 
PDF
How Attendance Management Software is Revolutionizing Education.pdf
Pikmykid
 
PDF
Notification System for Construction Logistics Application
Safe Software
 
PDF
Windows 10 Professional Preactivated.pdf
asghxhsagxjah
 
PDF
intro_to_cpp_namespace_robotics_corner.pdf
MohamedSaied877003
 
PDF
Introduction to Apache Iceberg™ & Tableflow
Alluxio, Inc.
 
PPTX
iaas vs paas vs saas :choosing your cloud strategy
CloudlayaTechnology
 
Meet in the Middle: Solving the Low-Latency Challenge for Agentic AI
Alluxio, Inc.
 
Function & Procedure: Function Vs Procedure in PL/SQL
Shani Tiwari
 
Australian Enterprises Need Project Service Automation
Navision India
 
Understanding the EU Cyber Resilience Act
ICS
 
Show Which Projects Support Your Strategy and Deliver Results with OnePlan df
OnePlan Solutions
 
Get Started with Maestro: Agent, Robot, and Human in Action – Session 5 of 5
klpathrudu
 
TexSender Pro 8.9.1 Crack Full Version Download
cracked shares
 
Odoo Migration Services by CandidRoot Solutions
CandidRoot Solutions Private Limited
 
How to get the licensing right for Microsoft Core Infrastructure Server Suite...
Q-Advise
 
Transforming Lending with IntelliGrow – Advanced Loan Software Solutions
Intelli grow
 
prodad heroglyph crack 2.0.214.2 Full Free Download
cracked shares
 
Cutting Optimization Pro 5.18.2 Crack With Free Download
cracked shares
 
Salesforce Experience Cloud Consultant.pdf
VALiNTRY360
 
ERP Consulting Services and Solutions by Contetra Pvt Ltd
jayjani123
 
How Attendance Management Software is Revolutionizing Education.pdf
Pikmykid
 
Notification System for Construction Logistics Application
Safe Software
 
Windows 10 Professional Preactivated.pdf
asghxhsagxjah
 
intro_to_cpp_namespace_robotics_corner.pdf
MohamedSaied877003
 
Introduction to Apache Iceberg™ & Tableflow
Alluxio, Inc.
 
iaas vs paas vs saas :choosing your cloud strategy
CloudlayaTechnology
 
Ad

Instance Space Analysis for Search Based Software Engineering

  • 1. Keynote On the Effectiveness of SBSE Techniques through Instance Space Analysis Aldeida Aleti Monash University, Australia @AldeidaAleti [email protected]
  • 2. Effectiveness of SBSE - Status Quo A large focus of SBSE research is in introducing new SBSE approaches As part of the evaluation process, usually a set of experiments are conducted - A benchmark is selected, e..g., Defects4J - The new approach is compared against the state of the art - Averages/medians are reported - Some statistical tests are conducted
  • 3. Instance Space Analysis 1. to understand and visualise the strengths and weaknesses of different approaches 2. to help with the objective assessment of different approaches a. Scrutinising how approaches perform under different conditions, and stress testing them
  • 4. Motivation 1: Are the problem instances adequate?
  • 5. Problem 1: How were the problem instances selected? Common benchmark problems are important for fair comparison, but are they - demonstrably diverse - unbiased - representative of a range of real world context, - challenging - discriminating
  • 6. ICSE 2022 review criteria
  • 7. Motivation 2: Reporting averages/medians obscures important information A. Perera, A. Aleti, M. Böhme and B. Turhan, "Defect Prediction Guided Search-Based Software Testing," 2020 35th IEEE/ACM International Conference on Automated Software Engineering (ASE), 2020, pp. 448-460.
  • 8. Problem 2: Performance is often problem dependent (NFT) - What are the strengths and weaknesses of the approaches? - Which are the problem instances where an approach performs really well and why? - Which are the problem instances where an approach struggles and why? - How do features of the problem instances affect the performance of the approaches? - Which features give an algorithm competitive advantage? - Given a problem instance with particular features, which approach should I use? Which algorithm is suitable for future problems?
  • 9. Example Which approach is better? SF110 C. Oliveira, A. Aleti, L. Grunske and K. Smith-Miles, "Mapping the Effectiveness of Automated Test Suite Generation Techniques," in IEEE Transactions on Reliability, vol. 67, no. 3, pp. 771-785, Sept. 2018, doi: 10.1109/TR.2018.2832072.
  • 11. Open Questions ● What impacts the effectiveness of SBSE techniques? ○ How can features of problem instances help us infer what are the strengths and weaknesses of different SBSE approaches? ○ How can we objectively assess different SBSE techniques ● How easy or hard are existing benchmarks? How diverse are they? Are they biased towards a particular technique? ● Can we select the most suitable SBSE technique given a problem with particular features?
  • 12. Empirical Review of Program Repair Tools: A Large-Scale Experiment on 2 141 Bugs and 23 551 Repair Attempts. T. Durieux, F. Madeiral, M. Martinez, R. Abreu. ESEC/FSE Foundations of Software Engineering (2019) doi: 10.1145/ 3338906.3338911.
  • 13. ISA K. Smith-Miles et al. / Computers & Operations Research 45 (2014) 12–24
  • 14. Steps of ISA 1. Create the metadata a. Features b. SBSE performances 2. Create instance space 3. Visualise footprints 4. Explain strengths/weaknesses
  • 16. Features (56) What makes the problem easy or hard?
  • 18. Performance measure ● Branch coverage. ● An approach is considered superior if its branch coverage is at least 1% higher than the other techniques; otherwise, we use the label “Equal.”
  • 19. Approaches ● Whole Test Suite with Archive (WSA) ● Many Objective Sorting Algorithm (MOSA) ● Random Testing (RT)
  • 20. Significant features ● coupling between object classes ○ the number of classes coupled to a given class (method calls, field accesses, inheritance, arguments, return types, and exceptions) ● response for a class ○ number of different methods that can be executed when a method is invoked for that object of a class
  • 26. E-APR
  • 30. (F1) MOA: Measure of Aggregation. (F2) CAM: Cohesion Among Methods (F3) AMC: Average Method Complexity (F4) PMC: Private Method Count (F5) AECSL: Atomic Expression Comparison Same Left indicates the number of statements with a binary expression that have more than an atomic expression (e.g., variable access). (F6) SPTWNG: Similar Primitive Type With Normal Guard indicates the number of statements that contain a variable (local or global) that is also used in another statement contained inside a guard (i.e., an If condition). (F7) CVNI: Compatible Variable Not Included is the number of local primitive type variables within the scope of a statement that involves primitive variables that are not part of that statement. (F8) VCTC: Variable Compatible Type in Condition measures the number of variables within an If condition that are compatible with another variable in the scope. (F9) PUIA: Primitive Used In Assignment - the number of primitive variables in assignments.
  • 32. ● Little overlap between IntroClassJava/Defects4J and the other datasets ● Bugs.jar has the most diverse bugs
  • 34. For ISA to reveal useful insights ● Diverse features ● Diverse instances ● Diverse approaches ● A good performance measure
  • 35. So what We have a responsibility to find the weaknesses of the approaches we develop We need to make sure that the chosen problem instances are demonstrably diverse, unbiased, representative of a range of real world context, challenging, discriminating of approach performance To understand which approach is suitable for future problems, we must understand which features impact its performance