After running as one of the most successful workshops at ICSE 2020, the International Workshop on Testing for Deep Learning and Deep Learning for Testing (DeepTest) returns once more as a co-located workshop at the ACM/IEEE International Conference on Software Engineering (ICSE) in 2021!
DeepTest is a high-quality workshop for research at the intersection of Machine Learning (ML) and software engineering (SE). ML is widely adopted in modern software systems, including safety-critical domains such as autonomous cars, medical diagnosis, and aircraft collision avoidance systems. Thus, it is crucial to rigorously test such applications to ensure high dependability. However, standard notions of software quality and reliability become irrelevant when considering ML systems, due to their non-deterministic nature and the lack of a transparent understanding of the models’ semantics. ML is also expected to revolutionize software development. Indeed, ML is being applied for devising novel program analysis and software testing techniques related to malware detection, fuzzy testing, bug-finding, and type-checking.
The workshop will combine academia and industry in a quest for well-founded practical solutions. The aim is to bring together an international group of researchers and practitioners with both ML and SE backgrounds to discuss their research, share datasets, and generally help the field to build momentum. The workshop will consist of invited talks, presentations based on research paper submissions, and one or more panel discussions, where all participants are invited to share their insights and ideas.
Sponsor Type:1; 9
Andrea Stocco
Organization Co-chair
Università della Svizzera italiana (USI)
Switzerland
Guy Barash
Organization Co-chair
Western Digital
Eitan Farchi
Organization Co-chair
IBM Haifa Research Lab
Israel
Gunel Jahangirova
Organization Co-chair
USI Lugano, Switzerland
Switzerland
Vincenzo Riccio
Organization Co-chair
USI Lugano, Switzerland
Switzerland
Diptikalyan Saha
Organization Co-chair
IBM Research India
India
Onn Shehory
Organization Co-chair
Bar Ilan University
Israel
Relevant topics include, but are not limited to:
Quality
Quality implication of ML algorithms on large-scale software systems
Application of classical statistics to ML systems quality
Training and payload data quality
Correctness of data abstraction, data trust
High-quality benchmarks for evaluating ML approaches
Testing and Verification
Test data synthesis for testing ML systems
White-box and black-box testing strategies
ML models for testing programs
Adversarial machine learning and adversary based learning
Test coverage
Vulnerability, sensitivity, and attacks against ML
Metamorphic testing as software quality assurance
New abstraction techniques for verification of ML systems
ML techniques for software verification
Dev-ops for ML
Fault Localization, Debugging, and Repairing
Quality Metrics for ML systems, e.g., Correctness, Accuracy, Fairness, Robustness, Explainability
Sensitivity to data distribution diversity and distribution drift
Failure explanation and automated debugging techniques
Runtime monitoring
Fault Localization and anomaly detection
Model repairing
The effect of labeling costs on solution quality (semi-supervised learning)
ML for fault prediction, localization, and repair
ML to aid program comprehension, program transformation, and program generation
06月01日
2021
会议日期
初稿截稿日期
注册截止日期
留言