Improve Model Performance with Error Analysis and Diagnostic Evaluation

model debugging n.w
1 / 36
Embed
Share

Learn how to enhance your model's performance by conducting error analysis, diagnostic evaluation, and interpretable evaluation. Explore common error patterns through manual analysis and gain insights into improving your model's accuracy and effectiveness in NLP tasks such as sentiment classification and NER. Discover essential resources to achieve your goal and identify weaknesses to refine your models effectively.

  • Model Diagnostic
  • Error Analysis
  • Diagnostic Evaluation
  • NLP Tasks
  • Model Improvement

Uploaded on | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

You are allowed to download the files provided on this website for personal or commercial use, subject to the condition that they are used lawfully. All files are the property of their respective owners.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author.

E N D

Presentation Transcript


  1. Model Debugging You ve implemented a nice model (or replicated a SOTA model) Your accuracy on the test set is bad What do I do? What do I do? Training/Test stage

  2. Another Typical Situation You ve implemented a nice model (or replicated a SOTA model) Your accuracy on the test set is good You want to know what your model is not good at? You want to know what your model is not good at?

  3. Model Diagnostic What is Model Diagnostic ? Identify the weaknesses (strengths) of your models Why do we need Model Diagnostic ? What Works? (Interpretability) What s Next? (Next step)

  4. Model Diagnostic How to further improve the performance? F1 Score of NER task F1 Score of NER task Performance of many NLP tasks (i.e. NER) has reached a plateau. reached a plateau.

  5. More Intuitively Model Diagnostic Model Debugging Assignment 4 Assignment 4 (improve the state-of- the-art) Assignment 3 Assignment 3 (achieve a state-of-the- art system)

  6. How to achieve this goal? Error Analysis Diagnostic Evaluation Interpretable Evaluation

  7. How to achieve this goal? Error Analysis (four must-read papers) Diagnostic Evaluation (four must-read papers) Interpretable Evaluation (two must-read papers)

  8. Error Analysis Manually Manually check test cases on which models make a wrong prediction (or unreasonable generation) Try to abstract commonalities commonalities of these error cases Model test samples outputs

  9. Error Analysis on Sentiment Classification Task The classifier will fail when Long-term Dependency Err-I: sentences with double negation I don t don t think this movie is not interesting not Err-II: sentences with subjunctive mood The movie could have could have been better. Err-III: sentences with annotation errors I like this movie -> negative

  10. Error Analysis on Sentiment Classification Task The classifier will fail when Err-I: sentences with double negation I don t don t think this movie is not interesting not Err-II: sentences with subjunctive mood Reasoning The movie could have could have been better. Err-III: sentences with annotation errors I like this movie -> negative

  11. Error Analysis on Sentiment Classification Task The classifier will fail when Err-I: sentences with double negation I don t don t think this movie is not interesting not Err-II: sentences with subjunctive mood The movie could have could have been better. Err-III: sentences with annotation errors De-noising I like this movie -> negative

  12. In Summary Na ve but super useful method Learning to perform error analysis is a good research habit Many solid ideas come from error analysis Improve yourself by error analysis Zero-distance with the data, get more domain knowledge

  13. Blind Spots of Error Analysis Err-I: sentences with double negation Err-II: sentence with subjunctive mood Err-III: sentence with annotation errors

  14. Blind Spots of Error Analysis What if there is no What if there is no Err Err- -II samples in II samples in the test set the test set Err-I,Err-III Model test samples outputs

  15. Blind Spots of Error Analysis What if there is no What if there is no Err Err- -II samples in II samples in the test set the test set Construct! Construct! Err-I,Err-III Model test samples outputs

  16. Diagnostic Evaluation Automatically Automatically construct a new current models will fail new set of test samples that Re-evaluate models using the newly-constructed data New test samples Model test samples outputs

  17. Diagnostic Evaluation Stress set Contrastive set Adversarial set Automatically Automatically construct a new current models will fail new set of test samples that Re-evaluate models using the newly-constructed data New test samples Model test samples outputs

  18. Confirmation bias in Diagnostic Evaluation How do we know what How do we know what types of samples to be types of samples to be constructed? constructed? New test samples Model test samples outputs

  19. Confirmation bias in Diagnostic Evaluation Assume that our Assume that our model will struggle at model will struggle at samples with some samples with some patters patters How do we know what How do we know what types of samples to be types of samples to be constructed? constructed? New test samples Model test samples outputs

  20. Interpretable Evaluation Motivation: a good evaluation metric can not only rank different systems but also tell their relative advantages (strengths and weaknesses) of them.

  21. How to achieve it? One sentence to summarize By partitioning partitioning the performance of test set different interpretable groups interpretable groups based on a pre- defined attribute attribute test set into Model test samples outputs

  22. How to achieve it? One sentence to summarize By partitioning partitioning the performance of test set into different interpretable groups interpretable groups based on a pre- defined attribute attribute Define Attributes Define Attributes Partition Test Samples Partition Test Samples Breakdown Performance Breakdown Performance

  23. Methodology Define attributes (e.g., entity length: eLen) Partition test samples Breakdown performance Performance Histogram Performance Histogram

  24. Attributes Different tasks could have different attributes Token-level, span-level, sentence-level Token-level: part-of-speech tag Span-level: span length Sentence-level: sentence length

  25. Performance Histogram Diagnostic for single system Better Worse

  26. Performance Histogram Diagnostic for two systems Performance Gap Performance Gap Histogram Histogram BERT BERT v.s v.s. . ELMo ELMo BERT-ELMo BERT ELMo 5 100 100 98 98 4 96 96 3 94 94 92 92 2 26 90 90 88 1 88 86 86 0 84 84 1 2 3 4 5 1 2 3 4 5 1 2 3 4 5 -1

  27. In Summary No need to construct new samples No need to think about potential error types But need attributes

  28. Model Diagnostic: Comparison Methodology Methodology Stage Stage Human Human effort effort Additional Additional test set test set Error Analysis test Diagnostic Evaluation test Interpretable Evaluation test

  29. Can we automate System Diagnostic? Require human efforts (more or less) Task-dependent

  30. Can we automate System Diagnostic? Methodology Methodology Stage Stage Human Human effort effort Additional Additional test set test set Error Analysis test Diagnostic Evaluation test Interpretable Evaluation test

  31. Compare-mt A diagnostic analysis toolkit for machine translation Calculates aggregate statistics about accuracy of aggregate statistics about accuracy of particular types of words or sentences particular types of words or sentences, finds salient test examples An example of this for quantitative analysis of language generation results (https://github.com/neulab/compare- mt)

  32. PBMT v.s. NMT Tips Tips: phrase phrase- -based translation and neural network network- -based based machine translation systems are two major paradigms over the past 20 years. based machine neural

  33. ExplainaBoard Next Generation of Leaderboard Track NLP progress Help researchers diagnose NLP systems

  34. LeaderBoard v.s. ExplainaBoard

  35. LeaderBoard v.s. ExplainaBoard Multiple Tasks Multiple Tasks Leaderboard and Leaderboard and Analysis Buttons Analysis Buttons Interpretable Evaluation Interpretable Evaluation Results Results

  36. ExplainaBoard Cover more tasks More functionalities Interpretability: Single system diagnosis Interactivity: System pair diagnosis Reliability: confidence interval, calibration value Github: https://github.com/neulab/InterpretEval

Related


More Related Content