About cookies on this site Our websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising. For more information, please review your options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.
Validation / testing
Let's start the discussion on how to test ML-based systems.
We'll cover an elementary example of sequential reasoning from section 4.2 in http://www.med.mcgill.ca/epidemiology/hanley/bios601/GaussianModel/JaynesProbabilityTheory.pdf.
As we discussed the use of concentration inequalities in our meeting on reinforcement learning, we provide a refresher on this subject and discuss Markov inequality.
We cover the Shapley value and recent application to data analysis. For a deep dive, review this paper. For the curious, slide picture was taken in Hummus Yosef.
At the end of August 2019, I presented our paper on testing ML applications in FSE. You may find this paper interesting.
In this chapter on how to test/validate ML based systems (under construction):
-
We'll cover how to create a non-parametric
confidence interval.
- We'll discuss the concept of empirical distribution to better motivate the non parametric confidence interval we have just discussed.
- In such ideal assumptions the central limit theorem can be used to create a confidence interval (see section 2).
- Bootstrapping is used to overcome budgets constraints.
- We'll discuss convergence in distribution.
- We'll revisit the bootstrapping example and cast it in the context of a ML learning example (example 2).
This is a Python example of a non parametric confidence interval with unlimited sampling.