Validation / testing
Let's start the discussion on how to test ML-based systems.
We'll cover an elementary example of sequential reasoning from section 4.2 in http://www.med.mcgill.ca/epidemiology/hanley/bios601/GaussianModel/JaynesProbabilityTheory.pdf.
As we discussed the use of concentration inequalities in our meeting on reinforcement learning, we provide a refresher on this subject and discuss Markov inequality.
We cover the Shapley value and recent application to data analysis. For a deep dive, review this paper. For the curious, slide picture was taken in Hummus Yosef.
Chebyshev's inequality:
At the end of August 2019, I presented our paper on testing ML applications in FSE. You may find this paper interesting.
In this chapter on how to test/validate ML based systems (under construction):
-
We'll cover how to create a non-parametric
confidence interval.
- We'll discuss the concept of empirical distribution to better motivate the non parametric confidence interval we have just discussed.
- In such ideal assumptions the central limit theorem can be used to create a confidence interval (see section 2).
- Bootstrapping is used to overcome budgets constraints.
- We'll discuss convergence in distribution.
- We'll revisit the bootstrapping example and cast it in the context of a ML learning example (example 2).
This is a Python example of a non parametric confidence interval with unlimited sampling.