Loading [MathJax]/extensions/TeX/boldsymbol.js

 

 

 

What are the basic ingredients?

Almost every problem in ML and data science starts with the same ingredients:

  • The dataset \mathbf{x} (could be some observable quantity of the system we are studying)
  • A model which is a function of a set of parameters \mathbf{\alpha} that relates to the dataset, say a likelihood function p(\mathbf{x}\vert \mathbf{\alpha}) or just a simple model f(\mathbf{\alpha})
  • A so-called cost function \mathcal{C} (\mathbf{x}, f(\mathbf{\alpha})) which allows us to decide how well our model represents the dataset.
We seek to minimize the function \mathcal{C} (\mathbf{x}, f(\mathbf{\alpha})) by finding the parameter values which minimize \mathcal{C} . This leads to various minimization algorithms. It may surprise many, but at the heart of all machine learning algortihms there is an optimization problem.