\(\renewcommand{\vec}[1]{\mathbf{#1}}\)

\(\newcommand{\vecw}{\vec{w}}\) \(\newcommand{\vecx}{\vec{x}}\) \(\newcommand{\vecu}{\vec{u}}\) \(\newcommand{\veca}{\vec{a}}\) \(\newcommand{\vecb}{\vec{b}}\)

\(\newcommand{\vecwi}{\vecw^{(i)}}\) \(\newcommand{\vecwip}{\vecw^{(i+1)}}\) \(\newcommand{\vecwim}{\vecw^{(i-1)}}\) \(\newcommand{\norm}[1]{\lVert #1 \rVert}\) \(\newcommand{\fhat}{\hat{f}}\)

Exploration 1.1: Different Settings of Machine Learning

Overview of Different Settings

There are many different types of learning settings:

Supervised Learning: A Closer Look

In supervised learning, the training examples are \(D = \{(\vecx^{(i)}, y^{(i)}) \mid i = 1 \ldots |D|\}\) where each label \(y^{(i)} = f(\vecx^{(i)})\) for some unknown function \(f\) which is the underlying function that generates data \(D\). The job of machine learning is to recover \(f\) from data \(D\), or to find a good approximation \(\fhat\) of \(f\) so that \(f(\vecx)\) and \(\fhat(\vecx)\) are close. This function \(\fhat\) is also known as the prediction rule learned from data.

We differentiate two types of function \(f\) which correspond to two subsettings of supervised learning, classification and regression:

classification examples: binary, multiclass, and structured classification

When Supervised Learning is Useful

Supervised learning is particularly useful in the following scenarios:

Videos