Cousera机器学习基石第四周笔记-Machine-Learning-Foundation-Week-4-Note-in-Cousera

Feasibility of Learning

Learning is Impossible?

Probability to the Rescue

Inferring Something Unknow

in sampleout sample

Possible versus Probable

Hoeffding’s Inequality

In big sample(N large),υ is probably close to u (within ϵ) ℙ[|vu|>ϵ] ≤ 2exp(−2ϵ2N) called Hoeffding’s Inequality, for marbles,coin,polling

the statement v = u is probably approximately correct(PAC)

  • valid for all N and ϵ

  • does not depend on u,no need to knowu

  • larger sample size N or looser gap ϵhigher probability for v = u

    if large N,can probably infer unknown u by know v

Connection to Learning

Added Components

for any fixed h, can probably infer unkown $E_out(h)=\underset{X\approx P}{\varepsilon}[h(x)\ne f(x)]$by known$ E_in(h)=^N_{n=1}[h(x)f(x)]$

The Formal Guarantee

if Ein(h) ≈ Eout(h) and$E_{in}(h)smallE_{out}(h)samllhf $with respect to P

Verification of One h

if Ein(h) small for the fixed h and A pick the h as g g=f PAC

if A force to pick THE h as g → Ein(h) almost always not small → g ≠ f PAC

real learning:

A shall make choices$\in \H$ (like PLA) rather than being forced to pick one h.

The ‘Verification’ Flow

Connection to Real Learning

BAD Sample and BAD Data

BAD Sample:$E_{out}=\frac{1}{2}$,but getting all heads(Ein = 0)

BAD Data for One h:Eout(h) and Einh far away

BAD data for Many h

BAD data for many h no freedom of choice by A there exists some h such that Eout(h) and Ein(h) far away

Bound of BAD Data

The Statistical Learning Flow

if |ℍ|= M finite, N large enough,for whatever g picked by A,Eout(g) ≈ Ein(g)

if A finds one g with Ein(g) ≈ 0,PAC guarantee forEout(g)⇒learning possible

M=? - see you in the next lectures~

吐槽

这个作业题是真的难啊,花了一个半小时才堪堪通过,尤其是最后几个写PLA和pocket算法的