【Machine Learning】Generalization Theory

本笔记基于清华大学《机器学习》的课程讲义中泛化理论相关部分,基本为笔者在考试前一两天所作的Cheat Sheet。内容较多,并不详细,主要作为复习和记忆的资料。

No free lunch

  • For algroithm A ′ A' A, exsits f f f that is perfect answer of D ∈ C × { 0 , 1 } D\in C\times\{0,1\} DC×{ 0,1}, such that L D ( f ) = 0 L_D(f)=0 LD(f)=0 and

E S ∼ D m [ L D ( A ′ ( S ) ) ] ≥ 1 4 E_{S\sim D^m}[L_D(A'(S))]\ge \frac{1}{4} ESDm[LD(A(S))]41

  • Then Pr ⁡ [ L D ( A ′ ( S ) ) ≥ 1 8 ] ≥ 1 7 \Pr[L_D(A'(S))\ge \frac{1}{8}]\ge \frac{1}{7} Pr[LD(A(S))81]71

  • Proof:
    max ⁡ i E S ∼ D i m [ L D ( A ′ ( S ) ) ] = max ⁡ i 1 k ∑ i = 1 k L D i ( A ′ ( S i ) ) ≥ 1 T ∑ j = 1 T 1 k ∑ i = 1 k L D j ( A ′ ( S i ) ) ≥ 1 k ∑ i = 1 k 1 T ∑ j = 1 T L D j ( A ′ ( S i ) ) ≥ min ⁡ S 1 T ∑ j = 1 T L D j ( A ′ ( S ) ) ≥ min ⁡ S 1 T ∑ j = 1 T 1 2 m ∑ i = 1 p 1 A ′  wrong at  v i ≥ min ⁡ S 1 T ∑ j = 1 T 1 2 p ∑ i = 1 p 1 A ′  wrong at  v i ≥ 1 2 min ⁡ S 1 T ∑ j = 1 T min ⁡ i 1 A ′  wrong at  v i ≥ 1 4 \begin{align*} \max_{i}E_{S\sim D_i^m}[L_D(A'(S))]&=\max_{i}\frac{1}{k}\sum_{i=1}^k L_{D_i}(A'(S_i))\\ &\ge \frac{1}{T}\sum_{j=1}^T\frac{1}{k}\sum_{i=1}^k L_{D_j}(A'(S_i))\\ &\ge \frac{1}{k}\sum_{i=1}^k\frac{1}{T}\sum_{j=1}^T L_{D_j}(A'(S_i))\\ &\ge \min_S\frac{1}{T}\sum_{j=1}^T L_{D_j}(A'(S))\\ &\ge \min_S\frac{1}{T}\sum_{j=1}^T \frac{1}{2m}\sum_{i=1}^p1_{A'\text{ wrong at }v_i}\\ &\ge \min_S\frac{1}{T}\sum_{j=1}^T \frac{1}{2p}\sum_{i=1}^p1_{A'\text{ wrong at }v_i}\\ &\ge \frac{1}{2}\min_S\frac{1}{T}\sum_{j=1}^T \min_{i} 1_{A'\text{ wrong at }v_i}\\ &\ge \frac{1}{4} \end{align*} imaxESDim[LD(A(S))]=imaxk1i=1kLDi(A(Si))T1j=1Tk1i=1kLDj(A(Si))k1i=1kT1j=1TLDj(A(Si))SminT1j=1TLDj(A(S))SminT1j=1T2m1i=1p1A wrong at viSminT1j=1T2p1i=1p1A wrong at vi21SminT1j=1Timin1A wrong at vi41

    • The last inequality is beause divide T T T into 2 2 2 parts. One pair f i , f i ′ f_i,f_{i'} fi,fionly differs at v i v_i vi.

ERM

  • With realizable assumption, the hypothesis class found by ERM is good enough with at least some samples

    • Consider the probability of bad samples L S ( h S ) = L S ( h ∗ ) = 0 L_S(h_S)=L_S(h^*)=0 LS(hS)=LS(h)=0 but L D , f ( h S ) > ϵ L_{D,f}(h_S)>\epsilon LD,f(hS)>ϵ. Then we need S S S to be the union(apply union bound) of misleading set L S ( h S ) = 0 L_S(h_S)=0 LS(hS)=0, each sample has probability ≤ 1 − ϵ \le 1-\epsilon 1ϵ. Then probability is ∣ H B ∣ ( 1 − ϵ ) m |H_B|(1-\epsilon)^m HB(1ϵ)m
  • PAC learnable: As sample number m ≥ m ( ϵ , δ ) m\ge m(\epsilon,\delta) mm(ϵ,δ), w.p. 1 − δ 1-\delta 1δ we can find a h h h such that L D , f ( h ) ≤ ϵ L_{D,f}(h)\le \epsilon LD,f(h)ϵ.

    • Agnostic PAC learnable: L D ( h ) ≤ L D ( h ∗ ) + ϵ L_{D}(h)\le L_{D}(h^*)+\epsilon LD(h)LD(h)+ϵ
  • VC dimension

Rademacher

  • Generalization:
    L D ( h ) − L S ( h ) ≤ 2 E S ′ ∼ D m R ( l ∘ H ∘ S ′ ) + c 2 ln ⁡ 2 δ m L_D(h)-L_S(h)\le 2E_{S'\sim D^m}R(l\circ H\circ S')+c\sqrt{\frac{2\ln\frac{2}{\delta}}{m}} LD(h)LS(h)2ESDmR(lHS)+cm2lnδ2

  • Massart Lemma:
    R ( A ) ≤ max ⁡ a ∈ A ∣ a − a ˉ ∣ 2 log ⁡ N m R(A)\le \max_{a\in A}|a-\bar{a}|\frac{\sqrt{2\log N}}{m} R(A)aAmaxaaˉm2logN

  • Contraction Lemma: If ϕ \phi ϕ is ρ \rho ρ-lipschitz, then
    R ( ϕ ∘ A ) ≤ ρ R ( A ) R(\phi\circ A)\le \rho R(A) R(ϕA)ρR(A)

相关推荐

最近更新

  1. TCP协议是安全的吗?

    2024-01-11 11:52:05       19 阅读
  2. 阿里云服务器执行yum,一直下载docker-ce-stable失败

    2024-01-11 11:52:05       19 阅读
  3. 【Python教程】压缩PDF文件大小

    2024-01-11 11:52:05       19 阅读
  4. 通过文章id递归查询所有评论(xml)

    2024-01-11 11:52:05       20 阅读

热门阅读

  1. golang一个轻量级基于内存的kv存储或缓存

    2024-01-11 11:52:05       32 阅读
  2. Python闭包与装饰

    2024-01-11 11:52:05       40 阅读
  3. Python 函数

    2024-01-11 11:52:05       32 阅读
  4. vue for循环不建议使用index作为key的原因

    2024-01-11 11:52:05       39 阅读
  5. python使用单例模式加载config.ini配置文件

    2024-01-11 11:52:05       34 阅读