Please wait, while the deck is loading…
a New Surrogate Risk for Learning from Weakly Labeled Data

- Authors: Valentina Zantedeschi, Rémi Emonet, Marc Sebban
- Date: 2017-04-10
- Lille - Magnet Team
Soft-Margin SVM
for a sample S of m instances (xi,yi)
argminθ,b21∥θ∥22+ci∑ξi
s.t.yi(θTμ(xi)+b)≥1−ξi,
ξi≥0
- θ, b the parameters of the linear separator
- μ a mapping function, so that μ(xi)Tμ(xj)=K(xi,xj)
Supervised Learning
In Brief
- Learning a classifier from a fully labeled set
Issues
- Label assignment is difficult and expensive:
- difficult: unique and reliable labels
- expensive: great amount of data, need for experts
- Datasets are generally noisy
How to handle the confidence in the labels?
Weak-Label Learning
labels are incorrect, missing or not unique
Sub-problems
- Semi-Supervised Learning
- Unsupervised Learning
- Label Proportions Learning
- Multi-Instance Learning
- Multi-Expert Learning
- Noisy-Tolerant Learning
Empirical Surrogate β-Risk
For any margin-based loss function Fϕ
Rϕβ(X,h)=mbϕi=1∑mσ∈−1,1∑βiσFϕ(σh(xi))
β degree of confidence / probability of labels
βi-1∈[0,1], βi+1∈[0,1]
βi-1+βi+1=1
Margin-based loss functions
Soft-Margin β-SVM
primal problem
argminθ,b21∥θ∥22+ci∑(βi-1ξi-1+βi+1ξi+1)
s.t.σ(θTμ(xi)+b)≥1−ξiσ,
ξiσ≥0
Lagrangian dual problem
αmax−21i,j∑σ,σ′∑αiσσαjσ′σ′K(xi,xj)+i∑σ∑αiσ
s.t.0≤αiσ≤cβiσ,
i=1∑mσ∑αiσσ=0
How the margin is affected

Relation with the Classical Risk
Let's rewrite the classical risk
Rϕ(X,Y,h) = Rϕβ(X,h) - m1i∑βi−yiyih(xi)
Rϕβ(X,h) is the β-risk
m1i∑βi−yiyih(xi) is a penality term on the missclassified instances
Iterative Algorithm
- Learn h
argminhN(h)+cRϕβ(X,h)
- Estimate y
∀i=1..m,yi=sign(βi+1−21)
- Learn β
argminβRϕβ(X,h)
s.t.i=1∑mβi-yiyih(xi)=0
βi-1+βi+1=1
Semi-Supervised Learning
with ml labeled instances and mu unlabeled instances
- Initialization of β
- ∀i=1..mlβiσ=1ifσ=yi,0otherwise
- ∀i=ml+1..muβiσ=0.5
- Iterative Algorithm
- Learning β of unlabeled set
Results

WellSVM:
Li Yu-Feng, Tsang Ivor W, Kwok James T, Zhou Zhi-Hua.
Convex and scalable weakly labeled SVMs
The Journal of Machine Learning Research, 2013.
Perspectives: Differential Privacy
How to accuratly learn while preserving the user privacy?
Learn on bags of instances:
Thanks for your attention
1 / 14 − Valentina Zantedeschi − a New Surrogate Risk for Learning from Weakly Labeled Data
←
→