Structured Output Support Vector Machines
总结自 Flexible discriminative learning with structured output support vector machines,非常好的一份 tutorial。
1. IntroPermalink
Structured output SVMs Extends SVMs to handle arbitrary output spaces, particularly ones with non-trivial structure (e.g. space of poses, textual translations, sentences in a grammar, etc.).
这一篇的符号:
叫 score 其实就是 就是 ,用 还标准些- 没有使用
,直接 称为 feature map- With a feature map, the nature of the input
is irrelevant (image, video, audio, …).
- With a feature map, the nature of the input
- 优化问题里使用了 hinge loss
- 0-1 classification 的情况下,
- 如果是 Support Vector Regression(直接用
来预测 的值),则 ,因为是 1 阶的,也称 error - hinge loss is a convex function
- 0-1 classification 的情况下,
- 然后它的优化问题是 minimize
2. SSVMPermalink
之前有说过:In structured output learning, the discriminant function becomes a function
进而 SSVM 变成一个搜索问题:
是不是有点 maximum likelihood 的意味?不过这个式子没有概率上的意义。
等式右边求
Standard SVMs can be easily interpreted as a structured SVMs:
- output space:
- joint feature map:
- inference:
3. The surrogate lossPermalink
类似 hinge loss,SSVM 也有一个 loss function
The key in the success of the structured SVMs is the existence of good surrogates. The aim is to make minimising
- Margin rescaling surrogate:
表示 “最小上界”
- Slack rescaling surrogate:
后面还讲了 Cutting plane algorithm
、BMRM: cutting planes with a regulariser
等内容以及 Matlab 的实现,这里就不展开了。
Comments