索引: Outline(00:00:08)
Challenging problems(00:00:19)
Data Mining(00:00:53)
Machine Learning(00:02:15)
Application in PR(00:03:14)
Difference(00:03:28)
Biometrics(00:04:04)
Bioinformatics(00:04:39)
ISI(00:05:08)
Confusion(00:05:34)
統(tǒng)計(jì)機(jī)器學(xué)習(xí)基礎(chǔ)研究(00:06:00)
Machine learning community(00:06:31)
學(xué)習(xí)(00:06:55)
Performance(00:08:15)
學(xué)習(xí)(00:08:19)
Performance(00:08:22)
More(00:08:53)
Theoretical Analysis(00:09:11)
Ian Hacking(00:09:44)
Statistical learning(00:10:28)
Andreas Buja(00:10:46)
Interpretation of Algorithms(00:11:22)
統(tǒng)計(jì)學(xué)習(xí)(00:11:58)
Main references(00:13:18)
Main kinds of theory(00:13:39)
Definition of Classifications(00:14:02)
統(tǒng)計(jì)學(xué)習(xí)(00:14:23)
Main kinds of theory(00:15:21)
Definition of Classifications(00:15:22)
Definition of regression(00:15:50)
Several well-known algorithms(00:16:27)
Framework of algorithms(00:17:02)
Designation of algorithms(00:17:58)
統(tǒng)計(jì)決策理論(00:18:39)
Bayesian:classification(00:19:26)
統(tǒng)計(jì)決策理論(00:20:10)
Bayesian:classification(00:20:13)
Bayesian: regression(00:20:18)
統(tǒng)計(jì)決策理論(00:20:55)
Bayesian:classification(00:21:00)
Bayesian: regression(00:21:17)
Estimating densities(00:21:25)
KNN(00:22:45)
Interpretation:KNN(00:23:20)
高維空間(00:24:15)
維數(shù)災(zāi)難(00:25:01)
維數(shù)災(zāi)難(00:25:50)
維數(shù)災(zāi)難:其它體現(xiàn)(00:26:45)
LMS(00:27:33)
Interpretation: LMS(00:29:57)
維數(shù)災(zāi)難(00:30:57)
KNN(00:30:58)
Designation of algorithms(00:30:59)
Designation of algorithms(00:31:00)
統(tǒng)計(jì)決策理論(00:31:01)
Estimating densities(00:31:18)
高維空間(00:31:19)
維數(shù)災(zāi)難:其它體現(xiàn)(00:31:20)
Interpretation: LMS(00:31:21)
Fisher Discriminant Analysis(00:31:40)
Interpretation: FDA(00:32:35)
FDA and LMS(00:33:04)
FDA: a novel interpretation(00:33:38)
FDA: parameters(00:34:24)
FDA: framework of algorithms(00:35:09)
Disadvantage(00:35:59)
Bias and variance analysis(00:36:44)
Bias-Variance Decomposition(00:37:17)
Bias-Variance Tradeoff(00:38:46)
Bias-Variance Decomposition(00:38:52)
Bias-Variance Tradeoff(00:39:05)
Interpretation: KNN(00:40:29)
Ridge regression(00:41:35)
Interpretation: ridge regression(00:42:03)
Ridge regression(00:42:43)
Interpretation: ridge regression(00:43:05)
Interpretation: parameter(00:43:28)
Interpretation: ridge regression(00:43:35)
Interpretation: parameter(00:43:37)
A note(00:44:32)
Other loss functions(00:45:39)
Interpretation: boosting(00:46:35)
Boosting方法的由來(00:47:22)
Boosting方法流程(AdaBoost)(00:48:18)
Interpretation: margin(00:48:47)
Interpretation: SVM(00:49:43)
SVM: experimental analysis(00:50:48)
Interpretation: base learners(00:51:57)
Disadvantage(00:52:38)
Generalization bound(00:53:15)
PAC Frame(00:54:16)
VC Theory and PAC Bounds(00:54:44)
PAC Bounds for Classification(00:55:38)
VC Dimension(00:55:51)
PAC Bounds for Classification(00:55:52)
VC Dimension(00:56:27)
A consistency problems(00:57:39)
Remarks on PAC+VC Bounds(00:58:33)
SVM: Linearly separable(00:59:21)
SVM: soft Margin(01:00:28)
SVM: Linearly separable(01:01:12)
SVM: soft Margin(01:01:22)
SVM: algorithms(01:01:59)
泛化能力的界(01:03:01)
Bound: VC Dimension(01:04:04)
Bound: VC dimension+errors(01:04:45)
Disadvantages of SRM(01:05:52)
Disadvantage: PAC+VC bound(01:06:52)
Several concepts(01:07:51)
Disadvantage: PAC+VC bound(01:08:00)
Several concepts(01:08:02)
Generalization Bound: margin(01:08:35)
Importance of Margin(01:09:48)
Generalization Bound: margin(01:10:29)
Importance of Margin(01:10:34)
Vapnik’s three periods(01:10:35)
Neural networks(01:11:51)
Interpretation: neural networks(01:12:55)
BP Algorithms(01:14:17)
Disadvantage(01:15:42)
The End(01:16:32)