网站大量收购闲置独家精品文档,联系QQ:2885784924

第十一章 分类与预测.pptVIP

  1. 1、本文档共69页,可阅读全部内容。
  2. 2、有哪些信誉好的足球投注网站(book118)网站文档一经付费(服务费),不意味着购买了该文档的版权,仅供个人/单位学习、研究之用,不得用于商业用途,未经授权,严禁复制、发行、汇编、翻译或者网络传播等,侵权必究。
  3. 3、本站所有内容均由合作方或网友上传,本站不对文档的完整性、权威性及其观点立场正确性做任何保证或承诺!文档内容仅供研究参考,付费前请自行鉴别。如您付费,意味着您自己接受本站规则且自行承担风险,本站不退款、不进行额外附加服务;查看《如何避免下载的几个坑》。如果您已付费下载过本站文档,您可以点击 这里二次下载
  4. 4、如文档侵犯商业秘密、侵犯著作权、侵犯人身权等,请点击“版权申诉”(推荐),也可以打举报电话:400-050-0827(电话支持时间:9:00-18:30)。
  5. 5、该文档为VIP文档,如果想要下载,成为VIP会员后,下载免费。
  6. 6、成为VIP后,下载本文档将扣除1次下载权益。下载后,不支持退款、换文档。如有疑问请联系我们
  7. 7、成为VIP后,您将拥有八大权益,权益包括:VIP文档下载权益、阅读免打扰、文档格式转换、高级专利检索、专属身份标志、高级客服、多端互通、版权登记。
  8. 8、VIP文档为合作方或网友上传,每下载1次, 网站将根据用户上传文档的质量评分、类型等,对文档贡献者给予高额补贴、流量扶持。如果你也想贡献VIP文档。上传文档
查看更多
第十一章 分类与预测

* SVM—Introduction Literature “Statistical Learning Theory” by Vapnik: extremely hard to understand, containing many errors too. C.?J.?C. Burges. A Tutorial on Support Vector Machines for Pattern Recognition. Knowledge Discovery and Data Mining, 2(2), 1998. Better than the Vapnik’s book, but still written too hard for introduction, and the examples are so not-intuitive The book “An Introduction to Support Vector Machines” by N. Cristianini and J. Shawe-Taylor Also written hard for introduction, but the explanation about the mercer’s theorem is better than above literatures The neural network book by Haykins Contains one nice chapter of SVM introduction * Lazy vs. Eager Learning Lazy vs. eager learning Lazy learning (e.g., instance-based learning): Simply stores training data (or only minor processing) and waits until it is given a test tuple Eager learning (the above discussed methods): Given a set of training set, constructs a classification model before receiving new (e.g., test) data to classify Lazy: less time in training but more time in predicting Accuracy Lazy method effectively uses a richer hypothesis space since it uses many local linear functions to form its implicit global approximation to the target function Eager: must commit to a single hypothesis that covers the entire instance space * Lazy Learner: Instance-Based Methods Instance-based learning: Store training examples and delay the processing (“lazy evaluation”) until a new instance must be classified Typical approaches k-nearest neighbor approach Instances represented as points in a Euclidean space. 近邻分类方法 近邻分类方法是基于实例的分类方法 不需要事先进行分类器的设计 直接使用训练集对未知类标号的数据样本进行分类 最近邻分类、k-近邻分类 * 最近邻法 最近邻法:nearest neighborhood classifier (nnc),将与测试样本最近邻样本的类别作为决策的结果。 对一个C类别问题,每类有Ni个样本,i=1,…,C,则第i类ωi的判别函数为: 决策规则: 最近邻法在原理上最直观,方法上也十分简单,明显的缺点就是计算量大,存储量大。 ‖·‖表示某种距离(相似性)度量,常用欧氏距离作为相似性度量。 * * The k-Nearest Neighbor Algorithm All instances correspond to points in the n-D space The nearest neighbor are defined in terms of

文档评论(0)

panguoxiang + 关注
实名认证
文档贡献者

该用户很懒,什么也没介绍

1亿VIP精品文档

相关文档