书签 分享 收藏 举报 版权申诉 / 26
上传文档赚钱

类型人工智能的课件CH18-Learning-from-observations.ppt

  • 上传人(卖家):三亚风情
  • 文档编号:2810324
  • 上传时间:2022-05-28
  • 格式:PPT
  • 页数:26
  • 大小:858KB
  • 【下载声明】
    1. 本站全部试题类文档,若标题没写含答案,则无答案;标题注明含答案的文档,主观题也可能无答案。请谨慎下单,一旦售出,不予退换。
    2. 本站全部PPT文档均不含视频和音频,PPT中出现的音频或视频标识(或文字)仅表示流程,实际无音频或视频文件。请谨慎下单,一旦售出,不予退换。
    3. 本页资料《人工智能的课件CH18-Learning-from-observations.ppt》由用户(三亚风情)主动上传,其收益全归该用户。163文库仅提供信息存储空间,仅对该用户上传内容的表现方式做保护处理,对上传内容本身不做任何修改或编辑。 若此文所含内容侵犯了您的版权或隐私,请立即通知163文库(点击联系客服),我们立即给予删除!
    4. 请根据预览情况,自愿下载本文。本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。
    5. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007及以上版本和PDF阅读器,压缩文件请下载最新的WinRAR软件解压。
    配套讲稿:

    如PPT文件的首页显示word图标,表示该PPT已包含配套word讲稿。双击word图标可打开word文档。

    特殊限制:

    部分文档作品中含有的国旗、国徽等图片,仅作为作品整体效果示例展示,禁止商用。设计者仅对作品中独创性部分享有著作权。

    关 键  词:
    人工智能 课件 CH18 Learning from observations
    资源描述:

    1、智能计算研究中心Learning from Observations(chapter18)Autumn 2012Instructor: Wang XiaolongHarbin Institute of Technology, Shenzhen Graduate SchoolIntelligent Computation Research Center(HITSGS ICRC)2Outlines Learning agents Inductive learning Decision tree learning Measuring learning performance3Learning Lea

    2、rning is essential for unknown environments, i.e., when designer lacks omniscience Learning is useful as a system construction method, i.e., expose the agent to reality rather than trying to write it down Learning modifies the agents decision mechanisms to improve performance4Learning agents5Learnin

    3、g element Design of a learning element is affected by Which components of the performance element are to be learned What feedback is available to learn these components What representation is used for the components Type of feedback: Supervised learning: involves learning a function from examples of

    4、 its input and outputs. Unsupervised learning: involves learning patterns in the input when no specific output values are supplied. Reinforcement learning: learn from rewards (reinforcement)6Inductive learning Simplest form: learn a function from examplesf is the target functionAn example is a pair

    5、(x, f(x)Problem: find a hypothesis hsuch that h fgiven a training set of examples(This is a highly simplified model of real learning: Ignores prior knowledge Assumes examples are given)7Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on

    6、 all examples)E.g., curve fitting:8Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on all examples)E.g., curve fitting:9Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on a

    7、ll examples)E.g., curve fitting:10Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on all examples)E.g., curve fitting:11Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on a

    8、ll examples)E.g., curve fitting:12Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on all examples)E.g., curve fitting:Ockhams razor: prefer the simplest hypothesis consistent with data - In Latin, it means “Entities are not to be multip

    9、lied beyond necessity”13Learning decision treesProblem: decide whether to wait for a table at a restaurant, based on the following attributes:1.Alternate: is there an alternative restaurant nearby?2.Bar: is there a comfortable bar area to wait in?3.Fri/Sat: is today Friday or Saturday?4.Hungry: are

    10、we hungry?5.Patrons: number of people in the restaurant (None, Some, Full)6.Price: price range ($, $, $)7.Raining: is it raining outside?8.Reservation: have we made a reservation?9.Type: kind of restaurant (French, Italian, Thai, Burger)10. WaitEstimate: estimated waiting time (0-10, 10-30, 30-60, 6

    11、0)14Attribute-based representationsExamples described by attribute values (Boolean, discrete, continuous)E.g., situations where I will/wont wait for a table:Classification of examples is positive (T) or negative (F)15Decision treesOne possible representation for hypothesesE.g., here is the “true” tr

    12、ee for deciding whether to wait:16ExpressivenessDecision trees can express any function of the input attributes.E.g., for Boolean functions, truth table row path to leaf:Trivially, there is a consistent decision tree for any training set with one path to leaf for each example (unless f nondeterminis

    13、tic in x) but it probably wont generalize to new examplesPrefer to find more compact decision trees17Hypothesis spacesHow many distinct decision trees with n Boolean attributes?= number of Boolean functions= number of distinct truth tables with 2n rows = 22nE.g., with 6 Boolean attributes, there are

    14、 18,446,744,073,709,551,616 treesMore expressive hypothesis space increases chance that target function can be expressed increases number of hypotheses consistent with training set may get worse predictions18Decision tree learningAim: find a small tree consistent with the training examplesIdea: (rec

    15、ursively) choose most significant attribute as root of (sub)tree19Choosing an attribute Idea: a good attribute splits the examples into subsets that are (ideally) all positive or all negativePatrons? is a better choice20Using information theoryInformation answers questionsThe more clueless I am abou

    16、t the answer initially, the more information is contained in the answerInformation Content (Entropy):I(P(v1), , P(vn) = i=1 -P(vi) log2 P(vi)For a training set containing p positive examples and n negative examples:npnnpnnppnppnpnnppI22loglog),(21Information gainA chosen attribute A divides the trai

    17、ning set E into subsets E1, , Ev according to their values for A, where A has v distinct values.Information Gain (IG) or reduction in entropy from the attribute test:Choose the attribute with the largest IGviiiiiiiiinpnnppInpnpAremainder1),()()(),()(AremaindernpnnppIAIG22Information gainFor the trai

    18、ning set, p = n = 6, I(6/12, 6/12) = 1 bitConsider the attributes Patrons and Type (and others too):Patrons has the highest IG of all attributes and so is chosen by the DTL algorithm as the rootbits 0)42,42(124)42,42(124)21,21(122)21,21(1221)(bits 0541.)64,62(126)0 , 1 (124) 1 , 0(1221)(IIIITypeIGII

    19、IPatronsIG23Example contd. Decision tree learned from the 12 examples: Substantially simpler than “true” tree-a more complex hypothesis isnt justified by small amount of data24Performance measurementHow do we know that h f ?1. Use theorems of computational/statistical learning theory2. Try h on a ne

    20、w test set of examples(use same distribution over example space as training set)Learning curve = % correct on test set as a function of training set size25SummaryLearning needed for unknown environments, lazy designersLearning agent = performance element + learning elementFor supervised learning, the aim is to find a simple hypothesis approximately consistent with training examplesDecision tree learning using information gain Learning performance = prediction accuracy measured on test set26Assignments Ex 18.3

    展开阅读全文
    提示  163文库所有资源均是用户自行上传分享,仅供网友学习交流,未经上传用户书面授权,请勿作他用。
    关于本文
    本文标题:人工智能的课件CH18-Learning-from-observations.ppt
    链接地址:https://www.163wenku.com/p-2810324.html

    Copyright@ 2017-2037 Www.163WenKu.Com  网站版权所有  |  资源地图   
    IPC备案号:蜀ICP备2021032737号  | 川公网安备 51099002000191号


    侵权投诉QQ:3464097650  资料上传QQ:3464097650
       


    【声明】本站为“文档C2C交易模式”,即用户上传的文档直接卖给(下载)用户,本站只是网络空间服务平台,本站所有原创文档下载所得归上传人所有,如您发现上传作品侵犯了您的版权,请立刻联系我们并提供证据,我们将在3个工作日内予以改正。

    163文库