您好,欢迎访问三七文档
当前位置:首页 > 电子/通信 > 综合/其它 > 机器学习-试卷-finals13
CS189Spring2013IntroductiontoMachineLearningFinal•Youhave3hoursfortheexam.•Theexamisclosedbook,closednotesexceptyourone-page(twosides)ortwo-page(oneside)cribsheet.•Pleaseusenon-programmablecalculatorsonly.•MarkyouranswersONTHEEXAMITSELF.Ifyouarenotsureofyouransweryoumaywishtoprovideabriefexplanation.AllshortanswersectionscanbesuccessfullyansweredinafewsentencesATMOST.•Fortrue/falsequestions,llintheTrue/Falsebubble.•Formultiple-choicequestions,llinthebubblesforALLCORRECTCHOICES(insomecases,theremaybemorethanone).Foraquestionwithppointsandkchoices,everyfalsepositivewilincurapenaltyofp=(k 1)points.•Forshortanswerquestions,unnecessarilylongexplanationsandextraneousdatawillbepenalized.Pleasetrytobeterseandpreciseanddothesidecalculationsonthescratchpapersprovided.•PleasedrawaboundingboxaroundyouranswerintheShortAnswerssection.Amissedanswerwithoutaboundingboxwillnotberegraded.FirstnameLastnameSIDForstauseonly:Q1.True/False/23Q2.MultipleChoiceQuestions/36Q3.ShortAnswers/26Total/851Q1.[23pts]True/False(a)[1pt]SolvinganonlinearseparationproblemwithahardmarginKernelizedSVM(GaussianRBFKernel)mightleadtoovertting.TrueFalse(b)[1pt]InSVMs,thesumoftheLagrangemultiplierscorrespondingtothepositiveexamplesisequaltothesumoftheLagrangemultiplierscorrespondingtothenegativeexamples.TrueFalse(c)[1pt]SVMsdirectlygiveustheposteriorprobabilitiesP(y=1jx)andP(y= 1jx).TrueFalse(d)[1pt]V(X)=E[X]2 E[X2]TrueFalse(e)[1pt]Inthediscriminativeapproachtosolvingclassicationproblems,wemodeltheconditionalprobabilityofthelabelsgiventheobservations.TrueFalse(f)[1pt]Inatwoclassclassicationproblem,apointontheBayesoptimaldecisionboundaryxalwayssatisesP(y=1jx)=P(y=0jx).TrueFalse(g)[1pt]AnylinearcombinationofthecomponentsofamultivariateGaussianisaunivariateGaussian.TrueFalse(h)[1pt]ForanytworandomvariablesXN(1;21)andYN(2;22),X+YN(1+2;21+22).TrueFalse(i)[1pt]StanfordandBerkeleystudentsaretryingtosolvethesamelogisticregressionproblemforadataset.TheStanfordgroupclaimsthattheirinitializationpointwillleadtoamuchbetteroptimumthanBerkeley'sinitializationpoint.Stanfordiscorrect.TrueFalse(j)[1pt]Inlogisticregression,wemodeltheoddsratio(p1 p)asalinearfunction.TrueFalse(k)[1pt]Randomforestscanbeusedtoclassifyinnitedimensionaldata.TrueFalse(l)[1pt]InboostingwestartwithaGaussianweightdistributionoverthetrainingsamples.TrueFalse(m)[1pt]InAdaboost,theerrorofeachhypothesisiscalculatedbytheratioofmisclassiedexamplestothetotalnumberofexamples.TrueFalse(n)[1pt]Whenk=1andN!1,thekNNclassicationrateisboundedabovebytwicetheBayeserrorrate.TrueFalse(o)[1pt]Asinglelayerneuralnetworkwithasigmoidactivationforbinaryclassicationwiththecrossentropylossisexactlyequivalenttologisticregression.TrueFalse2(p)[1pt]ThelossfunctionforLeNet5(theconvolutionalneuralnetworkbyLeCunetal.)isconvex.TrueFalse(q)[1pt]Convolutionisalinearoperationi.e.(f1+f2)g=f1g+f2g.TrueFalse(r)[1pt]Thek-meansalgorithmdoescoordinatedescentonanon-convexobjectivefunction.TrueFalse(s)[1pt]A1-NNclassierhashighervariancethana3-NNclassier.TrueFalse(t)[1pt]Thesinglelinkagglomerativeclusteringalgorithmgroupstwoclustersonthebasisofthemaximumdistancebetweenpointsinthetwoclusters.TrueFalse(u)[1pt]Thelargesteigenvectorofthecovariancematrixisthedirectionofminimumvarianceinthedata.TrueFalse(v)[1pt]TheeigenvectorsofAATandATAarethesame.TrueFalse(w)[1pt]Thenon-zeroeigenvaluesofAATandATAarethesame.TrueFalse3Q2.[36pts]MultipleChoiceQuestions(a)[4pts]Inlinearregression,wemodelP(yjx)N(wTx+w0;2).Theirreducibleerrorinthismodelis.2E[(y E[yjx])2jx]E[(y E[yjx])jx]E[yjx](b)[4pts]LetS1andS2bethesetofsupportvectorsandw1andw2bethelearntweightvectorsforalinearlyseparableproblemusinghardandsoftmarginlinearSVMsrespectively.Whichofthefollowingarecorrect?S1S2w1=w2S1maynotbeasubsetofS2w1maynotbeequaltow2.(c)[4pts]Ordinaryleast-squaresregressionisequivalenttoassumingthateachdatapointisgeneratedaccordingtoalinearfunctionoftheinputpluszero-mean,constant-varianceGaussiannoise.Inmanysystems,however,thenoisevarianceisitselfapositivelinearfunctionoftheinput(whichisassumedtobenon-negative,i.e.,x0).Whichofthefollowingfamiliesofprobabilitymodelscorrectlydescribesthissituationintheunivariatecase?P(yjx)=1p2xexp( (y (w0+w1x))22x2)P(yjx)=1p2exp( (y (w0+w1x))222)P(yjx)=1p2xexp( (y (w0+(w1+2)x))222)P(yjx)=1xp2exp( (y (w0+w1x))22x22)(d)[3pts]TheleftsingularvectorsofamatrixAcanbefoundin.EigenvectorsofAATEigenvectorsofATAEigenvectorsofA2EigenvaluesofAAT(e)[3pts]Averagingtheoutputofmultipledecisiontreeshelps.IncreasebiasDecreasebiasIncreasevarianceDecreasevariance(f)[4pts]LetAbeasymmetricmatrixandSbethematrixcontainingitseigenvectorsascolumnvectors,andDadiagonalmatrixcontainingthecorrespondingeigenvaluesonthediagonal.Whichofthefollowingaretrue:AS=SDAS=DSSA=DSAS=DST(g)[4pts]Considerthefollowingdataset:A=(0;2),B=(0;1)andC=(1;0).Thek-meansalgorithmisinitializedwithcentersatAandB.Uponconvergence,thetwocenterswillbeatAandCAandthemidpointofBCCandthemidpointofABAandB4(h)[3pts]Whichofthefollowinglossfunctionsareconvex?Misclassica
本文标题:机器学习-试卷-finals13
链接地址:https://www.777doc.com/doc-6863018 .html