Show simple item record

dc.contributor.authorXu, Isaac
dc.date.accessioned2022-08-04T17:48:57Z
dc.date.available2022-08-04T17:48:57Z
dc.date.issued2022-08-04
dc.identifier.urihttp://hdl.handle.net/10222/81773
dc.descriptionThe work presented looks to gain insight into notions of complexity and difficulty for a task by conducting an exercise in predicting the degree of difficulty a population of models may have on arbitrary classification tasks for a toy dataset. In establishing the differing model evaluation results from these tasks, an argument is made for a label-free means to evaluate models. Methods for evaluating learning without labels such as a clustering-based metrics and entropy are examined. Entropy was found to be the most effective measure to evaluate learning, but issues pertaining to learning methodology and early learning instability require further study.en_US
dc.description.abstractIn this work, we explore the viability of proposed label-free metrics to evaluate models. We begin by examining the effect on linear probe accuracy which different viable label schemes on an identical dataset may cause. We show that in a toy setting, a notion of “complexity” for distinguishing classes can have predictive capabilities for anticipating relative “difficulty” a population of models may encounter for a comparison between classification tasks. In establishing these arbitrary relative differences in valid formulations for an evaluation task, we justify the search for a label scheme independent means to evaluate learning. To this end, we examine label-free clustering-based metrics and entropy on representational spaces at progressive milestones during self-supervised learning and on pre-trained representational spaces. While clustering-based metrics show mixed success, entropy may be viable for monitoring learning and cross-architectural comparisons, despite displaying instability in early training and showing differing trends for certain learning methodologies.en_US
dc.language.isoenen_US
dc.subjectMachine Learningen_US
dc.subjectSelf-Supervised Learningen_US
dc.subjectClusteringen_US
dc.subjectComplexityen_US
dc.subjectInformation Theoryen_US
dc.titleTowards a Label-Free and Representation-Based Metric for Evaluating Machine Learning Modelsen_US
dc.date.defence2022-07-20
dc.contributor.departmentFaculty of Computer Scienceen_US
dc.contributor.degreeMaster of Computer Scienceen_US
dc.contributor.external-examinern/aen_US
dc.contributor.graduate-coordinatorDr. Michael McAllisteren_US
dc.contributor.thesis-readerDr. Malcolm Heywooden_US
dc.contributor.thesis-readerDr. Sageev Ooreen_US
dc.contributor.thesis-supervisorDr. Thomas Trappenbergen_US
dc.contributor.ethics-approvalNot Applicableen_US
dc.contributor.manuscriptsNot Applicableen_US
dc.contributor.copyright-releaseNot Applicableen_US
 Find Full text

Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record