{"id":805,"date":"2009-06-15T10:54:27","date_gmt":"2009-06-15T16:54:27","guid":{"rendered":"http:\/\/hunch.net\/?p=805"},"modified":"2009-06-17T00:20:05","modified_gmt":"2009-06-17T06:20:05","slug":"in-active-learning-the-question-changes","status":"publish","type":"post","link":"https:\/\/hunch.net\/?p=805","title":{"rendered":"In Active Learning, the question changes"},"content":{"rendered":"<p>A little over 4 years ago, <a href=\"http:\/\/www.cs.ucsd.edu\/~dasgupta\/\">Sanjoy<\/a> <a href=\"https:\/\/hunch.net\/?p=49\">made a post<\/a> saying roughly &#8220;we should  study active learning theoretically, because not much is understood&#8221;.   <\/p>\n<p>At the time, we did not understand basic things such as whether or not it was possible to PAC-learn with an active algorithm without making strong assumptions about the noise rate.  In other words, the fundamental question was &#8220;can we do it?&#8221;<\/p>\n<p>The nature of the question has fundamentally changed in my mind.   The answer is to the previous question is &#8220;yes&#8221;, both information theoretically and computationally, most places where supervised learning could be applied.  <\/p>\n<p>In many situation, the question has now changed to: &#8220;is it worth it?&#8221;  Is the programming and computational overhead low enough to make the label cost savings of active learning worthwhile?  Currently, there are situations where this question could go either way.  Much of the challenge for the future is in figuring out how to make active learning easier or more worthwhile.<\/p>\n<p>At the <a href=\"https:\/\/hunch.net\/~active_learning\">active learning tutorial<\/a>, I stated a set of somewhat more precise research questions that I don&#8217;t yet have answer to, and which I believe are worth answering.  Here is a bit of an expansion on those questions for those interested.<\/p>\n<ol>\n<li>Is active learning possible in a fully adversarial setting?  By fully adversarial, I mean when an adversary controls all the algorithms observations.  <a href=\"http:\/\/www.conflate.net\/icml\/paper\/2009\/289\">Some work<\/a> by <a href=\"http:\/\/www.dicom.uninsubria.it\/~cgentile\/\">Claudio<\/a> and <a href=\"http:\/\/homes.dsi.unimi.it\/~cesabian\/\">Nicolo<\/a> has moved in this direction, but there is not yet a solid answer.<\/li>\n<li>Is there an efficient and effective reduction of active learning to supervised learning?  The <a href=\"http:\/\/arxiv.org\/abs\/0812.4952\">bootstrap IWAL<\/a> approach is efficient but not effective in some situations where other approaches can succeed.  The  algorithm <a href=\"http:\/\/www.cs.ucsd.edu\/~dasgupta\/papers\/cal.pdf\">here<\/a> is a reduction to a special kind of supervised learning where you can specify both examples and constraints.  For many supervised learning algorithms, adding constraints seems problematic.<\/li>\n<li>Can active learning succeed with alternate labeling oracles?  The ones I see people trying to use in practice often differ because they can provide answers of varying specificity and cost, or because some oracles are good for some questions, but not good for others.<\/li>\n<li>At this point, there have been several successful applications of active learning, but that&#8217;s not the same thing as succeeding with more robust algorithms.  Can we succeed empirically with more robust algorithms?  And is the empirical cost of additional robustness worth the empirical peace-of-mind that your learning algorithm won&#8217;t go astray where other more aggressive approaches may do so?<\/li>\n<\/ol>\n","protected":false},"excerpt":{"rendered":"<p>A little over 4 years ago, Sanjoy made a post saying roughly &#8220;we should study active learning theoretically, because not much is understood&#8221;. At the time, we did not understand basic things such as whether or not it was possible to PAC-learn with an active algorithm without making strong assumptions about the noise rate. In &hellip; <\/p>\n<p class=\"link-more\"><a href=\"https:\/\/hunch.net\/?p=805\" class=\"more-link\">Continue reading<span class=\"screen-reader-text\"> &#8220;In Active Learning, the question changes&#8221;<\/span><\/a><\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[22,2],"tags":[],"class_list":["post-805","post","type-post","status-publish","format-standard","hentry","category-active","category-questions"],"_links":{"self":[{"href":"https:\/\/hunch.net\/index.php?rest_route=\/wp\/v2\/posts\/805","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/hunch.net\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/hunch.net\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/hunch.net\/index.php?rest_route=\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/hunch.net\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=805"}],"version-history":[{"count":0,"href":"https:\/\/hunch.net\/index.php?rest_route=\/wp\/v2\/posts\/805\/revisions"}],"wp:attachment":[{"href":"https:\/\/hunch.net\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=805"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/hunch.net\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=805"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/hunch.net\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=805"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}