{"id":1369,"date":"2017-02-14T15:45:18","date_gmt":"2017-02-14T19:45:18","guid":{"rendered":"http:\/\/dh.prattsils.org\/?p=1369"},"modified":"2017-02-14T15:45:18","modified_gmt":"2017-02-14T19:45:18","slug":"machine-learning-a-primer-a-dh-review","status":"publish","type":"post","link":"https:\/\/studentwork.prattsi.org\/dh\/2017\/02\/14\/machine-learning-a-primer-a-dh-review\/","title":{"rendered":"Machine Learning: A Primer: A DH Review"},"content":{"rendered":"<p style=\"text-align: center\"><strong>Machine Learning: A Primer<\/strong><\/p>\n<p style=\"text-align: center\"><strong>Speaker: Achim Koh<\/strong><\/p>\n<p style=\"text-align: center\"><strong>CUNY Graduate Center<\/strong><\/p>\n<hr \/>\n<p><span style=\"font-weight: 400\">Today, many critics push to make our algorithmic world transparent, to remove the cover on \u201cblack box\u201d machines and expose and deconstruct the systems by which these computers run. The scope of this feat, itself, appears daunting, algorithms lain in dense webs and expressed in arcane and protected jargon, yet the task becomes absurd when considering that many of these programs are not analyzable, dead, hard-coded, but organic, dynamic processes, executed by similar forces to what make human learning, and the agency that arises, so inscrutable.<\/span><\/p>\n<p><span style=\"font-weight: 400\">\u201cMachine Learning: A Primer,\u201d conducted by Achim Koh at CUNY\u2019s Graduate Center, introduced those that attended to not only the principles of machine learning (ML), but the issues that arise from the process. The talk split evenly in half, with the first hour devoted to the nature and ontology of machine learning, and the latter hour morphing into a discussion of its problems.<\/span><\/p>\n<p><span style=\"font-weight: 400\">Koh introduced the topic in comparison to more traditional algorithmic methods, where programmers code a certain set of functions to perform intended manipulations on some inputs, instead machine learners will train to infer output from data sources without the need for programmers to structure the functions themselves, allowing for statistical methods and computationally accelerated trial-and-error to come up with models that best fit datasets that may be too great for hard-coded formulations themselves. To illustrate this point, Koh showed a guide (found at: <\/span><a href=\"https:\/\/jalammar.github.io\/visual-interactive-guide-basics-neural-networks\/\"><span style=\"font-weight: 400\">https:\/\/jalammar.github.io\/visual-interactive-guide-basics-neural-networks\/<\/span><\/a><span style=\"font-weight: 400\">) where the viewer can discover the value of machine learning funamentals through interaction with a small dataset, line of \u201cbest fit\u201d, weights, and biases. Under \u201cTrain your Dragon,\u201d a minimal neural network can be hand-tweaked to reduce error in a similar manner to how a machine learner might do itself.<\/span><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-1370 size-full\" src=\"http:\/\/dh.prattsils.org\/wp-content\/uploads\/2017\/02\/trainyourdragon.png\" alt=\"trainyourdragon\" width=\"746\" height=\"639\" \/><\/p>\n<p><span style=\"font-weight: 400\">If you find this difficult and tedious, just imagine the effort undertaken in finding a line of best fit on millions of datapoints across multiple dimensions. Hence the power of machine learning, it systematizes guesswork (continuous evaluation based on observation) to accomplish objectives previously thought impossible by more formal methods.<\/span><\/p>\n<p><span style=\"font-weight: 400\">Thus, the goals of ML are threefold: pattern recognition, prediction, and decision-making, and the field is rife with terms to express these ends. Koh explained some of the lingo: the machine learns on \u201cvectors,\u201d or specific lines of data; \u201coptimization\u201d is the process by which machine increases \u201cprecision\u201d; this optimization is completed by \u201ctraining,\u201d enacted with the hope that the machine models the intended subset of reality with \u201caccuracy\u201d.<\/span><\/p>\n<p><span style=\"font-weight: 400\">Nevertheless, precision does not mean accuracy, as \u201coverfitting\u201d can arise where a machine matches datapoints so well that the eventual computer model represents the data and the noise present in the data, which contrarily increases in false negatives and positives.<\/span><\/p>\n<div id=\"attachment_1371\" style=\"width: 210px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" aria-describedby=\"caption-attachment-1371\" class=\"wp-image-1371 size-full\" src=\"http:\/\/dh.prattsils.org\/wp-content\/uploads\/2017\/02\/confusionmatrix.png\" alt=\"confusionmatrix\" width=\"200\" height=\"155\" \/><p id=\"caption-attachment-1371\" class=\"wp-caption-text\">(source:\u00a0https:\/\/upload.wikimedia.org\/wikipedia\/en\/a\/a6\/Binary_confusion_matrix.png)<\/p><\/div>\n<p><span style=\"font-weight: 400\">This \u201cconfusion matrix\u201d shows the possible predictive outcomes matched with the actual state of events in the world. The machine learner is in \u201cerror\u201d when it \u201cclassifies\u201d falsely, which leads to these false negatives and positives, \u201cclassification\u201d being when real world data is taken and processed by the algorithm that trained on a prior (training) set of data.<\/span><\/p>\n<p><span style=\"font-weight: 400\">The audience of the talk then learned about the types of machine learning, broken up most generally between \u201csupervised\u201d and \u201cunsupervised\u201d learning. I must admit, however, this distinction was unclearly presented. Supervised learning seems to deal \u201ctargets,\u201d where the machine seeks a function from input data to predetermined outcome, whereas unsupervised learning deals with pattern recognition of relationships between data points and clusters them accordingly. Other concepts such as \u201cneural networks\u201d, \u201cdeep learning\u201d, and \u201creinforcement learning\u201d were also briefly mentioned. <\/span><\/p>\n<p><span style=\"font-weight: 400\">Here, we glimpse the sense growth in the field of machine learning, and wonder, not only about its possibility, but its actuality, concealed by the facades of our present systems. The applications of machine learning are many and varied, and Koh mentioned \u201cnatural language processing,\u201d \u201ccomputer vision,\u201d \u201cdynamic pricing,\u201d \u201crecommendation and content curation,\u201d \u201crobotics,\u201d and \u201cart\u201d as the highlights. Of special interest, the art of machine learning allows us to gain a visual, aesthetic sense of the processes entailed by the technology, and Koh presented a number of sites that exhibit these projects (sources below).<\/span><\/p>\n<p><span style=\"font-weight: 400\">With such advancement in machine learning techniques, questions should arise as to their use. Machine learners are far from infallible. As with overfitting, the rote nature of algorithmic learning creates problems when dealing with heterogenic or complex data, in which minority classes in datasets might appear aberrant and mishandled by the program. This problem worsens when considering that we cannot account for these biases outright, as machine learning is opaque in its means. After a machine has trained off a dataset, programmers can only tweak parameters and adjust tolerances to get fewer false negative or positives, but never both at the same time. Artificial adjustment skews the entire system; a decrease in likelihood for false positives might increase the occurrence of false negatives, and vice versa.<\/span><\/p>\n<p><span style=\"font-weight: 400\">These issues appear in reality, extant in the penal system (<\/span><a href=\"https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing\"><span style=\"font-weight: 400\">https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing<\/span><\/a><span style=\"font-weight: 400\">). Computer learning algorithms exist in our justice system, assessing risk of re-offense by incarcerated individuals, across a number of factors (race not explicitly included). The results are staggering.<\/span><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-1372 size-full\" src=\"http:\/\/dh.prattsils.org\/wp-content\/uploads\/2017\/02\/predictionpropublica.png\" alt=\"predictionpropublica\" width=\"681\" height=\"241\" \/><\/p>\n<p><span style=\"font-weight: 400\">The study done by ProPublica finds that the algorithm assigns an overwhelming majority of inaccurate higher risk to the the black defendant population, in relation to the white counterpart. When risk scores see such wide use in all levels of the justice process, one would hope that these computational methods would help to alleviate bias, not hide it. This study evidences the contrary.<\/span><\/p>\n<p><span style=\"font-weight: 400\">As digital humanists, we have the obligation to see how the digital age affects its participants. The pervasiveness and opacity of machine learning requires further study and critique, and thus primers like the one run by Achim Koh on this discipline of Artificial Intelligence are both welcomed and necessary.<\/span><\/p>\n<p><span style=\"font-weight: 400\">Machine learning algorithms make digital life easier, as we can leverage computing on those most mundane yet awesome tasks, like analyzing a near continuous dataset to search for otherwise obscure patterns. The output are monumental accomplishments such as handwriting recognition and near-universal translation. Nevertheless, these neat displays belie the mess and diversity of the input data, and we wonder the what and how and why of what is left behind. I hope that machine learning improves and continues to provide innovative and unique solutions to today\u2019s problems, and I hope that critics continue to challenge its growth.<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p class=\"lead\">Machine Learning: A Primer Speaker: Achim Koh CUNY Graduate Center Today, many critics push to make our algorithmic world transparent, to remove the cover on \u201cblack box\u201d machines and expose and deconstruct the systems by which these computers run. The scope of this feat, itself, appears daunting, algorithms lain in dense webs and expressed in arcane and protected jargon, yet&hellip;<\/p>\n<p class=\"more-link-p\"><a class=\"btn btn-danger\" href=\"https:\/\/studentwork.prattsi.org\/dh\/2017\/02\/14\/machine-learning-a-primer-a-dh-review\/\">Read more &rarr;<\/a><\/p>\n","protected":false},"author":125,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[8],"tags":[47,48,51,103,130,144,146],"class_list":["post-1369","post","type-post","status-publish","format-standard","hentry","category-event-reviews","tag-dh","tag-dhweek","tag-digital-humanities","tag-machine-learning","tag-overfitting","tag-primer","tag-propublica"],"_links":{"self":[{"href":"https:\/\/studentwork.prattsi.org\/dh\/wp-json\/wp\/v2\/posts\/1369","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/studentwork.prattsi.org\/dh\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/studentwork.prattsi.org\/dh\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/studentwork.prattsi.org\/dh\/wp-json\/wp\/v2\/users\/125"}],"replies":[{"embeddable":true,"href":"https:\/\/studentwork.prattsi.org\/dh\/wp-json\/wp\/v2\/comments?post=1369"}],"version-history":[{"count":0,"href":"https:\/\/studentwork.prattsi.org\/dh\/wp-json\/wp\/v2\/posts\/1369\/revisions"}],"wp:attachment":[{"href":"https:\/\/studentwork.prattsi.org\/dh\/wp-json\/wp\/v2\/media?parent=1369"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/studentwork.prattsi.org\/dh\/wp-json\/wp\/v2\/categories?post=1369"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/studentwork.prattsi.org\/dh\/wp-json\/wp\/v2\/tags?post=1369"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}