{"id":845,"date":"2019-12-16T00:09:32","date_gmt":"2019-12-16T00:09:32","guid":{"rendered":"http:\/\/www.wellformedness.com\/blog\/?p=845"},"modified":"2019-12-19T15:33:23","modified_gmt":"2019-12-19T15:33:23","slug":"theory-error-analysis","status":"publish","type":"post","link":"https:\/\/www.wellformedness.com\/blog\/theory-error-analysis\/","title":{"rendered":"A theory of error analysis"},"content":{"rendered":"<p>Manual error analyses can help to identify the strengths and weaknesses of computational systems, ultimately suggesting future improvements and guiding development. However, they are often treated as an afterthought or neglected altogether. In three of my recent papers, we have been slowly developing what might be called a theory of error analysis. The systems evaluated include:<\/p>\n<ul>\n<li>number normalization (Gorman &amp; Sproat 2016); e.g., mapping <em>97000<\/em> onto <em>quatre vingt dix sept mille,<\/em><\/li>\n<li>inflection generation (Gorman et al. 2019); e.g., mapping pairs citation form and inflectional specification like (<em>aufbauen<\/em>, <tt>V;IND;PRS;2<\/tt>) onto inflected forms like <em>baust auf<\/em>, and<\/li>\n<li>grapheme-to-phoneme conversion (Lee et al. 2020); e.g., mapping orthographic forms like <em>almohadilla<\/em> onto phonemic or phonetic forms like \/almoa\u02c8di\u028ea\/ and [almoa\u02c8\u00f0i\u028ea].<\/li>\n<\/ul>\n<p>While these are rather different types of problems, the systems all have one thing in common: they <em>generate<\/em> linguistic representations. I discern three major classes of error such systems might make.<\/p>\n<ul>\n<li><strong>Target errors<\/strong> are only apparent errors; they arise when the gold data, the data to be predicted, is linguistically incorrect. This is particularly likely to arise with crowd-sourced data though such errors are also present in professionally annotated resources.<\/li>\n<li><strong>Linguistic errors <\/strong>are caused by misapplication of independently attested linguistic behaviors to the wrong input representations.\n<ul>\n<li>In the case of number normalization, these include using the wrong agreement affixes in Russian numbers; e.g., nom.sg. *\u0441\u0435\u043c\u044c\u0434\u0435\u0441\u044f\u0442 \u043c\u0438\u043b\u043b\u0438\u043e\u043d for gen.sg. \u0441\u0435\u043c\u044c\u0434\u0435\u0441\u044f\u0442 \u043c\u0438\u043b\u043b\u0438\u043e\u043d<strong>\u043e\u0432\u00a0<\/strong>&#8216;nine hundred million&#8217; (Gorman &amp; Sproat 2016:516)<\/li>\n<li>In inflection generation, these are what Gorman et al. 2019 call <em>allomorphy<\/em> errors; e.g., for instance, overapplying ablaut to the Dutch weak verb <em>printen<\/em> &#8216;to print&#8217; to produce a preterite *<em>pr<\/em><strong><em>o<\/em><\/strong><em>nt<\/em> instead of <em>printte\u00a0<\/em>(Gorman et al. 2019:144).<\/li>\n<li>In grapheme-to-phoneme conversion, these include failures to apply allophonic rules; e,g, in Korean, \uc775\uba85 &#8216;anonymity\u2019 is incorrectly transcribed as [ikmj\u028c\u0339\u014b] instead of [i\u014bmj\u028c\u0339\u014b], reflecting a failure to apply a rule of obstruent nasalization not indicated in the highly abstract <a href=\"https:\/\/en.wikipedia.org\/wiki\/Hangul\">hangul<\/a> orthography (Lee et al. under review).<\/li>\n<\/ul>\n<\/li>\n<li><strong>Silly errors<\/strong> are those errors which cannot be analyzed as either target errors or linguistic errors. These have long been noted as a feature of neural network models (e.g., Pinker &amp; Prince 1988, Sproat 1992:216f. for discussion of *<em>membled<\/em>) and occur even with modern neural network models.<\/li>\n<\/ul>\n<p>I propose that this tripartite distinction is a natural starting point when building an error taxonomy for many other language technology tasks, namely those that can be understood as generating linguistic sequences.<\/p>\n<h1>References<\/h1>\n<p>K. Gorman, A. D. McCarthy, R. Cotterell, E. Vylomova, M. Silfverberg, and M. Markowska (2019). Weird inflects but OK: making sense of morphological generation errors. In <em>CoNLL<\/em>, 140-151.<br \/>\nK. Gorman and R. Sproat (2016). Minimally supervised number normalization. <em>Transactions of the Association for Computational Linguistics<\/em> 4: 507-519.<br \/>\nJ. L. Lee, L. F.E. Ashby, M. E. Garza, Y. Lee-Sikka, S. Miller, A. Wong, A. D. McCarthy, and K. Gorman (under review). Massively multilingual pronunciation mining with WikiPron.<br \/>\nS. Pinker and A. Prince (1988). On language and connectionism: analysis of a parallel distributed processing model of language acquisition. <em>Cognition <\/em>28(1\u20132):73\u2013193.<br \/>\nR. Sproat (1992).\u00a0<em>Morphology and computation<\/em>. Cambridge: MIT Press.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Manual error analyses can help to identify the strengths and weaknesses of computational systems, ultimately suggesting future improvements and guiding development. However, they are often treated as an afterthought or neglected altogether. In three of my recent papers, we have been slowly developing what might be called a theory of error analysis. The systems evaluated &hellip; <a href=\"https:\/\/www.wellformedness.com\/blog\/theory-error-analysis\/\" class=\"more-link\">Continue reading<span class=\"screen-reader-text\"> &#8220;A theory of error analysis&#8221;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_crdt_document":"","footnotes":""},"categories":[4,5,6],"tags":[],"class_list":["post-845","post","type-post","status-publish","format-standard","hentry","category-language","category-nlp","category-phonology"],"_links":{"self":[{"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/posts\/845","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/comments?post=845"}],"version-history":[{"count":5,"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/posts\/845\/revisions"}],"predecessor-version":[{"id":850,"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/posts\/845\/revisions\/850"}],"wp:attachment":[{"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/media?parent=845"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/categories?post=845"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.wellformedness.com\/blog\/wp-json\/wp\/v2\/tags?post=845"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}