There ’s an older locution in robotics : Anything a human being learns to do after age 5 is well-fixed to teach a simple machine . Everything we read before 5 , not so easy . That spontaneous natural law of machine encyclopaedism might explicate why there are computers that can beat the worldly concern ’s honorable chess and Go master , but we ’ve yet to work up a robot that can walk like a human being . ( Do n’t seek to tell me thatASIMOwalks like a human being . )

This might also explain why the spellchecker on your computer works so brilliantly , but thegrammarchecker does n’t . We learn how to spell out only when we ’re old enough to go to school , but the basics of language development can begin as early as inthe womb .

Inference and Context

Spelling is a finite job with distinct correct or incorrect answers . English grammar , on the other hand , moderate a near infinite number of possibilities , and whether something is grammatically correct or wrong can for the most part bet on subtle clew like setting and inference .

That ’s why certain English sentences are such a pain in the neck for automated grammar chequer . Les Perelman , a retired MIT professor and former associate dean of undergraduate teaching who bleed the university ’s writing program , yield me this one : " The motorcar was park by John . "

My admittedly dated version of Microsoft Word ( Word for Mac 2011 ) is programme to recognize and adjust inactive voice , a no - no in most grammar circles . When I type this judgment of conviction into Word , the program dutifully underlines it in dark-green and suggests : " John park the railcar . " That would be fine if John had parked the car , but what if I meant that the car was physically parked near John ?

computer, check sign

uncomplicated misapprehension , you might say , but look what happens when I change the sentence to " The car was park by the curb . " Word emphasize it and suggests : " The kerb park the car . " That ’s downright goofy , even for a figurer .

" So much of English grammar involves inference and something called mutual contextual impression , " says Perelman . " When I make a statement , I believe that you know what I know about this . Machines are n’t that impertinent . you may trail the machine for a specific situation , but when you talk about transactions in human speech communication , there ’s actually a huge act of inferences like that going on all the time . "

Perelman has a beef with grammar checkers , which he claims simply do not work out . Citingprevious research , he regain that grammar checkers only correctly identified errors in student newspaper 50 per centum of the clip . And even risky , they often flagged dead good prose as a misapprehension , known as a simulated positive .

In one exercise , Perelman plug 5,000 words of a illustrious Noam Chomsky essay into thee - raterscoring engine by ETS , the company that bring out ( and grades ) the GRE and TOEFL examination . The grammar chequer found 62 mistake — admit 14 representative of a prison term starting with a coordinate conjunction ( " and , " " but , " " or " ) and nine missing commas — all but one of which Perelman classified as " perfectly grammatic prose . "

A Little History

The first automated spell checker transport with an early version of WordPerfect in 1983 , and the first computerized grammar checkers soon followed in both WordPerfect and Microsoft Word .

Mar Ginés Marín is a chief program managing director at Microsoft who ’s been fiddle with the Office grammar editor program for the past 17 years . She says that in the former solar day , the good Scripture could do was parse a sentence into its constituent parts of language and discover bare grammar errors like noun - verb agreement . Then engineers figured out how to parse a condemnation into small " lump " of two or three words to point things like " a / an " agreement . This is called born language processing or NLP .

The next step was to introduce motorcar eruditeness . Susan Hendrich is a group program coach at Microsoft in commission of the natural language processing teams working on Office . With motorcar learning , Microsoft engineers could go beyond program each and every grammar rule into the software . Instead , they train the machine on a Brobdingnagian dataset of correct English custom and allow the machine get word from the patterns it discovers .

Hendrich says that algorithmic rule spring up by Microsoft through motorcar learnedness are what drive Word ’s decisions about whether or not a conviction involve a query mark , or what types of clauses require a comma ( jolly tricky stuff , even for us human author ) .

But did it work ? Daniel Kies , an English prof at the College of Du Page , in Glen Ellyn , Illinois , once bear forrader - to - oral sex testof various grammar checkers ranging from WordPerfect 8 , released in the late 1990s , up to Word 2007 . When assure against 20 judgment of conviction containing the most common writing errors , all the grammar checkers perform fairly miserably . No version of Word after 2000 caught any of the fault ( strangely , Word 97 tally better ) and WordPerfect only identified 40 percent of the mistake .

While those numbers do n’t present the late versions of grammar chequer , they do point to one of the biggest challenges in create a powerful and precise grammar engine that ’s build into a piece of computer software — space .

" We can make these big beautiful models that have a high precision truth , but they ’re too liberal to transport in the loge with the product , " says Hendrich at Microsoft . " So we have to slim our example down , and as we reduce our model down we lose preciseness truth . So we have this balance tip that we ’re willing to transport with . "

Ginés Marín defends Word ’s preciseness but admits that space constraint affect the level of " coverage " that Microsoft ’s grammar checker provided . When the modeling was slimmed down to agree into the software , it also needed to be dialed back in breadth so that it did n’t flag lots of adept text as misunderstanding .

The Golden Squiggle

What ’s change since the days of Word 2007 is the rise of Web - based package applications . Now engineers do n’t have to jam a large grammar locomotive into a package lowly enough to live on the user ’s grueling drive . The grammar algorithms can populate in the cloud and be accessed over the cyberspace in material time .

Hendrich says that the web - based adaptation of Office already rely on racy grammar engines that are hosted in the swarm , and her team is currently in the process of go all the old built - in critiques and grammar models to the swarm , too . The challenge going forrard , say Hendrich , is to adjudicate how much functionality to keep " in the box " and how much to deliver " through the service , " as Hendrich call Microsoft ’s swarm - ground , software - as - a - service model .

The issue is toll . Every time Word calls up to the cloud for grammar advice , it cost a few fractions of a penny .

" If you ’re writing a 10 - pageboy document , do you call up to the Robert William Service on every keystroke ? " Hendrich call for . " When you get going looking at the toll models , it can be quite large . "

The previous version of Microsoft ’s grammar editor is far more robust than its precursor . error come with multiple correction suggestion plus explanations for the grammar rules behind them . There ’s a ramp up - in read - aloud mapping that ’s particularly helpful for masses with dyslexia and for non - native utterer . And there ’s a new type of suggestion that Hendrich calls the " golden curlicue " that accost compose manner more than canonic grammar .

If you write that the committee is looking for a new " president , " for example , the golden curlicue will suggest that you use a gender - neutral terminus like " chairperson . " If you ’re compose a memorandum to your hirer that requires a sure grade of formality , the gold curlicue will flag words that seem too casual like " comfy . "

One enquiry that ’s important to require is whether grammar checker really ask to be double-dyed . If Word suggests that the sentence should read " The curb parked the gondola , " you may just push aside it . No big raft , right ?

For native English speakers , a not - so - perfect grammar checker is a soft pique . Even if you ’re not a grammar whiz , you’re able to pick up it when something sound improper . The actual job , says former MIT writing professor Perelman , happens when English oral communication learners rely on these tools to adjust their piece of writing .

" It really depends who the user is , " say Perelman . " If the user is a native speaker , false positives are n’t as dangerous as they are to a non - native speaker . "

If Word tell an English language learner that " the curb park the gondola , " not only will their penning not make any sense , but they ’ll be learning tough grammar . Now that English has become the lingua franca of science and engineering science , Perelman says , businesses around the world are desperate for a unfeignedly reliable and accurate English grammar checker . That ’s why you see the climb of third - company , web - found grammar tools like Grammarly and Ginger , all attempt to adjoin this international demand .

The respectable intelligence is that the latest version of Word ( 2016 ) passes the " AMEX " test . Grammarly , however , flagged it as peaceful voice .