home overview techniques technology resources about
 
Techniques

Introduction

Remember all those times when you had to write an essay in Spanish or French and you were tempted to just write it in English and copy and paste it into BabelFish? And then, just to see, you tried a sentence and it came out completely garbled? In case you don’t, here’s an example.

English: We like natural language processing because it makes us intellectually excited.
French : Nous aimons le langage naturel traitant parce qu'il nous fait intellectuellement passionnants.
English again: We like the natural language treating parce qu'il makes us intellectually enthralling.

It may not be obvious at first why, at least at this point, getting a computer to understand language in a manner even marginally similar to the way a human does should be such a difficult task. In theory, after all, language is just a large amount of data that follows a set of rules. The rules are extensive, of course, and the data is abundant, but it seems that computers are tailored exactly to the task of performing complex manipulations on masses of data. What, then, is the problem?

This is the first topic that Martin Kay, Jean Mark Gawron, and Peter Norvig address in their book Verbmobil, which describes an attempt to make a portable language translation machine. They say that there are two problems:

1. The meaning that a word, a phrase, or a sentence conveys is determined not just by itself, but by other parts of the text, both preceding and following.
2. The meaning of a text as a whole is not determined by the words, phrases, and sentences that make it up, but by the situation in which it is used (Kay 13).