Posted by Alex Franz and Thorsten Brants, Google Machine Translation TeamHere at Google Research we have been using word n-gram models for a variety of R&D projects, such as statistical machine translation, speech recognition, spelling correction, entity detection, information extraction, and others. While such models have usually been estimated from training corpora containing at most a few billion words, we have been harnessing the vast power of Google's datacenters and distributed processing infrastructure to process larger and larger training...
Thursday, 3 August 2006
Subscribe to:
Posts (Atom)