Natural Language Understanding
For Machine Intelligence to truly be useful, it should excel at tasks that humans are good at, such as natural language understanding. The Google Brain team’s language understanding research focuses on developing learning algorithms that are capable of understanding language to enable machines to translate text, answer questions, summarize documents, or conversationally interact with humans.
Our research in this area started with neural language models and word vectors. Our work on word vectors, word2vec - which learns to map words to vectors, was opensourced in 2013 and has since gained widespread adoption in the research community and industry. Our work on language models has also made great strides (see this and this) in improving state-of-art prediction accuracies.
We also conduct fundamental research that leads to a series of advances using neural networks for end-to-end language (or language-related) tasks such as translation, parsing, speech recognition, image captioning and conversation modeling. The underlying technology is the seq2seq framework, which is also now used in SmartReply (and other products) at Google and is opensourced in TensorFlow.
Our recent research highlights are also in the areas of semi/unsupervised learning, multitask learning, learning to manipulate symbols and learning with augmented logic and arithmetic.
Some of Our Publications
- Distributed Representations of Words and Phrases and their Compositionality Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Corrado, Jeffrey Dean. NIPS, 2013 (1,848 citations)
- Efficient Estimation of Word Representations in Vector Space Tomas Mikolov, Kai Chen, Greg Corrado, Jeffrey Dean. ICLR, 2013 (1,820 citations)
- Sequence to Sequence Learning with Neural Networks Ilya Sutskever, Oriol Vinyals, Quoc V. Le. NIPS, 2014 (650 citations)
- Semi-supervised Sequence Learning Andrew M. Dai, Quoc V. Le. NIPS, 2015 (24 citations)
- Exploring the Limits of Language Modeling Rafal Jozefowicz, Oriol Vinyals, Mike Schuster, Noam Shazeer, Yonghui Wu. ArXiv, 2016 (22 citations)
- Neural Programmer: Inducing Latent Programs with Gradient Descent Arvind Neelakantan, Quoc V. Le, Ilya Sutskever. ICLR, 2016 (19 citations)
Publications by Year
- 2016
- Can Active Memory Replace Attention? Lukasz Kaiser and Samy Bengio. NIPS, 2016 (4 citations)
- Collective Entity Resolution with Multi-Focal Attention Amir Globerson, Nevena Lazic, Soumen Chakrabarti, Amarnag Subramanya, Michael Ringaard and Fernando Pereira. ACL 2016.
- Exploring the Limits of Language Modeling Rafal Jozefowicz, Oriol Vinyals, Mike Schuster, Noam Shazeer, Yonghui Wu. ArXiv, 2016 (75 citations)
- Generating Sentences from a Continuous Space Samuel R. Bowman, Luke Vilnis, Oriol Vinyals, Andrew M. Dai, Rafal Jozefowicz, Samy Bengio. ArXiv, 2016 (59 citations)
- Multi-task Sequence to Sequence Learning Minh-Thang Luong, Quoc V. Le, Ilya Sutskever, Oriol Vinyals, Lukasz Kaiser. ICLR, 2016 (64 citations)
- Multilingual Language Processing From Bytes Dan Gillick, Cliff Brunk, Oriol Vinyals, Amarnag Subramanya, NAACL, 2016 (28 citations)
- Neural Programmer: Inducing Latent Programs with Gradient Descent Arvind Neelakantan, Quoc V. Le, Ilya Sutskever. ICLR, 2016 (58 citations)
- Reward Augmented Maximum Likelihood for Neural Structured Prediction Mohammad Norouzi, Samy Bengio, Zhifeng Chen, Navdeep Jaitly, Mike Schuster, Yonghui Wu, Dale Schuurmans. NIPS, 2016 (10 citations)
- SmartReply: Automated Response Suggestion for Email Anjuli Kannan, Karol Kurach, Sujith Ravi, Tobias Kaufmann, Andrew Tomkins, Balint Miklos, Greg Corrado, Laszlo Lukacs, Marina Ganea, Peter Young, Vivek Ramavajjala. ArXiv, 2016 (18 citations)
- Virtual adversarial training on semi-supervised text classification Takeru Miyato, Andrew Dai and Ian Goodfellow. ArXiv, 2016 (4 citations)
- 2015
- A Neural Conversational Model Oriol Vinyals, Quoc V. Le. ICML Deep Learning Workshop, 2015. (127 citations)
- Addressing the Rare Word Problem in Neural Machine Translation Minh-Thang Luong, Ilya Sutskever, Quoc V. Le, Oriol Vinyals, Wojciech Zaremba. ACL, 2015 (155 citations)
- BilBOWA: Fast bilingual distributed representations without word alignments Stephan Gouws, Yoshua Bengio, Greg Corrado. ICML, 2015 (82 citations)
- Grammar as a Foreign Language Oriol Vinyals, Lukasz Kaiser, Terry Koo, Slav Petrov, Ilya Sutskever, Geoffrey Hinton. NIPS, 2015 (196 citations)
- Pointer Networks Oriol Vinyals, Meire Fortunato, Navdeep Jaitly. NIPS, 2015 (92 citations)
- Semi-supervised Sequence Learning Andrew M. Dai, Quoc V. Le. NIPS, 2015 (66 citations)
- Sentence Compression by Deletion with LSTMs Katja Filippova, Enrique Alfonseca, Carlos Colmenares, Lukasz Kaiser, Oriol Vinyals. Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP'15) (32 citations)
- Show and Tell: A Neural Image Caption Generator Oriol Vinyals, Alexander Toshev, Samy Bengio, Dumitru Erhan. CVPR, 2015 (766 citations)
- 2014
- Distributed Representations of Sentences and Documents Quoc V. Le, Tomas Mikolov. ICML, 2014 (530 citations)
- Grounded compositional semantics for finding and describing images with sentences Richard Socher, Andrej Karpathy, Quoc V. Le, Christopher D. Manning, Andrew Y. Ng. TACL, 2014 (183 citations)
- Sequence to Sequence Learning with Neural Networks Ilya Sutskever, Oriol Vinyals, Quoc V. Le. NIPS, 2014 (650 citations)
- 2013
- Distributed Representations of Words and Phrases and their Compositionality Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Corrado, Jeffrey Dean. NIPS, 2013 (3,581 citations)
- Document embedding with paragraph vectors Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Corrado, Jeffrey Dean. NIPS, 2013 (38 citations)
- Efficient Estimation of Word Representations in Vector Space Tomas Mikolov, Kai Chen, Greg Corrado, Jeffrey Dean. ICLR, 2013 (3,151 citations)
- Exploiting Similarities among Languages for Machine Translation Tomas Mikolov, Quoc V. Le, Ilya Sutskever. ArXiv, 2013 (270 citations)
- Listen, Attend and Spell William Chan, Navdeep Jaitly, Quoc V. Le, Oriol Vinyals. NIPS, 2013 (73 citations)
- One Billion Word Benchmark for Measuring Progress in Statistical Language Modeling Ciprian Chelba, Tomas Mikolov, Mike Schuster, Qi Ge, Thorsten Brants, Phillipp Koehn, Tony Robinson. ArXiv, 2013. (117 citations)