Nouvelles et reportages
Objects are the secret key to revealing the world between vision and language
| Chunyuan Li, Lei Zhang, et Jianfeng Gao
Humans perceive the world through many channels, such as images viewed by the eyes or voices heard by the ears. Though any individual channel might be incomplete or noisy, humans can naturally align and fuse the information collected from multiple…
A deep generative model trifecta: Three advances that work towards harnessing large-scale power
| Chunyuan Li et Jianfeng Gao
One of the core aspirations in artificial intelligence is to develop algorithms and techniques that endow computers with an ability to synthesize the observed data in our world. Every time researchers build a model to imitate this ability, this model…
Going deep on deep learning with Dr. Jianfeng Gao
Dr. Jianfeng Gao is a veteran computer scientist, an IEEE Fellow and the current head of the Deep Learning Group at Microsoft Research. He and his team are exploring novel approaches to advancing the state-of-the-art on deep learning in areas…
Next-generation architectures bridge gap between neural and symbolic representations with neural symbols
| Paul Smolensky
In both language and mathematics, symbols and their mutual relationships play a central role. The equation x = 1/y asserts the symbols x and y—that is, what they stand for—are related reciprocally; Kim saw the movie asserts that Kim and…
Expanding scene and language understanding with large-scale pre-training and a unified architecture
| Hamid Palangi
Making sense of the world around us is a skill we as human beings begin to learn from an early age. Though there is still much to know about the process, we can see that people learn a lot, both…
See what we mean – Visually grounded natural language navigation is going places
How do humans communicate efficiently? The common belief is that the words humans use to communicate – such as dog, for instance – invoke similar understanding of the physical concepts. Indeed, there exists a common conception about the physical appearance…
Dans l’actualité | Synced
ICLR 2019 | MILA, Microsoft, and MIT Share Best Paper Honours
Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks, from the Montreal Institute for Learning Algorithms (MILA) and the Microsoft Research Montréal lab, was one of two Best Paper winners at ICLR 2019.
Less pain, more gain: A simple method for VAE training with less of that KL-vanishing agony
| Chunyuan Li
There is a growing interest in exploring the use of variational auto-encoders (VAE), a deep latent variable model, for text generation. Compared to the standard RNN-based language model that generates sentences one word at a time without the explicit guidance…
Towards universal language embeddings
| Jianfeng Gao
Language embedding is a process of mapping symbolic natural language text (for example, words, phrases and sentences) to semantic vector representations. This is fundamental to deep learning approaches to natural language understanding (NLU). It is highly desirable to learn language…