Computational Linguistics
About

Yann LeCun

Yann LeCun (b. 1960) is a French-American computer scientist who pioneered convolutional neural networks and self-supervised learning, sharing the 2018 Turing Award and contributing foundational ideas to text classification and NLP through CNN-based methods.

CNN: h = f(W * x + b), where * denotes convolution

Yann LeCun is a French-American computer scientist who is VP and Chief AI Scientist at Meta and a professor at New York University. Together with Geoffrey Hinton and Yoshua Bengio, he shared the 2018 ACM A.M. Turing Award for contributions to deep learning. While his primary contributions are in computer vision, his development of convolutional neural networks and his advocacy for self-supervised learning have had substantial impact on text classification, character recognition, and representation learning for NLP.

Early Life and Education

Born in Paris, France, in 1960, LeCun studied at the Ecole Superieure d'Ingenieurs en Electrotechnique et Electronique (ESIEE) and earned his PhD from the Universite Pierre et Marie Curie in 1987. He worked at Bell Laboratories, where he developed LeNet for handwritten digit recognition, before joining NYU and later Facebook (now Meta) AI Research.

1960

Born in Paris, France

1987

Completed PhD at Universite Pierre et Marie Curie

1989

Published backpropagation applied to handwritten zip code recognition

1998

Published LeNet-5 and gradient-based learning for document recognition

2013

Became director of Facebook AI Research (FAIR)

2018

Received the ACM Turing Award with Hinton and Bengio

Key Contributions

LeCun's convolutional neural networks (CNNs) apply learned filters across input data using shared weights, dramatically reducing the number of parameters compared to fully connected networks. While developed for image recognition, CNNs have been widely adopted for NLP tasks: text classification using 1D convolutions over word embeddings (Kim, 2014), character-level models that learn features directly from characters, and sentence modelling using hierarchical convolutions.

LeCun has been a leading advocate for self-supervised learning, arguing that learning representations from unlabelled data through prediction tasks is the key to general intelligence. This philosophy directly aligns with the pre-training paradigm in NLP, where models like BERT learn from masked language modelling and GPT learns from next-word prediction — both forms of self-supervised learning on unlabelled text.

"Self-supervised learning is the cake, supervised learning is the icing on the cake, and reinforcement learning is the cherry on the cake." — Yann LeCun, on the importance of self-supervised learning

Legacy

LeCun's CNNs are used in text classification systems worldwide and his backpropagation techniques are fundamental to training all neural NLP models. His advocacy for self-supervised learning anticipated and inspired the pre-training revolution in NLP. As head of FAIR, he has overseen research that produced FastText, RoBERTa, and other influential NLP contributions. His emphasis on energy-based models and representation learning continues to influence the direction of both vision and language AI research.

Interactive Calculator

Enter a CSV of publications: year,title,citations_count. The calculator computes total citations, h-index, peak year, and a per-decade breakdown of scholarly output.

Click Calculate to see results, or Animate to watch the statistics update one record at a time.

Related Topics

References

  1. LeCun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278–2324. doi:10.1109/5.726791
  2. LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521(7553), 436–444. doi:10.1038/nature14539
  3. Kim, Y. (2014). Convolutional neural networks for sentence classification. Proceedings of EMNLP, 1746–1751.
  4. LeCun, Y. (2022). A path towards autonomous machine intelligence. OpenReview preprint.

External Links