Who is Yoshua Bengio?

Yoshua Bengio is a French-born Canadian computer scientist who is sometimes referred to as one of the “Godfathers of AI” and “Godfathers of Deep Learning”. 

Few people have done more to advance the field of artificial intelligence than Bengio, but the man himself tends to avoid public attention where possible and has openly stated that he is not a fan of the so-called “personalization of science”.

Education

Bengio spent nine years at McGill University in Montreal between 1982 and 1991. Initially enrolled as a Bachelor of Engineering student, Bengio’s Master’s thesis focused on Boltzmann machines and multi-layer neural networks for speech recognition.

In 1988, he started his Ph.D. in computer science with a particular focus on how convolutional neural networks (CNNs) and hidden Markov models (HMMs) could be combined for automatic speech recognition.

Over the 1980s, Bengio was one of a handful of computer scientists who proposed that computers could one day function in a similar way to the human brain.

Post-doctoral research

In 1991, Bengio undertook post-doctoral research in the brain and cognitive sciences department at the Massachusetts Institute of Technology (MIT). 

He subsequently joined AT&T Labs in a similar capacity in 1992 before leaving that company to become a Full Professor at the Université de Montréal in 1993. 

There, in 1998, he published the now pioneering paper Gradient-Based Learning Applied To Document Recognition. The paper posited that with the correct architecture, certain algorithms could recognize handwritten characters with more accuracy than conventional technology that used hand-designed heuristics.

Bengio has now been at the Université de Montréal for more than 30 years and has authored hundreds of academic papers that have been cited over 650,000 times

IVADO Labs

Bengio became Scientific Director of IVADO Labs in 2017, a company that helps other businesses realize the transformative potential of artificial intelligence. Under a unique agreement, IVADO receives support from the Canadian government and founding academic institutions and has access to the country’s network of AI R&D labs.

Whilst at the company, Bengio was named the winner of the $1 million A.M. Turing Award – considered the Nobel Prize of computing – and shared the prize with contemporaries Geoffrey Hinton and Yann LeCun. 

Association for Computing Machinery (ACM) president Cherri M. Pancake noted that “The recent growth of and interest in AI is due, in no small part, to the recent advances in deep learning for which Bengio, Hinton, and LeCun laid the foundation.”

MILA

Today, Bengio leads the Montreal Institute for Learning Algorithms (MILA) – a machine learning (ML) research lab he founded in 1993 as Laboratoire d’informatique des systèmes adaptatifs (LISA). Aside from Bengio, notable faculty members include Doina Precup and Joëlle Pineau.

Under Bengio’s leadership, MILA has, among other things, collaborated with IBM to accelerate AI and ML research with open-source software. The collaboration combines the Montreal-based university’s Orion open-source product with IBM’s Watson Machine Learning Accelerator – an AI model and inference tool the company sells to interested parties.

Key takeaways:

  • Yoshua Bengio is a French-born Canadian computer scientist who is sometimes referred to as one of the “Godfathers of AI” and “Godfathers of Deep Learning”. 
  • Over the 1980s, Bengio was one of a handful of computer scientists who proposed that computers could soon function in a similar way to the human brain. He completed post-doctoral research at AT&T and MIT before authoring a seminal paper in 1998.
  • Bengio became Scientific Director of IVADO Labs in 2017, a company that helps other businesses realize the transformative potential of artificial intelligence. He is also the founder and current leader of machine learning research lab MILA.

Key Highlights

  • Background and Contributions:
    • Yoshua Bengio is a prominent computer scientist known as one of the “Godfathers of AI” and “Godfathers of Deep Learning.”
    • He has made significant contributions to the field of artificial intelligence and deep learning.
  • Education and Early Career:
    • Bengio spent nine years at McGill University, focusing on topics such as Boltzmann machines and multi-layer neural networks for speech recognition.
    • His Ph.D. research centered on combining convolutional neural networks (CNNs) and hidden Markov models (HMMs) for automatic speech recognition.
    • In the 1980s, Bengio was among the early computer scientists proposing that computers could emulate human brain functions.
  • Post-doctoral Research and Academia:
    • Bengio undertook post-doctoral research at the Massachusetts Institute of Technology (MIT) in the brain and cognitive sciences department.
    • He became a Full Professor at the Université de Montréal in 1993 and authored the pioneering paper “Gradient-Based Learning Applied To Document Recognition” in 1998.
  • IVADO Labs and Turing Award:
    • Bengio became the Scientific Director of IVADO Labs in 2017, an organization that helps businesses harness the potential of artificial intelligence.
    • He was awarded the A.M. Turing Award, often referred to as the “Nobel Prize of computing,” along with Geoffrey Hinton and Yann LeCun for their contributions to deep learning.
  • MILA and Research Leadership:
    • Bengio leads the Montreal Institute for Learning Algorithms (MILA), a research lab he founded in 1993.
    • MILA collaborates with organizations like IBM to advance AI and machine learning research using open-source software.

Read Next: History of OpenAI, AI Business Models, AI Economy.

Connected Business Model Analyses

AI Paradigm

current-AI-paradigm

Pre-Training

pre-training

Large Language Models

large-language-models-llms
Large language models (LLMs) are AI tools that can read, summarize, and translate text. This enables them to predict words and craft sentences that reflect how humans write and speak.

Generative Models

generative-models

Prompt Engineering

prompt-engineering
Prompt engineering is a natural language processing (NLP) concept that involves discovering inputs that yield desirable or useful results. Like most processes, the quality of the inputs determines the quality of the outputs in prompt engineering. Designing effective prompts increases the likelihood that the model will return a response that is both favorable and contextual. Developed by OpenAI, the CLIP (Contrastive Language-Image Pre-training) model is an example of a model that utilizes prompts to classify images and captions from over 400 million image-caption pairs.

OpenAI Organizational Structure

openai-organizational-structure
OpenAI is an artificial intelligence research laboratory that transitioned into a for-profit organization in 2019. The corporate structure is organized around two entities: OpenAI, Inc., which is a single-member Delaware LLC controlled by OpenAI non-profit, And OpenAI LP, which is a capped, for-profit organization. The OpenAI LP is governed by the board of OpenAI, Inc (the foundation), which acts as a General Partner. At the same time, Limited Partners comprise employees of the LP, some of the board members, and other investors like Reid Hoffman’s charitable foundation, Khosla Ventures, and Microsoft, the leading investor in the LP.

OpenAI Business Model

how-does-openai-make-money
OpenAI has built the foundational layer of the AI industry. With large generative models like GPT-3 and DALL-E, OpenAI offers API access to businesses that want to develop applications on top of its foundational models while being able to plug these models into their products and customize these models with proprietary data and additional AI features. On the other hand, OpenAI also released ChatGPT, developing around a freemium model. Microsoft also commercializes opener products through its commercial partnership.

OpenAI/Microsoft

openai-microsoft
OpenAI and Microsoft partnered up from a commercial standpoint. The history of the partnership started in 2016 and consolidated in 2019, with Microsoft investing a billion dollars into the partnership. It’s now taking a leap forward, with Microsoft in talks to put $10 billion into this partnership. Microsoft, through OpenAI, is developing its Azure AI Supercomputer while enhancing its Azure Enterprise Platform and integrating OpenAI’s models into its business and consumer products (GitHub, Office, Bing).

Stability AI Business Model

how-does-stability-ai-make-money
Stability AI is the entity behind Stable Diffusion. Stability makes money from our AI products and from providing AI consulting services to businesses. Stability AI monetizes Stable Diffusion via DreamStudio’s APIs. While it also releases it open-source for anyone to download and use. Stability AI also makes money via enterprise services, where its core development team offers the chance to enterprise customers to service, scale, and customize Stable Diffusion or other large generative models to their needs.

Stability AI Ecosystem

stability-ai-ecosystem

About The Author

Scroll to Top
FourWeekMBA