What Is Adept AI?

  • Adept AI is a machine learning research and product lab. The company is building general intelligence based on the creative collaboration between humans and computers.
  • Adept AI wants to build a transformer that can handle all machine learning use cases. While existing models such as GPT-3 have impressive reading and writing skills, they cannot as yet act in the digital world or use initiative to perform tasks.
  • In addition to the three co-founders, the personnel who make up Adept’s team of nine leaders includes Kelsey Szot, Eric Elsen, Augustus Odena, Maxwell Nye, Anmol Gulati, and Fred Bertsch.

Adept AI was founded by a trio of individuals with extensive experience in large-scale neural networks. The first, David Luan, was head of Google’s large model efforts and also led the engineering team at OpenAI. Luan is now Adept’s CEO. 

Adept AI is a machine learning research and product lab. The company is building general intelligence based on the creative collaboration between humans and computers.

Adept AI history

Rounding out the trio are Chief Scientist Ashish Vaswani and CTO Niki Parmar – both former Google Brain employees who developed the transformer model architecture and precursor to applications such as GPT-3. 

Each of the three co-founders worked at Google with the goal to build a transformer that could handle all machine learning use cases. But there was a problem. While undoubtedly capable of reading and writing, models such as GPT-3 are not yet able to act in a digital world.

Emergence 

Adept AI emerged from stealth mode in April 2022 when it announced a Series A funding round worth $65 million.

Taking part in the round were a mixture of VC firms and angel investors including Root Ventures, Behance founder Scott Belsky, Airtable founder Howie Liu, and head of Tesla Autopilot Andrej Karpathy.

Luan noted that the funding would be used to move one step closer to general intelligence with a transformer that could act as well as read and write: “At Adept, we’re training a model to use every software tool and API that people use today.

Future vision

In keeping with the company’s goal to facilitate creative collaboration between humans and computers, staff will employ a natural language interface when using existing software such as Photoshop, Airtable, and Tableau.

At first, neural networks will be trained to perform very general tasks such as plotting data or generating a compliance report. The model will serve as what Adept calls an “overlay” as it interfaces between the user and their device or PC.

The company sees this strategy as one which puts humans in control, enabling them to offload tedious or time-consuming tasks to AI and spend more time on what they enjoy.

In the future, it hopes that AI will be able to use initiative and perform user commands without being prompted.

ACT-1

In September 2022, Adept released the Action Transformer (ACT-1) model to build “the next frontier of models that can take actions in the digital world” on almost any software tool, API, or website.

While this is an ambitious goal that will take some time to realize, the company acknowledges that ACT-1 is the first step in this direction.

ACT-1 was developed on Adept’s belief that general intelligence should be able to do anything that a human can on a computer.

To that end, it believes that a new era in personal computing will be characterized by natural language interfaces where humans can instruct computers on various tasks directly.

To support these assertions, Adept believes that in a few years:

  • The majority of human-computer interaction will occur via natural language (not GUI). At some point, today’s user interfaces will appear as archaic as landlines once did to smartphone users.
  • Advanced software will no longer be constrained “by the length of a drop-down menu. Features will become more accessible since no-one will have to learn how to perform specific tasks in a software program’s unique language. As a consequence, manuals, documentation, and FAQs will become the exclusive domain of models (not people).
  • With humans and AI working in tandem, innovation in most fields will accelerate. Collaboration will also enable people to become more creative and efficient while AI performs the heavy lifting in engineering and drug design, among many other fields.

What are ACT-1’s capabilities?

ACT-1 has been trained on numerous different tasks. But in the company’s most recent update, it explained that ACT-1 was currently connected to a Chrome extension so that it could learn how to use a web browser. 

In this scenario, the model observes the user as they click, type, and scroll on a custom “rendering” of the browser viewport which adapts to different websites. The action space – or the specific and finite set of actions the model can take in its environment – consists of all the UI elements on a web page. 

In one example, a user enters the prompt “find me a house in Houston that works for a family of 4. My budget is 600k”. Video footage then shows ACT-1 on the house platform Redfin where it enters “Houston” into the search box and uses filters to show homes with at least 4 bedrooms and a maximum price of $600,000.

In another, a user enters the prompt “add Max Nye at Adept as a new lead” and then “log a call with James Veel saying that he’s thinking about buying 100 widgets”. In this video, ACT-1 completes the tedious task of entering Nye’s details in multiple fields and then makes a note of the second request on Veel’s lead page. Both these actions occur within Salesforce.

Adept admits that the model needs to be made faster on both the software and modeling side. But it nevertheless expects that future versions will perform tasks with a latency that is mostly imperceptible to the user.

When will ACT-1 be released?

No release date has been confirmed, but interested parties can join a waitlist for the upcoming alpha release of the first product to be built around the ACT-1 model.

Adept AI team

Luan, Vaswani, and Parmar are joined by a small group of employees with previous experience at both Google and DeepMind. The personnel who make up Adept’s team of nine leaders includes:

  1. Kelsey Szot – a former McKinsey and Google employee who worked on the latter’s giant model production infrastructure.
  2. Eric Elsen – a former DeepMind research scientist who also held roles at Google Brain and Baidu.
  3. Augustus Odena – another Google Brain research scientist who co-created the Google Sheets SmartFill program synthesizer. Odena has also dabbled in neural network security and semi-supervised learning.
  4. Maxwell Nye – a recent deep learning Ph.D. graduate from MIT who worked at Facebook AI Research and Google Research. At Google, he utilized very large language models to write and then understand programs written in Python.
  5. Anmol Gulati – a speech recognition expert who researched large-scale speech and language modeling at Google Brain, and
  6. Fred Bertsch – another former Google Brain employee and software engineer who collaborated on Google Display Ads, Google Maps, Magenta, Stadia, and other products. Bertsch studied how people and product teams use and interact with generative models.

Key Highlights:

  • Founding Team: Adept AI was founded by David Luan, Ashish Vaswani, and Niki Parmar, all of whom have extensive experience in large-scale neural networks. David Luan, the CEO, was previously head of Google’s large model efforts and led the engineering team at OpenAI.
  • Company Overview: Adept AI is a machine learning research and product lab with a focus on building general intelligence through creative collaboration between humans and computers.
  • History and Emergence: The co-founders, including Ashish Vaswani and Niki Parmar from Google Brain, aimed to build a transformer that could not only read and write but also act in the digital world. Adept AI emerged from stealth mode in April 2022 with a Series A funding round of $65 million from VC firms and angel investors.
  • Future Vision: Adept AI envisions a future where natural language interfaces are used to interact with software tools and APIs, facilitating creative collaboration between humans and AI. The company believes this approach will empower users to offload repetitive tasks to AI, leading to increased efficiency and innovation.
  • ACT-1 Model: Adept AI introduced the Action Transformer (ACT-1) model, which aims to enable AI to take actions in the digital world using various software tools and websites. The model is designed to understand and execute user commands directly through natural language interfaces.
  • Capabilities of ACT-1: ACT-1 has been trained on various tasks and is connected to a Chrome extension to learn how to use a web browser. It can perform actions like searching for houses, adding leads in Salesforce, and more based on user prompts.
  • Release and Team: While there is no confirmed release date for ACT-1, interested parties can join a waitlist for the upcoming alpha release. Adept AI’s team includes experienced individuals from Google, DeepMind, and other prominent organizations in the field of AI and machine learning.

Read Next: History of OpenAI, AI Business Models, AI Economy.

Connected Business Model Analyses

AI Paradigm

current-AI-paradigm

Pre-Training

pre-training

Large Language Models

large-language-models-llms
Large language models (LLMs) are AI tools that can read, summarize, and translate text. This enables them to predict words and craft sentences that reflect how humans write and speak.

Generative Models

generative-models

Prompt Engineering

prompt-engineering
Prompt engineering is a natural language processing (NLP) concept that involves discovering inputs that yield desirable or useful results. Like most processes, the quality of the inputs determines the quality of the outputs in prompt engineering. Designing effective prompts increases the likelihood that the model will return a response that is both favorable and contextual. Developed by OpenAI, the CLIP (Contrastive Language-Image Pre-training) model is an example of a model that utilizes prompts to classify images and captions from over 400 million image-caption pairs.

OpenAI Organizational Structure

openai-organizational-structure
OpenAI is an artificial intelligence research laboratory that transitioned into a for-profit organization in 2019. The corporate structure is organized around two entities: OpenAI, Inc., which is a single-member Delaware LLC controlled by OpenAI non-profit, And OpenAI LP, which is a capped, for-profit organization. The OpenAI LP is governed by the board of OpenAI, Inc (the foundation), which acts as a General Partner. At the same time, Limited Partners comprise employees of the LP, some of the board members, and other investors like Reid Hoffman’s charitable foundation, Khosla Ventures, and Microsoft, the leading investor in the LP.

OpenAI Business Model

how-does-openai-make-money
OpenAI has built the foundational layer of the AI industry. With large generative models like GPT-3 and DALL-E, OpenAI offers API access to businesses that want to develop applications on top of its foundational models while being able to plug these models into their products and customize these models with proprietary data and additional AI features. On the other hand, OpenAI also released ChatGPT, developing around a freemium model. Microsoft also commercializes opener products through its commercial partnership.

OpenAI/Microsoft

openai-microsoft
OpenAI and Microsoft partnered up from a commercial standpoint. The history of the partnership started in 2016 and consolidated in 2019, with Microsoft investing a billion dollars into the partnership. It’s now taking a leap forward, with Microsoft in talks to put $10 billion into this partnership. Microsoft, through OpenAI, is developing its Azure AI Supercomputer while enhancing its Azure Enterprise Platform and integrating OpenAI’s models into its business and consumer products (GitHub, Office, Bing).

Stability AI Business Model

how-does-stability-ai-make-money
Stability AI is the entity behind Stable Diffusion. Stability makes money from our AI products and from providing AI consulting services to businesses. Stability AI monetizes Stable Diffusion via DreamStudio’s APIs. While it also releases it open-source for anyone to download and use. Stability AI also makes money via enterprise services, where its core development team offers the chance to enterprise customers to service, scale, and customize Stable Diffusion or other large generative models to their needs.

Stability AI Ecosystem

stability-ai-ecosystem

Discover more from FourWeekMBA

Subscribe now to keep reading and get access to the full archive.

Continue reading

Scroll to Top
FourWeekMBA