
What is Generative AI?
Generative Artificial Intelligence (AI) aims to understand and predict information
from a particular data set. It is important to know that generative AI is not totally
new. It is already used in applications like email via smart compose, which allows
the email program to finish a sentence started by a user. In many ways it is an
existing tool that has only recently started to ramp up.
Generative AI at an Inflection Point
Deep learning and predictive AI have also been in existence for some time,
however, recently there has been an incredible increase in model size and
complexity. Large Language Models (LLMs) exist today with hundreds of gigabytes
that can analyze huge amounts of data sets, although this analysis or “training”
takes a lot of computing power. The increase in model size has been made possible
through improvements in computing technique including Central Processing Units
(CPUs) and cloud computing, which allow customers to use thousands of Graphics
Processing Units (GPUs) from the cloud, as well as skyrocketing amounts of
available data. Creators of the models have also made them more “human friendly”
as they launch public applications, thereby making them more accessible.
Why are Transformers an Inflection Point for LLMs?
Transformers are deep learning models that use self-attention mechanisms to
weight the significance of each part of a given input data. Their use in LLMs
launched a chain of development in Natural Language Processing (NLP) — a
branch of AI aimed at aiding computers in understanding natural human language.
The transformer model, when used for NLP, can more efficiently train AI GPUs, thus
significantly driving down the costs of training versus alternative models. As the
CEO of NVIDIA, Jensen Huang, put it in 2022: “Transformers made self-supervised
learning possible, and AI jumped to warp speed.”
1
For a more in-depth explanation
of LLMs and Transformers, see commentary from Citi Global Insights on page 11.
OpenAI and ChatGPT
OpenAI started as a research lab in 2015 and is the AI research and deployment
company behind three generative AI models — ChatGPT, Codex, and DALL-E.
These models are trained to understand the structure of human language to create
text, code, and image content, as well as new types of data/insights from a training
set. The release of the models has become an inflection point in generative AI due
to improvements in compute, data availability, and public ability to test and further
refine the model. The third iteration of ChatGPT (Generated Pre-trained
Transformers or GPT-3) was launched in November 2022 as a human-like AI
platform capable of solving/answering prompts. What is different about ChatGPT for
search is it provides a conversational style response to an inquiry versus links to
suggested sites. Since its launch, it has become the fastest growing consumer
application in history, accumulating 100 million Monthly Average Users (MAUs) in
January 2023. For context, the prior record keepers for fastest growing application
are TikTok at 9 months and Instagram at 2.5 years.
2
1
GTC 2022 Spring Keynote with NVIDEO CEO Jensen Huang.
2
Dylan Patel and Afzal Ahmad, “The Inference Cost of Search Disruption — Large
Language Model Cost Analysis,” SemiAnalysis, February 9, 2023.
文档被以下合辑收录
相关文档
评论