Professional Documents
Culture Documents
Generative Ai
Generative Ai
Intelligence” which eventually became The Turing Test, which experts used to measure
computer intelligence. The term “artificial intelligence” was coined and came into
popular use.
Generative AI enables users to quickly generate new content based on a variety of
inputs. Inputs and outputs to these models can include text, images, sounds,
animation, 3D models, or other types of data.
This technology is no longer new since it entered the mainstream in late 2022. While
you may have played with (and enjoyed!) the likes of ChatGPT and Midjourney, they’re
barely more than surface-level distractions.
Corporate use for generative AI is far more sophisticated. If used to its full extent, it
will reduce product-development life cycle time, design drugs in months instead of
years, compose entirely new materials, generate synthetic data, optimize parts design,
automate creativity… In fact, experts predict that by 2025, 30% of outbound
marketing messages from large organization will be synthetically generated, and by
2030, a major blockbuster film will be released with 90% of the film generated by AI.
Going beyond the most headline-grabbing use cases, studies have shown that Gen. AI
increases productivity for a variety of tasks, with specific benefits for low-ability
workers and less experienced employees. Put simply, these tools will level the playing
field.
This is happening today, and will continue to happen, with increasing success, over
the coming decade. That is, if we can navigate the many risks associated with
generative AI. I’m particularly worried about deep fakes, copyright issues,
and malicious uses for fake news.
Generative AI models use neural networks to identify the patterns and structures
within existing data to generate new and original content.
One of the breakthroughs with generative AI models is the ability to leverage different
learning approaches, including unsupervised or semi-supervised learning for training.
This has given organizations the ability to more easily and quickly leverage a large
amount of unlabeled data to create foundation models. As the name suggests,
foundation models can be used as a base for AI systems that can perform multiple
tasks.
Examples of foundation models include GPT-3 and Stable Diffusion, which allow users
to leverage the power of language. For example, popular applications like ChatGPT,
which draws from GPT-3, allow users to generate an essay based on a short text
request. On the other hand, Stable Diffusion allows users to generate photorealistic
images given a text input.
1. Quality: Especially for applications that interact directly with users, having high-
quality generation outputs is key. For example, in speech generation, poor speech
quality is difficult to understand. Similarly, in image generation, the desired outputs
should be visually indistinguishable from natural images.
2. Diversity: A good generative model captures the minority modes in its data
distribution without sacrificing generation quality. This helps reduce undesired biases
in the learned models.
3. Speed: Many interactive applications require fast generation, such as real-time image
editing to allow use in content creation workflows.
There are multiple types of generative models, and combining the positive attributes of
each results in the ability to create even more powerful models.
Below is a breakdown:
Diffusion models: Also known as denoising diffusion probabilistic models (DDPMs),
diffusion models are generative models that determine vectors in latent space through
a two-step process during training. The two steps are forward diffusion and reverse
diffusion. The forward diffusion process slowly adds random noise to training data,
while the reverse process reverses the noise to reconstruct the data samples. Novel
data can be generated by running the reverse denoising process starting from entirely
random noise.
A diffusion model can take longer to train than a variational autoencoder (VAE) model,
but thanks to this two-step process, hundreds, if not an infinite amount, of layers can
be trained, which means that diffusion models generally offer the highest-quality
output when building generative AI models.
Additionally, diffusion models are also categorized as foundation models, because they
are large-scale, offer high-quality outputs, are flexible, and are considered best for
generalized use cases. However, because of the reverse sampling process, running
foundation models is a slow, lengthy process.
The two models are trained together and get smarter as the generator produces better
content and the discriminator gets better at spotting the generated content. This
procedure repeats, pushing both to continually improve after every iteration until the
generated content is indistinguishable from the existing content.
While GANs can provide high-quality samples and generate outputs quickly, the
sample diversity is weak, therefore making GANs better suited for domain-specific
data generation.
A self-attention layer assigns a weight to each part of an input. The weight signifies
the importance of that input in context to the rest of the input. Positional encoding is
a representation of the order in which input words occur.
Generative AI models can take inputs such as text, image, audio, video, and code and
generate new content into any of the modalities mentioned. For example, it can turn
text inputs into an image, turn an image into a song, or turn video into text.
The impact of generative models is wide-reaching, and its applications are only
growing. Listed are just a few examples of how generative AI is helping to advance and
transform the fields of transportation, natural sciences, and entertainment.
As an evolving space, generative models are still considered to be in their early stages,
giving them space for growth in the following areas.
Many companies such as NVIDIA, Cohere, and Microsoft have a goal to support the
continued growth and development of generative AI models with services and tools to
help solve these issues. These products and platforms abstract away the complexities
of setting up the models and running them at scale.
1. Generative AI algorithms can be used to create new, original content, such as images,
videos, and text, that’s indistinguishable from content created by humans. This can be
useful for applications such as entertainment, advertising, and creative arts.
2. Generative AI algorithms can be used to improve the efficiency and accuracy of
existing AI systems, such as natural language processing and computer vision. For
example, generative AI algorithms can be used to create synthetic data that can be
used to train and evaluate other AI algorithms.
3. Generative AI algorithms can be used to explore and analyze complex data in new
ways, allowing businesses and researchers to uncover hidden patterns and trends that
may not be apparent from the raw data alone.
4. Generative AI algorithms can help automate and accelerate a variety of tasks and
processes, saving time and resources for businesses and organizations.
Note: Demonstrating the capabilities of generative models, this section, “What are the
Benefits of Generative AI?” was written by the generative AI model ChatGPT.