What is Generative AI, How It work and Uses of Generative AI?
Generative AI is a relatively new form of AI that, unlike its predecessors, can create new content from its training data. Its extraordinary ability to produce human-like writing, images, audio, and video has captured the world’s imagination since the first consumer-grade generative AI chatbot was launched in the fall of 2022. A June 2023 report from McKinsey & Company estimated that generative AI has the potential to add between $6.1 and $7.9 trillion per year to the global economy by increasing labor productivity. To put that into context, the same research puts the annual economic potential of productivity gains from all AI technologies at between $17.1 and $25.6 trillion. So while generative AI has the spotlight in 2023, it’s still only a portion of AI’s full potential.
But every action has an equal and opposite reaction. So along with its remarkable productivity prospects, generative AI brings new potential business risks, such as inaccuracy, privacy violations, and intellectual property exposure, as well as the ability to cause large-scale economic and social disruption. For example, the productivity benefits of generative AI are unlikely to materialize without substantial worker retraining efforts, and even then, they will undoubtedly displace many people from their current jobs. Consequently, government policymakers around the world, and even some tech industry executives, are advocating for the rapid adoption of regulations regarding AI.
This article explores in depth the promise and peril of generative AI: how it works; its most immediate applications, use cases, and examples; its limitations; its potential business benefits and risks; best practices for its use; and an outlook for its future.
What is generative AI?
Generative artificial intelligence (GAI) is the name given to a subset of machine learning technologies that have recently developed the ability to rapidly create content in response to text prompts, which can be short and simple or very long and complex. Different generative AI tools can produce new audio, image, and video content, but it is text-oriented conversational AI that has sparked the imagination. In effect, people can converse with text-trained generative AI models in the same way they do with humans.
Generative AI took the world by storm in the months following the launch of ChatGPT, a chatbot based on OpenAI’s GPT-3.5 neural network model, on November 30, 2022. GPT stands for “generative pretrained transformer,” words that primarily describe the underlying architecture of the model’s neural network.
There are many earlier instances of conversational chatbots, starting with ELIZA from the Massachusetts Institute of Technology (MIT) in the mid-1960s. But most earlier chatbots, including ELIZA, were entirely or mostly rule-based, so they lacked contextual understanding. Their responses were limited to a set of predefined rules and templates. In contrast, the generative AI models that are emerging now have no such predefined rules or templates. Metaphorically speaking, they are blank, primitive brains (neural networks) that are exposed to the world through training with real-world information. They then independently develop intelligence—a representative model of how that world works—which they use to generate novel content in response to prompts. Even AI experts don’t know exactly how they do this, since the algorithms develop and adjust themselves as the system is trained.
Companies large and small should be excited about the potential of generative AI to bring the benefits of technological automation to knowledge work, which has so far largely resisted automation. Generative AI tools change the calculus of knowledge work automation; their ability to produce human-like writing, images, audio, or video in response to plain English prompts means they can collaborate with human partners to generate content that represents practical work.
"Over the next few years, many companies are going to be training their own large, specialized language models," said Larry Ellison, Oracle's president and chief technology officer, during the company's earnings call in June 2023.
Generative AI vs. AI
Artificial intelligence is a vast area of computer science, of which generative AI is only a small part, at least at present. Naturally, generative AI shares many characteristics with traditional AI. But there are also some important differences.
Another difference worth noting is that training foundational models for generative AI is “obscenely expensive,” to quote one AI researcher. Imagine, $100 million just for the hardware needed to get started, plus the equivalent costs of cloud services, since that’s where most AI development is done. Then there’s the cost of the huge volumes of data required.
Key findings
Generative Artificial Intelligence in detail
For companies large and small, the seemingly magical promise of generative AI is that it can bring the benefits of technological automation to knowledge work. Or, as a McKinsey report put it, “activities involving decision-making and collaboration, which previously had the least potential for automation.”
Historically, technology has been most effective at automating routine or repetitive tasks, for which decisions were already known or could be determined with a high level of confidence, based on specific and very clear rules. Think of manufacturing, with its precise repetition on the assembly line, or accounting, with its regulated principles set by industry associations. But generative AI has the potential to do much more sophisticated cognitive work. To suggest an admittedly extreme example, generative AI could help shape an organization’s strategies, responding to prompts asking for ideas and alternative scenarios from business managers in the midst of industry change.
In its report, McKinsey assessed 63 use cases across 16 business functions, concluding that 75% of the $1 trillion in potential value that could be achieved with generative AI will come from a subset of use cases in just four of those functions: customer operations, marketing and sales, software engineering, and research and development. Revenue growth prospects across industries were more evenly spread, though there were standouts: High-tech topped the list for potential boost, in terms of percentage of industry revenue, followed by banking, pharmaceuticals and medical, education, telecommunications, and healthcare.
Separately, a Gartner analysis agrees with McKinsey predictions — for example, that more than 30% of new drugs and materials will be discovered using generative AI techniques by 2025, up from zero today, and that 30% of large organizations’ outbound marketing messages will be synthetically generated by 2025, up from 2% in 2022. And in an online survey, Gartner found that customer experience and retention was the top response (38%) of 2,500 executives asked where their organizations were investing in generative AI.
What makes it possible for all this to happen so quickly is that unlike traditional AI, which has been quietly automating and adding value to business processes for decades, generative AI came into the world’s consciousness thanks to ChatGPT’s human-like conversational skill. That has also attracted people and shed light on generative AI technology, which focuses on other modalities. It seems like everyone is experimenting with writing text or creating music, images, and videos using one or more of the various models that specialize in each area. So with many organizations already experimenting with generative AI, its impact on business and society is likely to be colossal — and it will happen astonishingly quickly.
The obvious downside is that knowledge work will change. Individual roles will change, sometimes significantly, so workers will need to learn new skills. Some jobs will be lost. However, historically, big technological changes, such as generative AI, have always added more (and higher-value) jobs to the economy than they eliminate. But this is of little comfort to those whose jobs are eliminated.
How does it work?
There are two answers to the question of how generative AI models work. Empirically, we know how they work in detail because humans designed their various neural network implementations to do exactly what they do, iterating those designs over decades to get them ever better. AI developers know exactly how neurons are wired; they designed the training process for each model. In practice, however, no one knows exactly how generative AI models do what they do — that’s an embarrassing truth.
“We don’t know how they perform the actual creative task because what’s going on inside the layers of the neural network is too complex for us to figure out, at least today,” said Dean Thompson, former CTO of multiple AI startups that have been acquired over the years by companies like LinkedIn and Yelp, where he still works as a senior software engineer on large language models (LLMs). The ability of generative AI to produce new and original content seems to be an emergent property of what is known — namely, its structure and training. So while there’s a lot to explain in what we know, what a model like GPT-3.5 is actually doing internally — what it’s thinking, so to speak — has yet to be discovered. Some AI researchers are confident that this will be known in the next 5 to 10 years; others aren’t sure it will ever be fully understood.
Here's an overview of what we know about how generative AI works:
Why is it important?
A useful way to understand the importance of generative AI is to think of it as a calculator for creative, open-ended content. Much like a calculator automates routine, mundane math, freeing up a person to focus on higher-level tasks, generative AI has the potential to automate the more routine, mundane tasks that make up much of knowledge work, allowing people to focus on the higher-level parts of the work.
Consider the challenges marketers face in gaining actionable insights from the unstructured, inconsistent, and disconnected data they often encounter. Traditionally, they would need to consolidate that data as a first step, requiring a considerable amount of custom software engineering to give a common structure to disparate data sources such as social media, news, and customer feedback.
“But with LLMs, you can just plug information from different sources right into the application and then ask for key insights, or what feedback to prioritize, or ask for sentiment analysis, and it will just work,” said Basim Baig, a senior engineering manager specializing in AI and security at Duolingo. “The power of LLM here is that it allows you to skip that huge, expensive engineering step.”
Thinking further, Thompson suggests that product marketers could use LLMs to tag free text for analysis. For example, imagine you have a large database of social media mentions of your product. You could write software that applies an LLM and other technologies to:
You could then apply the results to:
Generative AI models
Generative AI represents a broad category of applications based on a growing range of neural network variations. While all generative AI fits the general description in the How does generative AI work? section, implementation techniques vary to support different media, such as images versus text, and to incorporate research and industry advances as they emerge.
Neural network models use repetitive patterns of artificial neurons and their interconnections. A neural network design, for any application including generative AI, often repeats the same pattern of neurons hundreds or thousands of times, usually reusing the same parameters. This is an essential part of what is called “neural network architecture.” The discovery of new architectures has been an important part of AI innovation since the 1980s, often driven by the goal of supporting a new medium. But then, once a new architecture has been invented, further progress is often made by using it in unexpected ways. Further innovation comes from combining elements of different architectures.
Two of the oldest and most common architectures are:
Although RNNs are still frequently used, successive efforts to improve RNNs led to a breakthrough:
Research, private industry, and open source efforts have created impactful models that innovate at higher levels of neural network architecture and application. For example, there have been crucial innovations in the training process, in how training feedback is incorporated to improve the model, and in how multiple models can be combined in generative AI applications. Here is a summary of some of the most important innovations in generative AI models:
What are the applications of generative AI?
While the world has only just begun to explore the potential of generative AI applications, it’s easy to see how businesses can benefit from applying it to their operations. Consider how generative AI could change key areas of customer interactions, sales and marketing, software engineering, and research and development.
In customer service, earlier AI technology automated processes and introduced self-service, but it also created new frustrations for customers. Generative AI promises to bring benefits to both customers and service reps, with chatbots that can adapt to different languages and regions, creating a more personalized and accessible customer experience. When human intervention is needed to resolve a customer issue, service reps can collaborate in real-time with generative AI tools to find effective strategies, improving the speed and accuracy of interactions. The speed with which generative AI can access a company’s full knowledge base and synthesize new solutions to customer complaints gives service staff a greater ability to effectively resolve specific customer issues, rather than relying on outdated phone systems and call transfers until an answer is found—or until the customer loses patience.
In marketing, generative AI can automate the integration and analysis of data from disparate sources, which should dramatically speed up the time to insights and drive better-informed decision-making and faster development of go-to-market strategies. Marketers can use this information alongside other AI-generated insights to create new, more targeted advertising campaigns. This reduces the time staff must spend collecting demographic and purchasing behavior data and provides more time to analyze results and generate new ideas.
Tom Stein, president and chief brand officer of B2B marketing agency Stein IAS, notes that every marketing agency, including his own, is exploring these opportunities at a rapid pace. Stein also notes that there are simpler, quicker wins for an agency’s internal processes.
“If we get an RFI [request for information], typically 70 to 80% of the RFI is going to ask for the same information as every other RFI, maybe with some contextual differences specific to that company’s situation,” says Stein, who also served as jury chair for the 2023 Cannes Lions Creative B2B Awards. “It’s not that complicated to put ourselves in a position to have any number of AI tools do that work for us… So if we get that 80% of our time back and can spend it adding value to the RFI and just making it sing, that’s a win all the way around. And there are a number of processes like that.”
In software development, collaborating with generative AI can simplify and speed up processes at every step, from planning to maintenance. During the initial creation phase, generative AI tools can analyze and organize large amounts of data and suggest multiple program configurations. Once coding begins, AI can test and troubleshoot, identify bugs, run diagnostics, and suggest solutions, both before and after release. Thompson notes that because many enterprise software projects incorporate multiple programming languages and disciplines, he and other software engineers have used AI to train themselves in unfamiliar areas much faster than they could previously. He has also used generative AI tools to explain unfamiliar code and identify specific problems.
In research and development (R&D), generative AI can increase the speed and depth of market research during the early phases of product design. AI programs, especially those with imaging capabilities, can then create detailed designs of potential products before simulating and testing them, giving workers the tools they need to make quick and effective adjustments throughout the R&D cycle.
Oracle founder Larry Ellison noted on the June earnings call that “specialized LLMs will accelerate the discovery of new, life-saving drugs.” Drug discovery is an R&D application that exploits generative models’ tendency to “hallucinate” incorrect or unverifiable information, but in a positive way: identifying new molecules and protein sequences in search of novel health treatments. Separately, Oracle subsidiary Cerner Enviza has partnered with the U.S. Food and Drug Administration (FDA) and John Snow Labs to apply AI tools to the challenge of “understanding the effects of drugs on large populations.” Oracle’s AI strategy is to make AI pervasive in its cloud applications and cloud infrastructure.
Use cases
Generative AI has the potential to speed up or completely automate a wide variety of tasks. Businesses should plan deliberate and specific ways to maximize the benefits it can bring to their operations. Here are some specific use cases:
What are the benefits of Generative Artificial Intelligence?
The benefits that generative AI can bring to a business stem primarily from three general attributes: knowledge synthesis, human-AI collaboration, and speed. While many of the benefits outlined below are similar to those promised in the past by AI models and automation tools, the presence of one or more of these three attributes can help businesses realize the advantages more quickly, easily, and effectively.
With generative AI, organizations can build custom models trained on their own institutional knowledge and intellectual property (IP), after which knowledge workers can ask the software to collaborate on a task in the same language they might use with a colleague. A specialized generative AI model can respond by synthesizing information from across the corporate knowledge base at astonishing speed. Not only does this approach reduce or eliminate the need for complex and often less effective and more expensive software engineering expertise to build programs specific to these tasks, it is also likely to uncover insights and connections that previous approaches could not.