Before we dive into the benefits of generative AI for the enterprise, it's important to understand the key terms in this field:
Natural Language Processing (NLP): A branch of AI that enables a computer to understand, analyze, and generate human language.
Transformer: A neural network architecture (introduced by Google in 2017) that can train significantly larger models on ever larger datasets than before. It allowed the emergence of recent LLMs to process and learn from sequences of words, recognizing patterns and relationships within the text.
Large Language Model (LLM): A deep learning model specialized in text generation. Today almost all LLMs are based on the transformer architecture (with variations from the original). As the name indicates, the real revolution is in the scale of these models, the very large number of parameters (typically in the 10s to 100s of billions), and the very large corpus of text used to train them.
Examples of commercial / closed-source models: OpenAI’s GPT3/3.5/4, GCP’s PaLM, AWS’s Bedrock, Anthropic’s Claude, etc.
Examples of open-source models: Google FLAN-T5 & FLAN-UL2, MPT-7B, Falcon, etc.
Pre-trained Model / Foundational Model: An LLM trained from scratch in a self-supervised way (without labels) on a very large dataset and made available for others to use. Foundational models are not intended to be used directly and need to be fine-tuned for specific tasks.
Fine-tuning: The process of taking a pre-trained model and adapting it to a specific task (summarization, text generation, question-answering, etc.) by training it in a supervised way (with labels). Fine-tuning requires much less data and compute power than the original pre-training. Well fine-tuned models often outperform much larger models.
Hallucination: A situation where the LLM generates a wrong output; the root cause is that the LLM uses its internal “knowledge” (what it has been trained on) and that knowledge is not applicable to the user query.
Multi-modal AI: a type of AI with the ability to process and understand inputs from different types of inputs such as text, speech, images, and videos.
Retrieval Model: a system (typically a Transformer) used to retrieve data from a source of information. Combining retrieval models with large language models partially addresses the hallucination problem by anchoring the LLM in a known corpus of data.
Vector Store: a type of data store specialized in managing vectorial representations of documents (known as embeddings), those stores are optimized to efficiently find nearest neighbors (according to various distance metrics), they are central architectural pieces of a Generative AI platform.
Enterprise AI: Providing predictive insights to expert users for improving performance of enterprise processes and systems. (See What Is Enterprise AI?)
Generative AI for Enterprise Search – Leveraging generative AI’s intuitive search bar to access the predictive insights and underlying data of enterprise AI applications, improving decision-making for all enterprise users.
Why Is Generative AI Transformational for the Enterprise?
Necessary cookies are absolutely essential for the website to function properly. This category only includes cookies that ensures basic functionalities and security features of the website. These cookies do not store any personal information.