💻Large Language Models (LLMs)
Last updated
Last updated
Large Language Models (LLMs) are sophisticated learning algorithms specially designed for the purpose of recognizing and comprehending language.
In other terms, a LLM is a type of AI model that uses deep learning techniques and vast amounts of data to perform a variety of 'Natural Language Processing' (NLP) tasks such as generating and classifying text, answering questions in a conversational manner, summarizing content, or predicting new content. NLP is a technology that enables machines to understand human speech in text or voice form to communicate with humans.
In fact, and without even knowing it, we experience the power of LLMs on a daily basis (e.g., Google Translate or Google Search). An important part of our favorite products uses LLMs to manipulate language better than ever before.
LLMs undergo a process called pre-training (costing millions of dollars), where they are exposed to vast repositories of text that provide the foundational knowledge the model uses. The size of the model, in terms of the number of parameters and layers, allows it to capture intricate relationships and patterns within the text.
Technically, the immense volume of textual data that these models train on is processed through what is known as neural networks. These are a type of artificial intelligence mechanism composed of intricate layers and nodes. The majority of LLMs, in their design, employ a specific kind of neural network architecture known as a transformer (this is why GPT stands for Generative Pre-trained Transformers). Transformers have shown a unique aptitude for processing language, allowing them to consider words in sentences not just in isolation but in relation to one another, offering a depth of comprehension. They keep track of the context of what is being written, and this is why the text that they write makes sense.
These models can be used in a variety of ways. The most famous models are generative prompt-based LLMs (like ChatGPT), which enable users to directly input instructions for the trained models to process and respond to. The resulting output is composed of information that attempts to follow the input text as instructions. Such models have the capacity to reasonably perform tasks it was never explicitly trained for.
One of the key recent improvement of these models was Reinforcement Learning from Human Feedback (RLHF). This process aligns the machine-generated outputs of a pre-trained LLM using human feedback.
RLHF enabled output to be of much higher coherence, resulting in highly believable and well-crafted responses. The idea behind Reinforcement Learning is that an AI will learn from the environment by interacting with it (through trial and error) and receiving rewards (negative or positive) as feedback for performing actions.
With technological advancements, LLMs continue to grow in complexity and capability.
For context, ChatGPT-4 possesses an astonishing 100 trillion parameters, a significant leap from its predecessor, ChatGPT-3.5, which had 175 million.
These parameters play a pivotal role in understanding the relationships between words, and determining how to weave them together coherently.