Skip to main content

Key terms

H2O LLM Studio uses several key terms across its documentation, and each, in turn, is explained in the sections below.

Prompt Engineering

Prompt engineering involves crafting precise and effective input queries to guide language models in generating desired outputs or responses.

Agents

Software entities or components that interact with data or perform tasks within a system.

ELO

An algorithm or method used to assess and rank the performance or accuracy of language models based on their proficiency in understanding and processing textual data.

Vector Store

A Vector Store stores numerical representations of text for fast access in language models.

Pre-training

The initial phase of training a machine learning model on a large dataset to learn general features before fine-tuning on a specific task.

Attention

A mechanism that enables models to focus on specific parts of input data relevant to the task at hand, enhancing their understanding and performance.

Embedding

Embedding refers to a mathematical representation of words or tokens in a numerical vector space, enabling machine learning models to understand and process language based on their context and relationships.

Language Model

A language model is an AI system that understands and generates human language, predicting and generating text based on patterns and context within a given sequence of words.

Transformer

A Transformer refers to a neural network architecture specifically designed for processing sequential data like text, using attention mechanisms to learn contextual relationships between words or tokens.

Encoders and Decoders

Encoders and decoders are vital parts of sequence-to-sequence models used in natural language processing. Encoders process input data into a fixed-size representation, while decoders generate an output sequence based on that representation.

Text generation

Text generation is the process of creating written content, such as sentences or paragraphs, using machine learning or AI algorithms based on patterns learned from existing text data.

In-context learning

In-context learning refers to the process where a machine learning model continuously improves and adapts by considering the context of new information within its existing knowledge, enhancing its accuracy and understanding over time.

Few-shot learning

Few-shot learning refers to a machine learning technique where a model can learn from a very small amount of labeled data to generalize and make predictions accurately on new, unseen data.

Summarization

Summarization is the process of condensing a larger piece of text into a shorter, coherent version while retaining its essential information.

Fine-tuning

Fine-tuning refers to adjusting and optimizing a pre-trained machine learning model using specific data to enhance its performance for a particular task.

GPT

GPT stands for "Generative Pre-trained Transformer," a type of language model that uses transformers to understand and generate human-like text based on vast amounts of training data.

GPU deployment

GPU deployment is the utilization of graphics processing units (GPUs) to execute and accelerate the computations involved in deploying machine learning models, improving speed and efficiency in model inference or training.

Tokenization

Tokenization is the process of breaking text into smaller units, typically words or phrases, to analyze or process them individually within a natural language processing system.


Feedback