Term of the Moment

DROID


Look Up Another Term


Definition: long-horizon context


The capability of an AI model to analyze a huge amount of information beyond the current "context window," which is essentially the amount of memory available to the system.

Using text as the training data, the largest number of tokens a model can analyze at one time is based on the context window. Using a book as a simple example, the context window might be the current chapter, but the long-horizon context would be the entire book.

Training and Inference
Long-horizon context is constantly being improved for both training and inference. For example, larger context windows see dependencies over a greater amount of training data. Improved long-horizon context ensures results are delivered faster when generating answers at the inference stage. See AI training vs. inference.

Context GPUs
Context GPUs are specially designed GPUs that handle large amounts of memory and function as a first processing stage. Generally optimized for inference, the data are passed onto the GPUs that perform massively parallel computations. See GPU.




Context and AI GPUs
NVIDIA combines its CPX context GPUs with AI GPUs (Rubin GPUs) on this compute tray. Along with switch trays, as many as 18 Vera Rubin trays are installed in one server rack (see inference engine and Vera Rubin). (Image courtesy of NVIDIA.)