Large language model

This topic explains large language models.

A Large Language Model (LLM) is a type of artificial intelligence that processes and generates human-like text based on vast amounts of data. These models are built using deep learning techniques, particularly neural networks with many layers, which enable them to understand and produce coherent language. LLMs are trained in diverse text, encompassing books, articles, websites, and more, allowing them to learn grammar, context, facts, and nuances of human language.

One of the key characteristics of LLMs is their ability to generate text that is contextually relevant and grammatically correct. They can perform a wide range of language-related tasks, including translation, summarization, question answering, and text completion. LLMs have billions of parameters, which are the weights in the neural network that the model adjusts during training to learn language patterns.

Supported models

This table shows the LLMs that are available to use within the GenAI application:

Name Description Token Window US East (Y/N) EU (Y/N)

Azure OpenAI – France Central

AWS Bedrock - Frankfurt

Claude Instant Optimized for rapid performance and competitive affordability. 100,000 Y N
Claude 2 Extensive context window to support larger prompts, such as document inputs. 100,000 Y Y
Claude 2.1 Claude 2 performance with additional support for hallucination reduction. 200,000 Y Y
Claude 3 Haiku Fast and compact model for near-instant responsiveness. Can answer simple queries and requests with speed as a priority. 200,000 Y N
Claude 3 Sonnet Strikes the ideal balance between intelligence and speed – particularly for enterprise workloads. Delivers strong performance at a lower cost. 200,000 Y N
Claude 3.5 Sonnet Twice the speed of Claude 3 Opus, ideal model for complex workflows within a cost-effective pricing. 200,000 Y N
Claude 3 Opus (pending) The most intelligent Claude 3 model, with best-in-market performance on highly complex tasks. 200,000 N N
Jurassic Mid Offers strong balance between text generation speed, cost, and output quality. 8,192 Y N
Jurassic Ultra Largest model of the Jurassic series, ideal for complex processing with optimal performance. 8,192 Y N
Titan Lite Designed for use-case specific generation tasks with affordability and speed as priorities. 4,096 Y N
Titan Express High-performance model for more complex tasks involving retrieval augmented generation (RAG). 8,192 Y Y
Llama 3 8B Ideal for limited computational power and resources, edge devices, and faster training times. 8,192 Y N
Llama 3 70B Ideal for content creation, conversational AI, language understanding, R&D, and Enterprise applications. 8,192 Y N