Large language model
A Large Language Model (LLM) is a type of artificial intelligence that processes and generates human-like text based on vast amounts of data. These models are built using deep learning techniques, particularly neural networks with many layers, which enable them to understand and produce coherent language. LLMs are trained in diverse text, encompassing books, articles, websites, and more, allowing them to learn grammar, context, facts, and nuances of human language.
One of the key characteristics of LLMs is their ability to generate text that is contextually relevant and grammatically correct. They can perform a wide range of language-related tasks, including translation, summarization, question answering, and text completion. LLMs have billions of parameters, which are the weights in the neural network that the model adjusts during training to learn language patterns.
Supported models
This table shows the LLMs that are available to use within the GenAI application:
Name | Description | Token Window | US East (Y/N) | EU (Y/N) Azure OpenAI – France Central AWS Bedrock - Frankfurt |
---|---|---|---|---|
Claude Instant | Optimized for rapid performance and competitive affordability. | 100,000 | Y | N |
Claude 2 | Extensive context window to support larger prompts, such as document inputs. | 100,000 | Y | Y |
Claude 2.1 | Claude 2 performance with additional support for hallucination reduction. | 200,000 | Y | Y |
Claude 3 Haiku | Fast and compact model for near-instant responsiveness. Can answer simple queries and requests with speed as a priority. | 200,000 | Y | N |
Claude 3 Sonnet | Strikes the ideal balance between intelligence and speed – particularly for enterprise workloads. Delivers strong performance at a lower cost. | 200,000 | Y | N |
Claude 3.5 Sonnet | Twice the speed of Claude 3 Opus, ideal model for complex workflows within a cost-effective pricing. | 200,000 | Y | N |
Claude 3 Opus (pending) | The most intelligent Claude 3 model, with best-in-market performance on highly complex tasks. | 200,000 | N | N |
Jurassic Mid | Offers strong balance between text generation speed, cost, and output quality. | 8,192 | Y | N |
Jurassic Ultra | Largest model of the Jurassic series, ideal for complex processing with optimal performance. | 8,192 | Y | N |
Titan Lite | Designed for use-case specific generation tasks with affordability and speed as priorities. | 4,096 | Y | N |
Titan Express | High-performance model for more complex tasks involving retrieval augmented generation (RAG). | 8,192 | Y | Y |
Llama 3 8B | Ideal for limited computational power and resources, edge devices, and faster training times. | 8,192 | Y | N |
Llama 3 70B | Ideal for content creation, conversational AI, language understanding, R&D, and Enterprise applications. | 8,192 | Y | N |