Getting started with Amazon SageMaker JumpStart

Overview

Amazon SageMaker JumpStart is a machine learning (ML) hub that can help you accelerate your ML journey. Explore how you can get started with built-in algorithms with pretrained models from model hubs, pretrained foundation models, and prebuilt solutions to solve common use cases. To get started, see documentation or example notebooks that you can quickly execute.

Product Type
Text Tasks
Vision Tasks
Tabular Tasks
Audio Tasks
Multimodal
Reinforced Learning
Showing results: 1-12
Total results: 567
  • Popularity
  • Featured First
  • A-Z Model Name
  • Z-A Model Name
  • foundation model

    Featured
    Text Generation

    Meta-Llama-3-70B-Instruct

    Meta
    70B instruction tuned variant of Llama 3 models. Llama 3 uses a decoder-only transformer architecture and new tokenizer that provides improved model performance.
    Fine-tunable
  • foundation model

    Featured
    Text Generation

    Meta-Llama-3.1-405B-FP8

    Meta
    405B variant of Llama 3.1 models. Llama 3.1 uses a decoder-only transformer architecture and new tokenizer that provides improved model performance.
    Fine-tunable
  • foundation model

    Featured
    Text Generation

    Meta-Llama-3.1-8B-Instruct

    Meta
    8B instruction tuned variant of Llama 3.1 models. Llama 3.1 uses a decoder-only transformer architecture and new tokenizer that provides improved model performance.
    Fine-tunable
  • foundation model

    Featured
    Text Generation

    Meta-Llama-3.1-70B-Instruct

    Meta
    70B instruction tuned variant of Llama 3.1 models. Llama 3.1 uses a decoder-only transformer architecture and new tokenizer that provides improved model performance.
    Fine-tunable
  • foundation model

    Featured
    Text Generation

    Meta Llama 3.2 3B Instruct

    Meta
    3B instruction tuned variant of Llama 3.2 models. Llama 3.2 uses a decoder-only transformer architecture and new tokenizer that provides improved model performance.
    Fine-tunable
  • foundation model

    Featured
    Text Generation

    Meta Llama 3.2 1B Instruct

    Meta
    1B instruction tuned variant of Llama 3.2 models. Llama 3.2 uses a decoder-only transformer architecture and new tokenizer that provides improved model performance.
    Fine-tunable
  • foundation model

    Featured
    Vision Language

    Meta Llama 3.2 11B Vision Instruct

    Meta
    11b instruction-tuned variants of Llama 3.2 models that supports both text and image as input.
    Deploy only
  • foundation model

    Featured
    Text Generation

    Llama 2 13B

    Meta
    13B variant of Llama 2 models. Llama 2 is an auto-regressive language model that uses an optimized transformer architecture. Llama 2 is intended for commercial and research use in English. It comes in a range of parameter sizes—7 billion, 13 billion, and 70 billion—as well as pre-trained and fine-tuned variations.
    Fine-tunable
  • foundation model

    Featured
    Text Generation

    Llama 3

    Meta

    Llama three from Meta comes in two parameter sizes — 8B and 70B with 8k context length — that can support a broad range of use cases with improvements in reasoning, code generation, and instruction following.

    Deploy Only
  • foundation model

    Featured
    Text Generation

    Llama 2 70B

    Meta
    70B variant of Llama 2 models. Llama 2 is an auto-regressive language model that uses an optimized transformer architecture. Llama 2 is intended for commercial and research use in English. It comes in a range of parameter sizes—7 billion, 13 billion, and 70 billion—as well as pre-trained and fine-tuned variations.
    Fine-tunable
  • foundation model

    Featured
    Text Generation

    Llama 2 7B

    Meta
    7B variant of Llama 2 models. Llama 2 is an auto-regressive language model that uses an optimized transformer architecture. Llama 2 is intended for commercial and research use in English. It comes in a range of parameter sizes—7 billion, 13 billion, and 70 billion—as well as pre-trained and fine-tuned variations.
    Fine-tunable
  • foundation model

    Featured
    Text Generation

    Llama 2 70B Chat

    Meta
    70B dialogue use case optimized variant of Llama 2 models. Llama 2 is an auto-regressive language model that uses an optimized transformer architecture. Llama 2 is intended for commercial and research use in English. It comes in a range of parameter sizes—7 billion, 13 billion, and 70 billion—as well as pre-trained and fine-tuned variations.
    Fine-tunable
1 48