Skip to main content

Fine-tuning Models

Well-supported LLMs

Fine-tuning is currently available for the following models:

Model NameParametersArchitectureLicenseContext WindowSupported Fine-Tuning Context Window
mistral-7b7 billionMistralApache 2.03276832768
mistral-7b-instruct7 billionMistralApache 2.03276832768
mistral-7b-instruct-v0-27 billionMistralApache 2.03276832768
mixtral-8x7b46.7 billionMixtralApache 2.0327687168
mixtral-8x7b-instruct-v0-146.7 billionMixtralApache 2.0327687168
llama-3-8b8 billionLlama-3Meta (request for commercial use)81928192
llama-3-8b-instruct8 billionLlama-3Meta (request for commercial use)81928192
llama-3-70b70 billionLlama-3Meta (request for commercial use)81928192
llama-3-70b-instruct70 billionLlama-3Meta (request for commercial use)81928192
llama-2-7b7 billionLlama-2Meta (request for commercial use)40964096
llama-2-7b-chat7 billionLlama-2Meta (request for commercial use)40964096
llama-2-13b13 billionLlama-2Meta (request for commercial use)40964096
llama-2-13b-chat13 billionLlama-2Meta (request for commercial use)40964096
llama-2-70b70 billionLlama-2Meta (request for commercial use)40964096
llama-2-70b-chat70 billionLlama-2Meta (request for commercial use)40964096
codellama-13b-instruct13 billionLlama-2Meta (request for commercial use)1638416384
codellama-70b-instruct70 billionLlama-2Meta (request for commercial use)40964096
gemma-2b2 billionGemmaGoogle81928192
gemma-2b-instruct2 billionGemmaGoogle81928192
gemma-7b7 billionGemmaGoogle81928192
gemma-7b-instruct7 billionGemmaGoogle81928192
zephyr-7b-beta7 billionMistralMIT3276832768
phi-22.78 billionPhiMicrosoft20482048

Many of the latest OSS models are released in two variants:

  • Base model (llama-2-7b, etc): These are models that are primarily trained on the objective of text completion.
  • Instruction-Tuned (llama-2-7b-chat, mistral-7b-instruct, etc): These are models that have been further trained on (instruction, output) pairs in order to better respond to human instruction-styled inputs. The instructions effectively constrains the model’s output to align with the response characteristics or domain knowledge.

Best-Effort LLMs (via HuggingFace)

Fine-tuning is also supported for any Huggingface LLM meeting the following criteria:

  • Has the "Text Generation" and "Transformer" tags
  • Does not have a "custom_code" tag
  • Are not post-quantized (ex. model containing a quantization method such as "AWQ" in the name)
  • Has text inputs and outputs

Note: Best-effort indicates we will try to support these models but it is not guaranteed.