diff --git a/src/Providers/TogetherAI/src/Predefined/AllModels.cs b/src/Providers/TogetherAI/src/Predefined/AllModels.cs
index 269a15e3..7e3f4672 100644
--- a/src/Providers/TogetherAI/src/Predefined/AllModels.cs
+++ b/src/Providers/TogetherAI/src/Predefined/AllModels.cs
@@ -215,6 +215,11 @@ public class Qwen15Chat4BModel(TogetherAiProvider provider) : TogetherAiModel(pr
public class MythomaxL213BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MythomaxL213B);
+///
+/// Open Router Provider Instance
+public class MetaLlama370BReferenceModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BReference);
+
+
///
/// Open Router Provider Instance
public class Mistral7BInstructModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Mistral7BInstruct);
@@ -230,11 +235,6 @@ public class Mistral7BInstructV02Model(TogetherAiProvider provider) : TogetherAi
public class OpenOrcaMistral7B8KModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.OpenOrcaMistral7B8K);
-///
-/// Open Router Provider Instance
-public class _01AiYiChat34BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds._01AiYiChat34B);
-
-
///
/// Open Router Provider Instance
public class NousHermesLlama27BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.NousHermesLlama27B);
@@ -290,9 +290,44 @@ public class Qwen15Chat110BModel(TogetherAiProvider provider) : TogetherAiModel(
public class Llama2Chat13BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Llama2Chat13B);
-///
+///
/// Open Router Provider Instance
-public class MetaLlama38BChatModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama38BChat);
+public class Gemma2Instruct27BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Gemma2Instruct27B);
+
+
+///
+/// Open Router Provider Instance
+public class _01AiYiChat34BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds._01AiYiChat34B);
+
+
+///
+/// Open Router Provider Instance
+public class MetaLlama370BInstructTurboModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BInstructTurbo);
+
+
+///
+/// Open Router Provider Instance
+public class MetaLlama38BInstructTurboModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama38BInstructTurbo);
+
+
+///
+/// Open Router Provider Instance
+public class MetaLlama370BInstructLiteModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BInstructLite);
+
+
+///
+/// Open Router Provider Instance
+public class Gemma2Instruct9BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Gemma2Instruct9B);
+
+
+///
+/// Open Router Provider Instance
+public class MetaLlama38BInstructLiteModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama38BInstructLite);
+
+
+///
+/// Open Router Provider Instance
+public class MetaLlama38BReferenceModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama38BReference);
///
@@ -320,11 +355,6 @@ public class CodeLlama70BModel(TogetherAiProvider provider) : TogetherAiModel(pr
public class CodeLlamaPython7BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.CodeLlamaPython7B);
-///
-/// Open Router Provider Instance
-public class MetaLlama370BChatModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BChat);
-
-
///
/// Open Router Provider Instance
public class ReservedDbrxInstructModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.ReservedDbrxInstruct);
@@ -350,11 +380,6 @@ public class MetaLlama38BInstructModel(TogetherAiProvider provider) : TogetherAi
public class MetaLlama370BInstructModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BInstruct);
-///
-/// Open Router Provider Instance
-public class Gemma2Instruct9BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Gemma2Instruct9B);
-
-
///
/// Open Router Provider Instance
public class Hermes2ThetaLlama370BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Hermes2ThetaLlama370B);
diff --git a/src/Providers/TogetherAI/src/TogetherAiModelIds.cs b/src/Providers/TogetherAI/src/TogetherAiModelIds.cs
index e2327779..1065953d 100644
--- a/src/Providers/TogetherAI/src/TogetherAiModelIds.cs
+++ b/src/Providers/TogetherAI/src/TogetherAiModelIds.cs
@@ -479,6 +479,17 @@ public enum TogetherAiModelIds
///
MythomaxL213B,
+ ///
+ /// Name: Meta Llama 3 70B Reference
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.9/MTok
+ /// Completion Cost: $0.9/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Llama-3-70b-chat-hf
+ ///
+ MetaLlama370BReference,
+
///
/// Name: Mistral (7B) Instruct
/// Organization: mistralai
@@ -512,17 +523,6 @@ public enum TogetherAiModelIds
///
OpenOrcaMistral7B8K,
- ///
- /// Name: 01-ai Yi Chat (34B)
- /// Organization: 01.AI
- /// Context Length: 4096
- /// Prompt Cost: $0.8/MTok
- /// Completion Cost: $0.8/MTok
- /// Description: The Yi series models are large language models trained from scratch by developers at 01.AI
- /// HuggingFace Url: https://huggingface.co/zero-one-ai/Yi-34B-Chat
- ///
- _01AiYiChat34B,
-
///
/// Name: Nous Hermes LLaMA-2 (7B)
/// Organization: NousResearch
@@ -645,7 +645,84 @@ public enum TogetherAiModelIds
Llama2Chat13B,
///
- /// Name: Meta Llama 3 8B Chat
+ /// Name: Gemma-2 Instruct (27B)
+ /// Organization: Google
+ /// Context Length: 8192
+ /// Prompt Cost: $0.8/MTok
+ /// Completion Cost: $0.8/MTok
+ /// Description: Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
+ /// HuggingFace Url: https://huggingface.co/google/gemma-2-27b-it
+ ///
+ Gemma2Instruct27B,
+
+ ///
+ /// Name: 01-ai Yi Chat (34B)
+ /// Organization: 01.AI
+ /// Context Length: 4096
+ /// Prompt Cost: $0.8/MTok
+ /// Completion Cost: $0.8/MTok
+ /// Description: The Yi series models are large language models trained from scratch by developers at 01.AI
+ /// HuggingFace Url: https://huggingface.co/zero-one-ai/Yi-34B-Chat
+ ///
+ _01AiYiChat34B,
+
+ ///
+ /// Name: Meta Llama 3 70B Instruct Turbo
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.88/MTok
+ /// Completion Cost: $0.88/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct-Turbo
+ ///
+ MetaLlama370BInstructTurbo,
+
+ ///
+ /// Name: Meta Llama 3 8B Instruct Turbo
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.18/MTok
+ /// Completion Cost: $0.18/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct-Turbo
+ ///
+ MetaLlama38BInstructTurbo,
+
+ ///
+ /// Name: Meta Llama 3 70B Instruct Lite
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.54/MTok
+ /// Completion Cost: $0.54/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct-Lite
+ ///
+ MetaLlama370BInstructLite,
+
+ ///
+ /// Name: Gemma-2 Instruct (9B)
+ /// Organization: google
+ /// Context Length: 8192
+ /// Prompt Cost: $0.3/MTok
+ /// Completion Cost: $0.3/MTok
+ /// Description: Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
+ /// HuggingFace Url: https://huggingface.co/google/gemma-2-9b-it
+ ///
+ Gemma2Instruct9B,
+
+ ///
+ /// Name: Meta Llama 3 8B Instruct Lite
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.1/MTok
+ /// Completion Cost: $0.1/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct-Lite
+ ///
+ MetaLlama38BInstructLite,
+
+ ///
+ /// Name: Meta Llama 3 8B Reference
/// Organization: Meta
/// Context Length: 8192
/// Prompt Cost: $0.2/MTok
@@ -653,7 +730,7 @@ public enum TogetherAiModelIds
/// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
/// HuggingFace Url: https://huggingface.co/meta-llama/Llama-3-8b-chat-hf
///
- MetaLlama38BChat,
+ MetaLlama38BReference,
///
/// Name: Mixtral-8x7B Instruct v0.1
@@ -710,17 +787,6 @@ public enum TogetherAiModelIds
///
CodeLlamaPython7B,
- ///
- /// Name: Meta Llama 3 70B Chat
- /// Organization: Meta
- /// Context Length: 8192
- /// Prompt Cost: $0.9/MTok
- /// Completion Cost: $0.9/MTok
- /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
- /// HuggingFace Url: https://huggingface.co/meta-llama/Llama-3-70b-chat-hf
- ///
- MetaLlama370BChat,
-
///
/// Name: Reserved - DBRX Instruct
/// Organization: Databricks
@@ -776,17 +842,6 @@ public enum TogetherAiModelIds
///
MetaLlama370BInstruct,
- ///
- /// Name: Gemma-2 Instruct (9B)
- /// Organization: google
- /// Context Length: 8192
- /// Prompt Cost: $0.3/MTok
- /// Completion Cost: $0.3/MTok
- /// Description: Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
- /// HuggingFace Url: https://huggingface.co/google/gemma-2-9b-it
- ///
- Gemma2Instruct9B,
-
///
/// Name: Hermes 2 Theta Llama-3 70B
/// Organization: NousResearch
diff --git a/src/Providers/TogetherAI/src/TogetherAiModelProvider.cs b/src/Providers/TogetherAI/src/TogetherAiModelProvider.cs
index e3a5acb6..6f09fee4 100644
--- a/src/Providers/TogetherAI/src/TogetherAiModelProvider.cs
+++ b/src/Providers/TogetherAI/src/TogetherAiModelProvider.cs
@@ -52,10 +52,10 @@ public static class TogetherAiModelProvider
{ TogetherAiModelIds.OlmoInstruct7B, ToMetadata("allenai/OLMo-7B-Instruct",2048,2.0000000000000002E-07,2.0000000000000002E-07)},
{ TogetherAiModelIds.Qwen15Chat4B, ToMetadata("Qwen/Qwen1.5-4B-Chat",32768,1.0000000000000001E-07,1.0000000000000001E-07)},
{ TogetherAiModelIds.MythomaxL213B, ToMetadata("Gryphe/MythoMax-L2-13b",4096,3E-07,3E-07)},
+ { TogetherAiModelIds.MetaLlama370BReference, ToMetadata("meta-llama/Llama-3-70b-chat-hf",8192,9.000000000000001E-07,9.000000000000001E-07)},
{ TogetherAiModelIds.Mistral7BInstruct, ToMetadata("mistralai/Mistral-7B-Instruct-v0.1",4096,2.0000000000000002E-07,2.0000000000000002E-07)},
{ TogetherAiModelIds.Mistral7BInstructV02, ToMetadata("mistralai/Mistral-7B-Instruct-v0.2",32768,2.0000000000000002E-07,2.0000000000000002E-07)},
{ TogetherAiModelIds.OpenOrcaMistral7B8K, ToMetadata("Open-Orca/Mistral-7B-OpenOrca",8192,2.0000000000000002E-07,2.0000000000000002E-07)},
- { TogetherAiModelIds._01AiYiChat34B, ToMetadata("zero-one-ai/Yi-34B-Chat",4096,8.000000000000001E-07,8.000000000000001E-07)},
{ TogetherAiModelIds.NousHermesLlama27B, ToMetadata("NousResearch/Nous-Hermes-llama-2-7b",4096,2.0000000000000002E-07,2.0000000000000002E-07)},
{ TogetherAiModelIds.Qwen15Chat32B, ToMetadata("Qwen/Qwen1.5-32B-Chat",32768,8.000000000000001E-07,8.000000000000001E-07)},
{ TogetherAiModelIds.Qwen2Instruct72B, ToMetadata("Qwen/Qwen2-72B-Instruct",32768,9.000000000000001E-07,9.000000000000001E-07)},
@@ -67,19 +67,24 @@ public static class TogetherAiModelProvider
{ TogetherAiModelIds.Mistral7BInstructV03, ToMetadata("mistralai/Mistral-7B-Instruct-v0.3",32768,2.0000000000000002E-07,2.0000000000000002E-07)},
{ TogetherAiModelIds.Qwen15Chat110B, ToMetadata("Qwen/Qwen1.5-110B-Chat",32768,1.8000000000000001E-06,1.8000000000000001E-06)},
{ TogetherAiModelIds.Llama2Chat13B, ToMetadata("meta-llama/Llama-2-13b-chat-hf",4096,2.2E-07,2.2E-07)},
- { TogetherAiModelIds.MetaLlama38BChat, ToMetadata("meta-llama/Llama-3-8b-chat-hf",8192,2.0000000000000002E-07,2.0000000000000002E-07)},
+ { TogetherAiModelIds.Gemma2Instruct27B, ToMetadata("google/gemma-2-27b-it",8192,8.000000000000001E-07,8.000000000000001E-07)},
+ { TogetherAiModelIds._01AiYiChat34B, ToMetadata("zero-one-ai/Yi-34B-Chat",4096,8.000000000000001E-07,8.000000000000001E-07)},
+ { TogetherAiModelIds.MetaLlama370BInstructTurbo, ToMetadata("meta-llama/Meta-Llama-3-70B-Instruct-Turbo",8192,8.8E-07,8.8E-07)},
+ { TogetherAiModelIds.MetaLlama38BInstructTurbo, ToMetadata("meta-llama/Meta-Llama-3-8B-Instruct-Turbo",8192,1.8E-07,1.8E-07)},
+ { TogetherAiModelIds.MetaLlama370BInstructLite, ToMetadata("meta-llama/Meta-Llama-3-70B-Instruct-Lite",8192,5.4E-07,5.4E-07)},
+ { TogetherAiModelIds.Gemma2Instruct9B, ToMetadata("google/gemma-2-9b-it",8192,3E-07,3E-07)},
+ { TogetherAiModelIds.MetaLlama38BInstructLite, ToMetadata("meta-llama/Meta-Llama-3-8B-Instruct-Lite",8192,1.0000000000000001E-07,1.0000000000000001E-07)},
+ { TogetherAiModelIds.MetaLlama38BReference, ToMetadata("meta-llama/Llama-3-8b-chat-hf",8192,2.0000000000000002E-07,2.0000000000000002E-07)},
{ TogetherAiModelIds.Mixtral8X7BInstructV01, ToMetadata("mistralai/Mixtral-8x7B-Instruct-v0.1",32768,6E-07,6E-07)},
{ TogetherAiModelIds.Llama2Chat7B, ToMetadata("meta-llama/Llama-2-7b-chat-hf",4096,2.0000000000000002E-07,2.0000000000000002E-07)},
{ TogetherAiModelIds.Llama2Chat70B, ToMetadata("meta-llama/Llama-2-70b-chat-hf",4096,9.000000000000001E-07,9.000000000000001E-07)},
{ TogetherAiModelIds.CodeLlama70B, ToMetadata("codellama/CodeLlama-70b-hf",16384,9.000000000000001E-07,9.000000000000001E-07)},
{ TogetherAiModelIds.CodeLlamaPython7B, ToMetadata("codellama/CodeLlama-7b-Python-hf",16384,2.0000000000000002E-07,2.0000000000000002E-07)},
- { TogetherAiModelIds.MetaLlama370BChat, ToMetadata("meta-llama/Llama-3-70b-chat-hf",8192,9.000000000000001E-07,9.000000000000001E-07)},
{ TogetherAiModelIds.ReservedDbrxInstruct, ToMetadata("medaltv/dbrx-instruct",0,1.2E-06,1.2E-06)},
{ TogetherAiModelIds.DbrxInstruct, ToMetadata("databricks/dbrx-instruct",32768,1.2E-06,1.2E-06)},
{ TogetherAiModelIds.UpstageSolarInstructV111BInt4, ToMetadata("togethercomputer/SOLAR-10.7B-Instruct-v1.0-int4",4096,3E-07,3E-07)},
{ TogetherAiModelIds.MetaLlama38BInstruct, ToMetadata("meta-llama/Meta-Llama-3-8B-Instruct",8192,2.0000000000000002E-07,2.0000000000000002E-07)},
{ TogetherAiModelIds.MetaLlama370BInstruct, ToMetadata("meta-llama/Meta-Llama-3-70B-Instruct",8192,9.000000000000001E-07,9.000000000000001E-07)},
- { TogetherAiModelIds.Gemma2Instruct9B, ToMetadata("google/gemma-2-9b-it",8192,3E-07,3E-07)},
{ TogetherAiModelIds.Hermes2ThetaLlama370B, ToMetadata("NousResearch/Hermes-2-Theta-Llama-3-70B",8192,0,0)},
{ TogetherAiModelIds.Llama370BInstructGradient1048K, ToMetadata("gradientai/Llama-3-70B-Instruct-Gradient-1048k",1048576,0,0)},