diff --git a/src/Providers/TogetherAI/src/Predefined/AllModels.cs b/src/Providers/TogetherAI/src/Predefined/AllModels.cs index 269a15e3..7e3f4672 100644 --- a/src/Providers/TogetherAI/src/Predefined/AllModels.cs +++ b/src/Providers/TogetherAI/src/Predefined/AllModels.cs @@ -215,6 +215,11 @@ public class Qwen15Chat4BModel(TogetherAiProvider provider) : TogetherAiModel(pr public class MythomaxL213BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MythomaxL213B); +/// +/// Open Router Provider Instance +public class MetaLlama370BReferenceModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BReference); + + /// /// Open Router Provider Instance public class Mistral7BInstructModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Mistral7BInstruct); @@ -230,11 +235,6 @@ public class Mistral7BInstructV02Model(TogetherAiProvider provider) : TogetherAi public class OpenOrcaMistral7B8KModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.OpenOrcaMistral7B8K); -/// -/// Open Router Provider Instance -public class _01AiYiChat34BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds._01AiYiChat34B); - - /// /// Open Router Provider Instance public class NousHermesLlama27BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.NousHermesLlama27B); @@ -290,9 +290,44 @@ public class Qwen15Chat110BModel(TogetherAiProvider provider) : TogetherAiModel( public class Llama2Chat13BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Llama2Chat13B); -/// +/// /// Open Router Provider Instance -public class MetaLlama38BChatModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama38BChat); +public class Gemma2Instruct27BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Gemma2Instruct27B); + + +/// +/// Open Router Provider Instance +public class _01AiYiChat34BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds._01AiYiChat34B); + + +/// +/// Open Router Provider Instance +public class MetaLlama370BInstructTurboModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BInstructTurbo); + + +/// +/// Open Router Provider Instance +public class MetaLlama38BInstructTurboModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama38BInstructTurbo); + + +/// +/// Open Router Provider Instance +public class MetaLlama370BInstructLiteModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BInstructLite); + + +/// +/// Open Router Provider Instance +public class Gemma2Instruct9BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Gemma2Instruct9B); + + +/// +/// Open Router Provider Instance +public class MetaLlama38BInstructLiteModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama38BInstructLite); + + +/// +/// Open Router Provider Instance +public class MetaLlama38BReferenceModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama38BReference); /// @@ -320,11 +355,6 @@ public class CodeLlama70BModel(TogetherAiProvider provider) : TogetherAiModel(pr public class CodeLlamaPython7BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.CodeLlamaPython7B); -/// -/// Open Router Provider Instance -public class MetaLlama370BChatModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BChat); - - /// /// Open Router Provider Instance public class ReservedDbrxInstructModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.ReservedDbrxInstruct); @@ -350,11 +380,6 @@ public class MetaLlama38BInstructModel(TogetherAiProvider provider) : TogetherAi public class MetaLlama370BInstructModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.MetaLlama370BInstruct); -/// -/// Open Router Provider Instance -public class Gemma2Instruct9BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Gemma2Instruct9B); - - /// /// Open Router Provider Instance public class Hermes2ThetaLlama370BModel(TogetherAiProvider provider) : TogetherAiModel(provider, TogetherAiModelIds.Hermes2ThetaLlama370B); diff --git a/src/Providers/TogetherAI/src/TogetherAiModelIds.cs b/src/Providers/TogetherAI/src/TogetherAiModelIds.cs index e2327779..1065953d 100644 --- a/src/Providers/TogetherAI/src/TogetherAiModelIds.cs +++ b/src/Providers/TogetherAI/src/TogetherAiModelIds.cs @@ -479,6 +479,17 @@ public enum TogetherAiModelIds /// MythomaxL213B, + /// + /// Name: Meta Llama 3 70B Reference
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.9/MTok
+ /// Completion Cost: $0.9/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Llama-3-70b-chat-hf + ///
+ MetaLlama370BReference, + /// /// Name: Mistral (7B) Instruct
/// Organization: mistralai
@@ -512,17 +523,6 @@ public enum TogetherAiModelIds ///
OpenOrcaMistral7B8K, - /// - /// Name: 01-ai Yi Chat (34B)
- /// Organization: 01.AI
- /// Context Length: 4096
- /// Prompt Cost: $0.8/MTok
- /// Completion Cost: $0.8/MTok
- /// Description: The Yi series models are large language models trained from scratch by developers at 01.AI
- /// HuggingFace Url: https://huggingface.co/zero-one-ai/Yi-34B-Chat - ///
- _01AiYiChat34B, - /// /// Name: Nous Hermes LLaMA-2 (7B)
/// Organization: NousResearch
@@ -645,7 +645,84 @@ public enum TogetherAiModelIds Llama2Chat13B, /// - /// Name: Meta Llama 3 8B Chat
+ /// Name: Gemma-2 Instruct (27B)
+ /// Organization: Google
+ /// Context Length: 8192
+ /// Prompt Cost: $0.8/MTok
+ /// Completion Cost: $0.8/MTok
+ /// Description: Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
+ /// HuggingFace Url: https://huggingface.co/google/gemma-2-27b-it + ///
+ Gemma2Instruct27B, + + /// + /// Name: 01-ai Yi Chat (34B)
+ /// Organization: 01.AI
+ /// Context Length: 4096
+ /// Prompt Cost: $0.8/MTok
+ /// Completion Cost: $0.8/MTok
+ /// Description: The Yi series models are large language models trained from scratch by developers at 01.AI
+ /// HuggingFace Url: https://huggingface.co/zero-one-ai/Yi-34B-Chat + ///
+ _01AiYiChat34B, + + /// + /// Name: Meta Llama 3 70B Instruct Turbo
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.88/MTok
+ /// Completion Cost: $0.88/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct-Turbo + ///
+ MetaLlama370BInstructTurbo, + + /// + /// Name: Meta Llama 3 8B Instruct Turbo
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.18/MTok
+ /// Completion Cost: $0.18/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct-Turbo + ///
+ MetaLlama38BInstructTurbo, + + /// + /// Name: Meta Llama 3 70B Instruct Lite
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.54/MTok
+ /// Completion Cost: $0.54/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct-Lite + ///
+ MetaLlama370BInstructLite, + + /// + /// Name: Gemma-2 Instruct (9B)
+ /// Organization: google
+ /// Context Length: 8192
+ /// Prompt Cost: $0.3/MTok
+ /// Completion Cost: $0.3/MTok
+ /// Description: Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
+ /// HuggingFace Url: https://huggingface.co/google/gemma-2-9b-it + ///
+ Gemma2Instruct9B, + + /// + /// Name: Meta Llama 3 8B Instruct Lite
+ /// Organization: Meta
+ /// Context Length: 8192
+ /// Prompt Cost: $0.1/MTok
+ /// Completion Cost: $0.1/MTok
+ /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
+ /// HuggingFace Url: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct-Lite + ///
+ MetaLlama38BInstructLite, + + /// + /// Name: Meta Llama 3 8B Reference
/// Organization: Meta
/// Context Length: 8192
/// Prompt Cost: $0.2/MTok
@@ -653,7 +730,7 @@ public enum TogetherAiModelIds /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
/// HuggingFace Url: https://huggingface.co/meta-llama/Llama-3-8b-chat-hf ///
- MetaLlama38BChat, + MetaLlama38BReference, /// /// Name: Mixtral-8x7B Instruct v0.1
@@ -710,17 +787,6 @@ public enum TogetherAiModelIds ///
CodeLlamaPython7B, - /// - /// Name: Meta Llama 3 70B Chat
- /// Organization: Meta
- /// Context Length: 8192
- /// Prompt Cost: $0.9/MTok
- /// Completion Cost: $0.9/MTok
- /// Description: Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
- /// HuggingFace Url: https://huggingface.co/meta-llama/Llama-3-70b-chat-hf - ///
- MetaLlama370BChat, - /// /// Name: Reserved - DBRX Instruct
/// Organization: Databricks
@@ -776,17 +842,6 @@ public enum TogetherAiModelIds ///
MetaLlama370BInstruct, - /// - /// Name: Gemma-2 Instruct (9B)
- /// Organization: google
- /// Context Length: 8192
- /// Prompt Cost: $0.3/MTok
- /// Completion Cost: $0.3/MTok
- /// Description: Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
- /// HuggingFace Url: https://huggingface.co/google/gemma-2-9b-it - ///
- Gemma2Instruct9B, - /// /// Name: Hermes 2 Theta Llama-3 70B
/// Organization: NousResearch
diff --git a/src/Providers/TogetherAI/src/TogetherAiModelProvider.cs b/src/Providers/TogetherAI/src/TogetherAiModelProvider.cs index e3a5acb6..6f09fee4 100644 --- a/src/Providers/TogetherAI/src/TogetherAiModelProvider.cs +++ b/src/Providers/TogetherAI/src/TogetherAiModelProvider.cs @@ -52,10 +52,10 @@ public static class TogetherAiModelProvider { TogetherAiModelIds.OlmoInstruct7B, ToMetadata("allenai/OLMo-7B-Instruct",2048,2.0000000000000002E-07,2.0000000000000002E-07)}, { TogetherAiModelIds.Qwen15Chat4B, ToMetadata("Qwen/Qwen1.5-4B-Chat",32768,1.0000000000000001E-07,1.0000000000000001E-07)}, { TogetherAiModelIds.MythomaxL213B, ToMetadata("Gryphe/MythoMax-L2-13b",4096,3E-07,3E-07)}, + { TogetherAiModelIds.MetaLlama370BReference, ToMetadata("meta-llama/Llama-3-70b-chat-hf",8192,9.000000000000001E-07,9.000000000000001E-07)}, { TogetherAiModelIds.Mistral7BInstruct, ToMetadata("mistralai/Mistral-7B-Instruct-v0.1",4096,2.0000000000000002E-07,2.0000000000000002E-07)}, { TogetherAiModelIds.Mistral7BInstructV02, ToMetadata("mistralai/Mistral-7B-Instruct-v0.2",32768,2.0000000000000002E-07,2.0000000000000002E-07)}, { TogetherAiModelIds.OpenOrcaMistral7B8K, ToMetadata("Open-Orca/Mistral-7B-OpenOrca",8192,2.0000000000000002E-07,2.0000000000000002E-07)}, - { TogetherAiModelIds._01AiYiChat34B, ToMetadata("zero-one-ai/Yi-34B-Chat",4096,8.000000000000001E-07,8.000000000000001E-07)}, { TogetherAiModelIds.NousHermesLlama27B, ToMetadata("NousResearch/Nous-Hermes-llama-2-7b",4096,2.0000000000000002E-07,2.0000000000000002E-07)}, { TogetherAiModelIds.Qwen15Chat32B, ToMetadata("Qwen/Qwen1.5-32B-Chat",32768,8.000000000000001E-07,8.000000000000001E-07)}, { TogetherAiModelIds.Qwen2Instruct72B, ToMetadata("Qwen/Qwen2-72B-Instruct",32768,9.000000000000001E-07,9.000000000000001E-07)}, @@ -67,19 +67,24 @@ public static class TogetherAiModelProvider { TogetherAiModelIds.Mistral7BInstructV03, ToMetadata("mistralai/Mistral-7B-Instruct-v0.3",32768,2.0000000000000002E-07,2.0000000000000002E-07)}, { TogetherAiModelIds.Qwen15Chat110B, ToMetadata("Qwen/Qwen1.5-110B-Chat",32768,1.8000000000000001E-06,1.8000000000000001E-06)}, { TogetherAiModelIds.Llama2Chat13B, ToMetadata("meta-llama/Llama-2-13b-chat-hf",4096,2.2E-07,2.2E-07)}, - { TogetherAiModelIds.MetaLlama38BChat, ToMetadata("meta-llama/Llama-3-8b-chat-hf",8192,2.0000000000000002E-07,2.0000000000000002E-07)}, + { TogetherAiModelIds.Gemma2Instruct27B, ToMetadata("google/gemma-2-27b-it",8192,8.000000000000001E-07,8.000000000000001E-07)}, + { TogetherAiModelIds._01AiYiChat34B, ToMetadata("zero-one-ai/Yi-34B-Chat",4096,8.000000000000001E-07,8.000000000000001E-07)}, + { TogetherAiModelIds.MetaLlama370BInstructTurbo, ToMetadata("meta-llama/Meta-Llama-3-70B-Instruct-Turbo",8192,8.8E-07,8.8E-07)}, + { TogetherAiModelIds.MetaLlama38BInstructTurbo, ToMetadata("meta-llama/Meta-Llama-3-8B-Instruct-Turbo",8192,1.8E-07,1.8E-07)}, + { TogetherAiModelIds.MetaLlama370BInstructLite, ToMetadata("meta-llama/Meta-Llama-3-70B-Instruct-Lite",8192,5.4E-07,5.4E-07)}, + { TogetherAiModelIds.Gemma2Instruct9B, ToMetadata("google/gemma-2-9b-it",8192,3E-07,3E-07)}, + { TogetherAiModelIds.MetaLlama38BInstructLite, ToMetadata("meta-llama/Meta-Llama-3-8B-Instruct-Lite",8192,1.0000000000000001E-07,1.0000000000000001E-07)}, + { TogetherAiModelIds.MetaLlama38BReference, ToMetadata("meta-llama/Llama-3-8b-chat-hf",8192,2.0000000000000002E-07,2.0000000000000002E-07)}, { TogetherAiModelIds.Mixtral8X7BInstructV01, ToMetadata("mistralai/Mixtral-8x7B-Instruct-v0.1",32768,6E-07,6E-07)}, { TogetherAiModelIds.Llama2Chat7B, ToMetadata("meta-llama/Llama-2-7b-chat-hf",4096,2.0000000000000002E-07,2.0000000000000002E-07)}, { TogetherAiModelIds.Llama2Chat70B, ToMetadata("meta-llama/Llama-2-70b-chat-hf",4096,9.000000000000001E-07,9.000000000000001E-07)}, { TogetherAiModelIds.CodeLlama70B, ToMetadata("codellama/CodeLlama-70b-hf",16384,9.000000000000001E-07,9.000000000000001E-07)}, { TogetherAiModelIds.CodeLlamaPython7B, ToMetadata("codellama/CodeLlama-7b-Python-hf",16384,2.0000000000000002E-07,2.0000000000000002E-07)}, - { TogetherAiModelIds.MetaLlama370BChat, ToMetadata("meta-llama/Llama-3-70b-chat-hf",8192,9.000000000000001E-07,9.000000000000001E-07)}, { TogetherAiModelIds.ReservedDbrxInstruct, ToMetadata("medaltv/dbrx-instruct",0,1.2E-06,1.2E-06)}, { TogetherAiModelIds.DbrxInstruct, ToMetadata("databricks/dbrx-instruct",32768,1.2E-06,1.2E-06)}, { TogetherAiModelIds.UpstageSolarInstructV111BInt4, ToMetadata("togethercomputer/SOLAR-10.7B-Instruct-v1.0-int4",4096,3E-07,3E-07)}, { TogetherAiModelIds.MetaLlama38BInstruct, ToMetadata("meta-llama/Meta-Llama-3-8B-Instruct",8192,2.0000000000000002E-07,2.0000000000000002E-07)}, { TogetherAiModelIds.MetaLlama370BInstruct, ToMetadata("meta-llama/Meta-Llama-3-70B-Instruct",8192,9.000000000000001E-07,9.000000000000001E-07)}, - { TogetherAiModelIds.Gemma2Instruct9B, ToMetadata("google/gemma-2-9b-it",8192,3E-07,3E-07)}, { TogetherAiModelIds.Hermes2ThetaLlama370B, ToMetadata("NousResearch/Hermes-2-Theta-Llama-3-70B",8192,0,0)}, { TogetherAiModelIds.Llama370BInstructGradient1048K, ToMetadata("gradientai/Llama-3-70B-Instruct-Gradient-1048k",1048576,0,0)},