Toggle navigation
MeasureThat.net
Create a benchmark
Tools
Feedback
FAQ
Register
Log In
Run results for:
Lodash cloneDeep vs structuredClone v1
https://developer.mozilla.org/en-US/docs/Web/API/structuredClone
Go to the benchmark
Embed
Embed Benchmark Result
Run details:
User agent:
Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/131.0.0.0 Safari/537.36
Browser:
Chrome 131
Operating system:
Mac OS X 10.15.7
Device Platform:
Desktop
Date tested:
one year ago
Test name
Executions per second
Lodash cloneDeep
1196.3 Ops/sec
Native structuredClone
1654.7 Ops/sec
HTML Preparation code:
<script src='https://cdnjs.cloudflare.com/ajax/libs/lodash.js/4.17.5/lodash.min.js'></script>
Script Preparation code:
var MyObject = { "variants": [ { "id": "019461ff-f31c-74b3-981c-dad1d891b8fb", "uri": "http://localhost:80/services/completion", "appId": "019461ff-f226-7a36-a9e1-53c11329a342", "baseId": "019461ff-f316-70de-84d1-14cda41b0250", "baseName": "app", "variantName": "app.key", "revision": 1, "configName": "key", "projectId": "01945fa2-db0d-78dc-9c67-462f129da09e", "appName": "test-arda-1", "parameters": { "agentaConfig": { "prompt": { "messages": [ { "role": "system", "content": "You are an expert in geography" }, { "role": "user", "content": "What is the capital of {country}?" } ], "input_keys": [ "country" ], "llm_config": { "model": "gpt-3.5-turbo", "top_p": 0.5, "max_tokens": 86, "temperature": 1, "frequency_penalty": 0 }, "user_prompt": "What is the capital of {country}?", "system_prompt": "You are an expert in geography", "template_format": "fstring" } } }, "isChat": false, "inputs": { "__id": "4a8e6964-bde9-4ee3-9735-56f332cc3087", "__metadata": { "type": "array", "title": "Input Rows", "description": "Input values for variant execution", "itemMetadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "nullable": false }, "value": [ { "__id": "318dd2f7-5475-4836-b888-9f76798c9370", "__metadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "__runs": {}, "country": { "__id": "b3c181ac-99ca-415e-a1bb-2ff44a0a6025", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true }, "value": "" } } ] }, "messages": {}, "name": "", "prompts": [ { "__id": "6f0105b2-d095-4a0f-ab2f-6508b5ba8686", "__metadata": { "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "properties": { "messages": { "type": "array", "title": "Messages", "itemMetadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "nullable": false }, "systemPrompt": { "type": "string", "title": "System Prompt", "nullable": true, "allowFreeform": true }, "userPrompt": { "type": "string", "title": "User Prompt", "nullable": true, "allowFreeform": true }, "templateFormat": { "type": "string", "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "options": [ { "label": "fstring", "value": "fstring" }, { "label": "jinja2", "value": "jinja2" }, { "label": "curly", "value": "curly" } ], "nullable": false, "allowFreeform": false }, "inputKeys": { "type": "array", "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted", "itemMetadata": { "type": "string", "nullable": false, "allowFreeform": true }, "nullable": true }, "llmConfig": { "type": "object", "title": "ModelConfig", "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false }, "temperature": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true, "isInteger": false }, "maxTokens": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "nullable": true, "isInteger": false }, "topP": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true, "isInteger": false }, "frequencyPenalty": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "presencePenalty": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "responseFormat": { "type": "object", "title": "ResponseFormatText", "description": "An object specifying the format that the model must output", "properties": { "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false } }, "nullable": true }, "stream": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true }, "tools": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "itemMetadata": { "type": "object", "properties": {} }, "nullable": true }, "toolChoice": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true, "allowFreeform": false } }, "nullable": false } } }, "messages": { "__id": "15958cdc-6895-48c8-b3c9-8ea6f3dbfc77", "__metadata": { "type": "array", "title": "Messages", "itemMetadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "nullable": false }, "value": [ { "__id": "c742a423-5b72-460a-8869-3def2d0d74e3", "__metadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "role": { "value": "system", "__id": "21bada93-1a22-45a9-9a62-1a28429da612", "__metadata": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false } }, "content": { "value": "You are an expert in geography", "__id": "859344ad-8450-4a7a-bdde-d57c511b07b5", "__metadata": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true } }, "name": { "value": null, "__id": "3efdfc74-08e4-48c8-9bbd-f5d2251eadbf", "__metadata": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true } }, "toolCalls": { "value": null, "__id": "c39ab97b-6dfd-40e1-876b-f3d416078f79", "__metadata": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true } }, "toolCallId": { "value": null, "__id": "5edd1560-1bd4-48b5-a13e-22e9ab8b5eb1", "__metadata": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, { "__id": "09706047-360b-4f4f-a96e-8cbc3a5a3d79", "__metadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "role": { "value": "user", "__id": "f13cf72e-fcd9-461c-913e-7a301c19beec", "__metadata": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false } }, "content": { "value": "What is the capital of {country}? {a1}", "__id": "2c8cce66-a25b-4c6a-b243-5cc93137d0fb", "__metadata": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true } }, "name": { "value": null, "__id": "d6577bea-c0f6-47b5-90d2-550ddc0b006f", "__metadata": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true } }, "toolCalls": { "value": null, "__id": "ab1c6c74-6ec5-433d-9d08-036c46dde8a7", "__metadata": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true } }, "toolCallId": { "value": null, "__id": "e344acff-152f-4e7b-a94b-14f20db5c81b", "__metadata": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } } ] }, "systemPrompt": { "value": null, "__id": "2161ab5b-fe2f-4586-b0be-c9fef3a43bb6", "__metadata": { "type": "string", "title": "System Prompt", "nullable": true } }, "userPrompt": { "value": null, "__id": "cef972cb-6a36-48c1-94ae-6ff1cbf75676", "__metadata": { "type": "string", "title": "User Prompt", "nullable": true } }, "templateFormat": { "value": "fstring", "__id": "e724bc26-8993-4edf-b760-2b68a8f793be", "__metadata": { "type": "string", "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "options": [ { "label": "fstring", "value": "fstring" }, { "label": "jinja2", "value": "jinja2" }, { "label": "curly", "value": "curly" } ], "nullable": false, "allowFreeform": false } }, "inputKeys": { "value": [ { "value": "country", "__id": "ce097afd-ec37-4ce2-96d6-1fdc563a4abd", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, { "value": "a1", "__id": "d411c317-79a2-4d5a-8447-5462bce97503", "__metadata": { "type": "string", "title": "a1", "nullable": false, "allowFreeform": true } } ], "__id": "c9083345-b2ff-48dd-a7c8-875049761511", "__metadata": { "type": "array", "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted", "nullable": true } }, "llmConfig": { "__id": "061a3297-217b-495d-8655-dc6fd95d6ac6", "__metadata": { "type": "object", "title": "ModelConfig", "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false }, "temperature": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true, "isInteger": false }, "maxTokens": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "nullable": true, "isInteger": false }, "topP": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true, "isInteger": false }, "frequencyPenalty": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "presencePenalty": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "responseFormat": { "type": "object", "title": "ResponseFormatText", "description": "An object specifying the format that the model must output", "properties": { "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false } }, "nullable": true }, "stream": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true }, "tools": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "itemMetadata": { "type": "object", "properties": {} }, "nullable": true }, "toolChoice": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true, "allowFreeform": false } } }, "model": { "value": "gpt-3.5-turbo", "__id": "c2f35d73-6c4f-4d16-8391-cf4dea75c0f5", "__metadata": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false } }, "temperature": { "value": 1, "__id": "a89f8a38-6ea7-49bc-993d-55ae3c11097b", "__metadata": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true } }, "maxTokens": { "value": 86, "__id": "de83333d-c938-4d71-983a-66d582d1cded", "__metadata": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "isInteger": true, "nullable": true } }, "topP": { "value": 0.5, "__id": "6b765570-edee-461e-89f5-2f8bbf02e126", "__metadata": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true } }, "frequencyPenalty": { "value": 0, "__id": "fff2b53b-78fc-4966-aa77-5760725f5327", "__metadata": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true } }, "presencePenalty": { "value": 0, "__id": "b013fe56-0a17-4b12-b514-87fb593bc8d6", "__metadata": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true } }, "responseFormat": { "value": null, "__id": "4843c191-bd33-4bec-b787-dc4fac1619d5", "__metadata": { "type": "compound", "title": "Response Format", "description": "An object specifying the format that the model must output", "nullable": true, "options": [ { "label": "ResponseFormatText", "value": "text", "config": { "type": "text" } }, { "label": "ResponseFormatJSONObject", "value": "json_object", "config": { "type": "json_object" } }, { "label": "ResponseFormatJSONSchema", "value": "json_schema", "config": { "type": "json_schema", "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } } } ] } }, "stream": { "value": null, "__id": "7a12a51c-8c24-4a28-9917-c0d842442347", "__metadata": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true } }, "tools": { "value": null, "__id": "e63c7ea1-231f-435e-8722-489a0a937a60", "__metadata": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "nullable": true } }, "toolChoice": { "value": null, "__id": "8c64f589-bb17-4359-83dc-0f364740fcfa", "__metadata": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true } } } } ] }, { "id": "01946478-6ca0-72bf-88ed-0c803c2f9f4f", "uri": "http://localhost:80/services/completion", "appId": "019461ff-f226-7a36-a9e1-53c11329a342", "baseId": "019461ff-f316-70de-84d1-14cda41b0250", "baseName": "app", "variantName": "app.arda-2", "revision": 5, "configName": "arda-2", "projectId": "01945fa2-db0d-78dc-9c67-462f129da09e", "appName": "test-arda-1", "parameters": { "agentaConfig": { "prompt": { "messages": [ { "role": "system", "content": "You are an expert in geography" }, { "role": "user", "content": "What is the capital of {country}?" } ], "input_keys": [ "country" ], "llm_config": { "model": "gpt-3.5-turbo", "top_p": 0.5, "max_tokens": 84, "temperature": 1, "frequency_penalty": 0 }, "template_format": "fstring" } } }, "isChat": false, "inputs": { "__id": "ff120363-2c11-47a6-95de-0f72828676f8", "__metadata": { "type": "array", "title": "Input Rows", "description": "Input values for variant execution", "itemMetadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "nullable": false }, "value": [ { "__id": "dde5a18c-1113-4bf3-bc35-b0e817f3d529", "__metadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "__runs": {}, "country": { "__id": "eb5189a5-8a51-4be5-8da0-76be00544b38", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true }, "value": "" } } ] }, "messages": {}, "name": "", "prompts": [ { "__id": "39d74b2b-53e2-46bc-8057-f8527836dfaf", "__metadata": { "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "properties": { "messages": { "type": "array", "title": "Messages", "itemMetadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "nullable": false }, "systemPrompt": { "type": "string", "title": "System Prompt", "nullable": true, "allowFreeform": true }, "userPrompt": { "type": "string", "title": "User Prompt", "nullable": true, "allowFreeform": true }, "templateFormat": { "type": "string", "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "options": [ { "label": "fstring", "value": "fstring" }, { "label": "jinja2", "value": "jinja2" }, { "label": "curly", "value": "curly" } ], "nullable": false, "allowFreeform": false }, "inputKeys": { "type": "array", "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted", "itemMetadata": { "type": "string", "nullable": false, "allowFreeform": true }, "nullable": true }, "llmConfig": { "type": "object", "title": "ModelConfig", "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false }, "temperature": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true, "isInteger": false }, "maxTokens": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "nullable": true, "isInteger": false }, "topP": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true, "isInteger": false }, "frequencyPenalty": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "presencePenalty": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "responseFormat": { "type": "object", "title": "ResponseFormatText", "description": "An object specifying the format that the model must output", "properties": { "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false } }, "nullable": true }, "stream": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true }, "tools": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "itemMetadata": { "type": "object", "properties": {} }, "nullable": true }, "toolChoice": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true, "allowFreeform": false } }, "nullable": false } } }, "messages": { "__id": "b9c9d862-3326-48ad-8549-379df5b7ccc7", "__metadata": { "type": "array", "title": "Messages", "itemMetadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "nullable": false }, "value": [ { "__id": "ee434a01-d7f3-494a-8584-370bf9e2f1fb", "__metadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "role": { "value": "system", "__id": "025033e7-9279-4e75-a949-b7f6791bfc51", "__metadata": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false } }, "content": { "value": "You are an expert in geography", "__id": "f44b4ef5-7f4f-4684-a9ff-7db90d481b01", "__metadata": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true } }, "name": { "value": null, "__id": "43b9a314-49c3-4d7d-aa38-44481a5586a8", "__metadata": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true } }, "toolCalls": { "value": null, "__id": "0f478f7e-080b-4bec-a965-c2fb52f3650e", "__metadata": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true } }, "toolCallId": { "value": null, "__id": "2f8662a1-7334-4692-80fd-2b1597506e48", "__metadata": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, { "__id": "d5d204e6-23fa-4b45-98a7-5b289a65262e", "__metadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "role": { "value": "user", "__id": "9eebfcc9-9cb9-4630-a9d6-07659d3cf3a3", "__metadata": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false } }, "content": { "value": "What is the capital of {country}?", "__id": "ae3c7fe6-a992-49ae-acde-232706e1c533", "__metadata": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true } }, "name": { "value": null, "__id": "052bf3ee-d78c-481c-a08c-f98600557ad3", "__metadata": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true } }, "toolCalls": { "value": null, "__id": "60a38c86-febe-482b-aa26-f73ea773d70a", "__metadata": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true } }, "toolCallId": { "value": null, "__id": "ab653a63-ba2d-4a17-bea6-6d21a4796c57", "__metadata": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } } ] }, "systemPrompt": { "value": null, "__id": "1b677c97-e3f6-43b4-93cd-d758ba52af82", "__metadata": { "type": "string", "title": "System Prompt", "nullable": true } }, "userPrompt": { "value": null, "__id": "4bb5f46c-8e56-4a84-8734-256f86d8b59f", "__metadata": { "type": "string", "title": "User Prompt", "nullable": true } }, "templateFormat": { "value": "fstring", "__id": "b1bdfb1a-3f66-4bab-a2a2-09f1f8adaa4b", "__metadata": { "type": "string", "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "options": [ { "label": "fstring", "value": "fstring" }, { "label": "jinja2", "value": "jinja2" }, { "label": "curly", "value": "curly" } ], "nullable": false, "allowFreeform": false } }, "inputKeys": { "value": [ { "value": "country", "__id": "10ee0960-60e9-4a17-8f08-bb7222a8ef1e", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } } ], "__id": "2fbc73b0-09ad-4458-a27d-285b34d150a6", "__metadata": { "type": "array", "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted", "nullable": true } }, "llmConfig": { "__id": "6ed1b3ff-9865-4658-a58e-3af989d6a8c3", "__metadata": { "type": "object", "title": "ModelConfig", "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false }, "temperature": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true, "isInteger": false }, "maxTokens": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "nullable": true, "isInteger": false }, "topP": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true, "isInteger": false }, "frequencyPenalty": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "presencePenalty": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "responseFormat": { "type": "object", "title": "ResponseFormatText", "description": "An object specifying the format that the model must output", "properties": { "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false } }, "nullable": true }, "stream": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true }, "tools": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "itemMetadata": { "type": "object", "properties": {} }, "nullable": true }, "toolChoice": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true, "allowFreeform": false } } }, "model": { "value": "gpt-3.5-turbo", "__id": "2643d1ad-b6a1-416b-b769-5c3add32e6a6", "__metadata": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false } }, "temperature": { "value": 1, "__id": "db5ad2fa-6e64-4da7-a4e1-4b06b4e8e22c", "__metadata": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true } }, "maxTokens": { "value": 84, "__id": "fe1f6aec-6d91-43cc-b0aa-6381b7759748", "__metadata": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "isInteger": true, "nullable": true } }, "topP": { "value": 0.5, "__id": "dde89181-bdda-451a-8dae-c350ebaa5b87", "__metadata": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true } }, "frequencyPenalty": { "value": 0, "__id": "7d63570a-dbcc-466c-8126-7cb1a143122f", "__metadata": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true } }, "presencePenalty": { "value": 0, "__id": "eb50375c-6e62-41c5-8da0-8c5f69a4bde9", "__metadata": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true } }, "responseFormat": { "value": null, "__id": "bbdd8e13-f0ac-416c-8c5f-fbdf0b70fbd0", "__metadata": { "type": "compound", "title": "Response Format", "description": "An object specifying the format that the model must output", "nullable": true, "options": [ { "label": "ResponseFormatText", "value": "text", "config": { "type": "text" } }, { "label": "ResponseFormatJSONObject", "value": "json_object", "config": { "type": "json_object" } }, { "label": "ResponseFormatJSONSchema", "value": "json_schema", "config": { "type": "json_schema", "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } } } ] } }, "stream": { "value": null, "__id": "d8c349a3-91f0-454b-936f-a2aca998980d", "__metadata": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true } }, "tools": { "value": null, "__id": "ac6b5383-423f-4b0e-a53b-633ad4f2ad54", "__metadata": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "nullable": true } }, "toolChoice": { "value": null, "__id": "d5381a5e-e988-4120-90a4-044e69c30964", "__metadata": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true } } } } ] } ], "selected": [ "019461ff-f31c-74b3-981c-dad1d891b8fb" ], "dirtyStates": {}, "generationData": { "__id": "f01fc583-3359-49c9-b871-5d6032af2995", "__metadata": { "type": "array", "title": "Input Rows", "description": "Input values for variant execution", "itemMetadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "nullable": false }, "value": [ { "__id": "fb5bfef1-be48-4469-9a0a-3a8614947334", "__metadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "__runs": {}, "country": { "__id": "690d7666-2349-4158-b238-a6115505a7a2", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true }, "value": "" } } ] }, "spec": { "openapi": "3.1.0", "info": { "title": "FastAPI", "version": "0.1.0" }, "paths": { "/health": { "get": { "summary": "Health", "operationId": "health_health_get", "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": {} } } } } } }, "/run": { "post": { "summary": "Generate", "operationId": "generate_run_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_run_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } }, "/generate_deployed": { "post": { "summary": "Generate", "operationId": "generate_generate_deployed_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_generate_deployed_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } }, "/test": { "post": { "summary": "Generate", "operationId": "generate_test_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_test_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } }, "/generate": { "post": { "summary": "Generate", "operationId": "generate_generate_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_generate_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } }, "/playground/run": { "post": { "summary": "Generate", "operationId": "generate_playground_run_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_playground_run_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } } }, "components": { "schemas": { "AgentaNodesResponse": {}, "BaseResponse": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" }, "Body_generate_generate_deployed_post": { "properties": { "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_generate_deployed_post" }, "Body_generate_generate_post": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_generate_post" }, "Body_generate_playground_run_post": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_playground_run_post" }, "Body_generate_run_post": { "properties": { "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_run_post" }, "Body_generate_test_post": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_test_post" }, "ExceptionDto": {}, "HTTPValidationError": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" }, "JSONSchema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" }, "LifecycleDto": {}, "LinkDto": {}, "Message": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "ModelConfig": { "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig", "description": "Configuration for model parameters" }, "MyConfig": { "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "NodeDto": {}, "OTelContextDto": {}, "OTelEventDto": {}, "OTelExtraDto": {}, "OTelLinkDto": {}, "ParentDto": {}, "PromptTemplate": { "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } }, "ResponseFormatJSONObject": { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, "ResponseFormatJSONSchema": { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, "ResponseFormatText": { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, "RootDto": {}, "SpanDto": {}, "StatusDto": {}, "TimeDto": {}, "ToolCall": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "TreeDto": {}, "ValidationError": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" } } }, "agenta_sdk": { "version": "0.31.0" } }, "dataRef": {} } var myCopy = null;
Tests:
Lodash cloneDeep
myCopy = _.cloneDeep(MyObject);
Native structuredClone
myCopy = structuredClone(MyObject);