Toggle navigation
MeasureThat.net
Create a benchmark
Tools
Feedback
FAQ
Register
Log In
Lodash cloneDeep vs structuredClone v1
(version: 1)
https://developer.mozilla.org/en-US/docs/Web/API/structuredClone
Comparing performance of:
Lodash cloneDeep vs Native structuredClone
Created:
one year ago
by:
Guest
Jump to the latest result
HTML Preparation code:
<script src='https://cdnjs.cloudflare.com/ajax/libs/lodash.js/4.17.5/lodash.min.js'></script>
Script Preparation code:
var MyObject = { "variants": [ { "id": "019461ff-f31c-74b3-981c-dad1d891b8fb", "uri": "http://localhost:80/services/completion", "appId": "019461ff-f226-7a36-a9e1-53c11329a342", "baseId": "019461ff-f316-70de-84d1-14cda41b0250", "baseName": "app", "variantName": "app.key", "revision": 1, "configName": "key", "projectId": "01945fa2-db0d-78dc-9c67-462f129da09e", "appName": "test-arda-1", "parameters": { "agentaConfig": { "prompt": { "messages": [ { "role": "system", "content": "You are an expert in geography" }, { "role": "user", "content": "What is the capital of {country}?" } ], "input_keys": [ "country" ], "llm_config": { "model": "gpt-3.5-turbo", "top_p": 0.5, "max_tokens": 86, "temperature": 1, "frequency_penalty": 0 }, "user_prompt": "What is the capital of {country}?", "system_prompt": "You are an expert in geography", "template_format": "fstring" } } }, "isChat": false, "inputs": { "__id": "4a8e6964-bde9-4ee3-9735-56f332cc3087", "__metadata": { "type": "array", "title": "Input Rows", "description": "Input values for variant execution", "itemMetadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "nullable": false }, "value": [ { "__id": "318dd2f7-5475-4836-b888-9f76798c9370", "__metadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "__runs": {}, "country": { "__id": "b3c181ac-99ca-415e-a1bb-2ff44a0a6025", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true }, "value": "" } } ] }, "messages": {}, "name": "", "prompts": [ { "__id": "6f0105b2-d095-4a0f-ab2f-6508b5ba8686", "__metadata": { "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "properties": { "messages": { "type": "array", "title": "Messages", "itemMetadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "nullable": false }, "systemPrompt": { "type": "string", "title": "System Prompt", "nullable": true, "allowFreeform": true }, "userPrompt": { "type": "string", "title": "User Prompt", "nullable": true, "allowFreeform": true }, "templateFormat": { "type": "string", "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "options": [ { "label": "fstring", "value": "fstring" }, { "label": "jinja2", "value": "jinja2" }, { "label": "curly", "value": "curly" } ], "nullable": false, "allowFreeform": false }, "inputKeys": { "type": "array", "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted", "itemMetadata": { "type": "string", "nullable": false, "allowFreeform": true }, "nullable": true }, "llmConfig": { "type": "object", "title": "ModelConfig", "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false }, "temperature": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true, "isInteger": false }, "maxTokens": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "nullable": true, "isInteger": false }, "topP": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true, "isInteger": false }, "frequencyPenalty": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "presencePenalty": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "responseFormat": { "type": "object", "title": "ResponseFormatText", "description": "An object specifying the format that the model must output", "properties": { "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false } }, "nullable": true }, "stream": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true }, "tools": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "itemMetadata": { "type": "object", "properties": {} }, "nullable": true }, "toolChoice": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true, "allowFreeform": false } }, "nullable": false } } }, "messages": { "__id": "15958cdc-6895-48c8-b3c9-8ea6f3dbfc77", "__metadata": { "type": "array", "title": "Messages", "itemMetadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "nullable": false }, "value": [ { "__id": "c742a423-5b72-460a-8869-3def2d0d74e3", "__metadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "role": { "value": "system", "__id": "21bada93-1a22-45a9-9a62-1a28429da612", "__metadata": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false } }, "content": { "value": "You are an expert in geography", "__id": "859344ad-8450-4a7a-bdde-d57c511b07b5", "__metadata": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true } }, "name": { "value": null, "__id": "3efdfc74-08e4-48c8-9bbd-f5d2251eadbf", "__metadata": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true } }, "toolCalls": { "value": null, "__id": "c39ab97b-6dfd-40e1-876b-f3d416078f79", "__metadata": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true } }, "toolCallId": { "value": null, "__id": "5edd1560-1bd4-48b5-a13e-22e9ab8b5eb1", "__metadata": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, { "__id": "09706047-360b-4f4f-a96e-8cbc3a5a3d79", "__metadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "role": { "value": "user", "__id": "f13cf72e-fcd9-461c-913e-7a301c19beec", "__metadata": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false } }, "content": { "value": "What is the capital of {country}? {a1}", "__id": "2c8cce66-a25b-4c6a-b243-5cc93137d0fb", "__metadata": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true } }, "name": { "value": null, "__id": "d6577bea-c0f6-47b5-90d2-550ddc0b006f", "__metadata": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true } }, "toolCalls": { "value": null, "__id": "ab1c6c74-6ec5-433d-9d08-036c46dde8a7", "__metadata": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true } }, "toolCallId": { "value": null, "__id": "e344acff-152f-4e7b-a94b-14f20db5c81b", "__metadata": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } } ] }, "systemPrompt": { "value": null, "__id": "2161ab5b-fe2f-4586-b0be-c9fef3a43bb6", "__metadata": { "type": "string", "title": "System Prompt", "nullable": true } }, "userPrompt": { "value": null, "__id": "cef972cb-6a36-48c1-94ae-6ff1cbf75676", "__metadata": { "type": "string", "title": "User Prompt", "nullable": true } }, "templateFormat": { "value": "fstring", "__id": "e724bc26-8993-4edf-b760-2b68a8f793be", "__metadata": { "type": "string", "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "options": [ { "label": "fstring", "value": "fstring" }, { "label": "jinja2", "value": "jinja2" }, { "label": "curly", "value": "curly" } ], "nullable": false, "allowFreeform": false } }, "inputKeys": { "value": [ { "value": "country", "__id": "ce097afd-ec37-4ce2-96d6-1fdc563a4abd", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, { "value": "a1", "__id": "d411c317-79a2-4d5a-8447-5462bce97503", "__metadata": { "type": "string", "title": "a1", "nullable": false, "allowFreeform": true } } ], "__id": "c9083345-b2ff-48dd-a7c8-875049761511", "__metadata": { "type": "array", "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted", "nullable": true } }, "llmConfig": { "__id": "061a3297-217b-495d-8655-dc6fd95d6ac6", "__metadata": { "type": "object", "title": "ModelConfig", "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false }, "temperature": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true, "isInteger": false }, "maxTokens": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "nullable": true, "isInteger": false }, "topP": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true, "isInteger": false }, "frequencyPenalty": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "presencePenalty": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "responseFormat": { "type": "object", "title": "ResponseFormatText", "description": "An object specifying the format that the model must output", "properties": { "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false } }, "nullable": true }, "stream": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true }, "tools": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "itemMetadata": { "type": "object", "properties": {} }, "nullable": true }, "toolChoice": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true, "allowFreeform": false } } }, "model": { "value": "gpt-3.5-turbo", "__id": "c2f35d73-6c4f-4d16-8391-cf4dea75c0f5", "__metadata": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false } }, "temperature": { "value": 1, "__id": "a89f8a38-6ea7-49bc-993d-55ae3c11097b", "__metadata": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true } }, "maxTokens": { "value": 86, "__id": "de83333d-c938-4d71-983a-66d582d1cded", "__metadata": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "isInteger": true, "nullable": true } }, "topP": { "value": 0.5, "__id": "6b765570-edee-461e-89f5-2f8bbf02e126", "__metadata": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true } }, "frequencyPenalty": { "value": 0, "__id": "fff2b53b-78fc-4966-aa77-5760725f5327", "__metadata": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true } }, "presencePenalty": { "value": 0, "__id": "b013fe56-0a17-4b12-b514-87fb593bc8d6", "__metadata": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true } }, "responseFormat": { "value": null, "__id": "4843c191-bd33-4bec-b787-dc4fac1619d5", "__metadata": { "type": "compound", "title": "Response Format", "description": "An object specifying the format that the model must output", "nullable": true, "options": [ { "label": "ResponseFormatText", "value": "text", "config": { "type": "text" } }, { "label": "ResponseFormatJSONObject", "value": "json_object", "config": { "type": "json_object" } }, { "label": "ResponseFormatJSONSchema", "value": "json_schema", "config": { "type": "json_schema", "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } } } ] } }, "stream": { "value": null, "__id": "7a12a51c-8c24-4a28-9917-c0d842442347", "__metadata": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true } }, "tools": { "value": null, "__id": "e63c7ea1-231f-435e-8722-489a0a937a60", "__metadata": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "nullable": true } }, "toolChoice": { "value": null, "__id": "8c64f589-bb17-4359-83dc-0f364740fcfa", "__metadata": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true } } } } ] }, { "id": "01946478-6ca0-72bf-88ed-0c803c2f9f4f", "uri": "http://localhost:80/services/completion", "appId": "019461ff-f226-7a36-a9e1-53c11329a342", "baseId": "019461ff-f316-70de-84d1-14cda41b0250", "baseName": "app", "variantName": "app.arda-2", "revision": 5, "configName": "arda-2", "projectId": "01945fa2-db0d-78dc-9c67-462f129da09e", "appName": "test-arda-1", "parameters": { "agentaConfig": { "prompt": { "messages": [ { "role": "system", "content": "You are an expert in geography" }, { "role": "user", "content": "What is the capital of {country}?" } ], "input_keys": [ "country" ], "llm_config": { "model": "gpt-3.5-turbo", "top_p": 0.5, "max_tokens": 84, "temperature": 1, "frequency_penalty": 0 }, "template_format": "fstring" } } }, "isChat": false, "inputs": { "__id": "ff120363-2c11-47a6-95de-0f72828676f8", "__metadata": { "type": "array", "title": "Input Rows", "description": "Input values for variant execution", "itemMetadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "nullable": false }, "value": [ { "__id": "dde5a18c-1113-4bf3-bc35-b0e817f3d529", "__metadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "__runs": {}, "country": { "__id": "eb5189a5-8a51-4be5-8da0-76be00544b38", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true }, "value": "" } } ] }, "messages": {}, "name": "", "prompts": [ { "__id": "39d74b2b-53e2-46bc-8057-f8527836dfaf", "__metadata": { "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "properties": { "messages": { "type": "array", "title": "Messages", "itemMetadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "nullable": false }, "systemPrompt": { "type": "string", "title": "System Prompt", "nullable": true, "allowFreeform": true }, "userPrompt": { "type": "string", "title": "User Prompt", "nullable": true, "allowFreeform": true }, "templateFormat": { "type": "string", "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "options": [ { "label": "fstring", "value": "fstring" }, { "label": "jinja2", "value": "jinja2" }, { "label": "curly", "value": "curly" } ], "nullable": false, "allowFreeform": false }, "inputKeys": { "type": "array", "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted", "itemMetadata": { "type": "string", "nullable": false, "allowFreeform": true }, "nullable": true }, "llmConfig": { "type": "object", "title": "ModelConfig", "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false }, "temperature": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true, "isInteger": false }, "maxTokens": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "nullable": true, "isInteger": false }, "topP": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true, "isInteger": false }, "frequencyPenalty": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "presencePenalty": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "responseFormat": { "type": "object", "title": "ResponseFormatText", "description": "An object specifying the format that the model must output", "properties": { "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false } }, "nullable": true }, "stream": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true }, "tools": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "itemMetadata": { "type": "object", "properties": {} }, "nullable": true }, "toolChoice": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true, "allowFreeform": false } }, "nullable": false } } }, "messages": { "__id": "b9c9d862-3326-48ad-8549-379df5b7ccc7", "__metadata": { "type": "array", "title": "Messages", "itemMetadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "nullable": false }, "value": [ { "__id": "ee434a01-d7f3-494a-8584-370bf9e2f1fb", "__metadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "role": { "value": "system", "__id": "025033e7-9279-4e75-a949-b7f6791bfc51", "__metadata": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false } }, "content": { "value": "You are an expert in geography", "__id": "f44b4ef5-7f4f-4684-a9ff-7db90d481b01", "__metadata": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true } }, "name": { "value": null, "__id": "43b9a314-49c3-4d7d-aa38-44481a5586a8", "__metadata": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true } }, "toolCalls": { "value": null, "__id": "0f478f7e-080b-4bec-a965-c2fb52f3650e", "__metadata": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true } }, "toolCallId": { "value": null, "__id": "2f8662a1-7334-4692-80fd-2b1597506e48", "__metadata": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, { "__id": "d5d204e6-23fa-4b45-98a7-5b289a65262e", "__metadata": { "type": "object", "title": "Message", "properties": { "role": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false }, "content": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true }, "name": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true }, "toolCalls": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true }, "toolCallId": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } }, "role": { "value": "user", "__id": "9eebfcc9-9cb9-4630-a9d6-07659d3cf3a3", "__metadata": { "type": "string", "title": "Role", "options": [ { "label": "system", "value": "system" }, { "label": "user", "value": "user" }, { "label": "assistant", "value": "assistant" }, { "label": "tool", "value": "tool" }, { "label": "function", "value": "function" } ], "nullable": false, "allowFreeform": false } }, "content": { "value": "What is the capital of {country}?", "__id": "ae3c7fe6-a992-49ae-acde-232706e1c533", "__metadata": { "type": "string", "title": "Content", "nullable": true, "allowFreeform": true } }, "name": { "value": null, "__id": "052bf3ee-d78c-481c-a08c-f98600557ad3", "__metadata": { "type": "string", "title": "Name", "nullable": true, "allowFreeform": true } }, "toolCalls": { "value": null, "__id": "60a38c86-febe-482b-aa26-f73ea773d70a", "__metadata": { "type": "array", "title": "Tool Calls", "itemMetadata": { "type": "object", "title": "ToolCall", "properties": { "id": { "type": "string", "title": "Id", "nullable": false, "allowFreeform": true }, "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false }, "function": { "type": "object", "title": "Function", "properties": {}, "additionalProperties": true, "nullable": false } } }, "nullable": true } }, "toolCallId": { "value": null, "__id": "ab653a63-ba2d-4a17-bea6-6d21a4796c57", "__metadata": { "type": "string", "title": "Tool Call Id", "nullable": true, "allowFreeform": true } } } ] }, "systemPrompt": { "value": null, "__id": "1b677c97-e3f6-43b4-93cd-d758ba52af82", "__metadata": { "type": "string", "title": "System Prompt", "nullable": true } }, "userPrompt": { "value": null, "__id": "4bb5f46c-8e56-4a84-8734-256f86d8b59f", "__metadata": { "type": "string", "title": "User Prompt", "nullable": true } }, "templateFormat": { "value": "fstring", "__id": "b1bdfb1a-3f66-4bab-a2a2-09f1f8adaa4b", "__metadata": { "type": "string", "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "options": [ { "label": "fstring", "value": "fstring" }, { "label": "jinja2", "value": "jinja2" }, { "label": "curly", "value": "curly" } ], "nullable": false, "allowFreeform": false } }, "inputKeys": { "value": [ { "value": "country", "__id": "10ee0960-60e9-4a17-8f08-bb7222a8ef1e", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } } ], "__id": "2fbc73b0-09ad-4458-a27d-285b34d150a6", "__metadata": { "type": "array", "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted", "nullable": true } }, "llmConfig": { "__id": "6ed1b3ff-9865-4658-a58e-3af989d6a8c3", "__metadata": { "type": "object", "title": "ModelConfig", "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false }, "temperature": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true, "isInteger": false }, "maxTokens": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "nullable": true, "isInteger": false }, "topP": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true, "isInteger": false }, "frequencyPenalty": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "presencePenalty": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true, "isInteger": false }, "responseFormat": { "type": "object", "title": "ResponseFormatText", "description": "An object specifying the format that the model must output", "properties": { "type": { "type": "string", "title": "Type", "nullable": false, "allowFreeform": false } }, "nullable": true }, "stream": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true }, "tools": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "itemMetadata": { "type": "object", "properties": {} }, "nullable": true }, "toolChoice": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true, "allowFreeform": false } } }, "model": { "value": "gpt-3.5-turbo", "__id": "2643d1ad-b6a1-416b-b769-5c3add32e6a6", "__metadata": { "type": "string", "title": "Model", "description": "ID of the model to use", "options": [ { "label": "Aleph Alpha", "options": [ { "label": "luminous-base", "value": "luminous-base", "group": "Aleph Alpha" }, { "label": "luminous-base-control", "value": "luminous-base-control", "group": "Aleph Alpha" }, { "label": "luminous-extended-control", "value": "luminous-extended-control", "group": "Aleph Alpha" }, { "label": "luminous-supreme", "value": "luminous-supreme", "group": "Aleph Alpha" } ] }, { "label": "Anthropic", "options": [ { "label": "anthropic/claude-3-5-sonnet-20240620", "value": "anthropic/claude-3-5-sonnet-20240620", "group": "Anthropic" }, { "label": "anthropic/claude-3-opus-20240229", "value": "anthropic/claude-3-opus-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-sonnet-20240229", "value": "anthropic/claude-3-sonnet-20240229", "group": "Anthropic" }, { "label": "anthropic/claude-3-haiku-20240307", "value": "anthropic/claude-3-haiku-20240307", "group": "Anthropic" }, { "label": "anthropic/claude-2.1", "value": "anthropic/claude-2.1", "group": "Anthropic" }, { "label": "anthropic/claude-2", "value": "anthropic/claude-2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1.2", "value": "anthropic/claude-instant-1.2", "group": "Anthropic" }, { "label": "anthropic/claude-instant-1", "value": "anthropic/claude-instant-1", "group": "Anthropic" } ] }, { "label": "Anyscale", "options": [ { "label": "anyscale/meta-llama/Llama-2-13b-chat-hf", "value": "anyscale/meta-llama/Llama-2-13b-chat-hf", "group": "Anyscale" }, { "label": "anyscale/meta-llama/Llama-2-70b-chat-hf", "value": "anyscale/meta-llama/Llama-2-70b-chat-hf", "group": "Anyscale" } ] }, { "label": "Cohere", "options": [ { "label": "cohere/command-light", "value": "cohere/command-light", "group": "Cohere" }, { "label": "cohere/command-r-plus", "value": "cohere/command-r-plus", "group": "Cohere" }, { "label": "cohere/command-nightly", "value": "cohere/command-nightly", "group": "Cohere" } ] }, { "label": "DeepInfra", "options": [ { "label": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-70b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "value": "deepinfra/meta-llama/Llama-2-13b-chat-hf", "group": "DeepInfra" }, { "label": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "value": "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "group": "DeepInfra" }, { "label": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "value": "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "group": "DeepInfra" }, { "label": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "value": "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1", "group": "DeepInfra" } ] }, { "label": "Gemini", "options": [ { "label": "gemini/gemini-1.5-pro-latest", "value": "gemini/gemini-1.5-pro-latest", "group": "Gemini" }, { "label": "gemini/gemini-1.5-flash", "value": "gemini/gemini-1.5-flash", "group": "Gemini" } ] }, { "label": "Groq", "options": [ { "label": "groq/llama3-8b-8192", "value": "groq/llama3-8b-8192", "group": "Groq" }, { "label": "groq/llama3-70b-8192", "value": "groq/llama3-70b-8192", "group": "Groq" }, { "label": "groq/llama2-70b-4096", "value": "groq/llama2-70b-4096", "group": "Groq" }, { "label": "groq/mixtral-8x7b-32768", "value": "groq/mixtral-8x7b-32768", "group": "Groq" }, { "label": "groq/gemma-7b-it", "value": "groq/gemma-7b-it", "group": "Groq" } ] }, { "label": "Mistral", "options": [ { "label": "mistral/mistral-tiny", "value": "mistral/mistral-tiny", "group": "Mistral" }, { "label": "mistral/mistral-small", "value": "mistral/mistral-small", "group": "Mistral" }, { "label": "mistral/mistral-medium", "value": "mistral/mistral-medium", "group": "Mistral" }, { "label": "mistral/mistral-large-latest", "value": "mistral/mistral-large-latest", "group": "Mistral" } ] }, { "label": "Open AI", "options": [ { "label": "gpt-3.5-turbo-1106", "value": "gpt-3.5-turbo-1106", "group": "Open AI" }, { "label": "gpt-3.5-turbo", "value": "gpt-3.5-turbo", "group": "Open AI" }, { "label": "gpt-4", "value": "gpt-4", "group": "Open AI" }, { "label": "gpt-4o", "value": "gpt-4o", "group": "Open AI" }, { "label": "gpt-4o-mini", "value": "gpt-4o-mini", "group": "Open AI" }, { "label": "gpt-4-1106-preview", "value": "gpt-4-1106-preview", "group": "Open AI" } ] }, { "label": "OpenRouter", "options": [ { "label": "openrouter/openai/gpt-3.5-turbo", "value": "openrouter/openai/gpt-3.5-turbo", "group": "OpenRouter" }, { "label": "openrouter/openai/gpt-3.5-turbo-16k", "value": "openrouter/openai/gpt-3.5-turbo-16k", "group": "OpenRouter" }, { "label": "openrouter/anthropic/claude-instant-v1", "value": "openrouter/anthropic/claude-instant-v1", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-chat-bison", "value": "openrouter/google/palm-2-chat-bison", "group": "OpenRouter" }, { "label": "openrouter/google/palm-2-codechat-bison", "value": "openrouter/google/palm-2-codechat-bison", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-13b-chat", "value": "openrouter/meta-llama/llama-2-13b-chat", "group": "OpenRouter" }, { "label": "openrouter/meta-llama/llama-2-70b-chat", "value": "openrouter/meta-llama/llama-2-70b-chat", "group": "OpenRouter" } ] }, { "label": "Perplexity AI", "options": [ { "label": "perplexity/pplx-7b-chat", "value": "perplexity/pplx-7b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-chat", "value": "perplexity/pplx-70b-chat", "group": "Perplexity AI" }, { "label": "perplexity/pplx-7b-online", "value": "perplexity/pplx-7b-online", "group": "Perplexity AI" }, { "label": "perplexity/pplx-70b-online", "value": "perplexity/pplx-70b-online", "group": "Perplexity AI" } ] }, { "label": "Together AI", "options": [ { "label": "together_ai/togethercomputer/llama-2-70b-chat", "value": "together_ai/togethercomputer/llama-2-70b-chat", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-70b", "value": "together_ai/togethercomputer/llama-2-70b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/LLaMA-2-7B-32K", "value": "together_ai/togethercomputer/LLaMA-2-7B-32K", "group": "Together AI" }, { "label": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "value": "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/llama-2-7b", "value": "together_ai/togethercomputer/llama-2-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/alpaca-7b", "value": "together_ai/togethercomputer/alpaca-7b", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "value": "together_ai/togethercomputer/CodeLlama-34b-Instruct", "group": "Together AI" }, { "label": "together_ai/togethercomputer/CodeLlama-34b-Python", "value": "together_ai/togethercomputer/CodeLlama-34b-Python", "group": "Together AI" }, { "label": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "value": "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "group": "Together AI" }, { "label": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "value": "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "group": "Together AI" }, { "label": "together_ai/Austism/chronos-hermes-13b", "value": "together_ai/Austism/chronos-hermes-13b", "group": "Together AI" } ] } ], "nullable": false, "allowFreeform": false } }, "temperature": { "value": 1, "__id": "db5ad2fa-6e64-4da7-a4e1-4b06b4e8e22c", "__metadata": { "type": "number", "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "min": 0, "max": 2, "nullable": true } }, "maxTokens": { "value": 84, "__id": "fe1f6aec-6d91-43cc-b0aa-6381b7759748", "__metadata": { "type": "number", "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "min": 0, "isInteger": true, "nullable": true } }, "topP": { "value": 0.5, "__id": "dde89181-bdda-451a-8dae-c350ebaa5b87", "__metadata": { "type": "number", "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "min": 0, "max": 1, "nullable": true } }, "frequencyPenalty": { "value": 0, "__id": "7d63570a-dbcc-466c-8126-7cb1a143122f", "__metadata": { "type": "number", "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "min": -2, "max": 2, "nullable": true } }, "presencePenalty": { "value": 0, "__id": "eb50375c-6e62-41c5-8da0-8c5f69a4bde9", "__metadata": { "type": "number", "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "min": -2, "max": 2, "nullable": true } }, "responseFormat": { "value": null, "__id": "bbdd8e13-f0ac-416c-8c5f-fbdf0b70fbd0", "__metadata": { "type": "compound", "title": "Response Format", "description": "An object specifying the format that the model must output", "nullable": true, "options": [ { "label": "ResponseFormatText", "value": "text", "config": { "type": "text" } }, { "label": "ResponseFormatJSONObject", "value": "json_object", "config": { "type": "json_object" } }, { "label": "ResponseFormatJSONSchema", "value": "json_schema", "config": { "type": "json_schema", "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } } } ] } }, "stream": { "value": null, "__id": "d8c349a3-91f0-454b-936f-a2aca998980d", "__metadata": { "type": "boolean", "title": "Stream", "description": "If set, partial message deltas will be sent", "nullable": true } }, "tools": { "value": null, "__id": "ac6b5383-423f-4b0e-a53b-633ad4f2ad54", "__metadata": { "type": "array", "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool", "nullable": true } }, "toolChoice": { "value": null, "__id": "d5381a5e-e988-4120-90a4-044e69c30964", "__metadata": { "type": "string", "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model", "options": [ { "label": "none", "value": "none" }, { "label": "auto", "value": "auto" } ], "nullable": true } } } } ] } ], "selected": [ "019461ff-f31c-74b3-981c-dad1d891b8fb" ], "dirtyStates": {}, "generationData": { "__id": "f01fc583-3359-49c9-b871-5d6032af2995", "__metadata": { "type": "array", "title": "Input Rows", "description": "Input values for variant execution", "itemMetadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "nullable": false }, "value": [ { "__id": "fb5bfef1-be48-4469-9a0a-3a8614947334", "__metadata": { "type": "object", "title": "Input Row", "description": "Single set of input values", "properties": { "country": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true } }, "nullable": false }, "__runs": {}, "country": { "__id": "690d7666-2349-4158-b238-a6115505a7a2", "__metadata": { "type": "string", "title": "country", "nullable": false, "allowFreeform": true }, "value": "" } } ] }, "spec": { "openapi": "3.1.0", "info": { "title": "FastAPI", "version": "0.1.0" }, "paths": { "/health": { "get": { "summary": "Health", "operationId": "health_health_get", "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": {} } } } } } }, "/run": { "post": { "summary": "Generate", "operationId": "generate_run_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_run_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } }, "/generate_deployed": { "post": { "summary": "Generate", "operationId": "generate_generate_deployed_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_generate_deployed_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } }, "/test": { "post": { "summary": "Generate", "operationId": "generate_test_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_test_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } }, "/generate": { "post": { "summary": "Generate", "operationId": "generate_generate_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_generate_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } }, "/playground/run": { "post": { "summary": "Generate", "operationId": "generate_playground_run_post", "requestBody": { "content": { "application/json": { "schema": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_playground_run_post" } } }, "required": true }, "responses": { "200": { "description": "Successful Response", "content": { "application/json": { "schema": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" } } } }, "422": { "description": "Validation Error", "content": { "application/json": { "schema": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" } } } } } } } }, "components": { "schemas": { "AgentaNodesResponse": {}, "BaseResponse": { "properties": { "version": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Version", "default": "3.1" }, "data": { "anyOf": [ { "type": "string" }, { "type": "object" }, { "type": "null" } ], "title": "Data" }, "content_type": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content Type", "default": "string" }, "tree": { "anyOf": [ {}, { "type": "null" } ] }, "tree_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tree Id" } }, "type": "object", "title": "BaseResponse" }, "Body_generate_generate_deployed_post": { "properties": { "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_generate_deployed_post" }, "Body_generate_generate_post": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_generate_post" }, "Body_generate_playground_run_post": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_playground_run_post" }, "Body_generate_run_post": { "properties": { "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_run_post" }, "Body_generate_test_post": { "properties": { "ag_config": { "default": { "prompt": { "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 }, "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "template_format": "fstring", "user_prompt": "What is the capital of {country}?" } }, "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "inputs": { "title": "Inputs" } }, "type": "object", "required": [ "inputs" ], "title": "Body_generate_test_post" }, "ExceptionDto": {}, "HTTPValidationError": { "properties": { "detail": { "items": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" }, "type": "array", "title": "Detail" } }, "type": "object", "title": "HTTPValidationError" }, "JSONSchema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" }, "LifecycleDto": {}, "LinkDto": {}, "Message": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "ModelConfig": { "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig", "description": "Configuration for model parameters" }, "MyConfig": { "properties": { "prompt": { "default": { "messages": [ { "content": "You are an expert in geography", "role": "system" }, { "content": "What is the capital of {country}?", "role": "user" } ], "system_prompt": "You are an expert in geography", "user_prompt": "What is the capital of {country}?", "template_format": "fstring", "llm_config": { "frequency_penalty": 0, "max_tokens": -1, "model": "gpt-3.5-turbo", "presence_penalty": 0, "temperature": 1, "top_p": 0.5 } }, "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } } }, "type": "object", "title": "MyConfig" }, "NodeDto": {}, "OTelContextDto": {}, "OTelEventDto": {}, "OTelExtraDto": {}, "OTelLinkDto": {}, "ParentDto": {}, "PromptTemplate": { "properties": { "messages": { "items": { "properties": { "role": { "type": "string", "enum": [ "system", "user", "assistant", "tool", "function" ], "title": "Role" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Content" }, "name": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Name" }, "tool_calls": { "anyOf": [ { "items": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "type": "array" }, { "type": "null" } ], "title": "Tool Calls" }, "tool_call_id": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Tool Call Id" } }, "type": "object", "required": [ "role" ], "title": "Message" }, "type": "array", "title": "Messages", "default": [ { "role": "system", "content": "" }, { "role": "user", "content": "" } ] }, "system_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "System Prompt" }, "user_prompt": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "User Prompt" }, "template_format": { "type": "string", "enum": [ "fstring", "jinja2", "curly" ], "title": "Template Format", "description": "Format type for template variables: fstring {var}, jinja2 {{ var }}, or curly {{var}}", "default": "fstring" }, "input_keys": { "anyOf": [ { "items": { "type": "string" }, "type": "array" }, { "type": "null" } ], "title": "Input Keys", "description": "Optional list of input keys for validation. If not provided, any inputs will be accepted" }, "llm_config": { "description": "Configuration for the model parameters", "properties": { "model": { "type": "string", "title": "Model", "description": "ID of the model to use", "default": "gpt-3.5-turbo", "choices": { "Aleph Alpha": [ "luminous-base", "luminous-base-control", "luminous-extended-control", "luminous-supreme" ], "Anthropic": [ "anthropic/claude-3-5-sonnet-20240620", "anthropic/claude-3-opus-20240229", "anthropic/claude-3-sonnet-20240229", "anthropic/claude-3-haiku-20240307", "anthropic/claude-2.1", "anthropic/claude-2", "anthropic/claude-instant-1.2", "anthropic/claude-instant-1" ], "Anyscale": [ "anyscale/meta-llama/Llama-2-13b-chat-hf", "anyscale/meta-llama/Llama-2-70b-chat-hf" ], "Cohere": [ "cohere/command-light", "cohere/command-r-plus", "cohere/command-nightly" ], "DeepInfra": [ "deepinfra/meta-llama/Llama-2-70b-chat-hf", "deepinfra/meta-llama/Llama-2-13b-chat-hf", "deepinfra/codellama/CodeLlama-34b-Instruct-hf", "deepinfra/mistralai/Mistral-7B-Instruct-v0.1", "deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1" ], "Gemini": [ "gemini/gemini-1.5-pro-latest", "gemini/gemini-1.5-flash" ], "Groq": [ "groq/llama3-8b-8192", "groq/llama3-70b-8192", "groq/llama2-70b-4096", "groq/mixtral-8x7b-32768", "groq/gemma-7b-it" ], "Mistral": [ "mistral/mistral-tiny", "mistral/mistral-small", "mistral/mistral-medium", "mistral/mistral-large-latest" ], "Open AI": [ "gpt-3.5-turbo-1106", "gpt-3.5-turbo", "gpt-4", "gpt-4o", "gpt-4o-mini", "gpt-4-1106-preview" ], "OpenRouter": [ "openrouter/openai/gpt-3.5-turbo", "openrouter/openai/gpt-3.5-turbo-16k", "openrouter/anthropic/claude-instant-v1", "openrouter/google/palm-2-chat-bison", "openrouter/google/palm-2-codechat-bison", "openrouter/meta-llama/llama-2-13b-chat", "openrouter/meta-llama/llama-2-70b-chat" ], "Perplexity AI": [ "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "perplexity/pplx-7b-online", "perplexity/pplx-70b-online" ], "Together AI": [ "together_ai/togethercomputer/llama-2-70b-chat", "together_ai/togethercomputer/llama-2-70b", "together_ai/togethercomputer/LLaMA-2-7B-32K", "together_ai/togethercomputer/Llama-2-7B-32K-Instruct", "together_ai/togethercomputer/llama-2-7b", "together_ai/togethercomputer/alpaca-7b", "together_ai/togethercomputer/CodeLlama-34b-Instruct", "together_ai/togethercomputer/CodeLlama-34b-Python", "together_ai/WizardLM/WizardCoder-Python-34B-V1.0", "together_ai/NousResearch/Nous-Hermes-Llama2-13b", "together_ai/Austism/chronos-hermes-13b" ] }, "x-parameter": "grouped_choice" }, "temperature": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": 0 }, { "type": "null" } ], "title": "Temperature", "description": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic", "default": 1 }, "max_tokens": { "anyOf": [ { "type": "integer", "minimum": 0 }, { "type": "null" } ], "title": "Max Tokens", "description": "The maximum number of tokens that can be generated in the chat completion", "default": -1 }, "top_p": { "anyOf": [ { "type": "number", "maximum": 1, "minimum": 0 }, { "type": "null" } ], "title": "Top P", "description": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass", "default": 0.5 }, "frequency_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Frequency Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far", "default": 0 }, "presence_penalty": { "anyOf": [ { "type": "number", "maximum": 2, "minimum": -2 }, { "type": "null" } ], "title": "Presence Penalty", "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far", "default": 0 }, "response_format": { "anyOf": [ { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, { "type": "null" } ], "title": "Response Format", "description": "An object specifying the format that the model must output" }, "stream": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Stream", "description": "If set, partial message deltas will be sent" }, "tools": { "anyOf": [ { "items": { "type": "object" }, "type": "array" }, { "type": "null" } ], "title": "Tools", "description": "A list of tools the model may call. Currently, only functions are supported as a tool" }, "tool_choice": { "anyOf": [ { "type": "string", "enum": [ "none", "auto" ] }, { "type": "object" }, { "type": "null" } ], "title": "Tool Choice", "description": "Controls which (if any) tool is called by the model" } }, "type": "object", "title": "ModelConfig" } }, "type": "object", "title": "PromptTemplate", "description": "A template for generating prompts with formatting capabilities", "x-parameters": { "prompt": "true" } }, "ResponseFormatJSONObject": { "properties": { "type": { "type": "string", "const": "json_object", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatJSONObject" }, "ResponseFormatJSONSchema": { "properties": { "type": { "type": "string", "const": "json_schema", "title": "Type" }, "json_schema": { "properties": { "name": { "type": "string", "title": "Name" }, "description": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "title": "Description" }, "schema": { "anyOf": [ { "type": "object" }, { "type": "null" } ], "title": "Schema" }, "strict": { "anyOf": [ { "type": "boolean" }, { "type": "null" } ], "title": "Strict" } }, "type": "object", "required": [ "name", "schema" ], "title": "JSONSchema" } }, "type": "object", "required": [ "type", "json_schema" ], "title": "ResponseFormatJSONSchema" }, "ResponseFormatText": { "properties": { "type": { "type": "string", "const": "text", "title": "Type" } }, "type": "object", "required": [ "type" ], "title": "ResponseFormatText" }, "RootDto": {}, "SpanDto": {}, "StatusDto": {}, "TimeDto": {}, "ToolCall": { "properties": { "id": { "type": "string", "title": "Id" }, "type": { "type": "string", "const": "function", "title": "Type", "default": "function" }, "function": { "additionalProperties": { "type": "string" }, "type": "object", "title": "Function" } }, "type": "object", "required": [ "id", "function" ], "title": "ToolCall" }, "TreeDto": {}, "ValidationError": { "properties": { "loc": { "items": { "anyOf": [ { "type": "string" }, { "type": "integer" } ] }, "type": "array", "title": "Location" }, "msg": { "type": "string", "title": "Message" }, "type": { "type": "string", "title": "Error Type" } }, "type": "object", "required": [ "loc", "msg", "type" ], "title": "ValidationError" } } }, "agenta_sdk": { "version": "0.31.0" } }, "dataRef": {} } var myCopy = null;
Tests:
Lodash cloneDeep
myCopy = _.cloneDeep(MyObject);
Native structuredClone
myCopy = structuredClone(MyObject);
Rendered benchmark preparation results:
Suite status:
<idle, ready to run>
Run tests (2)
Previous results
Fork
Test case name
Result
Lodash cloneDeep
Native structuredClone
Fastest:
N/A
Slowest:
N/A
Latest run results:
Run details:
(Test run date:
one year ago
)
User agent:
Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/131.0.0.0 Safari/537.36
Browser/OS:
Chrome 131 on Mac OS X 10.15.7
View result in a separate tab
Embed
Embed Benchmark Result
Test name
Executions per second
Lodash cloneDeep
1196.3 Ops/sec
Native structuredClone
1654.7 Ops/sec
Related benchmarks:
Lodash cloneDeep vs JSON Clone 7545236934650044
Lodash cloneDeep vs structuredClone BIG
Lodash cloneDeep vs structuredClone vs JSON Parse large set 2
Lodash cloneDeep vs structuredClone (specific use case)
Lodash cloneDeep vs structuredClone with large object
Lodash cloneDeep vs structuredClone vs JSON.stringify (big object)
JSON.stringify vs structuredClone test
comparision in all deep clone
Lodash cloneDeep vs structuredClone 300KB
Comments
Confirm delete:
Do you really want to delete benchmark?