|
{ |
|
"name": "Codettes", |
|
"description": "An advanced AI assistant designed to assist users with a wide range of tasks by providing insightful responses.", |
|
"strict": false, |
|
"parameters": { |
|
"type": "object", |
|
"store": true, |
|
"top_p": 1, |
|
"required": [ |
|
"Config" |
|
], |
|
"properties": { |
|
"Config": { |
|
"type": "object", |
|
"required": [ |
|
"max_input_length", |
|
"max_retries", |
|
"model_name", |
|
"perspectives", |
|
"safety_thresholds" |
|
], |
|
"properties": { |
|
"model_name": { |
|
"type": "string", |
|
"description": "The name of the model being used" |
|
}, |
|
"max_retries": { |
|
"type": "number", |
|
"description": "Maximum number of retries for processing requests" |
|
}, |
|
"perspectives": { |
|
"type": "array", |
|
"items": { |
|
"type": "string", |
|
"description": "Different perspectives for cognitive processing" |
|
}, |
|
"description": "Array of perspectives to utilize in processing queries" |
|
}, |
|
"max_input_length": { |
|
"type": "number", |
|
"description": "Maximum length of user input" |
|
}, |
|
"safety_thresholds": { |
|
"type": "object", |
|
"required": [ |
|
"memory", |
|
"cpu", |
|
"response_time" |
|
], |
|
"properties": { |
|
"cpu": { |
|
"type": "number", |
|
"description": "CPU usage threshold percentage" |
|
}, |
|
"memory": { |
|
"type": "number", |
|
"description": "Memory usage threshold percentage" |
|
}, |
|
"response_time": { |
|
"type": "number", |
|
"description": "Maximum acceptable response time in seconds" |
|
} |
|
}, |
|
"additionalProperties": false |
|
} |
|
}, |
|
"systemPrompt": { |
|
"type": "string", |
|
"description": "Initial prompt to set the behavior and capabilities of the AI assistant" |
|
}, |
|
"chatParameters": { |
|
"type": "object", |
|
"required": [ |
|
"deploymentName", |
|
"frequencyPenalty", |
|
"maxResponseLength", |
|
"pastMessagesToInclude", |
|
"presencePenalty", |
|
"temperature", |
|
"topProbablities", |
|
"stopSequences" |
|
], |
|
"properties": { |
|
"temperature": { |
|
"type": "number", |
|
"description": "Sampling temperature controlling randomness in responses" |
|
}, |
|
"stopSequences": { |
|
"type": "array", |
|
"items": { |
|
"type": "string", |
|
"description": "Sequence indicating completion of response" |
|
}, |
|
"description": "List of sequences to stop generating further tokens" |
|
}, |
|
"deploymentName": { |
|
"type": "string", |
|
"description": "Name of the deployment for the AI model" |
|
}, |
|
"presencePenalty": { |
|
"type": "number", |
|
"description": "Penalty applied to promote new topic introduction" |
|
}, |
|
"topProbablities": { |
|
"type": "number", |
|
"description": "Sampling parameter influencing response diversity" |
|
}, |
|
"frequencyPenalty": { |
|
"type": "number", |
|
"description": "Penalty for word repetition" |
|
}, |
|
"maxResponseLength": { |
|
"type": "number", |
|
"description": "Maximum length of the response that the assistant can generate" |
|
}, |
|
"pastMessagesToInclude": { |
|
"type": "string", |
|
"description": "Number of past messages to include in context for generating responses" |
|
} |
|
}, |
|
"additionalProperties": false |
|
}, |
|
"fewShotExamples": { |
|
"type": "array", |
|
"items": { |
|
"type": "object", |
|
"required": [ |
|
"input", |
|
"output" |
|
], |
|
"properties": { |
|
"input": { |
|
"type": "string", |
|
"description": "Input from the user" |
|
}, |
|
"output": { |
|
"type": "string", |
|
"description": "Assistant's response to the user input" |
|
} |
|
}, |
|
"additionalProperties": false |
|
}, |
|
"description": "Examples of interactions to aid in understanding function usage" |
|
}, |
|
"additionalProperties": false |
|
} |
|
}, |
|
"temperature": 1, |
|
"presence_penalty": 0, |
|
"frequency_penalty": 0, |
|
"additionalProperties": false, |
|
"max_completion_tokens": 8728 |
|
} |
|
} |
|
|