Executes a deployed Prompt and returns the result.
The ID of the Prompt Deployment. Must provide either this or prompt_deployment_name.
The name of the Prompt Deployment. Must provide either this or prompt_deployment_id.
Optionally specify a release tag if you want to pin to a specific release of the Prompt Deployment
Executes a deployed Workflow and streams back its results.
The ID of the Workflow Deployment. Must provide either this or workflow_deployment_name.
The name of the Workflow Deployment. Must provide either this or workflow_deployment_id.
Optionally specify a release tag if you want to pin to a specific release of the Workflow Deployment
The list of inputs defined in the Workflow's deployment with their corresponding values.
Generate a completion using a previously defined deployment.
Note: Uses a base url of https://predict.vellum.ai
.
The name of the deployment. Must provide either this or deployment_id.
The generation request to make. Bulk requests are no longer supported, this field must be an array of length 1.
Generate a stream of completions using a previously defined deployment.
Note: Uses a base url of https://predict.vellum.ai
.
The name of the deployment. Must provide either this or deployment_id.
The generation request to make. Bulk requests are no longer supported, this field must be an array of length 1.
Perform a search against a document index.
Note: Uses a base url of https://predict.vellum.ai
.
The ID of the index to search against. Must provide either this or index_name.
The name of the index to search against. Must provide either this or index_id.
Used to submit feedback regarding the quality of previously generated completions.
Note: Uses a base url of https://predict.vellum.ai
.
The name of the deployment. Must provide either this or deployment_id.
Used to submit feedback regarding the quality of previous workflow execution and its outputs.
**Note:** Uses a base url of `https://predict.vellum.ai`.
Feedback regarding the quality of an output on a previously executed workflow.
The Vellum-generated ID of a previously executed workflow. Must provide either this or external_id.
Used to retrieve a deployment given its ID or name.
The current status of the deployment
ACTIVE
- ActiveINACTIVE
- InactiveARCHIVED
- Archived
The environment this deployment is used in
DEVELOPMENT
- DevelopmentSTAGING
- StagingPRODUCTION
- Production
Creates a new document index.
The current status of the document index
ACTIVE
- ActiveARCHIVED
- Archived
The environment this document index is used in
DEVELOPMENT
- DevelopmentSTAGING
- StagingPRODUCTION
- Production
The current status of the document index
ACTIVE
- ActiveARCHIVED
- Archived
The environment this document index is used in
DEVELOPMENT
- DevelopmentSTAGING
- StagingPRODUCTION
- Production
Used to retrieve a Document Index given its ID or name.
The current status of the document index
ACTIVE
- ActiveARCHIVED
- Archived
The environment this document index is used in
DEVELOPMENT
- DevelopmentSTAGING
- StagingPRODUCTION
- Production
Used to list documents. Optionally filter on supported fields.
Update a Document, keying off of its Vellum-generated ID. Particularly useful for updating its metadata.
A human-readable label for the document. Defaults to the originally uploaded file's file name.
A human-readable label for the document. Defaults to the originally uploaded file's file name.
The current processing state of the document
QUEUED
- QueuedPROCESSING
- ProcessingPROCESSED
- ProcessedFAILED
- Failed
Used to retrieve a model version given its ID.
Which LLM provider this model version is associated with.
ANTHROPIC
- AnthropicAWS_BEDROCK
- AWS BedrockAZURE_OPENAI
- Azure OpenAICOHERE
- CohereGOOGLE
- GoogleHOSTED
- HostedMOSAICML
- MosaicMLOPENAI
- OpenAIHUGGINGFACE
- HuggingFaceMYSTIC
- MysticPYQ
- PyqREPLICATE
- Replicate
Registers a prompt within Vellum and creates associated Vellum entities. Intended to be used by integration partners, not directly by Vellum users.
Under the hood, this endpoint creates a new sandbox, a new model version, and a new deployment.
The initial LLM provider to use for this prompt
ANTHROPIC
- AnthropicAWS_BEDROCK
- AWS BedrockAZURE_OPENAI
- Azure OpenAICOHERE
- CohereGOOGLE
- GoogleHOSTED
- HostedMOSAICML
- MosaicMLOPENAI
- OpenAIHUGGINGFACE
- HuggingFaceMYSTIC
- MysticPYQ
- PyqREPLICATE
- Replicate
Upserts a new scenario for a sandbox, keying off of the optionally provided scenario id.
If an id is provided and has a match, the scenario will be updated. If no id is provided or no match is found, a new scenario will be appended to the end.
Note that a full replacement of the scenario is performed, so any fields not provided will be removed or overwritten with default values.
Upserts a new test case for a test suite, keying off of the optionally provided test case id.
If an id is provided and has a match, the test case will be updated. If no id is provided or no match is found, a new test case will be appended to the end.
Note that a full replacement of the test case is performed, so any fields not provided will be removed or overwritten with default values.
Deletes an existing test case for a test suite, keying off of the test case id.