Changelog | March, 2025

New Workflow Deployment Executions APIs

March 26th, 2025

We now have two new apis List Workflow Deployment Executions and Retrieve Workflow Deployment Execution for listing your Executions for a specific Workflow Deployment, and retrieving any specific Workflow Deployment Execution.

Support for Gemini 2.5 Pro Model

March 25th, 2025

We’ve added support for Gemini 2.5 Pro Experiemental (Version 03-25) with a whopping 1M input token context window and 64k output tokens via Google’s Gemini API

Select all rows in Evaluation Reports

March 25th, 2025

You can now select all rows on a page in the Evaluation Report by clicking the checkbox in the header. This selection persists across pages, with the checkbox applying only to the current page.

Workflow Execution Detail Cost Breakdown

March 24th, 2025

We are now displaying the aggregated cost for a Workflow Execution’s inside of our Workflow Execution’s Detail Page. This aggregation is available for all Workflow spans within your Workflow Execution, giving a holistic view of your total execution cost.

Cost Breakdown

Deployment Release Reviews

Inspired by Github PR reviews, you can now provide reviews on a Prompt or Workflow Deployment Release after it has been deployed. This is helpful if your company mandates a review process for all changes that make it to production.

To see Release Reviews in action, check out the demo video below:

Support for Llama 3.3 70B via Cerebras

March 21st, 2025

We’ve added support for Llama 3.3 70B via Cerebras AI

New Webhook and Datadog Events

March 20th, 2025

We are releasing 2 new events for Webhooks and Datadog integrations:

  • workflow.execution.initiated
  • workflow.execution.fulfilled

These events can be used for better visibility in the execution of a Workflow, such as retrieving the input values used to kick it off, as well as the output values that it produced. You might also use these events to track latency, data drift, and more.

Retry and Try Node adornments

March 22nd, 2025

Error handling and retry logic have both been traditionally quite cumbersome in Vellum. Often, you want to catch errors around a single node and “adorn” it with different error handling behavior. Today we are releasing Retry and Try Node Adornments.

SDK Node Adornments

Adornments are accessible in the Node sidepanel after clicking on a Node. They act as Nodes themselves, that reference the node it wraps as a single node Subworkflow:

  • Retry Node Adornments repeatedly invokes the Node until it either succeeds or the max number of attempts are hit.
  • Try Node Adornments attempts to invoke the Node, and continues with an Error output if it fails.

Monitoring views for Node Adornment invocations show as if the targeted Node was invoked as a single node Subworkflow:

Node Adornment Monitoring

Monitoring View Overhaul

March 22nd, 2025

We’re excited to introduce completely revamped Monitoring interfaces for Prompts and Workflows Deployments that bring significant improvements to how you visualize and analyze your AI system performance.

Revamped Monitoring View

Our new Grafana-based implementation loads data faster and handles metric processing more efficiently, making it easier than ever to monitor your AI applications at scale.

Key improvements include:

  • Faster dashboard load times
  • Improved date range and Release Tag selection
  • Zoom in on specific time ranges

Copy/Paste Prompt Variants

March 20th, 2025

It is now possible to copy a JSON representation of a Vellum prompt to your clipboard and then paste over another Prompt Variant. This is especially useful if you want to go back to a prior history state, copy just one Prompt, and then paste it in your current live draft.

Copy/Paste Prompt Variants

Updated Workflow Node Inputs

March 20th, 2025

Workflow Nodes used to use a chunky UI for configuring their inputs. We’ve now updated this UI to be simpler by making use of what we call “Expression Inputs.” These Node inputs are functionally the same as before, but tease soon-to-come functionality where you’ll be able to define more complex expressions for what gets passed to a Node.

Workflow Node Expression Inputs

Metric Setup Improvements

March 20th, 2025

We’ve significantly simplified the process of configuring and editing Evaluation Metrics in Vellum. You can now manage your evaluation metrics directly within an evaluation report – adding, editing, and removing metrics – all without leaving the Prompt/Workflow that you’re evaluating.

Add a Metric

You’ll also find a new Metric Settings button that opens a modal where you can configure your metrics directly

Metric Settings

Datadog Integration

March 20th, 2025

It’s now possible to receive real-time updates about actions taking place in Vellum using Datadog.

Datadog

From the organization settings page, you can configure a Datadog integration with a custom list of events you care about.

This is useful if your organization already uses Datadog, or you’d like to leverage Datadog’s monitoring, alerting, and BI capabilities using your Vellum data.

All New Workflows are SDK-Enabled

March 20th, 2025

About two months ago, we began the Beta period of SDK-enabled Workflows. These Workflows use the new Workflows SDK as the underlying engine and run these Workflows in a, secure, more performant, isolated environment. The Workflows SDK enables exciting new functionality including custom nodes, custom docker runtimes, new expression inputs, Node adornments, and so much more.

As of today, all new Workflows going forward will be SDK-enabled by default. We expect that all features from the old Workflows engine are supported, with the exception of “Run from Node”, which we hope to reenable later this month. To revert a Workflow that is SDK-enabled back to the legacy Workflows engine, simply toggle off this checkbox from the Workflow settings:

SDK Compatible Workflow

If you have an existing Workflow and would like to make it SDK-enabled, you can use the same toggle.

JSON Outputs for Prompt Nodes

March 19th, 2025

Until now in SDK enabled workflows, you only had string or array outputs for Prompt Nodes. With this update, you will now be able to reference a Prompt Node’s JSON output if the json_mode or json_schema parameter is enabled. This is useful as it enables you to not have to do any additional casting in a Templating Node or Code Execution Node.

prompt_node_json_output

PDF Support for Gemini 2.0 Flash Models

March 19th, 2025

The following Gemini 2.0 Flash models now allow for drag-and-drop PDF documents to be used within your Prompts.

  • Gemini 2.0 Flash Experimental
  • Gemini 2.0 Flash Experimental Thinking Mode
  • Gemini 2.0 Flash

Automatic Evaluations Setup

March 14th, 2025

Now, when navigating to the Evaluations tab of a Prompt/Workflow for the first time, Vellum will auto-generate an initial Test Suite for you. We’ll automatically create one Test Case per Scenario found in the Sandbox and prepare everything needed for you to add Metrics and Ground Truth data.

Auto Evaluation Report Initialization

File Hosting for Images and PDFs

March 11th, 2025

Until now, when you provided an image to a Prompt in Vellum, it’d either have to be a public URL or the base64 encoded representation of the image. With this update, we now support secure file hosting, such that when you drag-and-drop an image (and also now PDFs!) into a Prompt, Vellum securely hosts the file on your behalf.

The end result is that you can now provide much larger images and PDFs to LLMs within Vellum without worrying about a decrease in performance or page load times.

PDFs as a Prompt Input

March 11th, 2025

Recently, certain model providers like Anthropic have begun supporting PDFs as native LLM inputs using a special content type called document (check out their docs for details here.

This is similar to how you might provide a multi-modal model with an image as an input, but now you can provide a PDF as well.

Vellum now also supports passing PDFs as inputs to a Prompt for models that support it. You can do this by drag-and-dropping a PDF file into a Chat History variable in a Prompt. The mechanics are very similar to how you might work with images (see details here).

This is particularly useful for data extraction tasks, where you might want to extract structured data from a PDF and then use that data to power some downstream process.

Support for Qwen QwQ Models via Groq

March 11th, 2025

We’ve added support for a variety of Qwen’s newest QwQ 32B models via Groq’s preview models.

We’ve added the following models:

  • QwQ 32B
  • QwQ 2.5 Coder 32B
  • QWQ 2.5 32B

Support for Qwen QwQ 32B via Fireworks AI

March 11th, 2025

We’ve added support for Qwen’s newest QwQ 32B model via Fireworks AI.

Webhooks

March 10th, 2025

It’s now possible to receive real-time updates about actions taking place in Vellum using Webhooks.

Webhooks

From the organization settings page, you can configure a webhook endpoint with a custom list of events you care about. You can further customize it with your own auth configurations.

This is useful if you’d like to store Vellum monitoring data in your own external data stores. For example, you might save events to a Data Warehouse to power a custom health dashboard.

Keep an eye out, as more event types will be added soon!

Workflow Deployment Executions - Cost Column

March 7th, 2025

You can now see the total cost per Workflow Execution for a given Workflow Deployment in its Executions table. This toggle can be shown/hidden via the “columns” menu.

Cost Column

This is useful for getting a sense of how much a given AI use-case costs to support. This column will be populated for new Workflow Executions going forward and sums the costs associated with all Prompt invocations within the Workflow’s execution (included nested invocations within Subworkflow Nodes, Map Nodes, etc.).

We’ll be exposing more cost metrics throughout Vellum in the coming weeks. Stay tuned!

Prompt Sandbox Pagination

March 5th, 2025

We’ve added pagination to the Prompt Sandbox page. Now, when you have a large number of Scenarios in a Prompt Sandbox, they’ll be split across multiple pages. You can navigate between pages using the pagination controls at the bottom

Prompt Sandbox Pagination

This should result in performance improvements for those with large Prompt Sandboxes.

March 3rd, 2025

We’ve added an eagerly-awaited-for and long-overdue feature to Vellum – Global Search 🎉 You can now search across all your Prompts, Workflows, Document Indexes, and more using the new Search bar in the Vellum side nav.

Global Search Side Nav

Doing so will pull up a search bar where you can search for any resource in your Workspace. You can directly navigate to the resource from the search results.

Global Search Omnibox

You can also access Global Search from any page through the keyboard shortcut Cmd/Ctrl + K. Give it a try and let us know what you think!