Skip to Content
DocsStreamingexp.Overview

Streaming Overview

Experimental Feature

This is a new streaming implementation with support for multiple output formats (including AI SDK v5). It will replace `stream()` once battle-tested, and the API may change as we incorporate feedback.

Mastra supports real-time, incremental responses from agents and workflows, allowing users to see output as it’s generated instead of waiting for completion. This is useful for chat, long-form content, multi-step workflows, or any scenario where immediate feedback matters.

Getting started

Mastra currently supports two streaming methods, this page explains how to use streamVNext().

  1. .stream(): Current stable API, supports AI SDK v1.
  2. .streamVNext(): Experimental API, supports AI SDK v2.

Streaming with agents

You can pass a single string for simple prompts, an array of strings when providing multiple pieces of context, or an array of message objects with role and content for precise control over roles and conversational flows.

Using Agent.streamVNext()

A textStream breaks the response into chunks as it’s generated, allowing output to stream progressively instead of arriving all at once. Iterate over the textStream using a for await loop to inspect each stream chunk.

const testAgent = mastra.getAgent("testAgent"); const stream = await testAgent.streamVNext([ { role: "user", content: "Help me organize my day" }, ]); for await (const chunk of stream.textStream) { process.stdout.write(chunk); }

See Agent.streamVNext() for more information.

Output from Agent.streamVNext()

The output streams the generated response from the agent.

Of course! To help you organize your day effectively, I need a bit more information. Here are some questions to consider: ...

Agent stream properties

An agent stream provides access to various response properties:

  • stream.textStream: A readable stream that emits text chunks.
  • stream.text: Promise that resolves to the full text response.
  • stream.finishReason: The reason the agent stopped streaming.
  • stream.usage: Token usage information.

AI SDK v5 Compatibility

For integration with AI SDK v5, use format ‘aisdk’ to get an AISDKV5OutputStream:

const testAgent = mastra.getAgent("testAgent"); const stream = await testAgent.streamVNext( [{ role: "user", content: "Help me organize my day" }], { format: "aisdk" } ); for await (const chunk of stream.textStream) { process.stdout.write(chunk); }

Streaming with workflows

Streaming from a workflow returns a sequence of structured events describing the run lifecycle, rather than incremental text chunks. This event-based format makes it possible to track and respond to workflow progress in real time once a run is created using .createRunAsync().

Using Run.streamVNext()

This is the experimental API. It returns a ReadableStream of events directly.

const run = await testWorkflow.createRunAsync(); const stream = await run.streamVNext({ inputData: { value: "initial data" } }); for await (const chunk of stream) { console.log(chunk); }

See Run.streamVNext() for more information.

Output from Run.streamVNext()

The experimental API event structure includes runId and from at the top level, making it easier to identify and track workflow runs without digging into the payload.

// ... { type: 'step-start', runId: '1eeaf01a-d2bf-4e3f-8d1b-027795ccd3df', from: 'WORKFLOW', payload: { stepName: 'step-1', args: { value: 'initial data' }, stepCallId: '8e15e618-be0e-4215-a5d6-08e58c152068', startedAt: 1755121710066, status: 'running' } }

Workflow stream properties

A workflow stream provides access to various response properties:

  • stream.status: The status of the workflow run.
  • stream.result: The result of the workflow run.
  • stream.usage: The total token usage of the workflow run.