Vercel Workflows is now GA

Vercel| 00:08:39|Apr 16, 2026
Chapters7
Explains the shift from short-lived requests to long running, agent driven workflows and the need for scalable backend primitives.

Vercel Workflows is GA, enabling long-running, observable tasks with zero-setup queues and built-in streaming from a single function.

Summary

Zachary (Vercel) announces that Workflows, a feature shipped for a decade’s worth of serverless primitives, is now generally available. The spotlight isn’t just on more deployments (weekly deployments have doubled; a thousand-percent rise in coding agents) but on the type of software teams are building—agents that must reason, wait for humans, retry failures, and stream progress in real time. The demo centers on a birthday-card generator, illustrating how a single function with the Use workflow directive can suspend, resume, and orchestrate steps across prompts, image generation, and emails without managing a database or queues. You get end-to-end observability from the Vercel dashboard and the CLI, with real-time streaming and the ability to decrypt inputs for debugging. The workflow runs on serverless functions with automatic retries, no timeout limits, and persistent state across suspensions. Real-world customers like Mocks, Flora, and Durable are already using Workflows to run video pipelines, coordinate dozens of image-generation models, or ship new agents to millions of sites. The bottom line: developers can focus on product logic while Vercel handles reliability, durability, and observability at scale, from a laptop to millions of users. Workflows is GA today, inviting teams to build long-running, observable agents without rebuilding infrastructure.”

Key Takeaways

  • 100 million runs have been executed since Workflows entered beta, leading to GA today.
  • A single function can run for days, suspend between steps, and resume exactly where it left off using the Use workflow directive.
  • Observability is built in: both the dashboard and CLI provide real-time visibility into each step and its state.
  • Workflows automatically retries failed steps and queues each step behind the scenes on serverless functions.
  • There are no timeout limits for a workflow; it can sleep for seconds, days, or weeks as needed.
  • Streaming updates to clients are built in, allowing real-time progress without external databases.
  • Real-world usage includes Mocks (video generation pipelines), Flora (50 image-generation models), and Durable (3 million customer websites).

Who Is This For?

Essential viewing for frontend and backend developers building long-running, agent-based workflows who want durable, observable processes without managing queues or databases.

Notable Quotes

"We built workflows to solve."
Framing the purpose of Workflows as a solution to long-running, observable tasks.
"Workflows is generally available today, so go achieve something that works."
Announcement of GA and the call to action for builders.
"This is the entire thing. It's one function that you can just read top to bottom and reason about from end to end."
Emphasizing simplicity and readability of the workflow model.
"There are no timeout limits, so workflows can run as long as they need to."
Highlighting long-running capability without artificial constraints.
"The client can even disconnect and reconnect just fine. This is a persistent stream."
Showcasing real-time streaming and resilience to client disconnects.

Questions This Video Answers

  • How do Vercel Workflows handle long-running tasks without a database?
  • What does the Use workflow directive do in Vercel Workflows?
  • What are the real-world use cases for Vercel Workflows in production?
  • How does real-time observability work in Vercel Workflows (dashboard and CLI)?
Vercel WorkflowsGA releaseserverlesslong-running workflowsobservabilitystreamingCLIobservability dashboardephemeral webhooksretry queues
Full Transcript
Hey, good morning everyone. Thanks so much for  joining us today. I'm so excited for what we're   about to share. So for a decade now, code running  on Vercel was primarily built for immediate web   requests. But the most interesting software  our customers create today no longer fits that   model. And the number of these new projects  is exploding just in the last few months.  Weekly deployments in Vercel have doubled, and a  third of those new deployments come from coding   agents, a number that's gone up 1,000% in  just six months. But the most interesting   part to me isn't how much we're deploying, it's  what we're deploying. Those agents are shipping   agents and the agents that our customers  build. They don't just respond once and   wait. They need to reason. Called tools.  We run humans and then do it all again.   And each of those workflows needs a backend  that can keep up. So what I've seen in Vercel   is that the teams shipping the fastest right  now, they aren't rebuilding infrastructure and   coding queues themselves, or writing retry logic  from scratch, or even wiring of observability.  The stuff's been done. So they're using our  infrastructure primitives, the stuff we've   spent the last decade building so they can  focus all of their focus on their product.  And that's exactly the problem.  We built workflows to solve.  We launched it in beta six months ago,  and the response was incredible. And now,   100 million runs later, we're so excited  to announce that workflows is GA today.  Let me show you how it works.  The app I'm going to show you   is a birthday card generator that I built. Now it could be anything really, but this hits   all the things we just talked about. It doesn't  fit into a single request. The waits in humans, it   retries when things fail and it streams progress  in real time to the user. So here's how it works.  Let me send myself a card, and I'll start by  entering my email. Next, I'll pick a date.  And I'm going to pick the 23rd of April.  Now I like doing bonfires my birthday,   so I'll do a potion beach bonfire and  I'm sending it to myself. So for any.  And I'll enter my own email here just to  make the demo easy. So it's going to have   myself RSVP to my own birthday. Cool. Now I'm  going to hit Generate and Send and I'll see   the birthday card start generating. I'll head  over to the Rowsell dashboard where I can see   this run just came in. Let me click in. And I can see this in real time as a   generates prompts and then generates the  message and generates an image as well.  This is the observability you were talking  about. Every step is visible right here.  You can see exactly what's running,  what's completed and what's waiting.  Over. On the user side, I can see that the  card's been generated and the RSVP emails   have been sent. At this point, the workflow  is suspended. There's no compute actually   happening. It's waiting for a human to respond.  And I could go into my email, respond right now,   or even wait for a few hours or days and respond.  Then this will continue to track the progress.  Now, before I show you the email, I actually  want to show you the CLI workflow ships with. The   CLI on day one that lets you inspect everything  that's available on the observability dashboard.  This means that you and your agents have access to  the complete picture. When you're trying to debug   something that went wrong. Okay, over to my email. I should see the email come right in. There it is.   And I can. It's just asking me to  RSVP. I'll hit yes, that'll be there.  Back to the dashboard. I can see the hooks  resumed and I should kick off into a sleep.  Back to the user dashboard and I'll see the  same thing. We're no longer waiting for RSVP,   and now it's just waiting for the  23rd so I can send out the birthday.  Now I want to make you wait for  two weeks, so I'll head back to the   dashboard and wake up the sleep right here. That should immediately resume the workflow,   and we'll see the workflow complete when we  head back to the user dashboard. Postcards   been sent and over in my email I should  get a birthday card with the RSVP as well.  Cool. Now I want   to show you the code and remember the thing we  sent earlier. Teams shipping the fastest today   aren't dealing with queues and managing state and  infrastructure. That's the old way in workflow.   Everything we've seen today is a single function  and it starts with this directive. Use workflow.  This one directive is what lets this  function run for days, suspend between   steps and resume exactly where it left off. So when the workflow calls a step like generate   prompts or generate image. The workflow  suspends the function queues to work in   the background element, retry if it fails,  and then resume the workflow when it's done.   The entire time. I get full observability on the  Vercel dashboard of every step along the way.  I can even click into one of these and by default  on Russell, everything's encrypted, but I'll   decrypt the data so you can see the input. Yeah, this is available both   on the dashboard on the CLI. Workflows also come with built in streaming.  This is how I pushed real time updates  to the client to track progress live.  I didn't have to pull in some sort of  database just to pull this off. The   client can even disconnect and reconnect  just fine. This is a persistent stream.  All right. And here's the hook.  For every RSVP, I'm creating an ephemeral webhook  URL on the fly. I'm then sending in an email,   and then I can simply just wait for it down here. The workflow pauses right here until every   person clicks on their link. There was no  database involved, and the results even   just live in local JavaScript variables.  The workflow holds all of its own state.  This is the distributed systems machinery  that we said was just gone. You don't   have to think about this stuff anymore. All right, lastly, let's sleep at the end   to sleep for my birthday. It's right here. The sleep could be a couple of seconds,   days or even weeks like we just did. And while  it's sleeping, the workflow is just suspended   with zero compute running. When it wakes  up, it continues exactly where it left off.  Maintaining state along the way. And this is the entire thing. It's   one function that you can just read top to bottom  and reason about from end to end. Every step is   backed by a queue. It's a retry automatically  and runs on its own serverless function. There   are no timeout limits, so workflows can run  as long as they need to. This is what we mean   when we said Vercel handles everything underneath. You can just write the logic so we take care of   the rest. Now in this demo I showed you a simple  in use case, but our customers use workflows to   run agents and production at tremendous scale.  Mocks uses workflows to ensure video generation   pipelines run reliably every time. And Flora's orchestrating parallel   creative agents across 50 image  generation models all with workflow.  Durable is a small team that hosts 3 million  customer websites, and they use workflows to   ship new agents to their users every single day. Now I started this demo by saying that the   most interesting software our customers  build no longer fits in a single request,   and we're seeing more and more apps  like these deployed every single day.  Agents are shipping agents, and the teams  building them aren't thinking about cuz   they're focusing on their customer problems. And the shift isn't just happening inside   Vercel. Our customer deployments  are doubling and fewer engineers   than ever are hand rolling infrastructure.  So whatever app or agent you want to build,   workflows will make it reliable, long  running, durable, and observable.  From your laptop to million users, all  with the same application code. Workflows   is generally available today, so  go achieve something that works.

Get daily recaps from
Vercel

AI-powered summaries delivered to your inbox. Save hours every week while staying fully informed.