• Vercel Queues now in public beta

    Vercel Queues is a durable event streaming system built with Fluid compute, and is now available in public beta for all teams. Vercel Queues also powers Workflow: use Queues for direct message publishing and consumption, Workflow for ergonomic multi step orchestration.

    Functions need a reliable way to defer expensive work and guarantee that tasks complete even when functions crash or new deployments roll out. Queues makes it simple to process messages asynchronously with automatic retries and delivery guarantees, providing at-least-once delivery semantics.

    How it works:

    • Messages are sent to a durable topic

    • The queue fans messages out to subscribed consumer groups.

    • Each consumer group processes messages independently.

    • The queue redelivers messages to consumer groups until successfully processed or expired.

    Publish messages from any route handler:

    app/api/orders/route.ts
    import { send } from '@vercel/queue';
    export async function POST(request: Request) {
    const order = await request.json();
    const { messageId } = await send('orders', order);
    return Response.json({ messageId });
    }

    Create a consumer:

    app/api/queues/fulfill-order/route.ts
    import { handleCallback } from '@vercel/queue';
    export const POST = handleCallback(async (order, metadata) => {
    console.log('Fulfilling order', metadata.messageId, order);
    // await doAnythingAsync(order);
    });

    Configure the consumer group:

    vercel.json
    {
    "functions": {
    "app/api/queues/fulfill-order/route.ts": {
    "experimentalTriggers": [{ "type": "queue/v2beta", "topic": "orders" }]
    }
    }
    }

    Adding a trigger makes the route private: it has no public URL and only Vercel's queue infrastructure can invoke it.

    Vercel Queues is billed per API operation, starting at $0.60 per 1M operations, and includes:

    • Multiple AZ synchronous replication

    • At-least-once delivery

    • Customizable visibility timeout

    • Delayed delivery

    • Idempotency keys

    • Concurrency control

    • Per-deployment topic partitioning

    Functions invoked by Queues in push mode are charged at existing Fluid compute rates.

    Get started with the Queues documentation.

  • Chat SDK adds Telegram adapter support

    Chat SDK now supports Telegram, extending its single-codebase approach to Slack, Discord, GitHub, and Teams, with the new Telegram adapter.

    Teams can build bots that support mentions, message reactions, direct messages, and typing indicators.

    The adapter handles single file uploads and renders basic text cards, with buttons and link buttons that display as inline keyboard elements, allowing developers to create interactive workflows directly within Telegram chats.

    Get start with Telegram adapter setup:

    import { Chat } from "chat";
    import { createTelegramAdapter } from "@chat-adapter/telegram";
    const bot = new Chat({
    userName: "mybot",
    adapters: {
    telegram: createTelegramAdapter(),
    },
    });
    bot.onNewMention(async (thread, message) => {
    await thread.post(`You said: ${message.text}`);
    });

    Telegram does not expose full historical message APIs to bots, so message history relies on adapter-level caching. Additionally, callback data is limited to 64 bytes, and the platform does not currently support modals or ephemeral messages.

    Read the documentation to get started.

  • Developer role now available for Pro teams

    Pro teams can now assign the Developer role to their members. Previously only available for Enterprise teams, the Developer role gives Pro teams more granular access control.

    Developers can safely deploy to projects on a team, with more limited team-wide configuration control and environment variables visibility.

    Owners can assign the Developer role to any existing seat or invite new members from the team members settings.

    Learn more about team level roles.

  • New dashboard redesign is now the default

    Dash - DarkDash - Dark

    The new dashboard navigation is now the default experience for all Vercel users.

    Following a successful opt-in beta release in January it has now rolled out fully as of February 26, 2026, with several improvements made based on feedback.

    The redesigned navigation includes:

    • New sidebar with horizontal tabs moved to a resizable sidebar that can be hidden when not needed

    • Consistent tabs for unified navigations across both team and project levels

    • Improved order with navigation items prioritized the most common developer workflows

    • Projects as filters so you can switch between team and project versions of the same page in one click

    • Optimized for mobile with floating bottom bar optimized for one-handed use

    No action is required. The new navigation is available to all users automatically.

    Open your dashboard to see the updated experience.

  • Nano Banana 2 is live on AI Gateway

    Gemini 3.1 Flash Image Preview (Nano Banana 2) is now available on AI Gateway.

    This release improves visual quality while maintaining the generation speed and cost of flash-tier models.

    Nano Banana 2 can use Google Image Search to ground outputs in real-world imagery. This helps with rendering lesser-known landmarks and objects by retrieving live visual data. This model also introduces configurable thinking levels (Minimal and High) to let the model reason through complex prompts before rendering. New resolutions and new aspect ratios (512p, 1:4 and 1:8) are available alongside the existing options to expand to support more types of creative assets.

    To use this model, set model to google/gemini-3.1-flash-image-preview in the AI SDK. Nano Banana 2 is a multimodal model. Use `streamText` or `generateText` to generate images alongside text responses. This example shows how the model can use web search to find live data.

    import { streamText } from 'ai';
    const result = streamText({
    model: 'google/gemini-3.1-flash-image-preview',
    providerOptions: {
    google: { responseModalities: ['TEXT', 'IMAGE'] },
    },
    prompt: 'Generate an image of the 2026 Super Bowl at golden hour',
    });

    You can also change the thinking level: in this example, thinking is set to high for a more thorough response.

    import { streamText } from 'ai';
    const result = streamText({
    model: 'google/gemini-3.1-flash-image-preview',
    providerOptions: {
    google: {
    responseModalities: ['TEXT', 'IMAGE'],
    thinkingConfig: {
    includeThoughts: true,
    thinkingLevel: 'high',
    },
    },
    },
    prompt:
    `An exploded view diagram of a modern GPU, showing the die, HBM stacks, interposer,
    and cooling solution as separate floating layers with labeled callouts.`,
    });

    AI Gateway provides a unified API for calling models, tracking usage and cost, and configuring retries, failover, and performance optimizations for higher-than-provider uptime. It includes built-in observability, Bring Your Own Key support, and intelligent provider routing with automatic retries.

    Learn more about AI Gateway, view the AI Gateway model leaderboard or try it in our model playground.

  • GPT 5.3 Codex is now on AI Gateway

    GPT 5.3 Codex is now available on AI Gateway. GPT 5.3 Codex brings together the coding strengths of GPT-5.2-Codex and the reasoning depth of GPT-5.2 in a single model that's 25% faster and more token-efficient.

    Built for long-running agentic work, the model handles research, tool use, and multi-step execution across the full software lifecycle, from debugging and deployment to product documents and data analysis. Additionally, you can steer it mid-task without losing context. For web development, it better understands underspecified prompts and defaults to more functional, production-ready output.

    To use this model, set model to openai/gpt-5.3-codex in the AI SDK.

    import { streamText } from 'ai';
    const result = streamText({
    model: 'openai/gpt-5.3-codex',
    prompt:
    `Research our current API architecture, identify performance
    bottlenecks, refactor the slow endpoints, add monitoring,
    and deploy the changes to staging.`,
    });

    AI Gateway provides a unified API for calling models, tracking usage and cost, and configuring retries, failover, and performance optimizations for higher-than-provider uptime. It includes built-in observability, Bring Your Own Key support, and intelligent provider routing with automatic retries.

    Learn more about AI Gateway, view the AI Gateway model leaderboard or try it in our model playground.