Deploy Deep Search with DR-Tulu on Vast.ai

January 15, 2026
2 Min Read
By Team Vast
Share
Subscribe

DR-Tulu is AI2's open-source research agent—an 8B model that autonomously plans research strategies, searches the web, reads pages, and synthesizes comprehensive answers with citations. Unlike typical LLMs with tools bolted on, DR-Tulu was trained end-to-end with its MCP tools, making it a compelling open alternative to proprietary research APIs. This guide walks through deploying DR-Tulu on Vast.ai, where you can run the full agent stack on affordable GPU instances and integrate it into your own agentic applications.

What Makes DR-Tulu Special?

Key features include:

  • Native MCP Integration: Trained with web search and page reading tools from the start
  • Autonomous Research: Plans strategies, gathers sources, and synthesizes cited answers
  • Efficient Architecture: 8B parameters with strong research capabilities

What's in the New Guide

Our latest documentation walks you through deploying DR-Tulu on Vast.ai with a split architecture: vLLM for GPU inference on Vast.ai, MCP backend running locally for tool execution.

Complete Deployment Instructions

  • Instance selection with the Vast.ai CLI
  • vLLM configuration with the correct version (critical for compatibility)
  • MCP backend setup with required API keys
  • Workflow configuration to connect the components

Three Ways to Use DR-Tulu

  • Interactive chat for exploratory research
  • Batch evaluation against built-in datasets
  • Python API for programmatic integration

Why This Architecture?

The split deployment keeps GPU-intensive inference on Vast.ai while running tool orchestration locally. This means:

  • Your API keys (Serper, Jina) stay on your machine (or wherever you deploy the MCP backend)
  • Model inference doesn't require sending data to third-party APIs
  • You can modify the MCP backend without redeploying the model
  • Lower Vast.ai costs since you only need GPU for inference

Who Should Use This Guide?

This deployment guide is perfect for:

  • Researchers who need cited, source-backed answers at scale
  • Developers building applications that require web research capabilities
  • Teams evaluating open alternatives to proprietary research APIs

Get Started

The complete guide is now available in our documentation:

Read: Running DR-Tulu on Vast.ai →

Whether you're exploring agentic AI architectures or ready to deploy your own research assistant, this guide provides everything you need to get DR-Tulu running on Vast.ai infrastructure.

Ready to try it? Sign up for Vast.ai and follow the guide to deploy your first instance.

Vast AI

© 2026 Vast.ai. All rights reserved.

Vast.ai