Overview
Large language models are the most significant technology shift for VC infrastructure in the past decade. They’ve changed how you build tools, how you extract information from documents, how you analyze companies, and how fast you can ship features as a solo developer. If you’re building VC technology in 2026 without using LLMs, you’re working at a significant disadvantage. But LLMs are becoming baseline, not cutting-edge. Every fund is using ChatGPT. Every engineer is using Claude Code or Cursor. The competitive advantage isn’t that you’re using LLMs, it’s how you’re using them and what you’re building on top of them. This chapter covers emerging trends that matter for VC technology: connecting AI to your internal data through MCP, the evolution from single coding sessions to orchestrated agent workflows, why file-native agents are replacing RAG systems, and what hype to ignore.MCP: Connecting AI to Your Internal Data (or Just Use CLI Tools)
Model Context Protocol (MCP) is Anthropic’s standard for connecting AI assistants to data sources. Instead of copy-pasting data into Claude or building custom integrations for every tool, you build MCP servers that expose your data in a standardized way. Claude Code (and other MCP-compatible tools) can then query your internal systems directly. That’s the official story. Here’s the alternative take: Some developers argue MCPs are unnecessary abstraction and you should just write CLI tools instead. Claude Code can already execute bash commands and call CLI tools. A simple CLI script that queries your database or CRM is more universal than an MCP server (works with any tool that can run bash), simpler to build (no SDK required), and easier to maintain. The debate is ongoing, and it’s not clear which approach will win. For now, here’s practical guidance: if you’re building something simple (query your database, fetch data from your CRM), start with a CLI tool. If you need features MCP provides (structured resources, interactive prompts, complex tooling), build an MCP server. Both approaches work for connecting AI to your internal data. Why this matters for VC funds You have valuable data scattered across systems: companies in your CRM, research in your data warehouse, portfolio metrics in various dashboards, memos in Google Docs or Notion. When you’re building features or analyzing data, you currently need to manually pull information from each system, paste it into prompts, and context-switch constantly. MCP servers let AI tools access this data directly. You can ask Claude Code “show me all Series A companies in fintech we’ve talked to in the last 6 months” and it queries your CRM through an MCP server. You can ask “what are the latest metrics for our portfolio companies?” and it pulls from your data warehouse. The AI has the same access to data that you do, without you needing to be the intermediary. The broader trend: Connecting AI to internal data Whether through MCP servers, CLI tools, or other approaches, the key insight is that AI tools become dramatically more useful when they can access your internal data directly. Instead of being a general-purpose assistant, they become specialized tools that understand your fund’s portfolio, pipeline, and research. This could mean:- Querying your CRM to find companies or check deal status
- Running SQL against your data warehouse to analyze portfolio performance
- Searching through investment memos and research to find relevant context
- Pulling metrics from portfolio company dashboards
AI-Assisted Development: From Sessions to Orchestration
Claude Code and Cursor have already changed how you build software. You can implement features in hours that previously took days. You can build entire applications as a solo developer that previously required small teams. This is the current state, and it’s already transformative. But we’re at the beginning, not the end, of AI-assisted development. Current state: Single session, single developer Today, you start a Claude Code session, describe what you want to build, and Claude helps you write code, debug issues, and ship features. When the session ends (or you hit context limits), you start fresh. You’re still fundamentally working alone, just with a very capable assistant. This is already powerful. As covered in Choosing Your Stack, choosing popular technology stacks (Next.js, TypeScript) means AI coding tools work better and you ship faster. But there are limits: complex features that span multiple services, background work that takes hours to run, coordinating changes across many files. Near future: Agent orchestration The next evolution is multiple AI agents working in parallel on the same codebase through git worktrees. Instead of one Claude Code session, you might have ten agents simultaneously:- Agent 1 implements the frontend UI for a new feature
- Agent 2 builds the backend API endpoints
- Agent 3 writes tests for both
- Agent 4 updates documentation
- Agent 5 handles database migrations
- Agents 6-10 work on related features or refactoring
- Modal Sandboxes: Container-based execution with sub-second startup, network controls for restricting outbound access, and scaling to 50,000+ concurrent sandboxes.
- Vercel Sandbox: Ephemeral Linux VMs designed for AI agents and code generation, with SDK-first integration.
- Sprites: Hardware-isolated Firecracker VMs with persistent state, checkpointing, and layer 3 network policies.
File-Native Agents: Beyond RAG and Knowledge Graphs
For the past two years, the standard approach to helping AI systems work with large document collections has been RAG (Retrieval Augmented Generation): chunk documents into pieces, embed them as vectors, store in a vector database, retrieve relevant chunks based on query similarity, stuff them into context. This approach is becoming obsolete. Why RAG was necessary RAG existed as a workaround for limited context windows. If you could only fit 8K or 32K tokens into context, you couldn’t give an AI access to hundreds of documents. So you chunked documents, embedded them, and retrieved only the most relevant pieces for each query. Knowledge graphs were a similar workaround: extract entities and relationships, build a structured graph, query it to find relevant information. Both approaches created intermediate representations (vectors, graphs) because we couldn’t work with documents directly. What changed Context windows are now large enough and getting larger. AI agents have file system access and can use tools (grep, specialized readers). Context compaction techniques let agents maintain understanding across indefinitely long sessions. This means agents can work with files directly, like humans do. No chunking, no embedding, no graph extraction. Just “here’s a folder of investment memos, analyze the fintech companies we’ve evaluated.” File-native agents: Documents stay as documents Instead of transforming documents into vectors or graphs, file-native agents:- Have direct access to files in their native formats (PDFs, markdown, spreadsheets)
- Use tools to search and analyze (grep for keywords, specialized readers for PDFs)
- Maintain context through compaction (file system holds artifacts, compacted context holds insights)
- Store research as markdown files in git repositories
- Give AI agents file system access to these repositories
- Let agents use grep, read files, and search naturally
- Focus on context compaction and session management, not retrieval algorithms
What to Ignore
Don’t fine-tune models. Foundation models work fine for almost all VC use cases (extracting data, analyzing companies, answering questions). Fine-tuning requires training data, evaluation infrastructure, ongoing maintenance, and rarely produces better results for most VC tasks. Don’t add AI just to say you have AI. Build features that solve actual problems. “AI-powered market maps” that are just LLM text aren’t better than human-curated maps. Use AI where it creates real leverage: extracting information from documents, processing unstructured data, helping engineers build faster. Not for theater. AGI timelines don’t matter for your job. Whether AGI arrives in 3 or 30 years doesn’t change what you should build today. Build tools that work now and are maintainable by humans. Model selection is simpler than you think. We use Claude Sonnet for most tasks, Claude Opus for deep reasoning. Don’t optimize for tiny cost differences. Pick an LLM provider, stick with it, and focus on building features.Staying Current
Technology for VC infrastructure evolves quickly. What’s cutting-edge today becomes baseline within months. Here’s how to stay up to date without spending all your time chasing trends: General tech communities- X: Follow engineers building in the AI space, VC tech practitioners, and companies building tools for VCs. The signal-to-noise ratio is low, but you’ll see emerging tools and approaches before they’re widely adopted.
- Hacker News: The Show HN section surfaces new tools and libraries. The comments often contain practical wisdom from people who’ve tried things in production. Good for understanding what’s actually working versus what’s just hype.
- Data Driven VC: Community and resources specifically for people building data infrastructure at VC funds. Much better signal than general tech communities for VC-specific challenges.
- Vestberry VC Day: Conferences focused on VC operations and technology. Good for understanding what funds of all sizes are building and what tools are emerging in the ecosystem.