Skip to main content

Intro

Module 2 shows how to keep agents effective as projects scale: wiring up Multi Repo Search tool so they can browse related repositories and tuning model choices (plus custom API keys) to match each agent’s task.

Video lesson

Key takeaways

  • Use the Multi Repo Search tool when questions span multiple repositories, letting agents pull relationships and artifacts without disrupting the current VS Code workspace.
  • Prepare Multi Repo Search tool access by creating a fine-grained GitHub personal access token, adding a connection in the Zencoder dashboard, and registering every repo that agents should consult.
  • Remember that repositories must be indexed before agents can read them, and the dashboard’s indexing log confirms when each sync finishes.
  • Remove linked repositories before deleting a connection; once the connection is gone, you can rebuild it from scratch with fresh credentials.
  • Verify MultiRepoTool availability per agent in the dashboard so the correct tooling is enabled during chats.
  • Set agent models explicitly when needed—Auto is a safe default, but tailoring model choice can optimize for cost, latency, or capability.
  • Add custom API keys for providers like Anthropic or OpenAI when you want usage charged to your own account instead of the workspace allocation.
  • Reference docs at docs.zencoder.ai for current model availability, pricing multipliers, and step-by-step instructions on configuring premium or BYO keys.
  • Follow the module’s model guidance: Grok Code FAST1 for budget tasks, Gemini 2.5 Pro for huge context, Sonnet 4.5 Parallel Thinking for spec workflows, GPT-5 Codecs for specialized code gen, and OPUS for the hardest problems.
  • Collect system/model cards from providers, feed them to an LLM (e.g., Gemini 2.5 Pro’s 1M-token window), and have the agent synthesize comparison tables so you pick the best model per scenario.