#601

Global Rank · of 601 Skills

hosted-agents AI Agent Skill

View Source: guanyang/antigravity-skills

Medium

Installation

npx skills add guanyang/antigravity-skills --skill hosted-agents

59

Installs

Hosted Agent Infrastructure

Hosted agents run in remote sandboxed environments rather than on local machines. When designed well, they provide unlimited concurrency, consistent execution environments, and multiplayer collaboration. The critical insight is that session speed should be limited only by model provider time-to-first-token, with all infrastructure setup completed before the user starts their session.

When to Activate

Activate this skill when:

  • Building background coding agents that run independently of user devices
  • Designing sandboxed execution environments for agent workloads
  • Implementing multiplayer agent sessions with shared state
  • Creating multi-client agent interfaces (Slack, Web, Chrome extensions)
  • Scaling agent infrastructure beyond local machine constraints
  • Building systems where agents spawn sub-agents for parallel work

Core Concepts

Move agent execution to remote sandboxed environments to eliminate the fundamental limits of local execution: resource contention, environment inconsistency, and single-user constraints. Remote sandboxes unlock unlimited concurrency, reproducible environments, and collaborative workflows because each session gets its own isolated compute with a known-good environment image.

Design the architecture in three layers because each layer scales independently. Build sandbox infrastructure for isolated execution, an API layer for state management and client coordination, and client interfaces for user interaction across platforms. Keep these layers cleanly separated so sandbox changes do not ripple into clients.

Detailed Topics

Sandbox Infrastructure

The Core Challenge
Eliminate sandbox spin-up latency because users perceive anything over a few seconds as broken. Development environments require cloning repositories, installing dependencies, and running build steps -- do all of this before the user ever submits a prompt.

Image Registry Pattern
Pre-build environment images on a regular cadence (every 30 minutes works well) because this makes synchronization with the latest code a fast delta rather than a full clone. Include in each image:

  • Cloned repository at a known commit
  • All runtime dependencies installed
  • Initial setup and build commands completed
  • Cached files from running app and test suite once

When starting a session, spin up a sandbox from the most recent image. The repository is at most 30 minutes out of date, making the remaining git sync fast.

Snapshot and Restore
Take filesystem snapshots at key points to enable instant restoration for follow-up prompts without re-running setup:

  • After initial image build (base snapshot)
  • When agent finishes making changes (session snapshot)
  • Before sandbox exit for potential follow-up

Git Configuration for Background Agents
Configure git identity explicitly in every sandbox because background agents are not tied to a specific user during image builds:

  • Generate GitHub app installation tokens for repository access during clone
  • Set git config user.name and user.email when committing and pushing changes
  • Use the prompting user's identity for commits, not the app identity

Warm Pool Strategy
Maintain a pool of pre-warmed sandboxes for high-volume repositories because cold starts are the primary source of user frustration:

  • Keep sandboxes ready before users start sessions
  • Expire and recreate pool entries as new image builds complete
  • Start warming a sandbox as soon as a user begins typing (predictive warm-up)

Agent Framework Selection

Server-First Architecture
Structure the agent framework as a server first, with TUI and desktop apps as thin clients, because this prevents duplicating agent logic across surfaces:

  • Multiple custom clients share one agent backend
  • Consistent behavior across all interaction surfaces
  • Plugin systems extend functionality without client changes
  • Event-driven architectures deliver real-time updates to any connected client

Code as Source of Truth
Select frameworks where the agent can read its own source code to understand behavior. Prioritize this because having code as source of truth prevents the agent from hallucinating about its own capabilities -- an underrated failure mode in AI development.

Plugin System Requirements
Require a plugin system that supports runtime interception because this enables safety controls and observability without modifying core agent logic:

  • Listen to tool execution events (e.g., tool.execute.before)
  • Block or modify tool calls conditionally
  • Inject context or state at runtime

Speed Optimizations

Predictive Warm-Up
Start warming the sandbox as soon as a user begins typing their prompt, not when they submit it, because the typing interval (5-30 seconds) is enough to complete most setup:

  • Clone latest changes in parallel with user typing
  • Run initial setup before user hits enter
  • For fast spin-up, sandbox can be ready before user finishes typing

Parallel File Reading
Allow the agent to start reading files immediately even if sync from latest base branch is not complete, because in large repositories incoming prompts rarely touch recently-changed files:

  • Agent can research immediately without waiting for git sync
  • Block file edits (not reads) until synchronization completes
  • This separation is safe because read-time data staleness of 30 minutes rarely matters for research

Maximize Build-Time Work
Move everything possible to the image build step because build-time duration is invisible to users:

  • Full dependency installation
  • Database schema setup
  • Initial app and test suite runs (populates caches)

Self-Spawning Agents

Agent-Spawned Sessions
Build tools that allow agents to spawn new sessions because frontier models are capable of decomposing work and coordinating sub-tasks:

  • Research tasks across different repositories
  • Parallel subtask execution for large changes
  • Multiple smaller PRs from one major task

Expose three primitives: start a new session with specified parameters, read status of any session (check-in capability), and continue main work while sub-sessions run in parallel.

Prompt Engineering for Self-Spawning
Engineer prompts that guide when agents should spawn sub-sessions rather than doing work inline:

  • Research tasks that require cross-repository exploration
  • Breaking monolithic changes into smaller PRs
  • Parallel exploration of different approaches

API Layer

Per-Session State Isolation
Isolate state per session (SQLite per session works well) because cross-session interference is a subtle and hard-to-debug failure mode:

  • Dedicated database per session
  • No session can impact another's performance
  • Architecture handles hundreds of concurrent sessions

Real-Time Streaming
Stream all agent work in real-time because high-frequency feedback is critical for user trust:

  • Token streaming from model providers
  • Tool execution status updates
  • File change notifications

Use WebSocket connections with hibernation APIs to reduce compute costs during idle periods while maintaining open connections.

Synchronization Across Clients
Build a single state system that synchronizes across all clients (chat interfaces, Slack bots, Chrome extensions, web interfaces, VS Code instances) because users switch surfaces frequently and expect continuity. All changes sync to the session state, enabling seamless client switching.

Multiplayer Support

Why Multiplayer Matters
Design for multiplayer from day one because it is nearly free to add with proper synchronization architecture, and it unlocks high-value workflows:

  • Teaching non-engineers to use AI effectively
  • Live QA sessions with multiple team members
  • Real-time PR review with immediate changes
  • Collaborative debugging sessions

Implementation Requirements
Build the data model so sessions are not tied to single authors because multiplayer fails silently if authorship is hardcoded:

  • Pass authorship info to each prompt
  • Attribute code changes to the prompting user
  • Share session links for instant collaboration

Authentication and Authorization

User-Based Commits
Use GitHub authentication to open PRs on behalf of the user (not the app) because this preserves the audit trail and prevents users from approving their own AI-generated changes:

  • Obtain user tokens for PR creation
  • PRs appear as authored by the human, not the bot

Sandbox-to-API Flow
Follow this sequence because it keeps sandbox permissions minimal while letting the API handle sensitive operations:

  1. Sandbox pushes changes (updating git user config)
  2. Sandbox sends event to API with branch name and session ID
  3. API uses user's GitHub token to create PR
  4. GitHub webhooks notify API of PR events

Client Implementations

Slack Integration
Prioritize Slack as the first distribution channel for internal adoption because it creates a virality loop as team members see others using it:

  • No syntax required, natural chat interface
  • Build a classifier (fast model with repo descriptions) to determine which repository to work in
  • Include hints for common repositories; allow "unknown" for ambiguous cases

Web Interface
Build a web interface with these features because it serves as the primary power-user surface:

  • Real-time streaming of agent work on desktop and mobile
  • Hosted VS Code instance running inside sandbox
  • Streamed desktop view for visual verification
  • Before/after screenshots for PRs
  • Statistics page: sessions resulting in merged PRs (primary metric), usage over time, live "humans prompting" count

Chrome Extension
Build a Chrome extension for non-engineering users because DOM and React internals extraction gives higher precision than raw screenshots at lower token cost:

  • Sidebar chat interface with screenshot tool
  • Extract DOM/React internals instead of raw images
  • Distribute via managed device policy (bypasses Chrome Web Store)

Practical Guidance

Follow-Up Message Handling

Choose between queueing and inserting follow-up messages sent during execution. Prefer queueing because it is simpler to manage and lets users send thoughts on next steps while the agent works. Build a mechanism to stop the agent mid-execution when needed, because without it users feel trapped.

Metrics That Matter

Track these metrics because they indicate real value rather than vanity usage:

  • Sessions resulting in merged PRs (primary success metric)
  • Time from session start to first model response
  • PR approval rate and revision count
  • Agent-written code percentage across repositories

Adoption Strategy

Drive internal adoption through visibility rather than mandates because forced usage breeds resentment:

  • Work in public spaces (Slack channels) for visibility
  • Let the product create virality loops
  • Do not force usage over existing tools
  • Build to people's needs, not hypothetical requirements

Guidelines

  1. Pre-build environment images on regular cadence (30 minutes is a good default)
  2. Start warming sandboxes when users begin typing, not when they submit
  3. Allow file reads before git sync completes; block only writes
  4. Structure agent framework as server-first with clients as thin wrappers
  5. Isolate state per session to prevent cross-session interference
  6. Attribute commits to the user who prompted, not the app
  7. Track merged PRs as primary success metric
  8. Build for multiplayer from the start; it is nearly free with proper sync architecture

Gotchas

  1. Cold start latency: First sandbox spin-up takes 30-60s and users perceive this as broken. Use warm pools and predictive warm-up on keystroke to eliminate perceived wait time.
  2. Image staleness: Infrequent image rebuilds mean agents run with outdated dependencies or code. Set a 30-minute rebuild cadence and monitor image age; alert if builds fail silently.
  3. Sandbox cost runaway: Long-running agents without timeout or budget caps accumulate unexpected costs. Set hard timeout limits (default 4 hours) and per-session cost ceilings.
  4. Auth token expiration mid-session: Long tasks fail when GitHub tokens expire partway through. Implement token refresh logic and check token validity before sensitive operations like PR creation.
  5. Git config in sandboxes: Missing user.name or user.email causes commit failures in background agents. Always set git identity explicitly during sandbox configuration, never assume it carries over from the image.
  6. State loss on sandbox recycle: Agents lose completed work if the sandbox is recycled or times out before results are extracted. Always snapshot before termination and extract artifacts (branches, PRs, files) before letting the sandbox die.
  7. Oversubscribing warm pools: Maintaining too many warm sandboxes wastes money during low-traffic periods. Scale pool size based on traffic patterns and time-of-day; use autoscaling rather than fixed pool sizes.
  8. Missing output extraction: Agents complete work inside the sandbox but results never get pulled out to the user. Build explicit extraction steps (push branch, create PR, return file contents) into the session teardown flow.

Integration

This skill builds on multi-agent-patterns for agent coordination and tool-design for agent-tool interfaces. It connects to:

  • multi-agent-patterns - Self-spawning agents follow supervisor patterns
  • tool-design - Building tools for agent spawning and status checking
  • context-optimization - Managing context across distributed sessions
  • filesystem-context - Using filesystem for session state and artifacts

References

Internal reference:

  • Infrastructure Patterns - Read when: implementing sandbox lifecycle, image builds, or warm pool logic for the first time

Related skills in this collection:

  • multi-agent-patterns - Read when: designing self-spawning or supervisor coordination patterns
  • tool-design - Read when: building tools for agent session management or status checking
  • context-optimization - Read when: context windows fill up across distributed agent sessions

External resources:

  • Ramp - Read when: evaluating whether to build vs. buy background agent infrastructure
  • Modal Sandboxes - Read when: choosing a cloud sandbox provider or comparing isolation models
  • Cloudflare Durable Objects - Read when: designing per-session state management with WebSocket hibernation
  • OpenCode - Read when: selecting a server-first agent framework or studying plugin architectures

Skill Metadata

Created: 2026-01-12
Last Updated: 2026-03-17
Author: Agent Skills for Context Engineering Contributors
Version: 1.1.0

Installs

Installs 59
Global Rank #601 of 601

Security Audit

ath Safe
socket Safe
Alerts: 0 Score: 90
snyk Medium
EU EU-Hosted Inference API

Power your AI Agents with the best open-source models.

Drop-in OpenAI-compatible API. No data leaves Europe.

Explore Inference API

GLM

GLM 5

$1.00 / $3.20

per M tokens

Kimi

Kimi K2.5

$0.60 / $2.80

per M tokens

MiniMax

MiniMax M2.5

$0.30 / $1.20

per M tokens

Qwen

Qwen3.5 122B

$0.40 / $3.00

per M tokens

How to use this skill

1

Install hosted-agents by running npx skills add guanyang/antigravity-skills --skill hosted-agents in your project directory. Run the install command above in your project directory. The skill file will be downloaded from GitHub and placed in your project.

2

No configuration needed. Your AI agent (Claude Code, Cursor, Windsurf, etc.) automatically detects installed skills and uses them as context when generating code.

3

The skill enhances your agent's understanding of hosted-agents, helping it follow established patterns, avoid common mistakes, and produce production-ready output.

What you get

Skills are plain-text instruction files — not executable code. They encode expert knowledge about frameworks, languages, or tools that your AI agent reads to improve its output. This means zero runtime overhead, no dependency conflicts, and full transparency: you can read and review every instruction before installing.

Compatibility

This skill works with any AI coding agent that supports the skills.sh format, including Claude Code (Anthropic), Cursor, Windsurf, Cline, Aider, and other tools that read project-level context files. Skills are framework-agnostic at the transport level — the content inside determines which language or framework it applies to.

Data sourced from the skills.sh registry and GitHub. Install counts and security audits are updated regularly.

EU Made in Europe

Chat with 100+ AI Models in one App.

Use Claude, ChatGPT, Gemini alongside with EU-Hosted Models like Deepseek, GLM-5, Kimi K2.5 and many more.

Customer Support