- Added new environments for reinforcement learning, including `HermesSweEnv` for software engineering tasks and `TerminalTestEnv` for inline testing. - Introduced `ToolContext` for unrestricted access to tools during reward computation. - Updated `.gitignore` to exclude `wandb/` directory. - Enhanced `README.md` with detailed architecture and usage instructions for Atropos environments. - Added configuration files for SWE and terminal test environments to streamline setup. - Removed unnecessary compiled Python files from `__pycache__`.
29 lines
1007 B
Python
29 lines
1007 B
Python
"""
|
|
Hermes-Agent Atropos Environments
|
|
|
|
Provides a layered integration between hermes-agent's tool-calling capabilities
|
|
and the Atropos RL training framework.
|
|
|
|
Layers:
|
|
- agent_loop: Reusable multi-turn agent loop with standard OpenAI-spec tool calling
|
|
- tool_context: Per-rollout tool access handle for reward/verification functions
|
|
- hermes_base_env: Abstract base environment (BaseEnv subclass) for Atropos
|
|
- tool_call_parsers: Client-side tool call parser registry for Phase 2 (VLLM /generate)
|
|
|
|
Concrete environments:
|
|
- terminal_test_env: Simple file-creation tasks for testing the stack
|
|
- hermes_swe_env: SWE-bench style tasks with Modal sandboxes
|
|
"""
|
|
|
|
from environments.agent_loop import AgentResult, HermesAgentLoop
|
|
from environments.tool_context import ToolContext
|
|
from environments.hermes_base_env import HermesAgentBaseEnv, HermesAgentEnvConfig
|
|
|
|
__all__ = [
|
|
"AgentResult",
|
|
"HermesAgentLoop",
|
|
"ToolContext",
|
|
"HermesAgentBaseEnv",
|
|
"HermesAgentEnvConfig",
|
|
]
|