OpenHands/openhands/server/conversation_manager/docker_nested_conversation_manager.py
Rohit Malhotra 25d9cf2890
[Refactor]: Add LLMRegistry for llm services (#9589)
Co-authored-by: openhands <openhands@all-hands.dev>
Co-authored-by: Graham Neubig <neubig@gmail.com>
Co-authored-by: Engel Nyst <enyst@users.noreply.github.com>
2025-08-18 02:11:20 -04:00

583 lines
24 KiB
Python

from __future__ import annotations
import asyncio
import hashlib
import os
from base64 import urlsafe_b64encode
from dataclasses import dataclass, field
from types import MappingProxyType
from typing import Any, cast
import docker
import httpx
import socketio
from docker.models.containers import Container
from openhands.controller.agent import Agent
from openhands.core.config import OpenHandsConfig
from openhands.core.config.llm_config import LLMConfig
from openhands.core.logger import openhands_logger as logger
from openhands.events.action import MessageAction
from openhands.events.nested_event_store import NestedEventStore
from openhands.events.stream import EventStream
from openhands.experiments.experiment_manager import ExperimentManagerImpl
from openhands.integrations.provider import PROVIDER_TOKEN_TYPE, ProviderHandler
from openhands.runtime import get_runtime_cls
from openhands.runtime.impl.docker.docker_runtime import DockerRuntime
from openhands.server.config.server_config import ServerConfig
from openhands.server.constants import ROOM_KEY
from openhands.server.conversation_manager.conversation_manager import (
ConversationManager,
)
from openhands.server.data_models.agent_loop_info import AgentLoopInfo
from openhands.server.monitoring import MonitoringListener
from openhands.server.session.conversation import ServerConversation
from openhands.server.session.conversation_init_data import ConversationInitData
from openhands.server.session.session import Session
from openhands.storage.conversation.conversation_store import ConversationStore
from openhands.storage.data_models.conversation_metadata import ConversationMetadata
from openhands.storage.data_models.conversation_status import ConversationStatus
from openhands.storage.data_models.settings import Settings
from openhands.storage.files import FileStore
from openhands.storage.locations import get_conversation_dir
from openhands.utils.async_utils import call_sync_from_async
from openhands.utils.import_utils import get_impl
from openhands.utils.utils import create_registry_and_convo_stats
@dataclass
class DockerNestedConversationManager(ConversationManager):
"""ServerConversation manager where the agent loops exist inside the docker containers."""
sio: socketio.AsyncServer
config: OpenHandsConfig
server_config: ServerConfig
file_store: FileStore
docker_client: docker.DockerClient = field(default_factory=docker.from_env)
_conversation_store_class: type[ConversationStore] | None = None
_starting_conversation_ids: set[str] = field(default_factory=set)
_runtime_container_image: str | None = None
async def __aenter__(self):
runtime_cls = get_runtime_cls(self.config.runtime)
runtime_cls.setup(self.config)
async def __aexit__(self, exc_type, exc_value, traceback):
runtime_cls = get_runtime_cls(self.config.runtime)
runtime_cls.teardown(self.config)
async def attach_to_conversation(
self, sid: str, user_id: str | None = None
) -> ServerConversation | None:
# Not supported - clients should connect directly to the nested server!
raise ValueError('unsupported_operation')
async def detach_from_conversation(self, conversation: ServerConversation):
# Not supported - clients should connect directly to the nested server!
raise ValueError('unsupported_operation')
async def join_conversation(
self,
sid: str,
connection_id: str,
settings: Settings,
user_id: str | None,
) -> AgentLoopInfo:
# Not supported - clients should connect directly to the nested server!
raise ValueError('unsupported_operation')
async def get_running_agent_loops(
self, user_id: str | None = None, filter_to_sids: set[str] | None = None
) -> set[str]:
"""Get the running agent loops directly from docker."""
containers: list[Container] = self.docker_client.containers.list()
names = (container.name or '' for container in containers)
conversation_ids = {
name[len('openhands-runtime-') :]
for name in names
if name.startswith('openhands-runtime-')
}
if filter_to_sids is not None:
conversation_ids = {
conversation_id
for conversation_id in conversation_ids
if conversation_id in filter_to_sids
}
return conversation_ids
async def get_connections(
self, user_id: str | None = None, filter_to_sids: set[str] | None = None
) -> dict[str, str]:
# We don't monitor connections outside the nested server, though we could introduce an API for this.
results: dict[str, str] = {}
return results
async def maybe_start_agent_loop(
self,
sid: str,
settings: Settings,
user_id: str | None,
initial_user_msg: MessageAction | None = None,
replay_json: str | None = None,
) -> AgentLoopInfo:
if not await self.is_agent_loop_running(sid):
await self._start_agent_loop(
sid, settings, user_id, initial_user_msg, replay_json
)
nested_url = self._get_nested_url(sid)
session_api_key = self._get_session_api_key_for_conversation(sid)
return AgentLoopInfo(
conversation_id=sid,
url=nested_url,
session_api_key=session_api_key,
event_store=NestedEventStore(
base_url=nested_url,
sid=sid,
user_id=user_id,
session_api_key=session_api_key,
),
status=ConversationStatus.STARTING
if sid in self._starting_conversation_ids
else ConversationStatus.RUNNING,
)
async def _start_agent_loop(
self,
sid: str,
settings: Settings,
user_id: str | None,
initial_user_msg: MessageAction | None,
replay_json: str | None,
):
logger.info(f'starting_agent_loop:{sid}', extra={'session_id': sid})
await self.ensure_num_conversations_below_limit(sid, user_id)
runtime = await self._create_runtime(sid, user_id, settings)
self._starting_conversation_ids.add(sid)
try:
# Build the runtime container image if it is missing
await call_sync_from_async(runtime.maybe_build_runtime_container_image)
self._runtime_container_image = runtime.runtime_container_image
# check that the container already exists...
if await self._start_existing_container(runtime):
self._starting_conversation_ids.discard(sid)
return
# initialize the container but dont wait for it to start
await call_sync_from_async(runtime.init_container)
# Start the conversation in a background task.
asyncio.create_task(
self._start_conversation(
sid,
settings,
runtime,
initial_user_msg,
replay_json,
runtime.api_url,
)
)
except Exception:
self._starting_conversation_ids.discard(sid)
raise
async def _start_conversation(
self,
sid: str,
settings: Settings,
runtime: DockerRuntime,
initial_user_msg: MessageAction | None,
replay_json: str | None,
api_url: str,
):
try:
await call_sync_from_async(runtime.wait_until_alive)
await call_sync_from_async(runtime.setup_initial_env)
async with httpx.AsyncClient(
headers={
'X-Session-API-Key': self._get_session_api_key_for_conversation(sid)
}
) as client:
# setup the settings...
settings_json = settings.model_dump(context={'expose_secrets': True})
settings_json.pop('custom_secrets', None)
settings_json.pop('git_provider_tokens', None)
if settings_json.get('git_provider'):
settings_json['git_provider'] = settings_json['git_provider'].value
settings_json.pop('secrets_store', None) or {}
response = await client.post(
f'{api_url}/api/settings', json=settings_json
)
response.raise_for_status()
# Setup provider tokens
provider_handler = self._get_provider_handler(settings)
provider_tokens = provider_handler.provider_tokens
if provider_tokens:
provider_tokens_json = {
k.value: {
'token': v.token.get_secret_value(),
'user_id': v.user_id,
'host': v.host,
}
for k, v in provider_tokens.items()
if v.token
}
response = await client.post(
f'{api_url}/api/add-git-providers',
json={
'provider_tokens': provider_tokens_json,
},
)
response.raise_for_status()
# Setup custom secrets
custom_secrets = settings.custom_secrets # type: ignore
if custom_secrets:
for key, secret in custom_secrets.items():
response = await client.post(
f'{api_url}/api/secrets',
json={
'name': key,
'description': secret.description,
'value': secret.secret.get_secret_value(),
},
)
response.raise_for_status()
init_conversation: dict[str, Any] = {
'initial_user_msg': initial_user_msg,
'image_urls': [],
'replay_json': replay_json,
'conversation_id': sid,
}
if isinstance(settings, ConversationInitData):
init_conversation['repository'] = settings.selected_repository
init_conversation['selected_branch'] = settings.selected_branch
init_conversation['git_provider'] = (
settings.git_provider.value if settings.git_provider else None
)
# Create conversation
response = await client.post(
f'{api_url}/api/conversations', json=init_conversation
)
logger.info(
f'_start_agent_loop:{response.status_code}:{response.json()}'
)
response.raise_for_status()
finally:
self._starting_conversation_ids.discard(sid)
async def send_to_event_stream(self, connection_id: str, data: dict):
# Not supported - clients should connect directly to the nested server!
raise ValueError('unsupported_operation')
async def request_llm_completion(
self,
sid: str,
service_id: str,
llm_config: LLMConfig,
messages: list[dict[str, str]],
) -> str:
# Not supported - clients should connect directly to the nested server!
raise ValueError('unsupported_operation')
async def send_event_to_conversation(self, sid, data):
async with httpx.AsyncClient(
headers={
'X-Session-API-Key': self._get_session_api_key_for_conversation(sid)
}
) as client:
nested_url = self._get_nested_url(sid)
response = await client.post(
f'{nested_url}/api/conversations/{sid}/events', json=data
)
response.raise_for_status()
async def disconnect_from_session(self, connection_id: str):
# Not supported - clients should connect directly to the nested server!
raise ValueError('unsupported_operation')
async def close_session(self, sid: str):
# First try to graceful stop server.
try:
container = self.docker_client.containers.get(f'openhands-runtime-{sid}')
except docker.errors.NotFound:
return
try:
nested_url = self.get_nested_url_for_container(container)
async with httpx.AsyncClient(
headers={
'X-Session-API-Key': self._get_session_api_key_for_conversation(sid)
}
) as client:
# Stop conversation
response = await client.post(
f'{nested_url}/api/conversations/{sid}/stop'
)
response.raise_for_status()
# Check up to 3 times that client has closed
for _ in range(3):
response = await client.get(f'{nested_url}/api/conversations/{sid}')
response.raise_for_status()
if response.json().get('status') == 'STOPPED':
break
await asyncio.sleep(1)
except Exception as e:
logger.warning(
'error_stopping_container', extra={'sid': sid, 'error': str(e)}
)
container.stop()
async def get_agent_loop_info(
self, user_id: str | None = None, filter_to_sids: set[str] | None = None
) -> list[AgentLoopInfo]:
results = []
containers: list[Container] = self.docker_client.containers.list()
for container in containers:
if not container.name or not container.name.startswith(
'openhands-runtime-'
):
continue
conversation_id = container.name[len('openhands-runtime-') :]
if filter_to_sids is not None and conversation_id not in filter_to_sids:
continue
nested_url = self.get_nested_url_for_container(container)
if os.getenv('NESTED_RUNTIME_BROWSER_HOST', '') != '':
# This should be set to http://localhost if you're running OH inside a docker container
nested_url = nested_url.replace(
self.config.sandbox.local_runtime_url,
os.getenv('NESTED_RUNTIME_BROWSER_HOST', ''),
)
agent_loop_info = AgentLoopInfo(
conversation_id=conversation_id,
url=nested_url,
session_api_key=self._get_session_api_key_for_conversation(
conversation_id
),
event_store=NestedEventStore(
base_url=nested_url,
sid=conversation_id,
user_id=user_id,
),
status=ConversationStatus.STARTING
if conversation_id in self._starting_conversation_ids
else ConversationStatus.RUNNING,
)
results.append(agent_loop_info)
return results
@classmethod
def get_instance(
cls,
sio: socketio.AsyncServer,
config: OpenHandsConfig,
file_store: FileStore,
server_config: ServerConfig,
monitoring_listener: MonitoringListener,
) -> ConversationManager:
return DockerNestedConversationManager(
sio=sio,
config=config,
server_config=server_config,
file_store=file_store,
)
def get_agent_session(self, sid: str):
"""Get the agent session for a given session ID.
Args:
sid: The session ID.
Returns:
The agent session, or None if not found.
"""
raise ValueError('unsupported_operation')
async def _get_conversation_store(self, user_id: str | None) -> ConversationStore:
conversation_store_class = self._conversation_store_class
if not conversation_store_class:
self._conversation_store_class = conversation_store_class = get_impl(
ConversationStore, # type: ignore
self.server_config.conversation_store_class,
)
store = await conversation_store_class.get_instance(self.config, user_id)
return store
def _get_nested_url(self, sid: str) -> str:
container = self.docker_client.containers.get(f'openhands-runtime-{sid}')
return self.get_nested_url_for_container(container)
def get_nested_url_for_container(self, container: Container) -> str:
env = container.attrs['Config']['Env']
container_port = int(next(e[5:] for e in env if e.startswith('port=')))
container_name = container.name or ''
conversation_id = container_name[len('openhands-runtime-') :]
nested_url = f'{self.config.sandbox.local_runtime_url}:{container_port}/api/conversations/{conversation_id}'
return nested_url
def _get_session_api_key_for_conversation(self, conversation_id: str) -> str:
jwt_secret = self.config.jwt_secret.get_secret_value() # type:ignore
conversation_key = f'{jwt_secret}:{conversation_id}'.encode()
session_api_key = (
urlsafe_b64encode(hashlib.sha256(conversation_key).digest())
.decode()
.replace('=', '')
)
return session_api_key
async def ensure_num_conversations_below_limit(
self, sid: str, user_id: str | None
) -> None:
response_ids = await self.get_running_agent_loops(user_id)
if len(response_ids) >= self.config.max_concurrent_conversations:
logger.info(
f'too_many_sessions_for:{user_id or ""}',
extra={'session_id': sid, 'user_id': user_id},
)
# Get the conversations sorted (oldest first)
conversation_store = await self._get_conversation_store(user_id)
conversations = await conversation_store.get_all_metadata(response_ids)
conversations.sort(key=_last_updated_at_key, reverse=True)
while len(conversations) >= self.config.max_concurrent_conversations:
oldest_conversation_id = conversations.pop().conversation_id
logger.debug(
f'closing_from_too_many_sessions:{user_id or ""}:{oldest_conversation_id}',
extra={'session_id': oldest_conversation_id, 'user_id': user_id},
)
# Send status message to client and close session.
status_update_dict = {
'status_update': True,
'type': 'error',
'id': 'AGENT_ERROR$TOO_MANY_CONVERSATIONS',
'message': 'Too many conversations at once. If you are still using this one, try reactivating it by prompting the agent to continue',
}
await self.sio.emit(
'oh_event',
status_update_dict,
to=ROOM_KEY.format(sid=oldest_conversation_id),
)
await self.close_session(oldest_conversation_id)
def _get_provider_handler(self, settings: Settings) -> ProviderHandler:
provider_tokens = None
if isinstance(settings, ConversationInitData):
provider_tokens = settings.git_provider_tokens
provider_handler = ProviderHandler(
provider_tokens=provider_tokens
or cast(PROVIDER_TOKEN_TYPE, MappingProxyType({}))
)
return provider_handler
async def _create_runtime(
self, sid: str, user_id: str | None, settings: Settings
) -> DockerRuntime:
# This session is created here only because it is the easiest way to get a runtime, which
# is the easiest way to create the needed docker container
config: OpenHandsConfig = ExperimentManagerImpl.run_config_variant_test(
user_id, sid, self.config
)
llm_registry, convo_stats, config = create_registry_and_convo_stats(
config, sid, user_id, settings
)
session = Session(
sid=sid,
llm_registry=llm_registry,
convo_stats=convo_stats,
file_store=self.file_store,
config=config,
sio=self.sio,
user_id=user_id,
)
llm_registry.retry_listner = session._notify_on_llm_retry
agent_cls = settings.agent or config.default_agent
agent_config = config.get_agent_config(agent_cls)
agent = Agent.get_cls(agent_cls)(agent_config, llm_registry)
config = config.model_copy(deep=True)
env_vars = config.sandbox.runtime_startup_env_vars
env_vars['CONVERSATION_MANAGER_CLASS'] = (
'openhands.server.conversation_manager.standalone_conversation_manager.StandaloneConversationManager'
)
env_vars['SERVE_FRONTEND'] = '0'
env_vars['RUNTIME'] = 'local'
# TODO: In the long term we may come up with a more secure strategy for user management within the nested runtime.
env_vars['USER'] = 'root'
env_vars['SESSION_API_KEY'] = self._get_session_api_key_for_conversation(sid)
# We need to be able to specify the nested conversation id within the nested runtime
env_vars['ALLOW_SET_CONVERSATION_ID'] = '1'
env_vars['WORKSPACE_BASE'] = '/workspace'
env_vars['SANDBOX_CLOSE_DELAY'] = '0'
env_vars['SKIP_DEPENDENCY_CHECK'] = '1'
env_vars['INITIAL_NUM_WARM_SERVERS'] = '1'
volumes: list[str | None]
if not config.sandbox.volumes:
volumes = []
else:
volumes = [v.strip() for v in config.sandbox.volumes.split(',')]
conversation_dir = get_conversation_dir(sid, user_id)
# Set up mounted volume for conversation directory within workspace
if config.file_store == 'local':
# Resolve ~ from path as the docker container does not work otherwise
file_store_path = os.path.realpath(
os.path.expanduser(config.file_store_path)
)
volumes.append(
f'{file_store_path}/{conversation_dir}:/root/.openhands/{conversation_dir}:rw'
)
config.sandbox.volumes = ','.join([v for v in volumes if v is not None])
if not config.sandbox.runtime_container_image:
config.sandbox.runtime_container_image = self._runtime_container_image
# Currently this eventstream is never used and only exists because one is required in order to create a docker runtime
event_stream = EventStream(sid, self.file_store, user_id)
runtime = DockerRuntime(
config=config,
event_stream=event_stream,
sid=sid,
plugins=agent.sandbox_plugins,
headless_mode=False,
attach_to_existing=False,
main_module='openhands.server',
llm_registry=llm_registry,
)
# Hack - disable setting initial env.
runtime.setup_initial_env = lambda: None # type:ignore
return runtime
async def _start_existing_container(self, runtime: DockerRuntime) -> bool:
try:
container = self.docker_client.containers.get(runtime.container_name)
if container:
status = container.status
if status == 'exited':
await call_sync_from_async(container.start)
return True
return False
except docker.errors.NotFound:
return False
def _last_updated_at_key(conversation: ConversationMetadata) -> float:
last_updated_at = conversation.last_updated_at
if last_updated_at is None:
return 0.0
return last_updated_at.timestamp()