Revert "feat(sandbox): Add Jupyter Kernel for Interactive Python Interpreter for Sandbox (#1215)" (#1229)

This reverts commit 492feecb67e825e9cf27d363c1de27edd65accc7.
This commit is contained in:
Xingyao Wang 2024-04-20 00:49:24 +08:00 committed by GitHub
parent 34286fabcc
commit 871eefe801
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
8 changed files with 5 additions and 341 deletions

View File

@ -19,9 +19,3 @@ RUN apt-get update && apt-get install -y \
&& rm -rf /var/lib/apt/lists/*
RUN service ssh start
RUN pip install jupyterlab notebook jupyter_kernel_gateway
# Add common data science utils
RUN pip install transformers[torch]
RUN pip install torch --index-url https://download.pytorch.org/whl/cpu
RUN pip install boilerpy3 pandas datasets sympy scikit-learn matplotlib seaborn

View File

@ -120,9 +120,6 @@ class DockerExecBox(Sandbox):
return -1, f'Command: "{cmd}" timed out'
return exit_code, logs.decode('utf-8')
def execute_python(self, code: str) -> str:
raise NotImplementedError('execute_python is not supported in DockerExecBox')
def execute_in_background(self, cmd: str) -> BackgroundCommand:
result = self.container.exec_run(
self.get_exec_cmd(cmd), socket=True, workdir=SANDBOX_WORKSPACE_DIR

View File

@ -1,238 +0,0 @@
import re
import os
import tornado
import asyncio
from tornado.escape import json_encode, json_decode, url_escape
from tornado.websocket import websocket_connect, WebSocketClientConnection
from tornado.ioloop import PeriodicCallback
from tornado.httpclient import AsyncHTTPClient, HTTPRequest
from opendevin.logger import opendevin_logger as logger
from uuid import uuid4
def strip_ansi(o: str) -> str:
"""
Removes ANSI escape sequences from `o`, as defined by ECMA-048 in
http://www.ecma-international.org/publications/files/ECMA-ST/Ecma-048.pdf
# https://github.com/ewen-lbh/python-strip-ansi/blob/master/strip_ansi/__init__.py
>>> strip_ansi("\\033[33mLorem ipsum\\033[0m")
'Lorem ipsum'
>>> strip_ansi("Lorem \\033[38;25mIpsum\\033[0m sit\\namet.")
'Lorem Ipsum sit\\namet.'
>>> strip_ansi("")
''
>>> strip_ansi("\\x1b[0m")
''
>>> strip_ansi("Lorem")
'Lorem'
>>> strip_ansi('\\x1b[38;5;32mLorem ipsum\\x1b[0m')
'Lorem ipsum'
>>> strip_ansi('\\x1b[1m\\x1b[46m\\x1b[31mLorem dolor sit ipsum\\x1b[0m')
'Lorem dolor sit ipsum'
"""
# pattern = re.compile(r'/(\x9B|\x1B\[)[0-?]*[ -\/]*[@-~]/')
pattern = re.compile(r'\x1B\[\d+(;\d+){0,2}m')
stripped = pattern.sub('', o)
return stripped
class JupyterKernel:
def __init__(self, url_suffix, convid, lang='python'):
self.base_url = f'http://{url_suffix}'
self.base_ws_url = f'ws://{url_suffix}'
self.lang = lang
self.kernel_id = None
self.convid = convid
logger.info(
f'Jupyter kernel created for conversation {convid} at {url_suffix}'
)
self.heartbeat_interval = 10000 # 10 seconds
self.heartbeat_callback = None
async def initialize(self):
await self.execute(r'%colors nocolor')
# pre-defined tools
# self.tools_to_run = [
# # TODO: You can add code for your pre-defined tools here
# ]
# for tool in self.tools_to_run:
# # logger.info(f"Tool initialized:\n{tool}")
# await self.execute(tool)
async def _send_heartbeat(self):
if not hasattr(self, 'ws') or not self.ws:
return
try:
self.ws.ping()
logger.debug('Heartbeat sent...')
except tornado.iostream.StreamClosedError:
logger.info('Heartbeat failed, reconnecting...')
try:
await self._connect()
except ConnectionRefusedError:
logger.info(
'ConnectionRefusedError: Failed to reconnect to kernel websocket - Is the kernel still running?'
)
async def _connect(self):
if hasattr(self, 'ws') and self.ws:
self.ws.close()
self.ws: WebSocketClientConnection = None
client = AsyncHTTPClient()
if not self.kernel_id:
n_tries = 5
while n_tries > 0:
try:
response = await client.fetch(
'{}/api/kernels'.format(self.base_url),
method='POST',
body=json_encode({'name': self.lang}),
)
kernel = json_decode(response.body)
self.kernel_id = kernel['id']
break
except Exception as e:
logger.error('Failed to connect to kernel')
logger.exception(e)
logger.info('Retrying in 5 seconds...')
# kernels are not ready yet
n_tries -= 1
await asyncio.sleep(5)
if n_tries == 0:
raise ConnectionRefusedError('Failed to connect to kernel')
ws_req = HTTPRequest(
url='{}/api/kernels/{}/channels'.format(
self.base_ws_url, url_escape(self.kernel_id)
)
)
self.ws = await websocket_connect(ws_req)
logger.info('Connected to kernel websocket')
# Setup heartbeat
if self.heartbeat_callback:
self.heartbeat_callback.stop()
self.heartbeat_callback = PeriodicCallback(
self._send_heartbeat, self.heartbeat_interval
)
self.heartbeat_callback.start()
async def execute(self, code, timeout=60):
if not hasattr(self, 'ws') or not self.ws:
await self._connect()
msg_id = uuid4().hex
self.ws.write_message(
json_encode(
{
'header': {
'username': '',
'version': '5.0',
'session': '',
'msg_id': msg_id,
'msg_type': 'execute_request',
},
'parent_header': {},
'channel': 'shell',
'content': {
'code': code,
'silent': False,
'store_history': False,
'user_expressions': {},
'allow_stdin': False,
},
'metadata': {},
'buffers': {},
}
)
)
logger.info(f'EXECUTE REQUEST SENT:\n{code}')
outputs = []
async def wait_for_messages():
execution_done = False
while not execution_done:
msg = await self.ws.read_message()
msg = json_decode(msg)
msg_type = msg['msg_type']
parent_msg_id = msg['parent_header'].get('msg_id', None)
if parent_msg_id != msg_id:
continue
if os.environ.get('DEBUG', False):
logger.info(
f"MSG TYPE: {msg_type.upper()} DONE:{execution_done}\nCONTENT: {msg['content']}"
)
if msg_type == 'error':
traceback = '\n'.join(msg['content']['traceback'])
outputs.append(traceback)
execution_done = True
elif msg_type == 'stream':
outputs.append(msg['content']['text'])
elif msg_type in ['execute_result', 'display_data']:
outputs.append(msg['content']['data']['text/plain'])
if 'image/png' in msg['content']['data']:
# use markdone to display image (in case of large image)
# outputs.append(f"\n<img src=\"data:image/png;base64,{msg['content']['data']['image/png']}\"/>\n")
outputs.append(
f"![image](data:image/png;base64,{msg['content']['data']['image/png']})"
)
elif msg_type == 'execute_reply':
execution_done = True
return execution_done
async def interrupt_kernel():
client = AsyncHTTPClient()
interrupt_response = await client.fetch(
f'{self.base_url}/api/kernels/{self.kernel_id}/interrupt',
method='POST',
body=json_encode({'kernel_id': self.kernel_id}),
)
logger.info(f'Kernel interrupted: {interrupt_response}')
try:
execution_done = await asyncio.wait_for(wait_for_messages(), timeout)
except asyncio.TimeoutError:
await interrupt_kernel()
return f'[Execution timed out ({timeout} seconds).]'
if not outputs and execution_done:
ret = '[Code executed successfully with no output]'
else:
ret = ''.join(outputs)
# Remove ANSI
ret = strip_ansi(ret)
if os.environ.get('DEBUG', False):
logger.info(f'OUTPUT:\n{ret}')
return ret
async def shutdown_async(self):
if self.kernel_id:
client = AsyncHTTPClient()
await client.fetch(
'{}/api/kernels/{}'.format(self.base_url, self.kernel_id),
method='DELETE',
)
self.kernel_id = None
if self.ws:
self.ws.close()
self.ws = None

View File

@ -37,9 +37,6 @@ class LocalBox(Sandbox):
except subprocess.TimeoutExpired:
return -1, 'Command timed out'
def execute_python(self, code: str) -> str:
raise NotImplementedError('execute_python is not supported in LocalBox')
def execute_in_background(self, cmd: str) -> BackgroundCommand:
process = subprocess.Popen(
cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT,

View File

@ -131,10 +131,6 @@ class Sandbox(ABC):
def execute(self, cmd: str) -> Tuple[int, str]:
pass
@abstractmethod
def execute_python(self, code: str) -> str:
pass
@abstractmethod
def execute_in_background(self, cmd: str):
pass

View File

@ -4,7 +4,6 @@ import platform
import sys
import time
import uuid
import asyncio
from collections import namedtuple
from typing import Dict, List, Tuple, Union
@ -17,7 +16,6 @@ from opendevin.sandbox.sandbox import Sandbox, BackgroundCommand
from opendevin.schema import ConfigType
from opendevin.utils import find_available_tcp_port
from opendevin.exceptions import SandboxInvalidBackgroundCommandError
from opendevin.sandbox.jupyter_kernel import JupyterKernel
InputType = namedtuple('InputType', ['content'])
OutputType = namedtuple('OutputType', ['content'])
@ -267,11 +265,6 @@ class DockerSSHBox(Sandbox):
except docker.errors.NotFound:
return False
def _get_port_mapping(self):
return {
f'{self._ssh_port}/tcp': self._ssh_port
}
def restart_docker_container(self):
try:
self.stop_docker_container()
@ -286,10 +279,9 @@ class DockerSSHBox(Sandbox):
network_kwargs['network_mode'] = 'host'
else:
# FIXME: This is a temporary workaround for Mac OS
network_kwargs['ports'] = self._get_port_mapping()
network_kwargs['ports'] = {f'{self._ssh_port}/tcp': self._ssh_port}
logger.warning(
('Using port forwarding. '
('Using port forwarding for Mac OS. '
'Server started by OpenDevin will not be accessible from the host machine at the moment. '
'See https://github.com/OpenDevin/OpenDevin/issues/897 for more information.'
)
@ -348,56 +340,11 @@ class DockerSSHBox(Sandbox):
except docker.errors.NotFound:
pass
def execute_python(self, code: str) -> str:
raise NotImplementedError('execute_python is not supported in DockerSSHBox. Please use DockerSSHJupyterBox.')
class DockerSSHJupyterBox(DockerSSHBox):
_jupyter_port: int
def __init__(
self,
container_image: str | None = None,
timeout: int = 120,
sid: str | None = None,
):
self._jupyter_port = find_available_tcp_port()
super().__init__(container_image, timeout, sid)
self.setup_jupyter()
def _get_port_mapping(self):
return {
f'{self._ssh_port}/tcp': self._ssh_port,
'8888/tcp': self._jupyter_port,
}
def setup_jupyter(self):
# Setup Jupyter
self.jupyer_background_cmd = self.execute_in_background(
'jupyter kernelgateway --KernelGatewayApp.ip=0.0.0.0 --KernelGatewayApp.port=8888'
)
self.jupyter_kernel = JupyterKernel(
url_suffix=f'{SSH_HOSTNAME}:{self._jupyter_port}',
convid=self.instance_id,
)
logger.info(f'Jupyter Kernel Gateway started at {SSH_HOSTNAME}:{self._jupyter_port}: {self.jupyer_background_cmd.read_logs()}')
# initialize the kernel
logger.info('Initializing Jupyter Kernel Gateway...')
time.sleep(1) # wait for the kernel to start
loop = asyncio.get_event_loop()
loop.run_until_complete(self.jupyter_kernel.initialize())
logger.info('Jupyter Kernel Gateway initialized')
def execute_python(self, code: str) -> str:
loop = asyncio.get_event_loop()
return loop.run_until_complete(self.jupyter_kernel.execute(code))
if __name__ == '__main__':
try:
ssh_box = DockerSSHJupyterBox()
ssh_box = DockerSSHBox()
except Exception as e:
logger.exception('Failed to start Docker container: %s', e)
sys.exit(1)
@ -406,7 +353,7 @@ if __name__ == '__main__':
"Interactive Docker container started. Type 'exit' or use Ctrl+C to exit.")
bg_cmd = ssh_box.execute_in_background(
"while true; do echo 'dot ' && sleep 5; done"
"while true; do echo 'dot ' && sleep 1; done"
)
sys.stdout.flush()
@ -424,12 +371,6 @@ if __name__ == '__main__':
ssh_box.kill_background(bg_cmd.id)
logger.info('Background process killed')
continue
if user_input.startswith('py:'):
output = ssh_box.execute_python(user_input[3:])
logger.info(output)
sys.stdout.flush()
continue
print('JUPYTER LOG:', ssh_box.jupyer_background_cmd.read_logs())
exit_code, output = ssh_box.execute(user_input)
logger.info('exit code: %d', exit_code)
logger.info(output)

24
poetry.lock generated
View File

@ -3391,7 +3391,6 @@ optional = false
python-versions = ">=3.9"
files = [
{file = "pandas-2.2.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:90c6fca2acf139569e74e8781709dccb6fe25940488755716d1d354d6bc58bce"},
{file = "pandas-2.2.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:c7adfc142dac335d8c1e0dcbd37eb8617eac386596eb9e1a1b77791cf2498238"},
{file = "pandas-2.2.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4abfe0be0d7221be4f12552995e58723c7422c80a659da13ca382697de830c08"},
{file = "pandas-2.2.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8635c16bf3d99040fdf3ca3db669a7250ddf49c55dc4aa8fe0ae0fa8d6dcc1f0"},
{file = "pandas-2.2.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:40ae1dffb3967a52203105a077415a86044a2bea011b5f321c6aa64b379a3f51"},
@ -3412,7 +3411,6 @@ files = [
{file = "pandas-2.2.2-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:43498c0bdb43d55cb162cdc8c06fac328ccb5d2eabe3cadeb3529ae6f0517c32"},
{file = "pandas-2.2.2-cp312-cp312-win_amd64.whl", hash = "sha256:d187d355ecec3629624fccb01d104da7d7f391db0311145817525281e2804d23"},
{file = "pandas-2.2.2-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:0ca6377b8fca51815f382bd0b697a0814c8bda55115678cbc94c30aacbb6eff2"},
{file = "pandas-2.2.2-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:9057e6aa78a584bc93a13f0a9bf7e753a5e9770a30b4d758b8d5f2a62a9433cd"},
{file = "pandas-2.2.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:001910ad31abc7bf06f49dcc903755d2f7f3a9186c0c040b827e522e9cef0863"},
{file = "pandas-2.2.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:66b479b0bd07204e37583c191535505410daa8df638fd8e75ae1b383851fe921"},
{file = "pandas-2.2.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:a77e9d1c386196879aa5eb712e77461aaee433e54c68cf253053a73b7e49c33a"},
@ -5178,26 +5176,6 @@ typing-extensions = ">=4.8.0"
opt-einsum = ["opt-einsum (>=3.3)"]
optree = ["optree (>=0.9.1)"]
[[package]]
name = "tornado"
version = "6.4"
description = "Tornado is a Python web framework and asynchronous networking library, originally developed at FriendFeed."
optional = false
python-versions = ">= 3.8"
files = [
{file = "tornado-6.4-cp38-abi3-macosx_10_9_universal2.whl", hash = "sha256:02ccefc7d8211e5a7f9e8bc3f9e5b0ad6262ba2fbb683a6443ecc804e5224ce0"},
{file = "tornado-6.4-cp38-abi3-macosx_10_9_x86_64.whl", hash = "sha256:27787de946a9cffd63ce5814c33f734c627a87072ec7eed71f7fc4417bb16263"},
{file = "tornado-6.4-cp38-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f7894c581ecdcf91666a0912f18ce5e757213999e183ebfc2c3fdbf4d5bd764e"},
{file = "tornado-6.4-cp38-abi3-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e43bc2e5370a6a8e413e1e1cd0c91bedc5bd62a74a532371042a18ef19e10579"},
{file = "tornado-6.4-cp38-abi3-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f0251554cdd50b4b44362f73ad5ba7126fc5b2c2895cc62b14a1c2d7ea32f212"},
{file = "tornado-6.4-cp38-abi3-musllinux_1_1_aarch64.whl", hash = "sha256:fd03192e287fbd0899dd8f81c6fb9cbbc69194d2074b38f384cb6fa72b80e9c2"},
{file = "tornado-6.4-cp38-abi3-musllinux_1_1_i686.whl", hash = "sha256:88b84956273fbd73420e6d4b8d5ccbe913c65d31351b4c004ae362eba06e1f78"},
{file = "tornado-6.4-cp38-abi3-musllinux_1_1_x86_64.whl", hash = "sha256:71ddfc23a0e03ef2df1c1397d859868d158c8276a0603b96cf86892bff58149f"},
{file = "tornado-6.4-cp38-abi3-win32.whl", hash = "sha256:6f8a6c77900f5ae93d8b4ae1196472d0ccc2775cc1dfdc9e7727889145c45052"},
{file = "tornado-6.4-cp38-abi3-win_amd64.whl", hash = "sha256:10aeaa8006333433da48dec9fe417877f8bcc21f48dda8d661ae79da357b2a63"},
{file = "tornado-6.4.tar.gz", hash = "sha256:72291fa6e6bc84e626589f1c29d90a5a6d593ef5ae68052ee2ef000dfd273dee"},
]
[[package]]
name = "tqdm"
version = "4.66.2"
@ -5992,4 +5970,4 @@ testing = ["big-O", "jaraco.functools", "jaraco.itertools", "more-itertools", "p
[metadata]
lock-version = "2.0"
python-versions = "^3.11"
content-hash = "ba00c1217b404cad1139884cdaa2072ab9c416ff95aff1d5f525e562ee7d1350"
content-hash = "3a5ca3c8b47e0e43994032d1620d85a8d602c52a93790d192b9fdb3a8ac36d97"

View File

@ -24,7 +24,6 @@ numpy = "*"
json-repair = "*"
playwright = "*"
pexpect = "*"
tornado = "*"
[tool.poetry.group.llama-index.dependencies]
llama-index = "*"