| | |
| | """ |
| | Code Implementation MCP Server |
| | |
| | This MCP server provides core functions needed for paper code reproduction: |
| | 1. File read/write operations |
| | 2. Code execution and testing |
| | 3. Code search and analysis |
| | 4. Iterative improvement support |
| | |
| | Usage: |
| | python tools/code_implementation_server.py |
| | """ |
| |
|
| | import os |
| | import subprocess |
| | import json |
| | import sys |
| | import io |
| | from pathlib import Path |
| | import re |
| | from typing import Dict, Any, List |
| | import tempfile |
| | import shutil |
| | import logging |
| | from datetime import datetime |
| |
|
| | |
| | if sys.stdout.encoding != "utf-8": |
| | try: |
| | if hasattr(sys.stdout, "reconfigure"): |
| | sys.stdout.reconfigure(encoding="utf-8") |
| | sys.stderr.reconfigure(encoding="utf-8") |
| | else: |
| | sys.stdout = io.TextIOWrapper(sys.stdout.detach(), encoding="utf-8") |
| | sys.stderr = io.TextIOWrapper(sys.stderr.detach(), encoding="utf-8") |
| | except Exception as e: |
| | print(f"Warning: Could not set UTF-8 encoding: {e}") |
| |
|
| | |
| | from mcp.server.fastmcp import FastMCP |
| |
|
| | |
| | logging.basicConfig(level=logging.INFO) |
| | logger = logging.getLogger(__name__) |
| |
|
| | |
| | mcp = FastMCP("code-implementation-server") |
| |
|
| | |
| | WORKSPACE_DIR = None |
| | OPERATION_HISTORY = [] |
| | CURRENT_FILES = {} |
| |
|
| |
|
| | def initialize_workspace(workspace_dir: str = None): |
| | """ |
| | Initialize workspace |
| | |
| | By default, the workspace will be set by the workflow via the set_workspace tool to: |
| | {plan_file_parent}/generate_code |
| | |
| | Args: |
| | workspace_dir: Optional workspace directory path |
| | """ |
| | global WORKSPACE_DIR |
| | if workspace_dir is None: |
| | |
| | |
| | WORKSPACE_DIR = Path.cwd() / "generate_code" |
| | |
| | |
| | else: |
| | WORKSPACE_DIR = Path(workspace_dir).resolve() |
| | |
| | WORKSPACE_DIR.mkdir(parents=True, exist_ok=True) |
| | logger.info(f"Workspace initialized: {WORKSPACE_DIR}") |
| |
|
| |
|
| | def ensure_workspace_exists(): |
| | """Ensure workspace directory exists""" |
| | global WORKSPACE_DIR |
| | if WORKSPACE_DIR is None: |
| | initialize_workspace() |
| |
|
| | |
| | if not WORKSPACE_DIR.exists(): |
| | WORKSPACE_DIR.mkdir(parents=True, exist_ok=True) |
| | logger.info(f"Workspace directory created: {WORKSPACE_DIR}") |
| |
|
| |
|
| | def validate_path(path: str) -> Path: |
| | """Validate if path is within workspace""" |
| | if WORKSPACE_DIR is None: |
| | initialize_workspace() |
| |
|
| | full_path = (WORKSPACE_DIR / path).resolve() |
| | if not str(full_path).startswith(str(WORKSPACE_DIR)): |
| | raise ValueError(f"Path {path} is outside workspace scope") |
| | return full_path |
| |
|
| |
|
| | def log_operation(action: str, details: Dict[str, Any]): |
| | """Log operation history""" |
| | OPERATION_HISTORY.append( |
| | {"timestamp": datetime.now().isoformat(), "action": action, "details": details} |
| | ) |
| |
|
| |
|
| | |
| |
|
| |
|
| | @mcp.tool() |
| | async def read_file( |
| | file_path: str, start_line: int = None, end_line: int = None |
| | ) -> str: |
| | """ |
| | Read file content, supports specifying line number range |
| | |
| | Args: |
| | file_path: File path, relative to workspace |
| | start_line: Starting line number (1-based, optional) |
| | end_line: Ending line number (1-based, optional) |
| | |
| | Returns: |
| | JSON string of file content or error message |
| | """ |
| | try: |
| | full_path = validate_path(file_path) |
| |
|
| | if not full_path.exists(): |
| | result = {"status": "error", "message": f"File does not exist: {file_path}"} |
| | log_operation( |
| | "read_file_error", {"file_path": file_path, "error": "file_not_found"} |
| | ) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | with open(full_path, "r", encoding="utf-8") as f: |
| | lines = f.readlines() |
| |
|
| | |
| | if start_line is not None or end_line is not None: |
| | start_idx = (start_line - 1) if start_line else 0 |
| | end_idx = end_line if end_line else len(lines) |
| | lines = lines[start_idx:end_idx] |
| |
|
| | content = "".join(lines) |
| |
|
| | result = { |
| | "status": "success", |
| | "content": content, |
| | "file_path": file_path, |
| | "total_lines": len(lines), |
| | "size_bytes": len(content.encode("utf-8")), |
| | } |
| |
|
| | log_operation( |
| | "read_file", |
| | { |
| | "file_path": file_path, |
| | "start_line": start_line, |
| | "end_line": end_line, |
| | "lines_read": len(lines), |
| | }, |
| | ) |
| |
|
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Failed to read file: {str(e)}", |
| | "file_path": file_path, |
| | } |
| | log_operation("read_file_error", {"file_path": file_path, "error": str(e)}) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | @mcp.tool() |
| | async def read_multiple_files(file_requests: str, max_files: int = 5) -> str: |
| | """ |
| | Read multiple files in a single operation (for batch reading) |
| | |
| | Args: |
| | file_requests: JSON string with file requests, e.g., |
| | '{"file1.py": {}, "file2.py": {"start_line": 1, "end_line": 10}}' |
| | or simple array: '["file1.py", "file2.py"]' |
| | max_files: Maximum number of files to read in one operation (default: 5) |
| | |
| | Returns: |
| | JSON string of operation results for all files |
| | """ |
| | try: |
| | |
| | try: |
| | requests_data = json.loads(file_requests) |
| | except json.JSONDecodeError as e: |
| | return json.dumps( |
| | { |
| | "status": "error", |
| | "message": f"Invalid JSON format for file_requests: {str(e)}", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | }, |
| | ensure_ascii=False, |
| | indent=2, |
| | ) |
| |
|
| | |
| | if isinstance(requests_data, list): |
| | |
| | normalized_requests = {file_path: {} for file_path in requests_data} |
| | elif isinstance(requests_data, dict): |
| | normalized_requests = requests_data |
| | else: |
| | return json.dumps( |
| | { |
| | "status": "error", |
| | "message": "file_requests must be a JSON object or array", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | }, |
| | ensure_ascii=False, |
| | indent=2, |
| | ) |
| |
|
| | |
| | if len(normalized_requests) == 0: |
| | return json.dumps( |
| | { |
| | "status": "error", |
| | "message": "No files provided for reading", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | }, |
| | ensure_ascii=False, |
| | indent=2, |
| | ) |
| |
|
| | if len(normalized_requests) > max_files: |
| | return json.dumps( |
| | { |
| | "status": "error", |
| | "message": f"Too many files provided ({len(normalized_requests)}), maximum is {max_files}", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | }, |
| | ensure_ascii=False, |
| | indent=2, |
| | ) |
| |
|
| | |
| | results = { |
| | "status": "success", |
| | "message": f"Successfully processed {len(normalized_requests)} files", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | "files_processed": len(normalized_requests), |
| | "files": {}, |
| | "summary": { |
| | "successful": 0, |
| | "failed": 0, |
| | "total_size_bytes": 0, |
| | "total_lines": 0, |
| | "files_not_found": 0, |
| | }, |
| | } |
| |
|
| | |
| | for file_path, options in normalized_requests.items(): |
| | try: |
| | full_path = validate_path(file_path) |
| | start_line = options.get("start_line") |
| | end_line = options.get("end_line") |
| |
|
| | if not full_path.exists(): |
| | results["files"][file_path] = { |
| | "status": "error", |
| | "message": f"File does not exist: {file_path}", |
| | "file_path": file_path, |
| | "content": "", |
| | "total_lines": 0, |
| | "size_bytes": 0, |
| | "start_line": start_line, |
| | "end_line": end_line, |
| | } |
| | results["summary"]["failed"] += 1 |
| | results["summary"]["files_not_found"] += 1 |
| | continue |
| |
|
| | with open(full_path, "r", encoding="utf-8") as f: |
| | lines = f.readlines() |
| |
|
| | |
| | original_line_count = len(lines) |
| | if start_line is not None or end_line is not None: |
| | start_idx = (start_line - 1) if start_line else 0 |
| | end_idx = end_line if end_line else len(lines) |
| | lines = lines[start_idx:end_idx] |
| |
|
| | content = "".join(lines) |
| | size_bytes = len(content.encode("utf-8")) |
| | lines_count = len(lines) |
| |
|
| | |
| | results["files"][file_path] = { |
| | "status": "success", |
| | "message": f"File read successfully: {file_path}", |
| | "file_path": file_path, |
| | "content": content, |
| | "total_lines": lines_count, |
| | "original_total_lines": original_line_count, |
| | "size_bytes": size_bytes, |
| | "start_line": start_line, |
| | "end_line": end_line, |
| | "line_range_applied": start_line is not None |
| | or end_line is not None, |
| | } |
| |
|
| | |
| | results["summary"]["successful"] += 1 |
| | results["summary"]["total_size_bytes"] += size_bytes |
| | results["summary"]["total_lines"] += lines_count |
| |
|
| | |
| | log_operation( |
| | "read_file_multi", |
| | { |
| | "file_path": file_path, |
| | "start_line": start_line, |
| | "end_line": end_line, |
| | "lines_read": lines_count, |
| | "size_bytes": size_bytes, |
| | "batch_operation": True, |
| | }, |
| | ) |
| |
|
| | except Exception as file_error: |
| | |
| | results["files"][file_path] = { |
| | "status": "error", |
| | "message": f"Failed to read file: {str(file_error)}", |
| | "file_path": file_path, |
| | "content": "", |
| | "total_lines": 0, |
| | "size_bytes": 0, |
| | "start_line": options.get("start_line"), |
| | "end_line": options.get("end_line"), |
| | } |
| |
|
| | results["summary"]["failed"] += 1 |
| |
|
| | |
| | log_operation( |
| | "read_file_multi_error", |
| | { |
| | "file_path": file_path, |
| | "error": str(file_error), |
| | "batch_operation": True, |
| | }, |
| | ) |
| |
|
| | |
| | if results["summary"]["failed"] > 0: |
| | if results["summary"]["successful"] > 0: |
| | results["status"] = "partial_success" |
| | results["message"] = ( |
| | f"Read {results['summary']['successful']} files successfully, {results['summary']['failed']} failed" |
| | ) |
| | else: |
| | results["status"] = "failed" |
| | results["message"] = ( |
| | f"All {results['summary']['failed']} files failed to read" |
| | ) |
| |
|
| | |
| | log_operation( |
| | "read_multiple_files", |
| | { |
| | "files_count": len(normalized_requests), |
| | "successful": results["summary"]["successful"], |
| | "failed": results["summary"]["failed"], |
| | "total_size_bytes": results["summary"]["total_size_bytes"], |
| | "status": results["status"], |
| | }, |
| | ) |
| |
|
| | return json.dumps(results, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Failed to read multiple files: {str(e)}", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | "files_processed": 0, |
| | } |
| | log_operation("read_multiple_files_error", {"error": str(e)}) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | @mcp.tool() |
| | async def write_file( |
| | file_path: str, content: str, create_dirs: bool = True, create_backup: bool = False |
| | ) -> str: |
| | """ |
| | Write content to file |
| | |
| | Args: |
| | file_path: File path, relative to workspace |
| | content: Content to write to file |
| | create_dirs: Whether to create directories if they don't exist |
| | create_backup: Whether to create backup file if file already exists |
| | |
| | Returns: |
| | JSON string of operation result |
| | """ |
| | try: |
| | full_path = validate_path(file_path) |
| |
|
| | |
| | if create_dirs: |
| | full_path.parent.mkdir(parents=True, exist_ok=True) |
| |
|
| | |
| | backup_created = False |
| | if full_path.exists() and create_backup: |
| | backup_path = full_path.with_suffix(full_path.suffix + ".backup") |
| | shutil.copy2(full_path, backup_path) |
| | backup_created = True |
| |
|
| | |
| | with open(full_path, "w", encoding="utf-8") as f: |
| | f.write(content) |
| |
|
| | |
| | CURRENT_FILES[file_path] = { |
| | "last_modified": datetime.now().isoformat(), |
| | "size_bytes": len(content.encode("utf-8")), |
| | "lines": len(content.split("\n")), |
| | } |
| |
|
| | result = { |
| | "status": "success", |
| | "message": f"File written successfully: {file_path}", |
| | "file_path": file_path, |
| | "size_bytes": len(content.encode("utf-8")), |
| | "lines_written": len(content.split("\n")), |
| | "backup_created": backup_created, |
| | } |
| |
|
| | log_operation( |
| | "write_file", |
| | { |
| | "file_path": file_path, |
| | "size_bytes": len(content.encode("utf-8")), |
| | "lines": len(content.split("\n")), |
| | "backup_created": backup_created, |
| | }, |
| | ) |
| |
|
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Failed to write file: {str(e)}", |
| | "file_path": file_path, |
| | } |
| | log_operation("write_file_error", {"file_path": file_path, "error": str(e)}) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | @mcp.tool() |
| | async def write_multiple_files( |
| | file_implementations: str, |
| | create_dirs: bool = True, |
| | create_backup: bool = False, |
| | max_files: int = 5, |
| | ) -> str: |
| | """ |
| | Write multiple files in a single operation (for batch implementation) |
| | |
| | Args: |
| | file_implementations: JSON string mapping file paths to content, e.g., |
| | '{"file1.py": "content1", "file2.py": "content2"}' |
| | create_dirs: Whether to create directories if they don't exist |
| | create_backup: Whether to create backup files if they already exist |
| | max_files: Maximum number of files to write in one operation (default: 5) |
| | |
| | Returns: |
| | JSON string of operation results for all files |
| | """ |
| | try: |
| | |
| | try: |
| | files_dict = json.loads(file_implementations) |
| | except json.JSONDecodeError as e: |
| | return json.dumps( |
| | { |
| | "status": "error", |
| | "message": f"Invalid JSON format for file_implementations: {str(e)}", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | }, |
| | ensure_ascii=False, |
| | indent=2, |
| | ) |
| |
|
| | |
| | if not isinstance(files_dict, dict): |
| | return json.dumps( |
| | { |
| | "status": "error", |
| | "message": "file_implementations must be a JSON object mapping file paths to content", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | }, |
| | ensure_ascii=False, |
| | indent=2, |
| | ) |
| |
|
| | if len(files_dict) == 0: |
| | return json.dumps( |
| | { |
| | "status": "error", |
| | "message": "No files provided for writing", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | }, |
| | ensure_ascii=False, |
| | indent=2, |
| | ) |
| |
|
| | if len(files_dict) > max_files: |
| | return json.dumps( |
| | { |
| | "status": "error", |
| | "message": f"Too many files provided ({len(files_dict)}), maximum is {max_files}", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | }, |
| | ensure_ascii=False, |
| | indent=2, |
| | ) |
| |
|
| | |
| | results = { |
| | "status": "success", |
| | "message": f"Successfully processed {len(files_dict)} files", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | "files_processed": len(files_dict), |
| | "files": {}, |
| | "summary": { |
| | "successful": 0, |
| | "failed": 0, |
| | "total_size_bytes": 0, |
| | "total_lines": 0, |
| | "backups_created": 0, |
| | }, |
| | } |
| |
|
| | |
| | for file_path, content in files_dict.items(): |
| | try: |
| | full_path = validate_path(file_path) |
| |
|
| | |
| | if create_dirs: |
| | full_path.parent.mkdir(parents=True, exist_ok=True) |
| |
|
| | |
| | backup_created = False |
| | if full_path.exists() and create_backup: |
| | backup_path = full_path.with_suffix(full_path.suffix + ".backup") |
| | shutil.copy2(full_path, backup_path) |
| | backup_created = True |
| | results["summary"]["backups_created"] += 1 |
| |
|
| | |
| | with open(full_path, "w", encoding="utf-8") as f: |
| | f.write(content) |
| |
|
| | |
| | size_bytes = len(content.encode("utf-8")) |
| | lines_count = len(content.split("\n")) |
| |
|
| | |
| | CURRENT_FILES[file_path] = { |
| | "last_modified": datetime.now().isoformat(), |
| | "size_bytes": size_bytes, |
| | "lines": lines_count, |
| | } |
| |
|
| | |
| | results["files"][file_path] = { |
| | "status": "success", |
| | "message": f"File written successfully: {file_path}", |
| | "size_bytes": size_bytes, |
| | "lines_written": lines_count, |
| | "backup_created": backup_created, |
| | } |
| |
|
| | |
| | results["summary"]["successful"] += 1 |
| | results["summary"]["total_size_bytes"] += size_bytes |
| | results["summary"]["total_lines"] += lines_count |
| |
|
| | |
| | log_operation( |
| | "write_file_multi", |
| | { |
| | "file_path": file_path, |
| | "size_bytes": size_bytes, |
| | "lines": lines_count, |
| | "backup_created": backup_created, |
| | "batch_operation": True, |
| | }, |
| | ) |
| |
|
| | except Exception as file_error: |
| | |
| | results["files"][file_path] = { |
| | "status": "error", |
| | "message": f"Failed to write file: {str(file_error)}", |
| | "size_bytes": 0, |
| | "lines_written": 0, |
| | "backup_created": False, |
| | } |
| |
|
| | results["summary"]["failed"] += 1 |
| |
|
| | |
| | log_operation( |
| | "write_file_multi_error", |
| | { |
| | "file_path": file_path, |
| | "error": str(file_error), |
| | "batch_operation": True, |
| | }, |
| | ) |
| |
|
| | |
| | if results["summary"]["failed"] > 0: |
| | if results["summary"]["successful"] > 0: |
| | results["status"] = "partial_success" |
| | results["message"] = ( |
| | f"Processed {results['summary']['successful']} files successfully, {results['summary']['failed']} failed" |
| | ) |
| | else: |
| | results["status"] = "failed" |
| | results["message"] = ( |
| | f"All {results['summary']['failed']} files failed to write" |
| | ) |
| |
|
| | |
| | log_operation( |
| | "write_multiple_files", |
| | { |
| | "files_count": len(files_dict), |
| | "successful": results["summary"]["successful"], |
| | "failed": results["summary"]["failed"], |
| | "total_size_bytes": results["summary"]["total_size_bytes"], |
| | "status": results["status"], |
| | }, |
| | ) |
| |
|
| | return json.dumps(results, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Failed to write multiple files: {str(e)}", |
| | "operation_type": "multi_file", |
| | "timestamp": datetime.now().isoformat(), |
| | "files_processed": 0, |
| | } |
| | log_operation("write_multiple_files_error", {"error": str(e)}) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | |
| |
|
| |
|
| | @mcp.tool() |
| | async def execute_python(code: str, timeout: int = 30) -> str: |
| | """ |
| | Execute Python code and return output |
| | |
| | Args: |
| | code: Python code to execute |
| | timeout: Timeout in seconds |
| | |
| | Returns: |
| | JSON string of execution result |
| | """ |
| | try: |
| | |
| | with tempfile.NamedTemporaryFile( |
| | mode="w", suffix=".py", delete=False, encoding="utf-8" |
| | ) as f: |
| | f.write(code) |
| | temp_file = f.name |
| |
|
| | try: |
| | |
| | ensure_workspace_exists() |
| |
|
| | |
| | result = subprocess.run( |
| | [sys.executable, temp_file], |
| | cwd=WORKSPACE_DIR, |
| | capture_output=True, |
| | text=True, |
| | timeout=timeout, |
| | encoding="utf-8", |
| | ) |
| |
|
| | execution_result = { |
| | "status": "success" if result.returncode == 0 else "error", |
| | "return_code": result.returncode, |
| | "stdout": result.stdout, |
| | "stderr": result.stderr, |
| | "timeout": timeout, |
| | } |
| |
|
| | if result.returncode != 0: |
| | execution_result["message"] = "Python code execution failed" |
| | else: |
| | execution_result["message"] = "Python code execution successful" |
| |
|
| | log_operation( |
| | "execute_python", |
| | { |
| | "return_code": result.returncode, |
| | "stdout_length": len(result.stdout), |
| | "stderr_length": len(result.stderr), |
| | }, |
| | ) |
| |
|
| | return json.dumps(execution_result, ensure_ascii=False, indent=2) |
| |
|
| | finally: |
| | |
| | os.unlink(temp_file) |
| |
|
| | except subprocess.TimeoutExpired: |
| | result = { |
| | "status": "error", |
| | "message": f"Python code execution timeout ({timeout}秒)", |
| | "timeout": timeout, |
| | } |
| | log_operation("execute_python_timeout", {"timeout": timeout}) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Python code execution failed: {str(e)}", |
| | } |
| | log_operation("execute_python_error", {"error": str(e)}) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | @mcp.tool() |
| | async def execute_bash(command: str, timeout: int = 30) -> str: |
| | """ |
| | Execute bash command |
| | |
| | Args: |
| | command: Bash command to execute |
| | timeout: Timeout in seconds |
| | |
| | Returns: |
| | JSON string of execution result |
| | """ |
| | try: |
| | |
| | dangerous_commands = ["rm -rf", "sudo", "chmod 777", "mkfs", "dd if="] |
| | if any(dangerous in command.lower() for dangerous in dangerous_commands): |
| | result = { |
| | "status": "error", |
| | "message": f"Dangerous command execution prohibited: {command}", |
| | } |
| | log_operation( |
| | "execute_bash_blocked", |
| | {"command": command, "reason": "dangerous_command"}, |
| | ) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | |
| | ensure_workspace_exists() |
| |
|
| | |
| | result = subprocess.run( |
| | command, |
| | shell=True, |
| | cwd=WORKSPACE_DIR, |
| | capture_output=True, |
| | text=True, |
| | timeout=timeout, |
| | encoding="utf-8", |
| | ) |
| |
|
| | execution_result = { |
| | "status": "success" if result.returncode == 0 else "error", |
| | "return_code": result.returncode, |
| | "stdout": result.stdout, |
| | "stderr": result.stderr, |
| | "command": command, |
| | "timeout": timeout, |
| | } |
| |
|
| | if result.returncode != 0: |
| | execution_result["message"] = "Bash command execution failed" |
| | else: |
| | execution_result["message"] = "Bash command execution successful" |
| |
|
| | log_operation( |
| | "execute_bash", |
| | { |
| | "command": command, |
| | "return_code": result.returncode, |
| | "stdout_length": len(result.stdout), |
| | "stderr_length": len(result.stderr), |
| | }, |
| | ) |
| |
|
| | return json.dumps(execution_result, ensure_ascii=False, indent=2) |
| |
|
| | except subprocess.TimeoutExpired: |
| | result = { |
| | "status": "error", |
| | "message": f"Bash command execution timeout ({timeout} seconds)", |
| | "command": command, |
| | "timeout": timeout, |
| | } |
| | log_operation("execute_bash_timeout", {"command": command, "timeout": timeout}) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Failed to execute bash command: {str(e)}", |
| | "command": command, |
| | } |
| | log_operation("execute_bash_error", {"command": command, "error": str(e)}) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | @mcp.tool() |
| | async def read_code_mem(file_paths: List[str]) -> str: |
| | """ |
| | Check if file summaries exist in implement_code_summary.md for multiple files |
| | |
| | Args: |
| | file_paths: List of file paths to check for summary information in implement_code_summary.md |
| | |
| | Returns: |
| | Summary information for all requested files if available |
| | """ |
| | try: |
| | if not file_paths or not isinstance(file_paths, list): |
| | result = { |
| | "status": "error", |
| | "message": "file_paths parameter is required and must be a list", |
| | } |
| | log_operation( |
| | "read_code_mem_error", {"error": "missing_or_invalid_file_paths"} |
| | ) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | |
| | unique_file_paths = list(dict.fromkeys(file_paths)) |
| |
|
| | |
| | ensure_workspace_exists() |
| |
|
| | |
| | current_path = Path(WORKSPACE_DIR) |
| | summary_file_path = current_path.parent / "implement_code_summary.md" |
| |
|
| | if not summary_file_path.exists(): |
| | result = { |
| | "status": "no_summary", |
| | "file_paths": unique_file_paths, |
| | "message": "No summary file found.", |
| | "results": [], |
| | } |
| | log_operation( |
| | "read_code_mem", |
| | {"file_paths": unique_file_paths, "status": "no_summary_file"}, |
| | ) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | |
| | with open(summary_file_path, "r", encoding="utf-8") as f: |
| | summary_content = f.read() |
| |
|
| | if not summary_content.strip(): |
| | result = { |
| | "status": "no_summary", |
| | "file_paths": unique_file_paths, |
| | "message": "Summary file is empty.", |
| | "results": [], |
| | } |
| | log_operation( |
| | "read_code_mem", |
| | {"file_paths": unique_file_paths, "status": "empty_summary"}, |
| | ) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | |
| | results = [] |
| | summaries_found = 0 |
| |
|
| | for file_path in unique_file_paths: |
| | |
| | file_section = _extract_file_section_from_summary( |
| | summary_content, file_path |
| | ) |
| |
|
| | if file_section: |
| | file_result = { |
| | "file_path": file_path, |
| | "status": "summary_found", |
| | "summary_content": file_section, |
| | "message": f"Summary information found for {file_path}", |
| | } |
| | summaries_found += 1 |
| | else: |
| | file_result = { |
| | "file_path": file_path, |
| | "status": "no_summary", |
| | "summary_content": None, |
| | "message": f"No summary found for {file_path}", |
| | } |
| |
|
| | results.append(file_result) |
| |
|
| | |
| | if summaries_found == len(unique_file_paths): |
| | overall_status = "all_summaries_found" |
| | elif summaries_found > 0: |
| | overall_status = "partial_summaries_found" |
| | else: |
| | overall_status = "no_summaries_found" |
| |
|
| | result = { |
| | "status": overall_status, |
| | "file_paths": unique_file_paths, |
| | "total_requested": len(unique_file_paths), |
| | "summaries_found": summaries_found, |
| | "message": f"Found summaries for {summaries_found}/{len(unique_file_paths)} files", |
| | "results": results, |
| | } |
| |
|
| | log_operation( |
| | "read_code_mem", |
| | { |
| | "file_paths": unique_file_paths, |
| | "status": overall_status, |
| | "total_requested": len(unique_file_paths), |
| | "summaries_found": summaries_found, |
| | }, |
| | ) |
| |
|
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Failed to check code memory: {str(e)}", |
| | "file_paths": file_paths |
| | if isinstance(file_paths, list) |
| | else [str(file_paths)], |
| | "results": [], |
| | } |
| | log_operation( |
| | "read_code_mem_error", {"file_paths": file_paths, "error": str(e)} |
| | ) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | def _extract_file_section_from_summary( |
| | summary_content: str, target_file_path: str |
| | ) -> str: |
| | """ |
| | Extract the specific section for a file from the summary content |
| | |
| | Args: |
| | summary_content: Full summary content |
| | target_file_path: Path of the target file |
| | |
| | Returns: |
| | File-specific section or None if not found |
| | """ |
| | import re |
| |
|
| | |
| | normalized_target = _normalize_file_path(target_file_path) |
| |
|
| | |
| | section_pattern = r"={80}\s*\n## IMPLEMENTATION File ([^;]+); ROUND \d+\s*\n={80}(.*?)(?=\n={80}|\Z)" |
| |
|
| | matches = re.findall(section_pattern, summary_content, re.DOTALL) |
| |
|
| | for file_path_in_summary, section_content in matches: |
| | file_path_in_summary = file_path_in_summary.strip() |
| | section_content = section_content.strip() |
| |
|
| | |
| | normalized_summary_path = _normalize_file_path(file_path_in_summary) |
| |
|
| | |
| | if _paths_match( |
| | normalized_target, |
| | normalized_summary_path, |
| | target_file_path, |
| | file_path_in_summary, |
| | ): |
| | |
| | file_section = f"""================================================================================ |
| | ## IMPLEMENTATION File {file_path_in_summary}; ROUND [X] |
| | ================================================================================ |
| | |
| | {section_content} |
| | |
| | --- |
| | *Extracted from implement_code_summary.md*""" |
| | return file_section |
| |
|
| | |
| | return _extract_file_section_alternative(summary_content, target_file_path) |
| |
|
| |
|
| | def _normalize_file_path(file_path: str) -> str: |
| | """Normalize file path for comparison""" |
| | |
| | normalized = file_path.strip("/").lower() |
| | |
| | normalized = normalized.replace("\\", "/") |
| |
|
| | |
| | common_prefixes = ["src/", "./src/", "./", "core/", "lib/", "main/"] |
| | for prefix in common_prefixes: |
| | if normalized.startswith(prefix): |
| | normalized = normalized[len(prefix) :] |
| | break |
| |
|
| | return normalized |
| |
|
| |
|
| | def _paths_match( |
| | normalized_target: str, |
| | normalized_summary: str, |
| | original_target: str, |
| | original_summary: str, |
| | ) -> bool: |
| | """Check if two file paths match using multiple strategies""" |
| |
|
| | |
| | if normalized_target == normalized_summary: |
| | return True |
| |
|
| | |
| | target_basename = os.path.basename(original_target) |
| | summary_basename = os.path.basename(original_summary) |
| | if target_basename == summary_basename and len(target_basename) > 4: |
| | return True |
| |
|
| | |
| | target_suffix = _remove_common_prefixes(normalized_target) |
| | summary_suffix = _remove_common_prefixes(normalized_summary) |
| | if target_suffix == summary_suffix: |
| | return True |
| |
|
| | |
| | if normalized_target.endswith(normalized_summary) or normalized_summary.endswith( |
| | normalized_target |
| | ): |
| | return True |
| |
|
| | |
| | if len(normalized_target) > 10 and normalized_target in normalized_summary: |
| | return True |
| | if len(normalized_summary) > 10 and normalized_summary in normalized_target: |
| | return True |
| |
|
| | return False |
| |
|
| |
|
| | def _remove_common_prefixes(file_path: str) -> str: |
| | """Remove common prefixes from file path""" |
| | prefixes_to_remove = ["src/", "core/", "./", "lib/", "main/"] |
| | path = file_path |
| |
|
| | for prefix in prefixes_to_remove: |
| | if path.startswith(prefix): |
| | path = path[len(prefix) :] |
| |
|
| | return path |
| |
|
| |
|
| | def _extract_file_section_alternative( |
| | summary_content: str, target_file_path: str |
| | ) -> str: |
| | """Alternative method to extract file section using simpler pattern matching""" |
| |
|
| | |
| | target_basename = os.path.basename(target_file_path) |
| |
|
| | |
| | sections = summary_content.split("=" * 80) |
| |
|
| | for i, section in enumerate(sections): |
| | if "## IMPLEMENTATION File" in section: |
| | |
| | lines = section.strip().split("\n") |
| | for line in lines: |
| | if "## IMPLEMENTATION File" in line: |
| | |
| | try: |
| | file_part = line.split("File ")[1].split("; ROUND")[0].strip() |
| |
|
| | |
| | if ( |
| | _normalize_file_path(target_file_path) |
| | == _normalize_file_path(file_part) |
| | or target_basename == os.path.basename(file_part) |
| | or target_file_path in file_part |
| | or file_part.endswith(target_file_path) |
| | ): |
| | |
| | if i + 1 < len(sections): |
| | content_section = sections[i + 1].strip() |
| | return f"""================================================================================ |
| | ## IMPLEMENTATION File {file_part} |
| | ================================================================================ |
| | |
| | {content_section} |
| | |
| | --- |
| | *Extracted from implement_code_summary.md using alternative method*""" |
| | except (IndexError, AttributeError): |
| | continue |
| |
|
| | return None |
| |
|
| |
|
| | |
| |
|
| |
|
| | @mcp.tool() |
| | async def search_code( |
| | pattern: str, |
| | file_pattern: str = "*.json", |
| | use_regex: bool = False, |
| | search_directory: str = None, |
| | ) -> str: |
| | """ |
| | Search patterns in code files |
| | |
| | Args: |
| | pattern: Search pattern |
| | file_pattern: File pattern (e.g., '*.py') |
| | use_regex: Whether to use regular expressions |
| | search_directory: Specify search directory (optional, uses WORKSPACE_DIR if not specified) |
| | |
| | Returns: |
| | JSON string of search results |
| | """ |
| | try: |
| | |
| | if search_directory: |
| | |
| | if os.path.isabs(search_directory): |
| | search_path = Path(search_directory) |
| | else: |
| | |
| | search_path = Path.cwd() / search_directory |
| | else: |
| | |
| | ensure_workspace_exists() |
| | search_path = WORKSPACE_DIR |
| |
|
| | |
| | if not search_path.exists(): |
| | result = { |
| | "status": "error", |
| | "message": f"Search directory不存在: {search_path}", |
| | "pattern": pattern, |
| | } |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | import glob |
| |
|
| | |
| | file_paths = glob.glob(str(search_path / "**" / file_pattern), recursive=True) |
| |
|
| | matches = [] |
| | total_files_searched = 0 |
| |
|
| | for file_path in file_paths: |
| | try: |
| | with open(file_path, "r", encoding="utf-8") as f: |
| | lines = f.readlines() |
| |
|
| | total_files_searched += 1 |
| | relative_path = os.path.relpath(file_path, search_path) |
| |
|
| | for line_num, line in enumerate(lines, 1): |
| | if use_regex: |
| | if re.search(pattern, line): |
| | matches.append( |
| | { |
| | "file": relative_path, |
| | "line_number": line_num, |
| | "line_content": line.strip(), |
| | "match_type": "regex", |
| | } |
| | ) |
| | else: |
| | if pattern.lower() in line.lower(): |
| | matches.append( |
| | { |
| | "file": relative_path, |
| | "line_number": line_num, |
| | "line_content": line.strip(), |
| | "match_type": "substring", |
| | } |
| | ) |
| |
|
| | except Exception as e: |
| | logger.warning(f"Error searching file {file_path}: {e}") |
| | continue |
| |
|
| | result = { |
| | "status": "success", |
| | "pattern": pattern, |
| | "file_pattern": file_pattern, |
| | "use_regex": use_regex, |
| | "search_directory": str(search_path), |
| | "total_matches": len(matches), |
| | "total_files_searched": total_files_searched, |
| | "matches": matches[:50], |
| | } |
| |
|
| | if len(matches) > 50: |
| | result["note"] = f"显示前50个匹配,总共找到{len(matches)}个匹配" |
| |
|
| | log_operation( |
| | "search_code", |
| | { |
| | "pattern": pattern, |
| | "file_pattern": file_pattern, |
| | "use_regex": use_regex, |
| | "search_directory": str(search_path), |
| | "total_matches": len(matches), |
| | "files_searched": total_files_searched, |
| | }, |
| | ) |
| |
|
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Code search failed: {str(e)}", |
| | "pattern": pattern, |
| | } |
| | log_operation("search_code_error", {"pattern": pattern, "error": str(e)}) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | |
| |
|
| |
|
| | @mcp.tool() |
| | async def get_file_structure(directory: str = ".", max_depth: int = 5) -> str: |
| | """ |
| | Get directory file structure |
| | |
| | Args: |
| | directory: Directory path, relative to workspace |
| | max_depth: 最大遍历深度 |
| | |
| | Returns: |
| | JSON string of file structure |
| | """ |
| | try: |
| | ensure_workspace_exists() |
| |
|
| | if directory == ".": |
| | target_dir = WORKSPACE_DIR |
| | else: |
| | target_dir = validate_path(directory) |
| |
|
| | if not target_dir.exists(): |
| | result = { |
| | "status": "error", |
| | "message": f"Directory does not exist: {directory}", |
| | } |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | def scan_directory(path: Path, current_depth: int = 0) -> Dict[str, Any]: |
| | """Recursively scan directory""" |
| | if current_depth >= max_depth: |
| | return {"type": "directory", "name": path.name, "truncated": True} |
| |
|
| | items = [] |
| | try: |
| | for item in sorted(path.iterdir()): |
| | relative_path = os.path.relpath(item, WORKSPACE_DIR) |
| |
|
| | if item.is_file(): |
| | file_info = { |
| | "type": "file", |
| | "name": item.name, |
| | "path": relative_path, |
| | "size_bytes": item.stat().st_size, |
| | "extension": item.suffix, |
| | } |
| | items.append(file_info) |
| | elif item.is_dir() and not item.name.startswith("."): |
| | dir_info = scan_directory(item, current_depth + 1) |
| | dir_info["path"] = relative_path |
| | items.append(dir_info) |
| | except PermissionError: |
| | pass |
| |
|
| | return { |
| | "type": "directory", |
| | "name": path.name, |
| | "items": items, |
| | "item_count": len(items), |
| | } |
| |
|
| | structure = scan_directory(target_dir) |
| |
|
| | |
| | def count_items(node): |
| | if node["type"] == "file": |
| | return {"files": 1, "directories": 0} |
| | else: |
| | counts = {"files": 0, "directories": 1} |
| | for item in node.get("items", []): |
| | item_counts = count_items(item) |
| | counts["files"] += item_counts["files"] |
| | counts["directories"] += item_counts["directories"] |
| | return counts |
| |
|
| | counts = count_items(structure) |
| |
|
| | result = { |
| | "status": "success", |
| | "directory": directory, |
| | "max_depth": max_depth, |
| | "structure": structure, |
| | "summary": { |
| | "total_files": counts["files"], |
| | "total_directories": counts["directories"] |
| | - 1, |
| | }, |
| | } |
| |
|
| | log_operation( |
| | "get_file_structure", |
| | { |
| | "directory": directory, |
| | "max_depth": max_depth, |
| | "total_files": counts["files"], |
| | "total_directories": counts["directories"] - 1, |
| | }, |
| | ) |
| |
|
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Failed to get file structure: {str(e)}", |
| | "directory": directory, |
| | } |
| | log_operation( |
| | "get_file_structure_error", {"directory": directory, "error": str(e)} |
| | ) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | |
| |
|
| |
|
| | @mcp.tool() |
| | async def set_workspace(workspace_path: str) -> str: |
| | """ |
| | Set workspace directory |
| | |
| | Called by workflow to set workspace to: {plan_file_parent}/generate_code |
| | This ensures all file operations are executed relative to the correct project directory |
| | |
| | Args: |
| | workspace_path: Workspace path (Usually {plan_file_parent}/generate_code) |
| | |
| | Returns: |
| | JSON string of operation result |
| | """ |
| | try: |
| | global WORKSPACE_DIR |
| | new_workspace = Path(workspace_path).resolve() |
| |
|
| | |
| | new_workspace.mkdir(parents=True, exist_ok=True) |
| |
|
| | old_workspace = WORKSPACE_DIR |
| | WORKSPACE_DIR = new_workspace |
| |
|
| | logger.info(f"New Workspace: {WORKSPACE_DIR}") |
| |
|
| | result = { |
| | "status": "success", |
| | "message": f"Workspace setup successful: {workspace_path}", |
| | "new_workspace": str(WORKSPACE_DIR), |
| | } |
| |
|
| | log_operation( |
| | "set_workspace", |
| | { |
| | "old_workspace": str(old_workspace) if old_workspace else None, |
| | "new_workspace": str(WORKSPACE_DIR), |
| | "workspace_alignment": "plan_file_parent/generate_code", |
| | }, |
| | ) |
| |
|
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Failed to set workspace: {str(e)}", |
| | "workspace_path": workspace_path, |
| | } |
| | log_operation( |
| | "set_workspace_error", {"workspace_path": workspace_path, "error": str(e)} |
| | ) |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | @mcp.tool() |
| | async def get_operation_history(last_n: int = 10) -> str: |
| | """ |
| | Get operation history |
| | |
| | Args: |
| | last_n: Return the last N operations |
| | |
| | Returns: |
| | JSON string of operation history |
| | """ |
| | try: |
| | recent_history = ( |
| | OPERATION_HISTORY[-last_n:] if last_n > 0 else OPERATION_HISTORY |
| | ) |
| |
|
| | result = { |
| | "status": "success", |
| | "total_operations": len(OPERATION_HISTORY), |
| | "returned_operations": len(recent_history), |
| | "workspace": str(WORKSPACE_DIR) if WORKSPACE_DIR else None, |
| | "history": recent_history, |
| | } |
| |
|
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| | except Exception as e: |
| | result = { |
| | "status": "error", |
| | "message": f"Failed to get operation history: {str(e)}", |
| | } |
| | return json.dumps(result, ensure_ascii=False, indent=2) |
| |
|
| |
|
| | |
| |
|
| |
|
| | def main(): |
| | """Start MCP server""" |
| | print("🚀 Code Implementation MCP Server") |
| | print( |
| | "📝 Paper Code Implementation Tool Server / Paper Code Implementation Tool Server" |
| | ) |
| | print("") |
| | print("Available tools / Available tools:") |
| | |
| | print( |
| | " • read_code_mem - Read code summary from implement_code_summary.md / Read code summary from implement_code_summary.md" |
| | ) |
| | print(" • write_file - Write file contents / Write file contents") |
| | print(" • execute_python - Execute Python code / Execute Python code") |
| | print(" • execute_bash - Execute bash command / Execute bash commands") |
| | print(" • search_code - Search code patterns / Search code patterns") |
| | print(" • get_file_structure - Get file structure / Get file structure") |
| | print(" • set_workspace - Set workspace / Set workspace") |
| | print(" • get_operation_history - Get operation history / Get operation history") |
| | print("") |
| | print("🔧 Server starting...") |
| |
|
| | |
| | initialize_workspace() |
| |
|
| | |
| | mcp.run() |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|