Source code for camel.utils.commons

# ========= Copyright 2023-2024 @ CAMEL-AI.org. All Rights Reserved. =========
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ========= Copyright 2023-2024 @ CAMEL-AI.org. All Rights Reserved. =========
import functools
import importlib
import logging
import os
import platform
import re
import socket
import subprocess
import threading
import time
import zipfile
from functools import wraps
from http import HTTPStatus
from pathlib import Path
from typing import (
    Any,
    Callable,
    Dict,
    List,
    Mapping,
    Optional,
    Set,
    Tuple,
    Type,
    TypeVar,
    cast,
)
from urllib.parse import urlparse

import pydantic
import requests
from pydantic import BaseModel

from camel.types import TaskType

from .constants import Constants

F = TypeVar('F', bound=Callable[..., Any])

logger = logging.getLogger(__name__)






[docs] def get_prompt_template_key_words(template: str) -> Set[str]: r"""Given a string template containing curly braces {}, return a set of the words inside the braces. Args: template (str): A string containing curly braces. Returns: List[str]: A list of the words inside the curly braces. Example: >>> get_prompt_template_key_words('Hi, {name}! How are you {status}?') {'name', 'status'} """ return set(re.findall(r'{([^}]*)}', template))
[docs] def get_first_int(string: str) -> Optional[int]: r"""Returns the first integer number found in the given string. If no integer number is found, returns None. Args: string (str): The input string. Returns: int or None: The first integer number found in the string, or None if no integer number is found. """ match = re.search(r'\d+', string) if match: return int(match.group()) else: return None
[docs] def download_tasks(task: TaskType, folder_path: str) -> None: r"""Downloads task-related files from a specified URL and extracts them. This function downloads a zip file containing tasks based on the specified `task` type from a predefined URL, saves it to `folder_path`, and then extracts the contents of the zip file into the same folder. After extraction, the zip file is deleted. Args: task (TaskType): An enum representing the type of task to download. folder_path (str): The path of the folder where the zip file will be downloaded and extracted. """ # Define the path to save the zip file zip_file_path = os.path.join(folder_path, "tasks.zip") # Download the zip file from the Google Drive link response = requests.get( "https://huggingface.co/datasets/camel-ai/" f"metadata/resolve/main/{task.value}_tasks.zip" ) # Save the zip file with open(zip_file_path, "wb") as f: f.write(response.content) with zipfile.ZipFile(zip_file_path, "r") as zip_ref: zip_ref.extractall(folder_path) # Delete the zip file os.remove(zip_file_path)
[docs] def get_task_list(task_response: str) -> List[str]: r"""Parse the response of the Agent and return task list. Args: task_response (str): The string response of the Agent. Returns: List[str]: A list of the string tasks. """ new_tasks_list = [] task_string_list = task_response.strip().split('\n') # each task starts with #. for task_string in task_string_list: task_parts = task_string.strip().split(".", 1) if len(task_parts) == 2: task_id = ''.join(s for s in task_parts[0] if s.isnumeric()) task_name = re.sub(r'[^\w\s_]+', '', task_parts[1]).strip() if task_name.strip() and task_id.isnumeric(): new_tasks_list.append(task_name) return new_tasks_list
[docs] def check_server_running(server_url: str) -> bool: r"""Check whether the port refered by the URL to the server is open. Args: server_url (str): The URL to the server running LLM inference service. Returns: bool: Whether the port is open for packets (server is running). """ parsed_url = urlparse(server_url) url_tuple = (parsed_url.hostname, parsed_url.port) sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) result = sock.connect_ex(url_tuple) sock.close() # if the port is open, the result should be 0. return result == 0
[docs] def dependencies_required(*required_modules: str) -> Callable[[F], F]: r"""A decorator to ensure that specified Python modules are available before a function executes. Args: required_modules (str): The required modules to be checked for availability. Returns: Callable[[F], F]: The original function with the added check for required module dependencies. Raises: ImportError: If any of the required modules are not available. Example: :: @dependencies_required('numpy', 'pandas') def data_processing_function(): # Function implementation... """ def decorator(func: F) -> F: @wraps(func) def wrapper(*args: Any, **kwargs: Any) -> Any: missing_modules = [ m for m in required_modules if not is_module_available(m) ] if missing_modules: raise ImportError( f"Missing required modules: {', '.join(missing_modules)}" ) return func(*args, **kwargs) return cast(F, wrapper) return decorator
[docs] def is_module_available(module_name: str) -> bool: r"""Check if a module is available for import. Args: module_name (str): The name of the module to check for availability. Returns: bool: True if the module can be imported, False otherwise. """ try: importlib.import_module(module_name) return True except ImportError: return False
[docs] def api_keys_required( param_env_list: List[Tuple[Optional[str], str]], ) -> Callable[[F], F]: r"""A decorator to check if the required API keys are provided in the environment variables or as function arguments. Args: param_env_list (List[Tuple[Optional[str], str]]): A list of tuples where each tuple contains a function argument name (as the first element, or None) and the corresponding environment variable name (as the second element) that holds the API key. Returns: Callable[[F], F]: The original function wrapped with the added check for the required API keys. Raises: ValueError: If any of the required API keys are missing, either from the function arguments or environment variables. Example: :: @api_keys_required([ ('api_key_arg', 'API_KEY_1'), ('another_key_arg', 'API_KEY_2'), (None, 'API_KEY_3'), ]) def some_api_function(api_key_arg=None, another_key_arg=None): # Function implementation that requires API keys """ import inspect def decorator(func: F) -> F: @wraps(func) def wrapper(*args: Any, **kwargs: Any) -> Any: signature = inspect.signature(func) bound_arguments = signature.bind(*args, **kwargs) bound_arguments.apply_defaults() arguments = bound_arguments.arguments missing_keys = [] for param_name, env_var_name in param_env_list: if not isinstance(env_var_name, str): raise TypeError( f"Environment variable name must be a string, got" f" {type(env_var_name)}" ) value = None if ( param_name ): # If param_name is provided, check function argument first if not isinstance(param_name, str): raise TypeError( f"Parameter name must be a string, " f"got {type(param_name)}" ) value = arguments.get(param_name) # If we found a valid value in arguments, continue to next # item if value: continue # Check environment variable if no valid value found yet value = os.environ.get(env_var_name) if not value or value.strip() == "": missing_keys.append(env_var_name) if missing_keys: raise ValueError( "Missing or empty required API keys in " f"environment variables: {', '.join(missing_keys)}" ) return func(*args, **kwargs) return cast(F, wrapper) return decorator
[docs] def get_system_information(): r"""Gathers information about the operating system. Returns: dict: A dictionary containing various pieces of OS information. """ sys_info = { "OS Name": os.name, "System": platform.system(), "Release": platform.release(), "Version": platform.version(), "Machine": platform.machine(), "Processor": platform.processor(), "Platform": platform.platform(), } return sys_info
[docs] def to_pascal(snake: str) -> str: """Convert a snake_case string to PascalCase. Args: snake (str): The snake_case string to be converted. Returns: str: The converted PascalCase string. """ # Check if the string is already in PascalCase if re.match(r'^[A-Z][a-zA-Z0-9]*([A-Z][a-zA-Z0-9]*)*$', snake): return snake # Remove leading and trailing underscores snake = snake.strip('_') # Replace multiple underscores with a single one snake = re.sub('_+', '_', snake) # Convert to PascalCase return re.sub( '_([0-9A-Za-z])', lambda m: m.group(1).upper(), snake.title(), )
[docs] def get_pydantic_major_version() -> int: r"""Get the major version of Pydantic. Returns: int: The major version number of Pydantic if installed, otherwise 0. """ try: return int(pydantic.__version__.split(".")[0]) except ImportError: return 0
[docs] def get_pydantic_object_schema(pydantic_params: Type[BaseModel]) -> Dict: r"""Get the JSON schema of a Pydantic model. Args: pydantic_params (Type[BaseModel]): The Pydantic model class to retrieve the schema for. Returns: dict: The JSON schema of the Pydantic model. """ return pydantic_params.model_json_schema()
[docs] def func_string_to_callable(code: str): r"""Convert a function code string to a callable function object. Args: code (str): The function code as a string. Returns: Callable[..., Any]: The callable function object extracted from the code string. """ local_vars: Mapping[str, object] = {} exec(code, globals(), local_vars) func = local_vars.get(Constants.FUNC_NAME_FOR_STRUCTURED_OUTPUT) return func
[docs] def json_to_function_code(json_obj: Dict) -> str: r"""Generate a Python function code from a JSON schema. Args: json_obj (dict): The JSON schema object containing properties and required fields, and json format is follow openai tools schema Returns: str: The generated Python function code as a string. """ properties = json_obj.get('properties', {}) required = json_obj.get('required', []) if not properties or not required: raise ValueError( "JSON schema must contain 'properties' and 'required' fields" ) args = [] docstring_args = [] return_keys = [] prop_to_python = { 'string': 'str', 'number': 'float', 'integer': 'int', 'boolean': 'bool', } for prop in required: # if no description, return empty string description = properties[prop].get('description', "") prop_type = properties[prop]['type'] python_type = prop_to_python.get(prop_type, prop_type) args.append(f"{prop}: {python_type}") docstring_args.append( f" {prop} ({python_type}): {description}." ) return_keys.append(prop) # extract entity of schema args_str = ", ".join(args) docstring_args_str = "\n".join(docstring_args) return_keys_str = ", ".join(return_keys) # function template function_code = f''' def {Constants.FUNC_NAME_FOR_STRUCTURED_OUTPUT}({args_str}): r"""Return response with a specified json format. Args: {docstring_args_str} Returns: Dict: A dictionary containing {return_keys_str}. """ return {{{", ".join([f'"{prop}": {prop}' for prop in required])}}} ''' return function_code
[docs] def text_extract_from_web(url: str) -> str: r"""Get the text information from given url. Args: url (str): The website you want to search. Returns: str: All texts extract from the web. """ try: import requests from newspaper import Article # Request the target page article = Article(url) article.download() article.parse() text = article.text except requests.RequestException as e: text = f"Can't access {url}, error: {e}" except Exception as e: text = f"Can't extract text from {url}, error: {e}" return text
[docs] def create_chunks(text: str, n: int) -> List[str]: r"""Returns successive n-sized chunks from provided text. Split a text into smaller chunks of size n". Args: text (str): The text to be split. n (int): The max length of a single chunk. Returns: List[str]: A list of split texts. """ chunks = [] i = 0 while i < len(text): # Find the nearest end of sentence within a range of 0.5 * n # and 1.5 * n tokens j = min(i + int(1.2 * n), len(text)) while j > i + int(0.8 * n): # Decode the tokens and check for full stop or newline chunk = text[i:j] if chunk.endswith(".") or chunk.endswith("\n"): break j -= 1 # If no end of sentence found, use n tokens as the chunk size if j == i + int(0.8 * n): j = min(i + n, len(text)) chunks.append(text[i:j]) i = j return chunks
[docs] def is_docker_running() -> bool: r"""Check if the Docker daemon is running. Returns: bool: True if the Docker daemon is running, False otherwise. """ try: result = subprocess.run( ["docker", "info"], check=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) return result.returncode == 0 except (subprocess.CalledProcessError, FileNotFoundError): return False
try: if os.getenv("AGENTOPS_API_KEY") is not None: from agentops import ( ToolEvent, record, ) else: raise ImportError except (ImportError, AttributeError): ToolEvent = None
[docs] def agentops_decorator(func): r"""Decorator that records the execution of a function if ToolEvent is available. Parameters: func (callable): The function to be decorated. Returns: callable: The wrapped function which records its execution details. """ @wraps(func) def wrapper(*args, **kwargs): if ToolEvent: tool_event = ToolEvent(name=func.__name__, params=kwargs) result = func(*args, **kwargs) tool_event.returns = result record(tool_event) return result return func(*args, **kwargs) return wrapper
[docs] class AgentOpsMeta(type): r"""Metaclass that automatically decorates all callable attributes with the agentops_decorator, except for the 'get_tools' method. Methods: __new__(cls, name, bases, dct): Creates a new class with decorated methods. """ def __new__(cls, name, bases, dct): if ToolEvent: for attr, value in dct.items(): if callable(value) and attr != 'get_tools': dct[attr] = agentops_decorator(value) return super().__new__(cls, name, bases, dct)
[docs] def track_agent(*args, **kwargs): r"""Mock track agent decorator for AgentOps.""" def noop(f): return f return noop
[docs] def handle_http_error(response: requests.Response) -> str: r"""Handles the HTTP errors based on the status code of the response. Args: response (requests.Response): The HTTP response from the API call. Returns: str: The error type, based on the status code. """ if response.status_code == HTTPStatus.UNAUTHORIZED: return "Unauthorized. Check your access token." elif response.status_code == HTTPStatus.FORBIDDEN: return "Forbidden. You do not have permission to perform this action." elif response.status_code == HTTPStatus.NOT_FOUND: return "Not Found. The resource could not be located." elif response.status_code == HTTPStatus.TOO_MANY_REQUESTS: return "Too Many Requests. You have hit the rate limit." else: return "HTTP Error"
[docs] def retry_on_error( max_retries: int = 3, initial_delay: float = 1.0 ) -> Callable: r"""Decorator to retry function calls on exception with exponential backoff. Args: max_retries (int): Maximum number of retry attempts initial_delay (float): Initial delay between retries in seconds Returns: Callable: Decorated function with retry logic """ def decorator(func: Callable) -> Callable: @functools.wraps(func) def wrapper(*args, **kwargs): delay = initial_delay last_exception = None for attempt in range(max_retries + 1): try: return func(*args, **kwargs) except Exception as e: last_exception = e if attempt == max_retries: logger.error( f"Failed after {max_retries} retries: {e!s}" ) raise logger.warning( f"Attempt {attempt + 1} failed: {e!s}. " f"Retrying in {delay:.1f}s..." ) time.sleep(delay) delay *= 2 # Exponential backoff raise last_exception return wrapper return decorator
[docs] class BatchProcessor: r"""Handles batch processing with dynamic sizing and error handling based on system load. """ def __init__( self, max_workers: Optional[int] = None, initial_batch_size: Optional[int] = None, monitoring_interval: float = 5.0, cpu_threshold: float = 80.0, memory_threshold: float = 85.0, ): r"""Initialize the BatchProcessor with dynamic worker allocation. Args: max_workers: Maximum number of workers. If None, will be determined dynamically based on system resources. (default: :obj:`None`) initial_batch_size: Initial size of each batch. If `None`, defaults to `10`. (default: :obj:`None`) monitoring_interval: Interval in seconds between resource checks. (default: :obj:`5.0`) cpu_threshold: CPU usage percentage threshold for scaling down. (default: :obj:`80.0`) memory_threshold: Memory usage percentage threshold for scaling down. (default: :obj:`85.0`) """ import psutil self.monitoring_interval = monitoring_interval self.cpu_threshold = cpu_threshold self.memory_threshold = memory_threshold self.last_check_time = time.time() self.psutil = psutil # Initialize performance metrics self.total_processed = 0 self.total_errors = 0 self.processing_times: List = [] if max_workers is None: self.max_workers = self._calculate_optimal_workers() else: self.max_workers = max_workers self.batch_size = ( 10 if initial_batch_size is None else initial_batch_size ) self.min_batch_size = 1 self.max_batch_size = 20 self.backoff_factor = 0.8 self.success_factor = 1.2 # Initial resource check self._update_resource_metrics() def _calculate_optimal_workers(self) -> int: r"""Calculate optimal number of workers based on system resources.""" cpu_count = self.psutil.cpu_count() cpu_percent = self.psutil.cpu_percent(interval=1) memory = self.psutil.virtual_memory() # Base number of workers on CPU count and current load if cpu_percent > self.cpu_threshold: workers = max(1, cpu_count // 4) elif cpu_percent > 60: workers = max(1, cpu_count // 2) else: workers = max(1, cpu_count - 1) # Further reduce if memory is constrained if memory.percent > self.memory_threshold: workers = max(1, workers // 2) return workers def _update_resource_metrics(self) -> None: r"""Update current resource usage metrics.""" self.current_cpu = self.psutil.cpu_percent() self.current_memory = self.psutil.virtual_memory().percent self.last_check_time = time.time() def _should_check_resources(self) -> bool: r"""Determine if it's time to check resource usage again.""" return time.time() - self.last_check_time >= self.monitoring_interval
[docs] def adjust_batch_size( self, success: bool, processing_time: Optional[float] = None ) -> None: r"""Adjust batch size based on success/failure and system resources. Args: success (bool): Whether the last batch completed successfully processing_time (Optional[float]): Time taken to process the last batch. (default: :obj:`None`) """ # Update metrics self.total_processed += 1 if not success: self.total_errors += 1 if processing_time is not None: self.processing_times.append(processing_time) # Check system resources if interval has elapsed if self._should_check_resources(): self._update_resource_metrics() # Adjust based on resource usage if ( self.current_cpu > self.cpu_threshold or self.current_memory > self.memory_threshold ): self.batch_size = max( int(self.batch_size * self.backoff_factor), self.min_batch_size, ) self.max_workers = max(1, self.max_workers - 1) return # Adjust based on success/failure if success: self.batch_size = min( int(self.batch_size * self.success_factor), self.max_batch_size ) else: self.batch_size = max( int(self.batch_size * self.backoff_factor), self.min_batch_size )
[docs] def get_performance_metrics(self) -> Dict[str, Any]: r"""Get current performance metrics. Returns: Dict containing performance metrics including: - total_processed: Total number of batches processed - error_rate: Percentage of failed batches - avg_processing_time: Average time per batch - current_batch_size: Current batch size - current_workers: Current number of workers - current_cpu: Current CPU usage percentage - current_memory: Current memory usage percentage """ metrics = { "total_processed": self.total_processed, "error_rate": (self.total_errors / max(1, self.total_processed)) * 100, "avg_processing_time": sum(self.processing_times) / max(1, len(self.processing_times)), "current_batch_size": self.batch_size, "current_workers": self.max_workers, "current_cpu": self.current_cpu, "current_memory": self.current_memory, } return metrics
[docs] def download_github_subdirectory( repo: str, subdir: str, data_dir: Path, branch="main" ): r"""Download subdirectory of the Github repo of the benchmark. This function downloads all files and subdirectories from a specified subdirectory of a GitHub repository and saves them to a local directory. Args: repo (str): The name of the GitHub repository in the format "owner/repo". subdir (str): The path to the subdirectory within the repository to download. data_dir (Path): The local directory where the files will be saved. branch (str, optional): The branch of the repository to use. Defaults to "main". """ from tqdm import tqdm api_url = ( f"https://api.github.com/repos/{repo}/contents/{subdir}?ref={branch}" ) headers = {"Accept": "application/vnd.github.v3+json"} response = requests.get(api_url, headers=headers) response.raise_for_status() files = response.json() os.makedirs(data_dir, exist_ok=True) for file in tqdm(files, desc="Downloading"): file_path = data_dir / file["name"] if file["type"] == "file": file_url = file["download_url"] file_response = requests.get(file_url) with open(file_path, "wb") as f: f.write(file_response.content) elif file["type"] == "dir": download_github_subdirectory( repo, f'{subdir}/{file["name"]}', file_path, branch )
[docs] def generate_prompt_for_structured_output( response_format: Optional[Type[BaseModel]], user_message: str, ) -> str: """ This function generates a prompt based on the provided Pydantic model and user message. Args: response_format (Type[BaseModel]): The Pydantic model class. user_message (str): The user message to be used in the prompt. Returns: str: A prompt string for the LLM. """ if response_format is None: return user_message json_schema = response_format.model_json_schema() sys_prompt = ( "Given the user message, please generate a JSON response adhering " "to the following JSON schema:\n" f"{json_schema}\n" "Make sure the JSON response is valid and matches the EXACT structure " "defined in the schema. Your result should only be a valid json " "object, without any other text or comments.\n" ) user_prompt = f"User message: {user_message}\n" final_prompt = f""" {sys_prompt} {user_prompt} """ return final_prompt
[docs] def with_timeout(timeout=None): r"""Decorator that adds timeout functionality to functions. Executes functions with a specified timeout value. Returns a timeout message if execution time is exceeded. Args: timeout (float, optional): The timeout duration in seconds. If None, will try to get timeout from the instance's timeout attribute. (default: :obj:`None`) Example: >>> @with_timeout(5) ... def my_function(): ... return "Success" >>> my_function() >>> class MyClass: ... timeout = 5 ... @with_timeout() ... def my_method(self): ... return "Success" """ def decorator(func): @functools.wraps(func) def wrapper(*args, **kwargs): # Determine the effective timeout value effective_timeout = timeout if effective_timeout is None and args: effective_timeout = getattr(args[0], 'timeout', None) # If no timeout value is provided, execute function normally if effective_timeout is None: return func(*args, **kwargs) # Container to hold the result of the function call result_container = [] def target(): result_container.append(func(*args, **kwargs)) # Start the function in a new thread thread = threading.Thread(target=target) thread.start() thread.join(effective_timeout) # Check if the thread is still alive after the timeout if thread.is_alive(): return ( f"Function `{func.__name__}` execution timed out, " f"exceeded {effective_timeout} seconds." ) else: return result_container[0] return wrapper # Handle both @with_timeout and @with_timeout() usage if callable(timeout): # If timeout is passed as a function, apply it to the decorator func, timeout = timeout, None return decorator(func) return decorator