import os # LLM Token Settings MAX_CONTEXT_TOKENS = int(os.getenv("MAX_CONTEXT_TOKENS", 262144)) MAX_OUTPUT_TOKENS = int(os.getenv("MAX_OUTPUT_TOKENS", 8000)) SUMMARIZATION_MAX_TOKENS = MAX_CONTEXT_TOKENS - MAX_OUTPUT_TOKENS - 1000 # Agent and Shard Settings MAX_CACHED_AGENTS = int(os.getenv("MAX_CACHED_AGENTS", 50)) SHARD_COUNT = int(os.getenv("SHARD_COUNT", 16)) # Connection Settings MAX_CONNECTIONS_PER_HOST = int(os.getenv("MAX_CONNECTIONS_PER_HOST", 100)) MAX_CONNECTIONS_TOTAL = int(os.getenv("MAX_CONNECTIONS_TOTAL", 500)) KEEPALIVE_TIMEOUT = int(os.getenv("KEEPALIVE_TIMEOUT", 30)) CONNECT_TIMEOUT = int(os.getenv("CONNECT_TIMEOUT", 10)) TOTAL_TIMEOUT = int(os.getenv("TOTAL_TIMEOUT", 60)) # File Cache Settings FILE_CACHE_SIZE = int(os.getenv("FILE_CACHE_SIZE", 1000)) FILE_CACHE_TTL = int(os.getenv("FILE_CACHE_TTL", 300)) # API Settings BACKEND_HOST = os.getenv("BACKEND_HOST", "https://api-dev.gptbase.ai") MASTERKEY = os.getenv("MASTERKEY", "master") FASTAPI_URL = os.getenv('FASTAPI_URL', 'http://127.0.0.1:8001') # Project Settings PROJECT_DATA_DIR = os.getenv("PROJECT_DATA_DIR", "./projects/data") # Tokenizer Settings TOKENIZERS_PARALLELISM = os.getenv("TOKENIZERS_PARALLELISM", "true") # Embedding Model Settings SENTENCE_TRANSFORMER_MODEL = os.getenv("SENTENCE_TRANSFORMER_MODEL", "TaylorAI/gte-tiny") # Tool Output Length Control Settings TOOL_OUTPUT_MAX_LENGTH = int(SUMMARIZATION_MAX_TOKENS/4) TOOL_OUTPUT_TRUNCATION_STRATEGY = os.getenv("TOOL_OUTPUT_TRUNCATION_STRATEGY", "smart")