bfh-studadmin-assist / .env.example
awellis's picture
Implement proper async parallelism for multi-agent pipeline
e20e6e8
raw
history blame contribute delete
792 Bytes
# OpenAI Configuration (required)
OPENAI_API_KEY=your_openai_api_key_here
LLM_MODEL=gpt-5-nano
EMBEDDING_MODEL=text-embedding-3-small
LLM_TEMPERATURE=0.7
LLM_MAX_TOKENS=2000
# Document Processing Configuration
DOCUMENTS_PATH=assets/markdown
CHUNK_SIZE=300 # Back to default
CHUNK_OVERLAP=50
MIN_CHUNK_SIZE=100
# Retrieval Configuration
RETRIEVAL_TOP_K=5
BM25_WEIGHT=0.8 # Strongly favor keyword matching
VECTOR_WEIGHT=0.2 # Reduce semantic weight
MIN_RELEVANCE_SCORE=0.5 # Lower threshold (was too high)
# Application Configuration
DEBUG=false
SKIP_FACT_CHECK=false # Skip fact-checking for faster responses
# Async Performance Configuration
USE_PARALLEL=true # Enable parallel agent execution (saves 3-5s)
AGENT_TIMEOUT=30 # Timeout per agent in seconds