| | """
|
| | Configuration loading for problem runtime settings.
|
| |
|
| | Loads and parses problem config.yaml files, including runtime resources
|
| | and docker configuration.
|
| | """
|
| |
|
| | from dataclasses import dataclass, field
|
| | from pathlib import Path
|
| | from typing import Any, Dict, Optional
|
| |
|
| | import yaml
|
| |
|
| |
|
| | DEFAULT_DOCKER_IMAGE = "python:3.11-slim-trixie"
|
| |
|
| |
|
| | @dataclass
|
| | class ResourcesConfig:
|
| | """SkyPilot-compatible resources configuration."""
|
| |
|
| | accelerators: Optional[str] = None
|
| | instance_type: Optional[str] = None
|
| | cpus: Optional[str] = None
|
| | memory: Optional[str] = None
|
| | disk_size: Optional[int] = None
|
| | disk_tier: Optional[str] = None
|
| | cloud: Optional[str] = None
|
| | region: Optional[str] = None
|
| | image_id: Optional[str] = None
|
| |
|
| | def to_dict(self) -> Dict[str, Any]:
|
| | """Convert to dict, excluding None values."""
|
| | return {k: v for k, v in vars(self).items() if v is not None}
|
| |
|
| | @property
|
| | def has_gpu(self) -> bool:
|
| | return self.accelerators is not None
|
| |
|
| | @property
|
| | def gpu_type(self) -> Optional[str]:
|
| | """Extract GPU type from accelerators (e.g., 'L4:1' -> 'L4')."""
|
| | if not self.accelerators:
|
| | return None
|
| | return self.accelerators.split(":")[0]
|
| |
|
| |
|
| | @dataclass
|
| | class DockerConfig:
|
| | """Docker configuration for running evaluations."""
|
| |
|
| | image: str = DEFAULT_DOCKER_IMAGE
|
| | gpu: bool = False
|
| | dind: bool = False
|
| |
|
| | @classmethod
|
| | def from_dict(cls, data: Dict[str, Any]) -> "DockerConfig":
|
| | """Create DockerConfig from a dictionary."""
|
| | return cls(
|
| | image=data.get("image", DEFAULT_DOCKER_IMAGE),
|
| | gpu=bool(data.get("gpu", False)),
|
| | dind=bool(data.get("dind", False)),
|
| | )
|
| |
|
| |
|
| | @dataclass
|
| | class RuntimeConfig:
|
| | """Complete runtime configuration from config.yaml."""
|
| |
|
| | timeout_seconds: Optional[int] = None
|
| | requires_gpu: Optional[bool] = None
|
| | resources: ResourcesConfig = field(default_factory=ResourcesConfig)
|
| | docker: DockerConfig = field(default_factory=DockerConfig)
|
| | environment: Optional[str] = None
|
| |
|
| |
|
| | @dataclass
|
| | class ProblemConfig:
|
| | """Full problem configuration from config.yaml."""
|
| |
|
| | tag: Optional[str] = None
|
| | runtime: RuntimeConfig = field(default_factory=RuntimeConfig)
|
| | dependencies: Dict[str, Any] = field(default_factory=dict)
|
| | datasets: list = field(default_factory=list)
|
| |
|
| |
|
| | def load_problem_config(problem_path: Path) -> ProblemConfig:
|
| | """
|
| | Load full problem configuration from config.yaml.
|
| |
|
| | Example config.yaml:
|
| | ```yaml
|
| | tag: hpc
|
| | runtime:
|
| | timeout_seconds: 1800
|
| | docker:
|
| | image: andylizf/triton-tlx:tlx-nv-cu122
|
| | gpu: true
|
| | resources:
|
| | accelerators: "L4:1"
|
| | cpus: "8+"
|
| | ```
|
| | """
|
| | config_file = problem_path / "config.yaml"
|
| | problem_config = ProblemConfig()
|
| |
|
| | if not config_file.exists():
|
| | return problem_config
|
| |
|
| | try:
|
| | with open(config_file, "r", encoding="utf-8") as f:
|
| | config = yaml.safe_load(f) or {}
|
| | except Exception:
|
| | return problem_config
|
| |
|
| |
|
| | if config.get("tag"):
|
| | problem_config.tag = str(config["tag"])
|
| |
|
| |
|
| | problem_config.dependencies = config.get("dependencies", {})
|
| | problem_config.datasets = config.get("datasets", [])
|
| |
|
| |
|
| | runtime = config.get("runtime", {})
|
| | rt = problem_config.runtime
|
| |
|
| | if runtime.get("timeout_seconds"):
|
| | rt.timeout_seconds = int(runtime["timeout_seconds"])
|
| | if runtime.get("requires_gpu") is not None:
|
| | rt.requires_gpu = bool(runtime["requires_gpu"])
|
| | if runtime.get("environment"):
|
| | rt.environment = str(runtime["environment"])
|
| |
|
| |
|
| | docker = runtime.get("docker", {})
|
| | if docker:
|
| | rt.docker = DockerConfig.from_dict(docker)
|
| | elif runtime.get("requires_gpu"):
|
| |
|
| | rt.docker.gpu = True
|
| |
|
| |
|
| | resources = runtime.get("resources", {})
|
| | if resources:
|
| | res = rt.resources
|
| | for key in ["accelerators", "instance_type", "cpus", "memory", "disk_tier", "cloud", "region", "image_id"]:
|
| | if resources.get(key):
|
| | setattr(res, key, str(resources[key]))
|
| | if resources.get("disk_size"):
|
| | res.disk_size = int(resources["disk_size"])
|
| |
|
| | return problem_config
|
| |
|
| |
|
| | def load_runtime_config(problem_path: Path) -> RuntimeConfig:
|
| | """Load runtime configuration from problem's config.yaml."""
|
| | return load_problem_config(problem_path).runtime
|
| |
|
| |
|
| | def load_docker_config_from_yaml(problem_path: Path) -> DockerConfig:
|
| | """Load docker configuration from problem's config.yaml."""
|
| | return load_problem_config(problem_path).runtime.docker
|
| |
|
| |
|
| | def get_effective_gpu_type(runtime_config: RuntimeConfig) -> Optional[str]:
|
| | """
|
| | Get effective GPU type from runtime config.
|
| |
|
| | Priority:
|
| | 1. resources.accelerators (extract type, e.g., "L4:1" -> "L4")
|
| | 2. docker.gpu flag (returns default "L4" if True)
|
| | 3. requires_gpu flag (returns default "L4" if True)
|
| | 4. None (CPU only)
|
| | """
|
| |
|
| | if runtime_config.resources.accelerators:
|
| | return runtime_config.resources.gpu_type
|
| |
|
| |
|
| | if runtime_config.docker.gpu:
|
| | return "L4"
|
| |
|
| |
|
| | if runtime_config.requires_gpu:
|
| | return "L4"
|
| |
|
| | return None
|
| |
|