| | """
|
| | Unified evaluation API for Frontier-CS.
|
| |
|
| | Provides a single interface for evaluating both algorithmic and research problems,
|
| | with support for different backends (local Docker, SkyPilot cloud).
|
| | """
|
| |
|
| | from pathlib import Path
|
| | from typing import Iterator, List, Literal, Optional, Union
|
| |
|
| | from .runner import EvaluationResult, DockerRunner, AlgorithmicRunner
|
| | from .runner.base import Runner
|
| |
|
| |
|
| | TrackType = Literal["algorithmic", "research"]
|
| | BackendType = Literal["docker", "skypilot"]
|
| |
|
| |
|
| | class FrontierCSEvaluator:
|
| | """
|
| | Unified evaluator for Frontier-CS problems.
|
| |
|
| | Example usage:
|
| | evaluator = FrontierCSEvaluator()
|
| |
|
| | # Algorithmic problem
|
| | result = evaluator.evaluate("algorithmic", problem_id=1, code=cpp_code)
|
| |
|
| | # Research problem (local Docker)
|
| | result = evaluator.evaluate("research", problem_id="flash_attn", code=py_code)
|
| |
|
| | # Research problem (SkyPilot)
|
| | result = evaluator.evaluate("research", problem_id="flash_attn", code=py_code,
|
| | backend="skypilot")
|
| |
|
| | # Batch evaluation
|
| | results = evaluator.evaluate_batch("research",
|
| | problem_ids=["flash_attn", "cross_entropy"],
|
| | code=py_code)
|
| | """
|
| |
|
| | def __init__(
|
| | self,
|
| | backend: BackendType = "docker",
|
| | base_dir: Optional[Path] = None,
|
| | judge_url: str = "http://localhost:8081",
|
| | cloud: str = "gcp",
|
| | region: Optional[str] = None,
|
| | ):
|
| | """
|
| | Initialize FrontierCSEvaluator.
|
| |
|
| | Args:
|
| | backend: Default backend for research problems ("docker" or "skypilot")
|
| | base_dir: Base directory of Frontier-CS repo (auto-detected if None)
|
| | judge_url: URL of the algorithmic judge server
|
| | cloud: Cloud provider for SkyPilot ("gcp", "aws", "azure")
|
| | region: Cloud region for SkyPilot
|
| | """
|
| | self.default_backend = backend
|
| | self.base_dir = base_dir
|
| | self.judge_url = judge_url
|
| | self.cloud = cloud
|
| | self.region = region
|
| |
|
| |
|
| | self._algorithmic_runner: Optional[AlgorithmicRunner] = None
|
| | self._docker_runner: Optional[DockerRunner] = None
|
| | self._skypilot_runner: Optional[Runner] = None
|
| |
|
| | @property
|
| | def algorithmic_runner(self) -> AlgorithmicRunner:
|
| | """Get or create the algorithmic runner."""
|
| | if self._algorithmic_runner is None:
|
| | self._algorithmic_runner = AlgorithmicRunner(judge_url=self.judge_url)
|
| | return self._algorithmic_runner
|
| |
|
| | @property
|
| | def docker_runner(self) -> DockerRunner:
|
| | """Get or create the Docker runner."""
|
| | if self._docker_runner is None:
|
| | self._docker_runner = DockerRunner(base_dir=self.base_dir)
|
| | return self._docker_runner
|
| |
|
| | @property
|
| | def skypilot_runner(self) -> Runner:
|
| | """Get or create the SkyPilot runner."""
|
| | if self._skypilot_runner is None:
|
| | from .runner.skypilot import SkyPilotRunner
|
| | self._skypilot_runner = SkyPilotRunner(
|
| | base_dir=self.base_dir,
|
| | cloud=self.cloud,
|
| | region=self.region,
|
| | )
|
| | return self._skypilot_runner
|
| |
|
| | def _get_runner(self, track: TrackType, backend: Optional[BackendType] = None) -> Runner:
|
| | """Get the appropriate runner for a track and backend."""
|
| | if track == "algorithmic":
|
| | return self.algorithmic_runner
|
| |
|
| | effective_backend = backend or self.default_backend
|
| | if effective_backend == "skypilot":
|
| | return self.skypilot_runner
|
| | return self.docker_runner
|
| |
|
| | def evaluate(
|
| | self,
|
| | track: TrackType,
|
| | problem_id: Union[str, int],
|
| | code: str,
|
| | *,
|
| | backend: Optional[BackendType] = None,
|
| | timeout: Optional[int] = None,
|
| | unbounded: bool = False,
|
| | ) -> EvaluationResult:
|
| | """
|
| | Evaluate a solution for a single problem.
|
| |
|
| | Args:
|
| | track: Problem track ("algorithmic" or "research")
|
| | problem_id: Problem identifier (int for algorithmic, str for research)
|
| | code: Solution code (C++ for algorithmic, Python for research)
|
| | backend: Backend to use ("docker" or "skypilot"), defaults to init value
|
| | timeout: Optional timeout in seconds
|
| | unbounded: For algorithmic problems, use unbounded score (no clipping)
|
| |
|
| | Returns:
|
| | EvaluationResult with score and status
|
| | """
|
| | runner = self._get_runner(track, backend)
|
| |
|
| | if track == "algorithmic" and hasattr(runner, 'evaluate'):
|
| | return runner.evaluate(str(problem_id), code, timeout=timeout, unbounded=unbounded)
|
| | return runner.evaluate(str(problem_id), code, timeout=timeout)
|
| |
|
| | def evaluate_file(
|
| | self,
|
| | track: TrackType,
|
| | problem_id: Union[str, int],
|
| | solution_path: Path,
|
| | *,
|
| | backend: Optional[BackendType] = None,
|
| | timeout: Optional[int] = None,
|
| | ) -> EvaluationResult:
|
| | """
|
| | Evaluate a solution file for a single problem.
|
| |
|
| | Args:
|
| | track: Problem track
|
| | problem_id: Problem identifier
|
| | solution_path: Path to solution file
|
| | backend: Backend to use
|
| | timeout: Optional timeout in seconds
|
| |
|
| | Returns:
|
| | EvaluationResult with score and status
|
| | """
|
| | runner = self._get_runner(track, backend)
|
| | return runner.evaluate_file(str(problem_id), solution_path, timeout=timeout)
|
| |
|
| | def evaluate_batch(
|
| | self,
|
| | track: TrackType,
|
| | problem_ids: List[Union[str, int]],
|
| | code: str,
|
| | *,
|
| | backend: Optional[BackendType] = None,
|
| | timeout: Optional[int] = None,
|
| | ) -> List[EvaluationResult]:
|
| | """
|
| | Evaluate a solution against multiple problems.
|
| |
|
| | Args:
|
| | track: Problem track
|
| | problem_ids: List of problem identifiers
|
| | code: Solution code (same code for all problems)
|
| | backend: Backend to use
|
| | timeout: Optional timeout per problem
|
| |
|
| | Returns:
|
| | List of EvaluationResult, one per problem
|
| | """
|
| | runner = self._get_runner(track, backend)
|
| | results = []
|
| | for pid in problem_ids:
|
| | result = runner.evaluate(str(pid), code, timeout=timeout)
|
| | results.append(result)
|
| | return results
|
| |
|
| | def evaluate_batch_iter(
|
| | self,
|
| | track: TrackType,
|
| | problem_ids: List[Union[str, int]],
|
| | code: str,
|
| | *,
|
| | backend: Optional[BackendType] = None,
|
| | timeout: Optional[int] = None,
|
| | ) -> Iterator[EvaluationResult]:
|
| | """
|
| | Evaluate a solution against multiple problems, yielding results as they complete.
|
| |
|
| | Args:
|
| | track: Problem track
|
| | problem_ids: List of problem identifiers
|
| | code: Solution code
|
| | backend: Backend to use
|
| | timeout: Optional timeout per problem
|
| |
|
| | Yields:
|
| | EvaluationResult for each problem as it completes
|
| | """
|
| | runner = self._get_runner(track, backend)
|
| | for pid in problem_ids:
|
| | yield runner.evaluate(str(pid), code, timeout=timeout)
|
| |
|
| | def list_problems(self, track: TrackType) -> List[str]:
|
| | """
|
| | List all available problems for a track.
|
| |
|
| | Args:
|
| | track: Problem track
|
| |
|
| | Returns:
|
| | List of problem identifiers
|
| | """
|
| | if track == "algorithmic":
|
| |
|
| | try:
|
| | alg_base = self.docker_runner.base_dir / "algorithmic" / "problems"
|
| | except Exception:
|
| | return []
|
| |
|
| | if not alg_base or not alg_base.exists():
|
| | return []
|
| |
|
| | problems = []
|
| | for item in alg_base.iterdir():
|
| | if item.is_dir() and not item.name.startswith("."):
|
| | problems.append(item.name)
|
| |
|
| |
|
| | def sort_key(name):
|
| | try:
|
| | return (0, int(name))
|
| | except ValueError:
|
| | return (1, name)
|
| |
|
| | return sorted(problems, key=sort_key)
|
| |
|
| |
|
| | research_problems_dir = self.docker_runner.research_dir / "problems"
|
| | if not research_problems_dir.exists():
|
| | return []
|
| |
|
| | problems = []
|
| |
|
| |
|
| | poc_dir = research_problems_dir / "poc_generation"
|
| | if poc_dir.exists():
|
| |
|
| | problems.extend([
|
| | "research/poc_generation/heap_buffer_overflow",
|
| | "research/poc_generation/heap_use_after_free",
|
| | "research/poc_generation/stack_buffer_overflow",
|
| | "research/poc_generation/uninitialized_value"
|
| | ])
|
| |
|
| |
|
| | for evaluator_file in research_problems_dir.rglob("evaluator.py"):
|
| |
|
| | if "poc_generation" not in str(evaluator_file):
|
| |
|
| | problem_path = evaluator_file.parent.relative_to(research_problems_dir)
|
| | problems.append("research/" + str(problem_path))
|
| |
|
| |
|
| | try:
|
| | alg_base = self.docker_runner.base_dir / "algorithmic" / "problems"
|
| | except Exception:
|
| | alg_base = None
|
| |
|
| | if alg_base and alg_base.exists():
|
| | for item in sorted(alg_base.iterdir(), key=lambda p: p.name):
|
| | if item.is_dir() and not item.name.startswith("."):
|
| | problems.append(f"algorithmic/{item.name}")
|
| |
|
| | return sorted(problems)
|
| |
|
| | def get_problem_statement(
|
| | self,
|
| | track: TrackType,
|
| | problem_id: Union[str, int],
|
| | ) -> Optional[str]:
|
| | """
|
| | Get the problem statement/readme for a problem.
|
| |
|
| | Args:
|
| | track: Problem track
|
| | problem_id: Problem identifier
|
| |
|
| | Returns:
|
| | Problem statement text, or None if not found
|
| | """
|
| | if track == "algorithmic":
|
| | return self.algorithmic_runner.get_problem_statement(str(problem_id))
|
| |
|
| |
|
| | problem_path = self.docker_runner.get_problem_path(str(problem_id))
|
| | readme = problem_path / "readme"
|
| | if readme.exists():
|
| | return readme.read_text(encoding="utf-8")
|
| | return None
|
| |
|
| |
|
| |
|
| | def evaluate(
|
| | track: TrackType,
|
| | problem_id: Union[str, int],
|
| | code: str,
|
| | *,
|
| | backend: BackendType = "docker",
|
| | timeout: Optional[int] = None,
|
| | ) -> EvaluationResult:
|
| | """
|
| | Quick evaluation function.
|
| |
|
| | Example:
|
| | from frontier_cs import evaluate
|
| | result = evaluate("research", "flash_attn", solution_code)
|
| | print(f"Score: {result.score}")
|
| | """
|
| | evaluator = FrontierCSEvaluator(backend=backend)
|
| | return evaluator.evaluate(track, problem_id, code, timeout=timeout)
|
| |
|