| import argparse |
| import json |
| import os |
| import sys |
| import traceback |
| import importlib.util |
| import pandas as pd |
| import gc |
| from typing import Any, Mapping, Tuple, List, Optional |
|
|
|
|
| def _is_number(x: Any) -> bool: |
| try: |
| float(x) |
| return True |
| except (TypeError, ValueError): |
| return False |
|
|
|
|
| def _to_float(x: Any, default: Optional[float] = None) -> Optional[float]: |
| if x is None: |
| return default |
| try: |
| return float(x) |
| except (TypeError, ValueError): |
| return default |
|
|
|
|
| def _is_correct_value(v: Any) -> bool: |
| """ |
| Treat >0 numeric as correct; otherwise truthy strings/bools as correct. |
| """ |
| if _is_number(v): |
| try: |
| return float(v) > 0.0 |
| except Exception: |
| return False |
| return bool(v) |
|
|
|
|
| def compute_oracle_stats(query: Mapping[str, Any], |
| tier_to_model: Mapping[str, str]) -> Tuple[str, float]: |
| """ |
| Pick the lowest-cost model (among tier_to_model.values()) that answered correctly. |
| """ |
| best_model = "no_model_correct" |
| best_cost = float("inf") |
|
|
| for model in tier_to_model.values(): |
| if not _is_correct_value(query.get(model, 0.0)): |
| continue |
|
|
| cost_key = f"{model}|total_cost" |
| cost = _to_float(query.get(cost_key, None), default=None) |
| if cost is None: |
| continue |
|
|
| if cost < best_cost: |
| best_cost = cost |
| best_model = model |
|
|
| if best_model == "no_model_correct": |
| return "no_model_correct", 0.0 |
| return best_model, best_cost |
|
|
|
|
| class Evaluator: |
| def __init__(self, problem_dir: str): |
| self.problem_dir = problem_dir |
| self.resources_dir = os.path.join(problem_dir, "resources") |
| |
| mounted_datasets_dir = "/datasets/llm_router" |
| if os.path.exists(mounted_datasets_dir) and os.listdir(mounted_datasets_dir): |
| self.datasets_dir = mounted_datasets_dir |
| else: |
| |
| self.datasets_dir = os.path.join(self.resources_dir, "datasets") |
| ordered_names = ["routerbench_0shot_test.csv"] |
| self.trace_files = [ |
| os.path.join(self.datasets_dir, name) |
| for name in ordered_names |
| if os.path.exists(os.path.join(self.datasets_dir, name)) |
| ] |
|
|
| def evaluate(self, solution_module_path: str) -> dict: |
| LAMBDA = 150.0 |
| CANDIDATE_MODELS = ["cheap", "mid", "expensive"] |
|
|
| |
| TIER_TO_MODEL = { |
| "cheap": "mistralai/mistral-7b-chat", |
| "mid": "mistralai/mixtral-8x7b-chat", |
| "expensive": "gpt-4-1106-preview", |
| } |
| |
| |
| spec = importlib.util.spec_from_file_location("solution", solution_module_path) |
| solution = importlib.util.module_from_spec(spec) |
| spec.loader.exec_module(solution) |
|
|
| if not hasattr(solution, "Solution"): |
| return {"score": 0.0, "runs_successfully": 0.0, "error": "Missing Solution"} |
|
|
| solver = solution.Solution() |
| |
| total_queries = 0 |
| total_correct, oracle_correct = 0, 0 |
| total_cost, oracle_cost = 0.0, 0.0 |
| for csv_path in self.trace_files: |
| dataset_name = os.path.basename(csv_path) |
| |
| |
| df = pd.read_csv(csv_path, low_memory=False) |
| |
| for _, row in df.iterrows(): |
| |
| |
| chosen_tier = solver.solve( |
| query=row["prompt"], |
| eval_name=row["eval_name"], |
| candidate_models=CANDIDATE_MODELS, |
| ) |
| if chosen_tier not in CANDIDATE_MODELS: |
| chosen_tier = "cheap" |
| |
| |
| model_name = TIER_TO_MODEL[chosen_tier] |
| cost_col = f"{model_name}|total_cost" |
| try: |
| cost = float(row[cost_col]) |
| except Exception: |
| cost = 0.0 |
| |
| |
| |
| |
| oracle, oracle_cost_sample = compute_oracle_stats(row, TIER_TO_MODEL) |
| if isinstance(oracle, str) and oracle.strip() == "no_model_correct": |
| correct = 0 |
| cost = 0.0 |
| oracle_correct_sample = 0 |
| else: |
| oracle_correct_sample = 1 |
| try: |
| correct = int(float(row[model_name])) |
| except Exception: |
| correct = 0 |
| |
| total_queries += 1 |
| total_correct += correct |
| oracle_correct += oracle_correct_sample |
| total_cost += cost |
| oracle_cost += oracle_cost_sample |
| |
| |
| del df |
| gc.collect() |
|
|
| |
| if total_queries == 0: |
| return {"score": 0.0, "runs_successfully": 0.0, "error": "Empty dataset"} |
| accuracy = total_correct / total_queries |
| avg_cost = total_cost / total_queries |
| raw_score = accuracy - (LAMBDA * avg_cost) |
| oracle_accuracy = oracle_correct / total_queries |
| oracle_avg_cost = oracle_cost / total_queries |
| oracle_raw_score = oracle_accuracy - (LAMBDA * oracle_avg_cost) |
| score = (raw_score / oracle_raw_score) * 100 if oracle_raw_score > 0 else 0.0 |
|
|
| return {"runs_successfully": 1.0, |
| "total_queries": total_queries, |
| "score": score, |
| "raw_score": raw_score, |
| "accuracy": accuracy, |
| "avg_cost": avg_cost, |
| "oracle_raw_score": oracle_raw_score, |
| "oracle_accuracy": oracle_accuracy, |
| "oracle_avg_cost": oracle_avg_cost, |
| "lambda": LAMBDA, |
| } |
|
|
|
|
| def main(): |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--solution", required=True) |
| parser.add_argument("--out", required=True) |
| args = parser.parse_args() |
|
|
| try: |
| problem_dir = os.path.dirname(os.path.abspath(__file__)) |
| result = Evaluator(problem_dir).evaluate(args.solution) |
| except Exception as e: |
| print(f"[evaluator] ERROR: {e}", file=sys.stderr) |
| print(traceback.format_exc(), file=sys.stderr) |
| result = {"score": 0.0, "runs_successfully": 0.0, "error": str(e)} |
|
|
| os.makedirs(os.path.dirname(args.out) or ".", exist_ok=True) |
| with open(args.out, "w") as f: |
| json.dump(result, f) |
| print(json.dumps(result)) |
| return 0 |
|
|
|
|
| if __name__ == "__main__": |
| raise SystemExit(main()) |