diff --git a/benchmarks/commit0/run_infer.py b/benchmarks/commit0/run_infer.py index 36af180a5..552e2fe5c 100644 --- a/benchmarks/commit0/run_infer.py +++ b/benchmarks/commit0/run_infer.py @@ -594,7 +594,7 @@ def evaluate_instance( def main() -> None: prompt_dir = (Path(__file__).parent / "prompts").resolve() - choices = [str(p.relative_to(Path.cwd())) for p in prompt_dir.glob("*.j2")] + choices = [str(p) for p in prompt_dir.glob("*.j2")] default_prompt_path = prompt_dir / "default.j2" assert default_prompt_path.exists(), ( f"Default prompt {default_prompt_path} not found" diff --git a/benchmarks/multiswebench/run_infer.py b/benchmarks/multiswebench/run_infer.py index ec5137a2d..7280ee485 100644 --- a/benchmarks/multiswebench/run_infer.py +++ b/benchmarks/multiswebench/run_infer.py @@ -405,7 +405,7 @@ def evaluate_instance( def main() -> None: prompt_dir = (Path(__file__).parent / "prompts").resolve() - choices = [str(p.relative_to(Path.cwd())) for p in prompt_dir.glob("*.j2")] + choices = [str(p) for p in prompt_dir.glob("*.j2")] default_prompt_path = prompt_dir / "default.j2" assert default_prompt_path.exists(), ( f"Default prompt {default_prompt_path} not found" diff --git a/benchmarks/swebench/run_infer.py b/benchmarks/swebench/run_infer.py index 231de93e9..e410ca8bd 100644 --- a/benchmarks/swebench/run_infer.py +++ b/benchmarks/swebench/run_infer.py @@ -350,7 +350,7 @@ def evaluate_instance( def main() -> None: prompt_dir = (Path(__file__).parent / "prompts").resolve() - choices = [str(p.relative_to(Path.cwd())) for p in prompt_dir.glob("*.j2")] + choices = [p.name for p in prompt_dir.glob("*.j2")] default_prompt_path = prompt_dir / "default.j2" assert default_prompt_path.exists(), ( f"Default prompt {default_prompt_path} not found" diff --git a/benchmarks/swebenchmultimodal/run_infer.py b/benchmarks/swebenchmultimodal/run_infer.py index 6a1b533a7..cf2daf713 100644 --- a/benchmarks/swebenchmultimodal/run_infer.py +++ b/benchmarks/swebenchmultimodal/run_infer.py @@ -403,7 +403,7 @@ def evaluate_instance( def main() -> None: prompt_dir = (Path(__file__).parent / "prompts").resolve() - choices = [str(p.relative_to(Path.cwd())) for p in prompt_dir.glob("*.j2")] + choices = [str(p) for p in prompt_dir.glob("*.j2")] default_prompt_path = prompt_dir / "default.j2" assert default_prompt_path.exists(), ( f"Default prompt {default_prompt_path} not found" diff --git a/benchmarks/swefficiency/run_infer.py b/benchmarks/swefficiency/run_infer.py index cf5a902d2..b07b11de6 100644 --- a/benchmarks/swefficiency/run_infer.py +++ b/benchmarks/swefficiency/run_infer.py @@ -396,7 +396,7 @@ def evaluate_instance( def main() -> None: prompt_dir = (Path(__file__).parent / "prompts").resolve() - choices = [str(p.relative_to(Path.cwd())) for p in prompt_dir.glob("*.j2")] + choices = [str(p) for p in prompt_dir.glob("*.j2")] default_prompt_path = prompt_dir / "default.j2" assert default_prompt_path.exists(), ( f"Default prompt {default_prompt_path} not found" diff --git a/benchmarks/swtbench/run_infer.py b/benchmarks/swtbench/run_infer.py index 36b78f265..93f890116 100644 --- a/benchmarks/swtbench/run_infer.py +++ b/benchmarks/swtbench/run_infer.py @@ -339,7 +339,7 @@ def evaluate_instance( def main() -> None: """Main entry point for SWT-bench evaluation.""" prompt_dir = (Path(__file__).parent / "prompts").resolve() - choices = [str(p.relative_to(Path.cwd())) for p in prompt_dir.glob("*.j2")] + choices = [str(p) for p in prompt_dir.glob("*.j2")] default_prompt_path = prompt_dir / "default.j2" assert default_prompt_path.exists(), ( f"Default prompt {default_prompt_path} not found"