diff --git a/sagemaker-train/src/sagemaker/train/evaluate/benchmark_evaluator.py b/sagemaker-train/src/sagemaker/train/evaluate/benchmark_evaluator.py index d6bad422c6..5d37e53f8c 100644 --- a/sagemaker-train/src/sagemaker/train/evaluate/benchmark_evaluator.py +++ b/sagemaker-train/src/sagemaker/train/evaluate/benchmark_evaluator.py @@ -301,7 +301,7 @@ class BenchMarkEvaluator(BaseEvaluator): benchmark: _Benchmark subtasks: Optional[Union[str, List[str]]] = None - evaluate_base_model: bool = False + evaluate_base_model: bool = True _hyperparameters: Optional[Any] = None diff --git a/sagemaker-train/src/sagemaker/train/evaluate/custom_scorer_evaluator.py b/sagemaker-train/src/sagemaker/train/evaluate/custom_scorer_evaluator.py index 78d297006c..290a6f80ba 100644 --- a/sagemaker-train/src/sagemaker/train/evaluate/custom_scorer_evaluator.py +++ b/sagemaker-train/src/sagemaker/train/evaluate/custom_scorer_evaluator.py @@ -137,7 +137,7 @@ class CustomScorerEvaluator(BaseEvaluator): _hyperparameters: Optional[Any] = None # Template-required fields - evaluate_base_model: bool = False + evaluate_base_model: bool = True @validator('dataset', pre=True) def _resolve_dataset(cls, v): diff --git a/sagemaker-train/src/sagemaker/train/evaluate/llm_as_judge_evaluator.py b/sagemaker-train/src/sagemaker/train/evaluate/llm_as_judge_evaluator.py index 8438b65688..98e1c50c48 100644 --- a/sagemaker-train/src/sagemaker/train/evaluate/llm_as_judge_evaluator.py +++ b/sagemaker-train/src/sagemaker/train/evaluate/llm_as_judge_evaluator.py @@ -123,7 +123,7 @@ class LLMAsJudgeEvaluator(BaseEvaluator): custom_metrics: Optional[str] = None # Template-required fields - evaluate_base_model: bool = False + evaluate_base_model: bool = True @validator('dataset', pre=True) def _resolve_dataset(cls, v):