data_juicer_sandbox.evaluators module#
- class data_juicer_sandbox.evaluators.AccuracyEvaluator(eval_config: dict)[源代码]#
-
A simple evaluator to compare the labels in the predicted ones and ground truth.
The config file for this evaluator should at least include the following items: 1. type: must be "accuracy". 2. predicted_dataset_path: Required. The dataset path to the data that stores the predicted labels. 3. ground_truth_dataset_path: The dataset path to the data that stores the ground truth labels. If it's None,
we assume that the ground truth labels are already in the predicted_dataset_path.
predicted_label_key: the key name to store the predicted labels. '.' operator is allowed.
ground_truth_label_key: the key name to store the ground truth labels. '.' operator is allowed.
- class data_juicer_sandbox.evaluators.MSEEvaluator(eval_config: dict)[源代码]#
-
A simple evaluator to compute the MSE between the predicted values and ground truth.
The config file for this evaluator should at least include the following items: 1. type: must be "mse". 2. predicted_dataset_path: Required. The dataset path to the data that stores the predicted labels. 3. ground_truth_dataset_path: The dataset path to the data that stores the ground truth labels. If it's None,
we assume that the ground truth labels are already in the predicted_dataset_path.
predicted_value_key: the key name to store the predicted values. '.' operator is allowed.
ground_truth_value_key: the key name to store the ground truth values. '.' operator is allowed.