mirror of
https://github.com/Doctorado-ML/benchmark.git
synced 2025-08-15 23:45:54 +00:00
add score names to help in scripts
This commit is contained in:
@@ -14,7 +14,8 @@ def parse_arguments():
|
|||||||
action=EnvDefault,
|
action=EnvDefault,
|
||||||
envvar="score",
|
envvar="score",
|
||||||
required=True,
|
required=True,
|
||||||
help="score name {accuracy, f1_micro, f1_macro, all}",
|
help="score name {accuracy, f1-micro, f1-macro, f1-weighted, "
|
||||||
|
"roc-auc-ovr, all}",
|
||||||
)
|
)
|
||||||
args = ap.parse_args()
|
args = ap.parse_args()
|
||||||
return (args.score,)
|
return (args.score,)
|
||||||
@@ -22,7 +23,13 @@ def parse_arguments():
|
|||||||
|
|
||||||
(score,) = parse_arguments()
|
(score,) = parse_arguments()
|
||||||
|
|
||||||
all_metrics = ["accuracy", "f1-macro", "f1-micro"]
|
all_metrics = [
|
||||||
|
"accuracy",
|
||||||
|
"f1-macro",
|
||||||
|
"f1-micro",
|
||||||
|
"f1-weighted",
|
||||||
|
"roc-auc-ovr",
|
||||||
|
]
|
||||||
|
|
||||||
metrics = all_metrics if score == "all" else [score]
|
metrics = all_metrics if score == "all" else [score]
|
||||||
|
|
||||||
|
@@ -18,7 +18,8 @@ def parse_arguments():
|
|||||||
envvar="score",
|
envvar="score",
|
||||||
type=str,
|
type=str,
|
||||||
required=True,
|
required=True,
|
||||||
help="score name {accuracy, f1_macro, ...}",
|
help="score name {accuracy, f1-micro, f1-macro, f1-weighted, "
|
||||||
|
"roc-auc-ovr, all}",
|
||||||
)
|
)
|
||||||
ap.add_argument(
|
ap.add_argument(
|
||||||
"-P",
|
"-P",
|
||||||
|
Reference in New Issue
Block a user