mirror of
https://github.com/Doctorado-ML/benchmark.git
synced 2025-08-16 16:05:54 +00:00
Fix lexic mistake
This commit is contained in:
@@ -1162,14 +1162,14 @@ class Summary:
|
|||||||
|
|
||||||
|
|
||||||
class PairCheck:
|
class PairCheck:
|
||||||
def __init__(self, score, model_a, model_b, winners=False, loosers=False):
|
def __init__(self, score, model_a, model_b, winners=False, losers=False):
|
||||||
self.score = score
|
self.score = score
|
||||||
self.model_a = model_a
|
self.model_a = model_a
|
||||||
self.model_b = model_b
|
self.model_b = model_b
|
||||||
self.show_winners = winners
|
self.show_winners = winners
|
||||||
self.show_loosers = loosers
|
self.show_losers = losers
|
||||||
self.winners = []
|
self.winners = []
|
||||||
self.loosers = []
|
self.losers = []
|
||||||
self.tie = []
|
self.tie = []
|
||||||
|
|
||||||
def compute(self):
|
def compute(self):
|
||||||
@@ -1194,23 +1194,23 @@ class PairCheck:
|
|||||||
if result > 0:
|
if result > 0:
|
||||||
self.winners.append(result_a["dataset"])
|
self.winners.append(result_a["dataset"])
|
||||||
elif result < 0:
|
elif result < 0:
|
||||||
self.loosers.append(result_a["dataset"])
|
self.losers.append(result_a["dataset"])
|
||||||
else:
|
else:
|
||||||
self.tie.append(result_a["dataset"])
|
self.tie.append(result_a["dataset"])
|
||||||
|
|
||||||
def report(self):
|
def report(self):
|
||||||
print(f"{'Model':<20} {'File':<70} {'Score':<10} Win Tie Loose")
|
print(f"{'Model':<20} {'File':<70} {'Score':<10} Win Tie Lose")
|
||||||
print("=" * 20 + " " + "=" * 70 + " " + "=" * 10 + " === === =====")
|
print("=" * 20 + " " + "=" * 70 + " " + "=" * 10 + " === === ====")
|
||||||
print(f"{self.model_a:<20} {self.file_a:<70} {self.score_a:10.5f}")
|
print(f"{self.model_a:<20} {self.file_a:<70} {self.score_a:10.5f}")
|
||||||
print(
|
print(
|
||||||
f"{self.model_b:<20} {self.file_b:<70} "
|
f"{self.model_b:<20} {self.file_b:<70} "
|
||||||
f"{self.score_b:10.5f} "
|
f"{self.score_b:10.5f} "
|
||||||
f"{TextColor.GREEN}{len(self.winners):3d} {TextColor.YELLOW}"
|
f"{TextColor.GREEN}{len(self.winners):3d} {TextColor.YELLOW}"
|
||||||
f"{len(self.tie):3d} {TextColor.RED}{len(self.loosers):5d}"
|
f"{len(self.tie):3d} {TextColor.RED}{len(self.losers):4d}"
|
||||||
)
|
)
|
||||||
if self.show_winners:
|
if self.show_winners:
|
||||||
print(TextColor.GREEN + "Winners:")
|
print(TextColor.GREEN + "Winners:")
|
||||||
print(self.winners)
|
print(self.winners)
|
||||||
if self.show_loosers:
|
if self.show_losers:
|
||||||
print(TextColor.RED + "Loosers:")
|
print(TextColor.RED + "losers:")
|
||||||
print(self.loosers)
|
print(self.losers)
|
||||||
|
@@ -42,11 +42,11 @@ def parse_arguments():
|
|||||||
)
|
)
|
||||||
ap.add_argument(
|
ap.add_argument(
|
||||||
"-l",
|
"-l",
|
||||||
"--loose",
|
"--lose",
|
||||||
type=bool,
|
type=bool,
|
||||||
default=False,
|
default=False,
|
||||||
required=False,
|
required=False,
|
||||||
help="show loose results",
|
help="show lose results",
|
||||||
)
|
)
|
||||||
args = ap.parse_args()
|
args = ap.parse_args()
|
||||||
return (
|
return (
|
||||||
@@ -54,7 +54,7 @@ def parse_arguments():
|
|||||||
args.model1,
|
args.model1,
|
||||||
args.model2,
|
args.model2,
|
||||||
args.win,
|
args.win,
|
||||||
args.loose,
|
args.lose,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@@ -64,8 +64,8 @@ if __name__ == "__main__":
|
|||||||
model1,
|
model1,
|
||||||
model2,
|
model2,
|
||||||
win_results,
|
win_results,
|
||||||
loose_results,
|
lose_results,
|
||||||
) = parse_arguments()
|
) = parse_arguments()
|
||||||
pair_check = PairCheck(score, model1, model2, win_results, loose_results)
|
pair_check = PairCheck(score, model1, model2, win_results, lose_results)
|
||||||
pair_check.compute()
|
pair_check.compute()
|
||||||
pair_check.report()
|
pair_check.report()
|
||||||
|
Reference in New Issue
Block a user