diff --git a/benchmark/tests/Benchmark_test.py b/benchmark/tests/Benchmark_test.py index c2a1b13..7636521 100644 --- a/benchmark/tests/Benchmark_test.py +++ b/benchmark/tests/Benchmark_test.py @@ -1,19 +1,14 @@ import os -import unittest import shutil from io import StringIO from unittest.mock import patch from openpyxl import load_workbook +from .TestBase import TestBase from ..Utils import Folders, Files from ..Results import Benchmark -from .Excel_test import ExcelTest -class BenchmarkTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) - super().__init__(*args, **kwargs) - +class BenchmarkTest(TestBase): def tearDown(self) -> None: benchmark = Benchmark("accuracy", visualize=False) files = [ @@ -41,29 +36,25 @@ class BenchmarkTest(unittest.TestCase): benchmark = Benchmark("accuracy", visualize=False) benchmark.compile_results() benchmark.save_results() - with open(benchmark.get_result_file_name()) as f: - computed = f.readlines() - with open(os.path.join("test_files", "exreport_csv.test")) as f_exp: - expected = f_exp.readlines() - self.assertEqual(computed, expected) + self.check_file_file( + benchmark.get_result_file_name(), "exreport_csv.test" + ) def test_exreport_report(self): benchmark = Benchmark("accuracy", visualize=False) benchmark.compile_results() benchmark.save_results() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: benchmark.report(tex_output=False) - with open(os.path.join("test_files", "exreport_report.test")) as f: - expected = f.read() - self.assertEqual(fake_out.getvalue(), expected) + self.check_output_file(fake_out, "exreport_report.test") def test_exreport(self): benchmark = Benchmark("accuracy", visualize=False) benchmark.compile_results() benchmark.save_results() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: benchmark.exreport() - with open(os.path.join("test_files", "exreport.test")) as f: + with open(os.path.join(self.test_files, "exreport.test")) as f: expected_t = f.read() computed_t = fake_out.getvalue() computed_t = computed_t.split("\n") @@ -80,7 +71,7 @@ class BenchmarkTest(unittest.TestCase): benchmark = Benchmark("accuracy", visualize=False) benchmark.compile_results() benchmark.save_results() - with patch("sys.stdout", new=StringIO()): + with patch(self.output, new=StringIO()): benchmark.exreport() self.assertFalse(os.path.exists(Files.exreport_pdf)) self.assertFalse(os.path.exists(Folders.report)) @@ -89,43 +80,34 @@ class BenchmarkTest(unittest.TestCase): benchmark = Benchmark("unknown", visualize=False) benchmark.compile_results() benchmark.save_results() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: benchmark.exreport() - computed = fake_out.getvalue() - with open(os.path.join("test_files", "exreport_error.test")) as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "exreport_error.test") def test_tex_output(self): benchmark = Benchmark("accuracy", visualize=False) benchmark.compile_results() benchmark.save_results() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: benchmark.report(tex_output=True) - with open(os.path.join("test_files", "exreport_report.test")) as f: + with open(os.path.join(self.test_files, "exreport_report.test")) as f: expected = f.read() self.assertEqual(fake_out.getvalue(), expected) self.assertTrue(os.path.exists(benchmark.get_tex_file())) - with open(benchmark.get_tex_file()) as f: - computed = f.read() - with open(os.path.join("test_files", "exreport_tex.test")) as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_file_file(benchmark.get_tex_file(), "exreport_tex.test") def test_excel_output(self): benchmark = Benchmark("accuracy", visualize=False) benchmark.compile_results() benchmark.save_results() - with patch("sys.stdout", new=StringIO()): + with patch(self.output, new=StringIO()): benchmark.exreport() benchmark.excel() file_name = benchmark.get_excel_file_name() book = load_workbook(file_name) for sheet_name in book.sheetnames: sheet = book[sheet_name] - ExcelTest.check_excel_sheet( - self, sheet, f"exreport_excel_{sheet_name}.test" - ) + self.check_excel_sheet(sheet, f"exreport_excel_{sheet_name}.test") # ExcelTest.generate_excel_sheet( # self, sheet, f"exreport_excel_{sheet_name}.test" # ) diff --git a/benchmark/tests/BestResults_test.py b/benchmark/tests/BestResults_test.py index 17e7024..761752c 100644 --- a/benchmark/tests/BestResults_test.py +++ b/benchmark/tests/BestResults_test.py @@ -1,13 +1,9 @@ import os -import unittest +from .TestBase import TestBase from ..Experiments import BestResults, Datasets -class BestResultTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) - super().__init__(*args, **kwargs) - +class BestResultTest(TestBase): def test_load(self): expected = { "balance-scale": [ diff --git a/benchmark/tests/Dataset_test.py b/benchmark/tests/Dataset_test.py index b1d99e2..2e1e819 100644 --- a/benchmark/tests/Dataset_test.py +++ b/benchmark/tests/Dataset_test.py @@ -1,20 +1,16 @@ -import os import shutil -import unittest - +from .TestBase import TestBase from ..Experiments import Randomized, Datasets -class DatasetTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) +class DatasetTest(TestBase): + def setUp(self): self.datasets_values = { "balance-scale": (625, 4, 3), "balloons": (16, 4, 2), "iris": (150, 4, 3), "wine": (178, 13, 3), } - super().__init__(*args, **kwargs) def tearDown(self) -> None: self.set_env(".env.dist") diff --git a/benchmark/tests/Excel_test.py b/benchmark/tests/Excel_test.py index 2e06647..1b092b4 100644 --- a/benchmark/tests/Excel_test.py +++ b/benchmark/tests/Excel_test.py @@ -1,17 +1,12 @@ import os -import csv -import unittest from openpyxl import load_workbook from xlsxwriter import Workbook +from .TestBase import TestBase from ..Results import Excel from ..Utils import Folders -class ExcelTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) - super().__init__(*args, **kwargs) - +class ExcelTest(TestBase): def tearDown(self) -> None: files = [ "results_accuracy_STree_iMac27_2021-09-30_11:42:07_0.xlsx", @@ -24,29 +19,6 @@ class ExcelTest(unittest.TestCase): os.remove(file_name) return super().tearDown() - @staticmethod - def generate_excel_sheet(test, sheet, file_name): - with open(os.path.join("test_files", file_name), "w") as f: - for row in range(1, sheet.max_row + 1): - for col in range(1, sheet.max_column + 1): - value = sheet.cell(row=row, column=col).value - if value is not None: - print(f'{row};{col};"{value}"', file=f) - - @staticmethod - def check_excel_sheet(test, sheet, file_name): - with open(os.path.join("test_files", file_name), "r") as f: - expected = csv.reader(f, delimiter=";") - for row, col, value in expected: - if value.isdigit(): - value = int(value) - else: - try: - value = float(value) - except ValueError: - pass - test.assertEqual(sheet.cell(int(row), int(col)).value, value) - def test_report_excel_compared(self): file_name = "results_accuracy_STree_iMac27_2021-09-30_11:42:07_0.json" report = Excel(file_name, compare=True) @@ -54,7 +26,7 @@ class ExcelTest(unittest.TestCase): file_output = report.get_file_name() book = load_workbook(file_output) sheet = book["STree"] - self.check_excel_sheet(self, sheet, "excel_compared.test") + self.check_excel_sheet(sheet, "excel_compared.test") def test_report_excel(self): file_name = "results_accuracy_STree_iMac27_2021-09-30_11:42:07_0.json" @@ -63,7 +35,7 @@ class ExcelTest(unittest.TestCase): file_output = report.get_file_name() book = load_workbook(file_output) sheet = book["STree"] - self.check_excel_sheet(self, sheet, "excel.test") + self.check_excel_sheet(sheet, "excel.test") def test_Excel_Add_sheet(self): file_name = "results_accuracy_STree_iMac27_2021-10-27_09:40:40_0.json" @@ -79,6 +51,6 @@ class ExcelTest(unittest.TestCase): book.close() book = load_workbook(os.path.join(Folders.results, excel_file_name)) sheet = book["STree"] - self.check_excel_sheet(self, sheet, "excel_add_STree.test") + self.check_excel_sheet(sheet, "excel_add_STree.test") sheet = book["ODTE"] - self.check_excel_sheet(self, sheet, "excel_add_ODTE.test") + self.check_excel_sheet(sheet, "excel_add_ODTE.test") diff --git a/benchmark/tests/Experiment_test.py b/benchmark/tests/Experiment_test.py index 72e0e9b..610c1ca 100644 --- a/benchmark/tests/Experiment_test.py +++ b/benchmark/tests/Experiment_test.py @@ -1,14 +1,12 @@ import os import json -import unittest +from .TestBase import TestBase from ..Experiments import Experiment, Datasets -class ExperimentTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) +class ExperimentTest(TestBase): + def setUp(self): self.exp = self.build_exp() - super().__init__(*args, **kwargs) def build_exp(self, hyperparams=False, grid=False): params = { diff --git a/benchmark/tests/GridSearch_test.py b/benchmark/tests/GridSearch_test.py index 659f51f..ec1f7c7 100644 --- a/benchmark/tests/GridSearch_test.py +++ b/benchmark/tests/GridSearch_test.py @@ -1,14 +1,12 @@ import os import json -import unittest +from .TestBase import TestBase from ..Experiments import GridSearch, Datasets -class GridSearchTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) +class GridSearchTest(TestBase): + def setUp(self): self.grid = self.build_exp() - super().__init__(*args, **kwargs) def tearDown(self) -> None: grid = self.build_exp() diff --git a/benchmark/tests/Models_test.py b/benchmark/tests/Models_test.py index 3483031..8bffb2d 100644 --- a/benchmark/tests/Models_test.py +++ b/benchmark/tests/Models_test.py @@ -1,4 +1,3 @@ -import unittest import warnings from sklearn.exceptions import ConvergenceWarning from sklearn.tree import DecisionTreeClassifier, ExtraTreeClassifier @@ -12,10 +11,11 @@ from sklearn.datasets import load_wine from stree import Stree from wodt import Wodt from odte import Odte +from .TestBase import TestBase from ..Models import Models -class ModelTest(unittest.TestCase): +class ModelTest(TestBase): def test_Models(self): test = { "STree": Stree, diff --git a/benchmark/tests/PairCheck_test.py b/benchmark/tests/PairCheck_test.py index 4397772..f4c7b5b 100644 --- a/benchmark/tests/PairCheck_test.py +++ b/benchmark/tests/PairCheck_test.py @@ -1,15 +1,11 @@ import os -import unittest from io import StringIO from unittest.mock import patch +from .TestBase import TestBase from ..Results import PairCheck -class PairCheckTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) - super().__init__(*args, **kwargs) - +class PairCheckTest(TestBase): def build_model( self, score="accuracy", @@ -23,46 +19,35 @@ class PairCheckTest(unittest.TestCase): def test_pair_check(self): report = self.build_model(model1="ODTE", model2="STree") report.compute() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() computed = fake_out.getvalue() - with open(os.path.join("test_files", "paircheck.test"), "r") as f: + with open(os.path.join(self.test_files, "paircheck.test"), "r") as f: expected = f.read() self.assertEqual(computed, expected) def test_pair_check_win(self): report = self.build_model(win=True) report.compute() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() - computed = fake_out.getvalue() - with open(os.path.join("test_files", "paircheck_win.test"), "r") as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "paircheck_win.test") def test_pair_check_lose(self): report = self.build_model( model1="RandomForest", model2="STree", lose=True ) report.compute() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() - computed = fake_out.getvalue() - with open(os.path.join("test_files", "paircheck_lose.test"), "r") as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "paircheck_lose.test") def test_pair_check_win_lose(self): report = self.build_model(win=True, lose=True) report.compute() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() - computed = fake_out.getvalue() - with open( - os.path.join("test_files", "paircheck_win_lose.test"), "r" - ) as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "paircheck_win_lose.test") def test_pair_check_store_result(self): report = self.build_model(win=True, lose=True) diff --git a/benchmark/tests/Report_test.py b/benchmark/tests/Report_test.py index f6c8302..cdc1ef4 100644 --- a/benchmark/tests/Report_test.py +++ b/benchmark/tests/Report_test.py @@ -1,16 +1,11 @@ -import os -import unittest from io import StringIO from unittest.mock import patch +from .TestBase import TestBase from ..Results import Report, BaseReport, ReportBest from ..Utils import Symbols -class ReportTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) - super().__init__(*args, **kwargs) - +class ReportTest(TestBase): def test_BaseReport(self): with patch.multiple(BaseReport, __abstractmethods__=set()): file_name = ( @@ -27,22 +22,18 @@ class ReportTest(unittest.TestCase): file_name="results/results_accuracy_STree_iMac27_2021-09-30_11:" "42:07_0.json" ) - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() - with open("test_files/report.test", "r") as f: - expected = f.read() - self.assertEqual(fake_out.getvalue(), expected) + self.check_output_file(fake_out, "report.test") def test_report_without_folder(self): report = Report( file_name="results_accuracy_STree_iMac27_2021-09-30_11:42:07_0" ".json" ) - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() - with open("test_files/report.test", "r") as f: - expected = f.read() - self.assertEqual(fake_out.getvalue(), expected) + self.check_output_file(fake_out, "report.test") def test_report_compared(self): report = Report( @@ -50,11 +41,9 @@ class ReportTest(unittest.TestCase): ".json", compare=True, ) - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() - with open("test_files/report_compared.test", "r") as f: - expected = f.read() - self.assertEqual(fake_out.getvalue(), expected) + self.check_output_file(fake_out, "report_compared.test") def test_compute_status(self): file_name = "results_accuracy_STree_iMac27_2021-10-27_09:40:40_0.json" @@ -62,7 +51,7 @@ class ReportTest(unittest.TestCase): file_name=file_name, compare=True, ) - with patch("sys.stdout", new=StringIO()): + with patch(self.output, new=StringIO()): report.report() res = report._compute_status("balloons", 0.99) self.assertEqual(res, Symbols.better_best) @@ -75,25 +64,18 @@ class ReportTest(unittest.TestCase): def test_report_best(self): report = ReportBest("accuracy", "STree", best=True, grid=False) - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() - with open("test_files/report_best.test", "r") as f: - expected = f.read() - self.assertEqual(fake_out.getvalue(), expected) + self.check_output_file(fake_out, "report_best.test") def test_report_grid(self): report = ReportBest("accuracy", "STree", best=False, grid=True) - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() - with open("test_files/report_grid.test", "r") as f: - expected = f.read() - self.assertEqual(fake_out.getvalue(), expected) + self.check_output_file(fake_out, "report_grid.test") def test_report_best_both(self): report = ReportBest("accuracy", "STree", best=True, grid=True) - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.report() - with open("test_files/report_best.test", "r") as f: - expected = f.read() - - self.assertEqual(fake_out.getvalue(), expected) + self.check_output_file(fake_out, "report_best.test") diff --git a/benchmark/tests/SQL_test.py b/benchmark/tests/SQL_test.py index f0e89c8..7acf9bf 100644 --- a/benchmark/tests/SQL_test.py +++ b/benchmark/tests/SQL_test.py @@ -1,14 +1,10 @@ import os -import unittest +from .TestBase import TestBase from ..Results import SQL from ..Utils import Folders -class SQLTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) - super().__init__(*args, **kwargs) - +class SQLTest(TestBase): def tearDown(self) -> None: files = [ "results_accuracy_ODTE_Galgo_2022-04-20_10:52:20_0.sql", @@ -26,9 +22,4 @@ class SQLTest(unittest.TestCase): file_name = os.path.join( Folders.results, file_name.replace(".json", ".sql") ) - - with open(file_name, "r") as file: - computed = file.read() - with open(os.path.join("test_files", "sql.test")) as f_exp: - expected = f_exp.read() - self.assertEqual(computed, expected) + self.check_file_file(file_name, "sql.test") diff --git a/benchmark/tests/Summary_test.py b/benchmark/tests/Summary_test.py index be09a9a..bd5b0b7 100644 --- a/benchmark/tests/Summary_test.py +++ b/benchmark/tests/Summary_test.py @@ -1,15 +1,10 @@ -import os -import unittest from io import StringIO from unittest.mock import patch +from .TestBase import TestBase from ..Results import Summary -class SummaryTest(unittest.TestCase): - def __init__(self, *args, **kwargs): - os.chdir(os.path.dirname(os.path.abspath(__file__))) - super().__init__(*args, **kwargs) - +class SummaryTest(TestBase): def test_summary_without_model(self): report = Summary() report.acquire() @@ -135,85 +130,57 @@ class SummaryTest(unittest.TestCase): def test_summary_list_results_model(self): report = Summary() report.acquire() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.list_results(model="STree") - computed = fake_out.getvalue() - with open( - os.path.join("test_files", "summary_list_model.test"), "r" - ) as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "summary_list_model.test") def test_summary_list_results_score(self): report = Summary() report.acquire() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.list_results(score="accuracy") - computed = fake_out.getvalue() - with open( - os.path.join("test_files", "summary_list_score.test"), "r" - ) as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "summary_list_score.test") def test_summary_list_results_n(self): report = Summary() report.acquire() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.list_results(score="accuracy", number=3) - computed = fake_out.getvalue() - with open(os.path.join("test_files", "summary_list_n.test"), "r") as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "summary_list_n.test") def test_summary_list_hiden(self): report = Summary(hidden=True) report.acquire() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.list_results(score="accuracy") - computed = fake_out.getvalue() - with open( - os.path.join("test_files", "summary_list_hidden.test"), "r" - ) as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "summary_list_hidden.test") def test_show_result_no_title(self): report = Summary() report.acquire() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: title = "" best = report.best_result( criterion="model", value="STree", score="accuracy" ) report.show_result(data=best, title=title) - computed = fake_out.getvalue() - with open( - os.path.join("test_files", "summary_show_results.test"), "r" - ) as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "summary_show_results.test") def test_show_result_title(self): report = Summary() report.acquire() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: title = "**Title**" best = report.best_result( criterion="model", value="STree", score="accuracy" ) report.show_result(data=best, title=title) - computed = fake_out.getvalue() - with open( - os.path.join("test_files", "summary_show_results_title.test"), "r" - ) as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "summary_show_results_title.test") def test_show_result_no_data(self): report = Summary() report.acquire() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: title = "**Test**" report.show_result(data={}, title=title) computed = fake_out.getvalue() @@ -245,11 +212,6 @@ class SummaryTest(unittest.TestCase): def test_show_top(self): report = Summary() report.acquire() - with patch("sys.stdout", new=StringIO()) as fake_out: + with patch(self.output, new=StringIO()) as fake_out: report.show_top() - computed = fake_out.getvalue() - with open( - os.path.join("test_files", "summary_show_top.test"), "r" - ) as f: - expected = f.read() - self.assertEqual(computed, expected) + self.check_output_file(fake_out, "summary_show_top.test") diff --git a/benchmark/tests/TestBase.py b/benchmark/tests/TestBase.py new file mode 100644 index 0000000..a30cba5 --- /dev/null +++ b/benchmark/tests/TestBase.py @@ -0,0 +1,44 @@ +import os +import csv +import unittest + + +class TestBase(unittest.TestCase): + def __init__(self, *args, **kwargs): + os.chdir(os.path.dirname(os.path.abspath(__file__))) + self.test_files = "test_files" + self.output = "sys.stdout" + super().__init__(*args, **kwargs) + + def generate_excel_sheet(self, sheet, file_name): + with open(os.path.join(self.test_files, file_name), "w") as f: + for row in range(1, sheet.max_row + 1): + for col in range(1, sheet.max_column + 1): + value = sheet.cell(row=row, column=col).value + if value is not None: + print(f'{row};{col};"{value}"', file=f) + + def check_excel_sheet(self, sheet, file_name): + with open(os.path.join(self.test_files, file_name), "r") as f: + expected = csv.reader(f, delimiter=";") + for row, col, value in expected: + if value.isdigit(): + value = int(value) + else: + try: + value = float(value) + except ValueError: + pass + self.assertEqual(sheet.cell(int(row), int(col)).value, value) + + def check_output_file(self, output, file_name): + with open(os.path.join(self.test_files, file_name)) as f: + expected = f.read() + self.assertEqual(output.getvalue(), expected) + + def check_file_file(self, computed_file, expected_file): + with open(computed_file) as f: + computed = f.read() + with open(os.path.join(self.test_files, expected_file)) as f: + expected = f.read() + self.assertEqual(computed, expected) diff --git a/benchmark/tests/Util_test.py b/benchmark/tests/Util_test.py index c4b7b5a..b3d87f1 100644 --- a/benchmark/tests/Util_test.py +++ b/benchmark/tests/Util_test.py @@ -1,11 +1,11 @@ import os import sys -import unittest import argparse +from .TestBase import TestBase from ..Utils import Folders, Files, Symbols, TextColor, EnvData, EnvDefault -class UtilTest(unittest.TestCase): +class UtilTest(TestBase): def test_Folders(self): self.assertEqual("results", Folders.results) self.assertEqual("hidden_results", Folders.hidden_results)