| @@ -29,7 +29,7 @@ class NLITest(unittest.TestCase): | |||||
| f'sentence1: {self.sentence1}\nsentence2: {self.sentence2}\n' | f'sentence1: {self.sentence1}\nsentence2: {self.sentence2}\n' | ||||
| f'pipeline1: {pipeline2(input=(self.sentence1, self.sentence2))}') | f'pipeline1: {pipeline2(input=(self.sentence1, self.sentence2))}') | ||||
| @unittest.skipUnless(test_level() >= 0, 'skip test in current test level') | |||||
| @unittest.skipUnless(test_level() >= 2, 'skip test in current test level') | |||||
| def test_run_with_model_from_modelhub(self): | def test_run_with_model_from_modelhub(self): | ||||
| model = Model.from_pretrained(self.model_id) | model = Model.from_pretrained(self.model_id) | ||||
| tokenizer = NLIPreprocessor(model.model_dir) | tokenizer = NLIPreprocessor(model.model_dir) | ||||
| @@ -37,7 +37,7 @@ class NLITest(unittest.TestCase): | |||||
| task=Tasks.nli, model=model, preprocessor=tokenizer) | task=Tasks.nli, model=model, preprocessor=tokenizer) | ||||
| print(pipeline_ins(input=(self.sentence1, self.sentence2))) | print(pipeline_ins(input=(self.sentence1, self.sentence2))) | ||||
| @unittest.skipUnless(test_level() >= 2, 'skip test in current test level') | |||||
| @unittest.skipUnless(test_level() >= 0, 'skip test in current test level') | |||||
| def test_run_with_model_name(self): | def test_run_with_model_name(self): | ||||
| pipeline_ins = pipeline(task=Tasks.nli, model=self.model_id) | pipeline_ins = pipeline(task=Tasks.nli, model=self.model_id) | ||||
| print(pipeline_ins(input=(self.sentence1, self.sentence2))) | print(pipeline_ins(input=(self.sentence1, self.sentence2))) | ||||