|
3 | 3 | import shutil
|
4 | 4 | import tempfile
|
5 | 5 | from typing import Type, Tuple, List, Any, Dict
|
6 |
| -from packaging import version |
| 6 | +import onnx |
7 | 7 | import torch
|
8 | 8 | import diffusers
|
9 |
| -import onnxruntime as ort |
10 | 9 | import optimum.onnxruntime
|
11 | 10 | from modules import shared
|
12 | 11 | from modules.paths import models_path
|
@@ -245,7 +244,7 @@ def run_olive(self, submodels: List[str], in_dir: os.PathLike, out_dir: os.PathL
|
245 | 244 | optimized_model_paths = {}
|
246 | 245 |
|
247 | 246 | for submodel in submodels:
|
248 |
| - log.info(f"\nProcessing {submodel}") |
| 247 | + print(f"\nProcessing {submodel}") |
249 | 248 |
|
250 | 249 | with open(os.path.join(sd_configs_path, "olive", 'sdxl' if self._is_sdxl else 'sd', f"{submodel}.json"), "r", encoding="utf-8") as config_file:
|
251 | 250 | olive_config: Dict[str, Dict[str, Dict]] = json.load(config_file)
|
@@ -282,7 +281,7 @@ def run_olive(self, submodels: List[str], in_dir: os.PathLike, out_dir: os.PathL
|
282 | 281 | **processor_final_pass_footprint["model_config"]["config"]
|
283 | 282 | ).model_path
|
284 | 283 |
|
285 |
| - log.info(f"Olive: Successfully processed model: submodel={submodel}") |
| 284 | + print(f"Olive: Successfully processed model: submodel={submodel}") |
286 | 285 |
|
287 | 286 | for submodel in submodels:
|
288 | 287 | src_path = optimized_model_paths[submodel]
|
|
0 commit comments