From e8f0f82db05244fc07253ddb3b643ca8fabba89b Mon Sep 17 00:00:00 2001 From: Davide <22103897+dvdblg@users.noreply.github.com> Date: Sun, 17 Mar 2024 18:48:59 +0100 Subject: [PATCH] feat(ml): add cache_dir option to OpenVINO EP (#8018) * add cache_dir option to OpenVINO EP * update provider options test to include cache_dir * use forward slash instead of string concatenation * fix cache_dir placement in provider options assertion --- machine-learning/app/models/base.py | 2 +- machine-learning/app/test_main.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/machine-learning/app/models/base.py b/machine-learning/app/models/base.py index 496012270c0..16adc159ad2 100644 --- a/machine-learning/app/models/base.py +++ b/machine-learning/app/models/base.py @@ -186,7 +186,7 @@ class InferenceModel(ABC): case "CPUExecutionProvider" | "CUDAExecutionProvider": option = {"arena_extend_strategy": "kSameAsRequested"} case "OpenVINOExecutionProvider": - option = {"device_type": "GPU_FP32"} + option = {"device_type": "GPU_FP32", "cache_dir": (self.cache_dir / "openvino").as_posix()} case _: option = {} options.append(option) diff --git a/machine-learning/app/test_main.py b/machine-learning/app/test_main.py index d79da0cbd5e..22038a65445 100644 --- a/machine-learning/app/test_main.py +++ b/machine-learning/app/test_main.py @@ -88,7 +88,7 @@ class TestBase: encoder = OpenCLIPEncoder("ViT-B-32__openai", providers=["OpenVINOExecutionProvider", "CPUExecutionProvider"]) assert encoder.provider_options == [ - {"device_type": "GPU_FP32"}, + {"device_type": "GPU_FP32", "cache_dir": (encoder.cache_dir / "openvino").as_posix()}, {"arena_extend_strategy": "kSameAsRequested"}, ]