From eccc3cfac6576a4c5e5f1ae5066ed3a8e4805a89 Mon Sep 17 00:00:00 2001 From: Iman Tabrizian Date: Mon, 12 Jun 2023 11:25:31 -0400 Subject: [PATCH] Fix jetson --- qa/L0_backend_python/python_test.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/qa/L0_backend_python/python_test.py b/qa/L0_backend_python/python_test.py index b28b0c48b3..71f3e68876 100644 --- a/qa/L0_backend_python/python_test.py +++ b/qa/L0_backend_python/python_test.py @@ -38,7 +38,6 @@ import os from tritonclient.utils import * -import tritonclient.utils.cuda_shared_memory as cuda_shared_memory import tritonclient.http as httpclient TEST_JETSON = bool(int(os.environ.get('TEST_JETSON', 0))) @@ -155,9 +154,11 @@ def test_growth_error(self): self._infer_help(model_name, shape, dtype) # GPU tensors are not supported on jetson + # CUDA Shared memory is not supported on jetson if not TEST_JETSON: - # CUDA Shared memory is not supported on jetson + def test_gpu_tensor_error(self): + import tritonclient.utils.cuda_shared_memory as cuda_shared_memory model_name = 'identity_bool' with httpclient.InferenceServerClient("localhost:8000") as client: input_data = np.array([[True] * 1000], dtype=bool) @@ -184,6 +185,7 @@ def test_gpu_tensor_error(self): cuda_shared_memory.destroy_shared_memory_region(shm0_handle) def test_dlpack_tensor_error(self): + import tritonclient.utils.cuda_shared_memory as cuda_shared_memory model_name = 'dlpack_identity' with httpclient.InferenceServerClient("localhost:8000") as client: input_data = np.array([[1] * 1000], dtype=np.float32)