diff --git a/qa/L0_tftrt_optimization/test.sh b/qa/L0_tftrt_optimization/test.sh index da0b3d7e93..04dcdc2f65 100755 --- a/qa/L0_tftrt_optimization/test.sh +++ b/qa/L0_tftrt_optimization/test.sh @@ -53,8 +53,6 @@ source ../common/util.sh RET=0 -# FIXME [DLIS-6660] Update TF image models to support dynamic batching -# Currently, models cannot load with TF-TRT optimization due to the removal of implicit batch support in TensorRT v10 onwards for MODEL in \ graphdef_float32_float32_float32 \ savedmodel_float32_float32_float32; do diff --git a/qa/L0_warmup/test.sh b/qa/L0_warmup/test.sh index 28c34e0b2f..aeed873b25 100755 --- a/qa/L0_warmup/test.sh +++ b/qa/L0_warmup/test.sh @@ -288,10 +288,6 @@ for BACKEND in ${BACKENDS}; do fi if [ "$BACKEND" == "graphdef" ]; then - # FIXME [DLIS-6660] Update TF image models to support dynamic batching - # Currently, models cannot load with TF-TRT optimization due to the removal of implicit batch support in TensorRT v10 onwards - continue - # Show effect of warmup by using a TF model with TF-TRT optimization which is # known to be slow on first inference. # Note: model can be obatined via the fetching script in docs/example