From 808ad82ccb9ad50a480e2434d68ea882e09eba4f Mon Sep 17 00:00:00 2001 From: yechank <161688079+yechank-nvidia@users.noreply.github.com> Date: Sun, 19 Apr 2026 09:02:50 -0700 Subject: [PATCH 1/2] revert backend Signed-off-by: yechank <161688079+yechank-nvidia@users.noreply.github.com> --- tensorrt_llm/_torch/models/modeling_radio.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tensorrt_llm/_torch/models/modeling_radio.py b/tensorrt_llm/_torch/models/modeling_radio.py index 9286554e0da4..5054f730e3ee 100644 --- a/tensorrt_llm/_torch/models/modeling_radio.py +++ b/tensorrt_llm/_torch/models/modeling_radio.py @@ -1010,7 +1010,7 @@ class RADIOVisionModel(PreTrainedModel): def __init__(self, model_config: model_config_lib.ModelConfig, disable_quantization: bool = True, - vision_attn_backend: Optional[str] = "FLASHINFER"): + vision_attn_backend: Optional[str] = "TRTLLM"): """ Args: model_config: Model configuration. From 9b4e323a11a27284db8edc9581751bf8b1ce404e Mon Sep 17 00:00:00 2001 From: yechank <161688079+yechank-nvidia@users.noreply.github.com> Date: Sun, 19 Apr 2026 09:20:03 -0700 Subject: [PATCH 2/2] revert reference data Signed-off-by: yechank <161688079+yechank-nvidia@users.noreply.github.com> --- .../_torch/modeling/test_modeling_nemotron_nano_v2_vl.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/tests/unittest/_torch/modeling/test_modeling_nemotron_nano_v2_vl.py b/tests/unittest/_torch/modeling/test_modeling_nemotron_nano_v2_vl.py index 9ef7b39cd0d6..d535f93cff34 100644 --- a/tests/unittest/_torch/modeling/test_modeling_nemotron_nano_v2_vl.py +++ b/tests/unittest/_torch/modeling/test_modeling_nemotron_nano_v2_vl.py @@ -179,13 +179,13 @@ def test_nemotron_nano_v2_vl_model_sanity_check( reference_data_dict = { "image": { "single": torch.tensor( - [-8.9814e-01, -1.5258e-01, -7.6061e-04, -6.3735e-01, -3.1303e-02] + [-8.5795e-01, -1.5373e-01, -7.2846e-04, -6.3667e-01, -3.1307e-02] ), - "multiple": torch.tensor([-0.4717, -0.7776, -0.0251, -1.2290, -1.0705]), + "multiple": torch.tensor([-0.5846, -0.6330, -0.0124, -0.1146, -0.0172]), }, "video": { - "single": torch.tensor([-1.4745, -0.0674, -1.4121, -0.2152, -1.6297]), - "multiple": torch.tensor([-0.9425, -0.2328, -0.0083, -1.6257, -0.6572]), + "single": torch.tensor([-0.5612, -0.0334, -0.8856, -0.4056, -0.6041]), + "multiple": torch.tensor([-0.4943, -0.9333, -0.0096, -1.2496, -0.9441]), }, } prompts = data_dict_fixture[modality][condition]["prompts"]