From f171021615fa8442184f924655f2e00cb4c0d892 Mon Sep 17 00:00:00 2001 From: "yingdong.han" Date: Mon, 30 Jun 2025 19:56:25 +0800 Subject: [PATCH] update doc --- deployment/tensorrt_llm/ReadMe.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/deployment/tensorrt_llm/ReadMe.md b/deployment/tensorrt_llm/ReadMe.md index eb51810..8950f9a 100644 --- a/deployment/tensorrt_llm/ReadMe.md +++ b/deployment/tensorrt_llm/ReadMe.md @@ -4,7 +4,7 @@ ## ✅ Introduction The Dolphin model employs a **Swin Encoder + MBart Decoder** architecture. In the HuggingFace Transformers [Config](https://huggingface.co/ByteDance/Dolphin/blob/main/config.json), -its architectures field is specified as "VisionEncoderDecoderModel". Dolphin, Nougat, and Donut share the same model architecture. TensorRT-LLM has already supported the Nougat model. +its architectures field is specified as "VisionEncoderDecoderModel". **Dolphin**, **[Nougat](https://huggingface.co/docs/transformers/model_doc/nougat)**, and **[Donut](https://huggingface.co/docs/transformers/model_doc/donut)** share the same model architecture. TensorRT-LLM has already supported the Nougat model. Following Nougat's conversion script, we have successfully implemented Dolphin on TensorRT-LLM. **Note:** [prompt_ids](./dolphin_runner.py#L120) MUST be of **int32** type, otherwise TensorRT-LLM will produce incorrect results.