You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Currently lookhead has already supported multimodal inputs while eagle has not. As eagle's input for embedding layer is not 2-D so when batch size > 1, it cannot expand correctly.
Hello, I have already hacked this feature under the help of TRT-LLM team member Ruoqian. We modified the implementation of expand function and changed the embedding layer of base model to be promptembedding when building the engine.
However, I found in TRT-LLM 0.19.0, the behaviour and generation quality is much worse than TRT-LLM 0.17.0, no matter it is eagle 1 or eagle 2. May I know if any other users report this?
Hi,
Currently lookhead has already supported multimodal inputs while eagle has not. As eagle's input for embedding layer is not 2-D so when batch size > 1, it cannot expand correctly.
TensorRT-LLM/tensorrt_llm/layers/embedding.py
Line 182 in c6f7d42
May I request for new feature that eagle supports multimodal input.
The text was updated successfully, but these errors were encountered: