You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I encountered multiple issues when running the Janus-1.3B model inference script, both with and without enabling FlashAttention. These errors prevent successful execution of the model in a standard environment, such as Google Colab. Below are the details of the issues:
Problem 1: Error Without FlashAttention
When running the inference script without FlashAttention, the following error is raised during the generate function call: NameError: name '_flash_supports_window_size' is not defined
Problem 2: FlashAttention 2.0 Unsupported
To address the above error, I attempted to install FlashAttention-2 and enable it via the attn_implementation="flash_attention_2" argument. However, this raises the following error: ValueError: MultiModalityCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co/deepseek-ai/Janus-1.3B/discussions/new
Steps to Reproduce:
Install FlashAttention 2.0 via: pip install flash-attn --no-build-isolation
2.Modify the script to include:
I encountered multiple issues when running the Janus-1.3B model inference script, both with and without enabling FlashAttention. These errors prevent successful execution of the model in a standard environment, such as Google Colab. Below are the details of the issues:
Problem 1: Error Without FlashAttention
When running the inference script without FlashAttention, the following error is raised during the generate function call:
NameError: name '_flash_supports_window_size' is not defined
Problem 2: FlashAttention 2.0 Unsupported
To address the above error, I attempted to install FlashAttention-2 and enable it via the attn_implementation="flash_attention_2" argument. However, this raises the following error:
ValueError: MultiModalityCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co/deepseek-ai/Janus-1.3B/discussions/new
Steps to Reproduce:
pip install flash-attn --no-build-isolation
2.Modify the script to include:
Thank you for your assistance! Let me know if you require additional details or logs to reproduce the issues.
The text was updated successfully, but these errors were encountered: