Updated FA info
Browse files
README.md
CHANGED
@@ -65,7 +65,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
65 |
model_name,
|
66 |
torch_dtype=torch.bfloat16,
|
67 |
device_map="auto",
|
68 |
-
attn_implementation="flash_attention_2", # Use "
|
69 |
use_cache=True,
|
70 |
trust_remote_code=True,
|
71 |
)
|
|
|
65 |
model_name,
|
66 |
torch_dtype=torch.bfloat16,
|
67 |
device_map="auto",
|
68 |
+
attn_implementation="flash_attention_2", # Use "eager" (or omit) if flash_attn is not installed
|
69 |
use_cache=True,
|
70 |
trust_remote_code=True,
|
71 |
)
|