Add flash-attention 2 for windows (#4235)
This commit is contained in:
parent
258d046218
commit
3345da2ea4
10 changed files with 139 additions and 110 deletions
|
@ -1,4 +1,5 @@
|
|||
import random
|
||||
import traceback
|
||||
from pathlib import Path
|
||||
|
||||
import torch
|
||||
|
@ -24,6 +25,9 @@ except ModuleNotFoundError:
|
|||
'https://github.com/Dao-AILab/flash-attention#installation-and-features'
|
||||
)
|
||||
pass
|
||||
except Exception:
|
||||
logger.warning('Failed to load flash-attention due to the following error:\n')
|
||||
traceback.print_exc()
|
||||
|
||||
|
||||
class Exllamav2Model:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue