Remove entmax dep
Browse files- requirements.txt +0 -1
- setup.py +0 -1
- tortoise/models/xtransformers.py +1 -2
requirements.txt
CHANGED
@@ -6,6 +6,5 @@ inflect
|
|
6 |
progressbar
|
7 |
einops
|
8 |
unidecode
|
9 |
-
entmax
|
10 |
scipy
|
11 |
librosa
|
|
|
6 |
progressbar
|
7 |
einops
|
8 |
unidecode
|
|
|
9 |
scipy
|
10 |
librosa
|
setup.py
CHANGED
@@ -21,7 +21,6 @@ setuptools.setup(
|
|
21 |
'progressbar',
|
22 |
'einops',
|
23 |
'unidecode',
|
24 |
-
'entmax',
|
25 |
'scipy',
|
26 |
'librosa',
|
27 |
'transformers',
|
|
|
21 |
'progressbar',
|
22 |
'einops',
|
23 |
'unidecode',
|
|
|
24 |
'scipy',
|
25 |
'librosa',
|
26 |
'transformers',
|
tortoise/models/xtransformers.py
CHANGED
@@ -10,7 +10,6 @@ from collections import namedtuple
|
|
10 |
from einops import rearrange, repeat, reduce
|
11 |
from einops.layers.torch import Rearrange
|
12 |
|
13 |
-
from entmax import entmax15
|
14 |
from torch.utils.checkpoint import checkpoint
|
15 |
|
16 |
DEFAULT_DIM_HEAD = 64
|
@@ -556,7 +555,7 @@ class Attention(nn.Module):
|
|
556 |
self.sparse_topk = sparse_topk
|
557 |
|
558 |
# entmax
|
559 |
-
self.attn_fn =
|
560 |
|
561 |
# add memory key / values
|
562 |
self.num_mem_kv = num_mem_kv
|
|
|
10 |
from einops import rearrange, repeat, reduce
|
11 |
from einops.layers.torch import Rearrange
|
12 |
|
|
|
13 |
from torch.utils.checkpoint import checkpoint
|
14 |
|
15 |
DEFAULT_DIM_HEAD = 64
|
|
|
555 |
self.sparse_topk = sparse_topk
|
556 |
|
557 |
# entmax
|
558 |
+
self.attn_fn = F.softmax
|
559 |
|
560 |
# add memory key / values
|
561 |
self.num_mem_kv = num_mem_kv
|