Spaces:
Running
on
A10G
Running
on
A10G
Commit
·
b27e8d4
1
Parent(s):
540e4d3
re-encode
Browse files
app.py
CHANGED
@@ -99,7 +99,7 @@ class BottleneckT5Autoencoder:
|
|
99 |
top_p=top_p,
|
100 |
num_return_sequences=1,
|
101 |
min_new_tokens=min_new_tokens,
|
102 |
-
#
|
103 |
)
|
104 |
return self.tokenizer.decode(output[0], skip_special_tokens=True)
|
105 |
|
@@ -118,6 +118,7 @@ def generate(prompt, in_embs=None,):
|
|
118 |
print('From embeds.')
|
119 |
in_embs = in_embs / in_embs.abs().max() * .15
|
120 |
text = autoencoder.generate_from_latent(in_embs.to('cuda').to(dtype=torch.bfloat16), temperature=.3, top_p=.99, min_new_tokens=5)
|
|
|
121 |
return text, in_embs.to('cpu')
|
122 |
|
123 |
|
|
|
99 |
top_p=top_p,
|
100 |
num_return_sequences=1,
|
101 |
min_new_tokens=min_new_tokens,
|
102 |
+
#num_beams=4,
|
103 |
)
|
104 |
return self.tokenizer.decode(output[0], skip_special_tokens=True)
|
105 |
|
|
|
118 |
print('From embeds.')
|
119 |
in_embs = in_embs / in_embs.abs().max() * .15
|
120 |
text = autoencoder.generate_from_latent(in_embs.to('cuda').to(dtype=torch.bfloat16), temperature=.3, top_p=.99, min_new_tokens=5)
|
121 |
+
in_embs = autoencoder.embed(text).to('cuda')
|
122 |
return text, in_embs.to('cpu')
|
123 |
|
124 |
|