Safetensors

Commit History

4:3 aspect ratio
fbb3e9e

jamesr66a commited on

Fluminato
d391c5d

jamesr66a commited on

Fix issue where cublas linear not installed causing TypeError
56c313c

aredden commited on

Fix unbound local error when pad_tokens=False
9e376d8

aredden commited on

fix unloading bug
7954c3b

aredden commited on

update readme
fad72ff

aredden commited on

Make lora loading api endpoint functional
fb3cdc4

aredden commited on

Improve lora implementation
71170f1

aredden commited on

lora unloading initial
d33c2e4

aredden commited on

add benchmarks numbers for rtx4000ada (non-sff)
d316f04

flowpoint commited on

Update README.md, Fix license link
49c776c
unverified

aredden commited on

Improved precision / reduced frequency of nan outputs, allow bf16 t5, f32 rmsnorm, larger clamp
f708e90

aredden commited on

Create LICENSE
3cc2f3f
unverified

aredden commited on

Merge pull request #11 from ClashLuke/patch-1
138a97c
unverified

aredden commited on

remove wrongly added lines
6fbe1d5
unverified

Lucas Nestler commited on

add h100
c541d34
unverified

Lucas Nestler commited on

remove torchao dependency, quantize entirely via linear
d45a331

aredden commited on

Fix issues with loading F8Linear from state dict when init_scale not initialized & loaded from meta device
3ddaa67

aredden commited on

Fix issue where lora alpha is not correct if lora from transformers checkpoint
7a7b2c1

aredden commited on

Small fix for issue where f16 CublasLinear layers weren't being used even when available.
6d82dcc

aredden commited on

Ensure repo only accesses CublasLinear lazily
00f5d2c

aredden commited on

Fix issue loading loras
fee1af5

aredden commited on

Add lora loading
a71da07

aredden commited on

Merge pull request #3 from aredden/improved_precision
af20799
unverified

aredden commited on

Update readme
49f2076

aredden commited on

Add quantize embedders/modulation to argparse options
604f17d

aredden commited on

Remove f8 flux, instead configure at load, improved quality & corrected configs
1f9e684

aredden commited on

Fix issue where torch.dtype throws error when converting to dtype
fb7df61

aredden commited on

Move compile out of FluxPipeline init
ac049be

aredden commited on

Dynamic swap with cublas linear / optional improved precision with vram drawback
37bd8c1

aredden commited on

Allow overriding config values from load_pipeline_from_config_path
25ae92b

aredden commited on

Merge pull request #2 from XmYx/main
b84f35e
unverified

aredden commited on

python 3.10 compatibility
9b84867
unverified

mixy89 commited on

Merge pull request #1 from dsingal0/main
c247326
unverified

aredden commited on

Remove unnecessary tokenization (still needs work)
44e4014

aredden commited on

Add case where seed is string & try/catch if invalid
7cec457

aredden commited on

Remove unnecessary synchronize, add more universal seeding & limit if run on windows
ffa6ff7

aredden commited on

Remove more unnecessary code, fix small typing hickup
6d0762c

aredden commited on

Remove unnecessary code, hide prints behind debug flag, hide warnings
0f3134f

aredden commited on

Update README.md - Fix config path names
4a2503e
unverified

aredden commited on

Small README update for clarification
42be379

aredden commited on

Add all relevent args to argparser & update readme
9dc5b0b

aredden commited on

Merge branch 'main' of https://github.com/aredden/flux-fp16-acc-api into main
58082af

aredden commited on

Adding device specific configs & more input image type options + small model spec from args change
e81fa57

aredden commited on

Update README.md
d9b39c9
unverified

aredden commited on

make table more clear- again
21a2bf7

aredden commited on

make table more clear
50c4b9f

aredden commited on

Add speed comparison
efb953a

aredden commited on

Add uvicorn requirement
6bd57ab

aredden commited on