Commit History

fix unloading bug
cc15333

aredden commited on

update readme
61a425a

aredden commited on

Make lora loading api endpoint functional
e8041a4

aredden commited on

Improve lora implementation
e21ae14

aredden commited on

lora unloading initial
264acad

aredden commited on

Merge pull request #14 from flowpoint/rtx4000ada_bench
e6814a2
unverified

aredden commited on

add benchmarks numbers for rtx4000ada (non-sff)
d316f04

flowpoint commited on

Update README.md, Fix license link
49c776c
unverified

aredden commited on

Improved precision / reduced frequency of nan outputs, allow bf16 t5, f32 rmsnorm, larger clamp
f708e90

aredden commited on

Create LICENSE
3cc2f3f
unverified

aredden commited on

Merge pull request #11 from ClashLuke/patch-1
138a97c
unverified

aredden commited on

remove wrongly added lines
6fbe1d5
unverified

Lucas Nestler commited on

add h100
c541d34
unverified

Lucas Nestler commited on

remove torchao dependency, quantize entirely via linear
d45a331

aredden commited on

Fix issues with loading F8Linear from state dict when init_scale not initialized & loaded from meta device
3ddaa67

aredden commited on

Fix issue where lora alpha is not correct if lora from transformers checkpoint
7a7b2c1

aredden commited on

Small fix for issue where f16 CublasLinear layers weren't being used even when available.
6d82dcc

aredden commited on

Ensure repo only accesses CublasLinear lazily
00f5d2c

aredden commited on

Fix issue loading loras
fee1af5

aredden commited on

Add lora loading
a71da07

aredden commited on

Merge pull request #3 from aredden/improved_precision
af20799
unverified

aredden commited on

Update readme
49f2076

aredden commited on

Add quantize embedders/modulation to argparse options
604f17d

aredden commited on

Remove f8 flux, instead configure at load, improved quality & corrected configs
1f9e684

aredden commited on

Fix issue where torch.dtype throws error when converting to dtype
fb7df61

aredden commited on

Move compile out of FluxPipeline init
ac049be

aredden commited on

Dynamic swap with cublas linear / optional improved precision with vram drawback
37bd8c1

aredden commited on

Allow overriding config values from load_pipeline_from_config_path
25ae92b

aredden commited on

Merge pull request #2 from XmYx/main
b84f35e
unverified

aredden commited on

python 3.10 compatibility
9b84867
unverified

mixy89 commited on

Merge pull request #1 from dsingal0/main
c247326
unverified

aredden commited on

Remove unnecessary tokenization (still needs work)
44e4014

aredden commited on

Add case where seed is string & try/catch if invalid
7cec457

aredden commited on

Remove unnecessary synchronize, add more universal seeding & limit if run on windows
ffa6ff7

aredden commited on

Remove more unnecessary code, fix small typing hickup
6d0762c

aredden commited on

Remove unnecessary code, hide prints behind debug flag, hide warnings
0f3134f

aredden commited on

Update README.md - Fix config path names
4a2503e
unverified

aredden commited on

Small README update for clarification
42be379

aredden commited on

Add all relevent args to argparser & update readme
9dc5b0b

aredden commited on

Merge branch 'main' of https://github.com/aredden/flux-fp16-acc-api into main
58082af

aredden commited on

Adding device specific configs & more input image type options + small model spec from args change
e81fa57

aredden commited on

Update README.md
d9b39c9
unverified

aredden commited on

make table more clear- again
21a2bf7

aredden commited on

make table more clear
50c4b9f

aredden commited on

Add speed comparison
efb953a

aredden commited on

Add uvicorn requirement
6bd57ab

aredden commited on

Add fields to configs, fix issue with offload from bnb, remove extra random text code
340f0a0

aredden commited on

Make README.md more comprehensive
1626546

aredden commited on

Fix for nightly
0aa9861

aredden commited on