You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
First of all, thanks for an amazing project! This runs on average 30% faster than flux on ComfyUI. I was wondering if there's any planned support for different schedulers and samplers like how you can choose them in the sample and scheduler nodes in ComfyUI?
It would be great to test out different combination of schedulers as people have mentioned and shared that using different schedulers can improve/change generation style and quality.
The text was updated successfully, but these errors were encountered:
Also love this project and would love to help contribute and collaborate where possible. Have been using flux on both comfy and the excellent https://github.com/aredden/flux-fp8-api project on my CUDA machine, and excited to see this in native MLX.
Generation speeds are a bit slower for me (at FP8 runtime quant, which is what I typically do in comfy).
Was looking into mlx compilation and trying a few quick things this morning on the code base, but seems like adding compilation will require some more time reading the code and documenting the data flow. Sounds like a job for Claude later on unless someone's already done something like this (I'm a data guy and it helps to see how it all flows).
Also highly interested in assisting with LoRA training if you get there. There's no real standards on this, but I like what kijai is doing in comfy (https://github.com/kijai/ComfyUI-FluxTrainer). Most are using ai-toolkit (github.com/ostris/ai-toolkit/), which I believe uses diffusers and kohya scripts (have tried to trace all the code here, and ostris did an awesome job opening the door on all this, but haven't been able to trace the flux-specific code end-to-end yet.
I've trained a few LoRAs just for tinkering purposes on single layers / full layers / several layers using a few different methods. If trained with runtime quant at FP8, the LoRA wont work unless it's also inferenced at runtime-quant FP8 (pre-quant doesn't work on this project or any others, which is good to see - consistency).
Looking forward to seeing this project grow, and using it as a reference educational tool for learning MLX deeper. Appreciate you open sourcing this, @filipstrand .
First of all, thanks for an amazing project! This runs on average 30% faster than flux on ComfyUI. I was wondering if there's any planned support for different schedulers and samplers like how you can choose them in the sample and scheduler nodes in ComfyUI?
It would be great to test out different combination of schedulers as people have mentioned and shared that using different schedulers can improve/change generation style and quality.
The text was updated successfully, but these errors were encountered: