Thread #108575392
File: highlights_g_108569503_1775839092_1.jpg (1.6 MB)
1.6 MB JPG
Discussion and Development of Local Image and Video Models
Previous: >>108569503
https://rentry.org/ldg-lazy-getting-started-guide
>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows
>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe
>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo
>Anima
https://huggingface.co/circlestone-labs/Anima
https://tagexplorer.github.io/
>Qwen
https://huggingface.co/collections/Qwen/qwen-image
>Klein
https://huggingface.co/collections/black-forest-labs/flux2
>LTX-2
https://huggingface.co/Lightricks/LTX-2
>Wan
https://github.com/Wan-Video/Wan2.2
>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46
>Illustrious
https://rentry.org/comfyui_guide_1girl
>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Collage: https://rentry.org/ldgcollage
>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg
>Local Text
>>>/g/lmg
>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
305 RepliesView Thread
>>
>mfw Resource news
04/10/2026
>JoyAI-Image-Edit now supports ComfyUI
https://github.com/jd-opensource/JoyAI-Image#-news
>Two Front Doors: Civitai.com, Civitai.red, and What's Next
https://civitai.com/articles/28369/two-front-doors-civitaicom-civitair ed-and-whats-next
>Uni-ViGU: Towards Unified Video Generation and Understanding via A Diffusion-Based Video Generator
https://fr0zencrane.github.io/uni-vigu-page
>PrivFedTalk: Privacy-Aware Federated Diffusion with Identity-Stable Adapters for Personalized Talking-Head Generation
https://github.com/mazumdarsoumya/PrivFedTalk
>AVGen-Bench: A Task-Driven Benchmark for Multi-Granular Evaluation of Text-to-Audio-Video Generation
http://aka.ms/avgenbench
>Cross-Modal Emotion Transfer for Emotion Editing in Talking Face Video
https://chanhyeok-choi.github.io/C-MET
>ChenkinNoob-XL-V0.5
https://modelscope.ai/models/ChenkinNoob/ChenkinNoob-XL-V0.5
>Control Order & Free Memory: Controls the order of node execution with device-agnostic memory management
https://github.com/mkim87404/ComfyUI-ControlOrder-FreeMemory
>DMax: Aggressive Parallel Decoding for dLLMs
https://github.com/czg1225/DMax
04/09/2026
>MAR-GRPO: Stabilized GRPO for AR-diffusion Hybrid Image Generation
https://github.com/AMAP-ML/mar-grpo
>HybridScorer: Score, sort, and cut large sets down fast with GPU-accelerated AI review
https://github.com/vangel76/HybridScorer
04/08/2026
>OrthoFuse: Training-free Riemannian Fusion of Orthogonal Style-Concept Adapters for Diffusion Models
https://github.com/ControlGenAI/OrthoFuse
>MIRAGE: Benchmarking and Aligning Multi-Instance Image Editing
https://github.com/ZiqianLiu666/MIRAGE
>Few-Shot Semantic Segmentation Meets SAM3
https://github.com/WongKinYiu/FSS-SAM3
>PoM: A Linear-Time Replacement for Attention with the Polynomial Mixer
https://github.com/davidpicard/pom
>>
File: o_00118_.png (1.2 MB)
1.2 MB PNG
>>
>mfw Research news
04/10/2026
>MegaStyle: Constructing Diverse and Scalable Style Dataset via Consistent Text-to-Image Style Mapping
https://jeoyal.github.io/MegaStyle
>Multimodal Large Language Models for Multi-Subject In-Context Image Generation
https://arxiv.org/abs/2604.07422
>Personalizing Text-to-Image Generation to Individual Taste
https://arxiv.org/abs/2604.07427
>EditCaption: Human-Aligned Instruction Synthesis for Image Editing via Supervised Fine-Tuning and Direct Preference Optimization
https://arxiv.org/abs/2604.08213
>LPM 1.0: Video-based Character Performance Model
https://large-performance-model.github.io
>Lighting-grounded Video Generation with Renderer-based Agent Reasoning
https://arxiv.org/abs/2604.07966
>Guiding a Diffusion Model by Swapping Its Tokens
https://arxiv.org/abs/2604.08048
>ImVideoEdit: Image-learning Video Editing via 2D Spatial Difference Attention Blocks
https://arxiv.org/abs/2604.07958
>FlowGuard: Towards Lightweight In-Generation Safety Detection for Diffusion Models via Linear Latent Decoding
https://arxiv.org/abs/2604.07879
>Phantom: Physics-Infused Video Generation via Joint Modeling of Visual and Latent Physical Dynamics
https://arxiv.org/abs/2604.08503
>Mitigating Entangled Steering in LVLMs for Hallucination Reduction
https://arxiv.org/abs/2604.07914
>SAT: Selective Aggregation Transformer for Image Super-Resolution
https://arxiv.org/abs/2604.07994
>RewardFlow: Generate Images by Optimizing What You Reward
https://plan-lab.github.io/rewardflow
>LINE: LLM-based Iterative Neuron Explanations for Vision Models
https://arxiv.org/abs/2604.08039
>SOLAR: Communication-Efficient Model Adaptation via Subspace-Oriented Latent Adapter Reparametrization
https://arxiv.org/abs/2604.08368
>GaussiAnimate: Reconstruct and Rig Animatable Categories with Level of Dynamics
https://cookmaker.cn/gaussianimate
>DSCA: Dynamic Subspace Concept Alignment for Lifelong VLM Editing
https://arxiv.org/abs/2604.07965
>>
>>
File: ComfyUI_temp_puiqp_00034_.png (1.5 MB)
1.5 MB PNG
>>
>>108575407
>>108575425
Go back you're not welcome here schizo
>>
From old fred
>>108575356
>adjust the denoise
At 1.0? I can't give it more denoise. Everything lower stays closer to the original, at least in K9B.
>>108575358
I actually didn't want it to do prompt like in >>108565108 this example, but i2i photo edit. Prompting of course can run on any model capable enough to understand the resulting prompt.
>>
>>
File: ComfyUI_temp_puiqp_00039_.png (1.5 MB)
1.5 MB PNG
>>
>>
>>108575486
the problem with i2i is that you need the first image to drive the second image, low denoise and you are stuck in anime land, high denoise and you are relying on prompts because it will ignore the anime image.
feeding the image into latent space as a reference might be what you want, but those results will probably be hit or miss and you will be relying heavily on prompting.
>>
File: ComfyUI_temp_puiqp_00040_.png (1.3 MB)
1.3 MB PNG
>>
>>
>>
File: o_00119_.png (893.9 KB)
893.9 KB PNG
>>
POST-NOOB ANIME MODEL GRAVEYARD:
>Illustrious-Lumina
>PonyV7
>Neta
>NetaYume
>Chroma (yes, it was originally intended to also do anime)
>Newbie
>Rouwei-Gemma
>Mugen
>Chenkin
>Nekofantasia
ALIVE AND KICKING:
>Anima
What happened?
>>
>>108575587
It's more about the fact that it can correctly cover a real song and replace the lyrics. Default Ace-Step's cover mode will fuck around to prevent you from infringing copyright. acestep.cpp has a cover mode that lets you have fun.
>>
>>108575486
you can feed flux2smol a reference latent but you need custom node to adjust strength, see here
https://github.com/capitan01R/ComfyUI-Flux2Klein-Enhancer
'FLUX.2 Klein Ref Latent Controller' would be the node
you need to tinker with it
>>108575613
mighty fucking worked up about the whole thing aren't you
>>
File: ComfyUI_temp_puiqp_00053_.png (3.8 MB)
3.8 MB PNG
>>
File: 235808074807204870845.jpg (86.5 KB)
86.5 KB JPG
>>108575774
skyrim looks like that?
>>
>>
>>
>>
File: ComfyUI_temp_puiqp_00057_.png (3.2 MB)
3.2 MB PNG
>>
>>
>>108575774
NOW WE ARE TALKING *opens beer*
I DENOUCE SMALL TITS. HALE HORTLER
>>108575836
nigga purple horses are gay
>>
>>
>>
>>
>>
File: ComfyUI_temp_puiqp_00060_.png (3.1 MB)
3.1 MB PNG
>>
File: 1746991891250979.png (194.3 KB)
194.3 KB PNG
>>108575863
>When is Z-Edit / Z-Omni happening?
you got chinese cultur'ed anon
>>
>>
>>
>>
>>
File: bruh.png (407.2 KB)
407.2 KB PNG
>>108575943
>Wan2.1's vae
>>
>>
>>
File: that's right.png (222.8 KB)
222.8 KB PNG
>>108576053
>>108576043
the good one would be no vae desu, pixel space ftw!
>>
>>
File: o_00123_.png (1.8 MB)
1.8 MB PNG
>>
>>
>>
File: o_00125_.png (1.8 MB)
1.8 MB PNG
>>
>>
File: ComfyUI_temp_puiqp_00065_.png (2.4 MB)
2.4 MB PNG
>>
File: ComfyUI_temp_puiqp_00066_.png (2.9 MB)
2.9 MB PNG
>>
File: kh1o4udxcd6e1.jpg (68.6 KB)
68.6 KB JPG
Does google's turboquant have any applications for local image gen? Seems like the whole field hasn't really done much better than illustrious
>>
>>108576507
>Does google's turboquant have any applications for local image gen?
I guess so? diffusion models also use KV cache
>Seems like the whole field hasn't really done much better than illustrious
yeah, the LLM fags can't stop improving the architecture while we're still finetuning SDXL for the 14th billion time, sad
>>
File: ComfyUI_temp_puiqp_00071_.png (3.9 MB)
3.9 MB PNG
>>
File: z_00098_.jpg (591.8 KB)
591.8 KB JPG
>>108576529
>>
File: ComfyUI_temp_orvnr_00004_.png (2.9 MB)
2.9 MB PNG
>>
File: ComfyUI_temp_orvnr_00006_.png (2.3 MB)
2.3 MB PNG
>>
File: ComfyUI_20720.png (2.1 MB)
2.1 MB PNG
>>108573407
That sucks. I had something similar happen to me when I forgot to resize some 4K screenshots one time. I turned the monitor on after waking up to check my progress and it was like 2% done with five DAYS left. So all I did for like 10hrs was turn electricity into heat. It took me a little while to figure out what happened too.
>>108576549
What's Robocop investigating here?
>>
>>108576688
>What's Robocop investigating here?
Damn you spotted the dataset quick https://www.youtube.com/watch?v=0nREpGt9_UY
>>
File: o_00136_.png (386.5 KB)
386.5 KB PNG
>>108576116
ayami
>>
>>
File: o_00143_.png (1.4 MB)
1.4 MB PNG
>>
>>
File: deWA_zi_00045_.png (2.6 MB)
2.6 MB PNG
a secret local debo model for debo is in training. will be released in a debo or so
>>
>>
File: please let it be true.png (235 KB)
235 KB PNG
>>108577143
>>
>>108575392
>using AI to help debug a training model that isn't working
>already spent 8+ hours of actual GPU work for nothing
>ask it to just find me working configs and use that to find out if anything is wrong with mine (which it has been helping me build since yesterday)
>literally gave me the wrong fucking parameter names from the beginning, and also recommended a patched python file to download
>try it, suddenly get errors I've never seen before
>"well, this confirms the problem is with the python file so you should download the patched version"
>replace it with the original python file and tell the AI
>"well, that's the right call, the community patched version is actually not fixing anything"
>finally after adjusting the settings some more, it looks like it MIGHT actually work (still waiting on the sample images)
This is like trying to solve a problem with a retard on your team.
>>
>>
File: ComfyUI_20780.png (1.6 MB)
1.6 MB PNG
>try to use VibeVoice with the latest Comfy
>one gen works before I need to fully restart Comfy because it wants all the memory and storage
Holy hell... did anyone let ComfyAnon know that if you're gonna take over memory operations you should know how to manage memory? This behavior is asinine.
Fix your shit, goddammit!
>>108576771
Love me some Robocop. Just watched the workprint a few weeks ago. It was fun seeing the placeholder footage and graphics.
>>
>>108575407
>>108575425
thanks, news are always welcome here
>>
>>
>>
>>
File: spaaaaace.jpg (1.9 MB)
1.9 MB JPG
>>
>>
>>
>>
>>108577980
Really? Fuck, the end of an era: ntrMIX, Noob, Animagine now WAI. I don't feel comfortable with Anima; the results are meh tier. I was testing some back and forth with the same tags between Anima and NovelAI, and maybe NovelAI only handles tags, but it's much more creative and fun in the results. Anima is boring and aesthetically years behind of 2024 Noob.
>>
File: 7 prompts for seedance.mp4 (3.6 MB)
3.6 MB MP4
>>
>>
>>
>>
File: ComfyUI_Anima_00046_.png (614.7 KB)
614.7 KB PNG
>>108576353
That is just power metal. Turbo XL can do it just fine
https://vocaroo.com/16iY9WL6UpXp
There really isn't that big a difference in quality between it and what cloud users would claim is "SOTA". What's more, this is the raw model, without a LoRA, obviously it gets much better with one, and you can't train anything like LoRAs with cloud.
>>108575559
>>108575623
This. Anons are really underestimating the meme potential of this model, especially when used as a tool
https://vocaroo.com/19RJdI0fBJ45
>>
>>
>>
Training anime models must currently be one of the most difficult things for any model. These realism models should actually be one of the easiest things. Imagine the fact of training millions of characters, each one with their particular clothing design, each one with their style, each one with their haircut, and so on for millions. These are millimetric details that the model must always keep in mind. To this we add the concepts, the scenarios, the everyday objects, and other things.
>>
File: spaceeeee.png (647 KB)
647 KB PNG
>>
>>
>>
>>
File: Settings.png (144.1 KB)
144.1 KB PNG
>>108578056
Here you go. I think they were mostly defaults. I reworked my captions three times before I was happy with the results though. I settled around 192 words per caption after creating detailed JSON captions and then distilling those down to natural language. The goal was only visual information with no filler (e.g., "appears to", "this or that", etc - you know how LLMs like to waffle about).
I preferred Prodigy with other trainers, but that wasn't an option on this last run pictured. I think it is now, I haven't updated in forever.
>>
>>
>>
File: wtf bruh.gif (2.8 MB)
2.8 MB GIF
https://xcancel.com/FunkletonJazz/status/2042755948947623977#m
>>
>>108578249
I think they should let us fine tune or make loras of their cloud models by paying for cloud storage monthly. If I don't pay for one month, they would delete my cloud loras and fine tunes.
@comfyanonymous What do you think of this idea? ComfyCloud with a ComfyDrive service to store my Seedance LoRAs and SeeDream fine tunes.
>>
File: 1767588376082735.jpg (628.6 KB)
628.6 KB JPG
this one came out good :)
>>
File: open saars it.png (154.7 KB)
154.7 KB PNG
OPEN THE WEIGHTS NOW!!! WE ARE BETRAYED!!
>>
File: based cockroach.png (66.1 KB)
66.1 KB PNG
>>108578429
>Sylvain Filoni
who the fuck is this guy? lol
https://xcancel.com/fffiloni/status/2042693042721718639#m
>>
>>
File: deWA_zi_00047_.png (2.4 MB)
2.4 MB PNG
>>
>>108578148
a psycho prompt gave some interesting results but a side effect was a bake effect. i just used seedvr2 for a clean upscale, no resample. tiled z-img upscaling tends to hallucinate pretty badly with extreme closeups of faces, not done dialing it in yet tho. maybe like a 0.1 denoise pass after vr2, tiled or not, I dunno.
>>
>>
>>
>>108578131
>That is just power metal. Turbo XL can do it just fine
>https://vocaroo.com/16iY9WL6UpXp
Fucking kek, the guitar turns into synth at 1:25 confirms this model can only do fucking electronica. They must have trained on like 90% synths and maybe 10% real instruments. Probably an amazing model if you're into electronica.
>>
>>
>>108578735
Perceived weaknesses with instruments are nothing a LoRA can't fix
https://xcancel.com/ostrisai/status/2041926198599807079#m
Just find 10 target songs of whatever genre you want, train on that, then issue is gone.
>>
>>108578778
>sft model at 100 steps
For text2music or cover-nosfq? For txt2music the cover strength has to be lowered to 0.5 for turbo, and supposedly even lower for SFT (0.06-0.2 according to the dev), though I never got it to work properly for that on the songs I tested. Though 100 steps of SFT helps against 50, I think it may be excessive, and I have found XL Turbo to be slightly better in audio quality than untuned SFT in my tests, 8 is fine for it as well.
>>
>>
>>
>>108578830
It's possible, plus you can also overbake on a single song if that's your desire. Though the cover feature is more useful for the purpose of recreating a song, even if you need it to have changes (just lower cover strength and do some prompt engineering).
>>
File: 1761719241240763.jpg (572.1 KB)
572.1 KB JPG
me and my wife
>>
>>108577931
>>108576688
based jenner how do your gens keep getting better and better
>>
File: deWA_zi_00049_.png (2.1 MB)
2.1 MB PNG
>>108578928
she's cheating on you with some steerage peasant boy
>>
>>
File: 00003-4252499356.png (3.9 MB)
3.9 MB PNG
>>
File: 1640378554508.jpg (10.7 KB)
10.7 KB JPG
>>108578928
looks like she's tasted a lot of chocolate before you. if your son is tanned, it's not because of the sun, LOL
>>
>>
File: 00004-4111254592.png (3.8 MB)
3.8 MB PNG
>>
>>
>>
>>
>>108575774
>>108575836
>>108575897
chroma sucks
>>
File: deWA_zi_00055_.png (2.4 MB)
2.4 MB PNG
>>108579181
>ComfyUI Prebuilt Docker Images
https://github.com/radiatingreverberations/comfyui-docker
>>
>>
>>
>>
>>
File: deWA_zi_00058_.png (2 MB)
2 MB PNG
>>108579494
z-image-turbo
>>
>>108579501
Small details in general because of VAE. Also the cloning of details. It tried to give them both blue gems on their chests even though it only makes sense on the woman and it duplicated the heart so it's on both her chest and necklace even though that looks stupid and redundant.
>>
>>
File: deWA_zi_00060_.png (2.4 MB)
2.4 MB PNG
>>108579592
>never tried action with z
its pretty hit-and-miss, but thats maybe because I have a lot of wildcards floating around
>any prompt?
you can skim through the workflow if you want
https://files.catbox.moe/tjpqyi.png
>>
>>108579501
one inpaint session and you can't tell anymore if the base gen is solid. easiest giveaway is often the background logic and lighting. we are one generation away from 'indistinguishable'. and hey look, suddenly hardware prices skyrocket
>>108579608
for the love of god, give it a rest, catjack
>>
>>
>>108579624
idgaf
>>
>>
File: file.png (45.6 KB)
45.6 KB PNG
>>108578380
imagine thinking this is "good" in 2026
>>
>>
>>
File: paper-zit-2026-04-11_00076_.png (3.1 MB)
3.1 MB PNG
>>108579897
more toes more better. uh i guess...
>>
>>
>>
File: paper-zit-2026-04-11_00127_.png (3.1 MB)
3.1 MB PNG
>>
File: 08906795795.jpg (1.9 MB)
1.9 MB JPG
>>
>>
File: 0c57dd7917f693a02801b7b4e2a032a4.png (253.5 KB)
253.5 KB PNG
What causes this mental illness? These people need to be euthanized.
>>
>>
File: ComfyUI_20953.png (2 MB)
2 MB PNG
>>108578959
>https://github.com/capitan01R/Capitan-ConditioningEnhancer
This really helps me attack the negatives during the low-step ZIM pass without causing problems. Everything else on the ZIT pass is light touches.
>>108580419
How many unique downloads?
>>
>>
>>
File: comfy__502.jpg (1.2 MB)
1.2 MB JPG
>>108579880
Feet are for gays and nutters. Real men prefer characterful faces.
>>
>>
>>
File: my_choice_alt.png (1012.6 KB)
1012.6 KB PNG
my choice
Something certain
Drawing me in
https://suno.com/s/keTP3Z6jaGyvBRNw
https://youtu.be/ttTHrH3Tb3U
>>
File: paper-zit-2026-04-11_00185_.png (2.8 MB)
2.8 MB PNG
>>
>>
File: 8HzFYXqJGyCWUGfF.mp4 (399.9 KB)
399.9 KB MP4
>The live-action conversion in Seedance 2.0 is exquisitely crafted. Motion capture replacement seems to be advancing as a partial alternative. It estimates movements from existing footage, regenerates them into live-action textures, and eliminates the need for a studio. Rather than just a filter, it appears to re-synthesize motion by motion + stabilize the timeline. I really feel that if you keep denying AI, you'll get left behind
>>
File: 1773940801849528.png (958 KB)
958 KB PNG
>>108580866
>I really feel that if you keep denying AI, you'll get left behind
deep down no one is denying AI, they're all pretending it's just a toy that won't improve further and has no real life usecase, but they seethe so much about it they know it's not true, AI is the future,
Just today, someone tried to kill Sam Altman, that's how serious it got
https://xcancel.com/sama/status/2042789312400363702#m
holy shit
>>
>>
>>108580866
>please use our slop :(
>>108580888
good
>>
>>
>>
>>
>>
>Seedance 2.0 just got released as an API to the whole world
>Literally every major studio on the planet is integrating it into their workflow
Crazy how powerful SaaS models are. You can generate Hollywood-level visuals in under $10.
>>
>>
>>
>>108581025
it can do any style you want, if you want to do 90s anime go for it
https://xcancel.com/KabookiAI/status/2042674460881207691#m
>>
>>
>>
>>
>>108581050
>the people who train the models, write and build custom nodes and spaghetti workflows are the are luddites!
and when asked what makes saas so good one of the first answere you get it
>you don't have to deal with spaghetti, custom nodes or complex workflows. you just type words on your phone.
>>
>>
>>
>>108581086
based. localbrowns started to embrace 'prompt engineering' and now they're panicking just like inkcels were. crazy how quickly the tables turned, the same people laughing at artists saying you no longer need to waste 10 years learning to draw a face are now the ones coping and crying about how fine-tuning the CFG Rescale node will keep them from being replaced.
>>
>>108581071
amazon and ebay are storefronts, youtube is a mostly free video streaming service.
there is a laundry list of complaints against steam, the two biggest being 'stop killing games' and digital inheritability.
after seeing the meltdown over shit like sora getting killed, i assume we will see "stop killing models" in a few years.
>>
>>
>>
>>108581130
only a brown person would think people give a fuck about a few dollars a month for a service.
the problem is being forced to go along with whatever the service provider decides, or having them straight up kill a model you were using.
we've seen them kill or lobotomize popular models several times already.
>>
>>108580888
Some local government official had their house shot up recently too and left a "No Data Centers" note on their porch. It's gonna be hilarious when these corporate goons are losing $20-40 billion dollar Data Centers left and right to people sick of their bullshit.
It'll be easier to build a new nuclear reactor than data center in a few years.
>>
File: glory to the CCP.png (414.9 KB)
414.9 KB PNG
>>108581210
America will kill itself because of luddites and China will keep advancing anyways
>>
>>108575836
>>108575897
kids these days and their plebeian love of realism
Vanilla Skyrim textures look like paintings, a much better look.
>>
The fuck is going on with comfy? This shit keeps happening and I have to toggle on and off the nodes 2.0 toggle. And yes the wires fuck up in both 2.0 and 1.0.
Is this some not yet fixed bug or me being retarded? Shit is infuriating.
>>
>>
File: pixel-0000-1726027384.png (432.4 KB)
432.4 KB PNG
>>
>>
>>
>>
>>
>>
File: Untitledsfdsdf.mp4 (3.6 MB)
3.6 MB MP4
>>108581691
Nice style, I haven't been able to replicate this. Wanted to see if wan can handle it without turning it into 3d, worked nice. But my latent upscaling added random flowers..
>>
File: 1775220060084641.png (214.8 KB)
214.8 KB PNG
>>108581783
I hope I'll get a local model as good as seedance 2.0 in my living...
>>
>>
>>
>>108581955
such a nice specimen ruined by a thong. whoever added that shit to the training data needs to be burnt on a stake.
>>108580419
horse fuckers have many fetishes.
>>
File: Slop Machine.png (583.2 KB)
583.2 KB PNG
Slop Machine status? How many nodes can ComfyUI support before it starts to break? I want to add ControlNet to it, but I’m worried ComfyUI might break because of the number of nodes and connections.
I’m notinto subgraphs or GET SET, I want to see all the connections and nodes at all times and wire everything manually.
>>
File: j3WiPS2FLVA.jpg (295.6 KB)
295.6 KB JPG
>108582051
>>
>>
File: file.png (427.7 KB)
427.7 KB PNG
>>108575392
Prompt for this?
>>
>>
File: o_00148_.png (1.1 MB)
1.1 MB PNG
https://files.catbox.moe/b1pc1n.png
>>
>>
>>
>>
>>
>>
>>
File: comfy__505.jpg (610.1 KB)
610.1 KB JPG
>>108580942
no titties no clickies
>>
>>
>>108582051
you can build monster workflows. I like working modular tho, with sd prompt reader nodes to transfer stuff between the workflows. for reference, check out this one here, used to be called "AP workflow" from some italian dude. & same, I like shit out in the open. getset cancer, nope
>>
File: 856836734732.jpg (2.1 MB)
2.1 MB JPG
>>
>>
File: 9764848536863.jpg (2.5 MB)
2.5 MB JPG
>>
https://www.runcomfy.com/comfyui-workflows/ap-workflow-for-comfyui-rea dy-to-use-online-solution
Imported it and my PC started running much slower lol. But at least now I know I can use a lot more nodes, I’m going to add ControlNet without fear. Thanks, anon!
>>
File: flux-2-klein-9b-Q8_0.gguf-20260411_164034-fix_00001.jpg (601.8 KB)
601.8 KB JPG
>>108582658
sorry I forgot the link, that guy had a different webpage last time I checked, like 2 years ago. don't forget rgthree "fast groups bypasser" for one click bypass solution
>>
>>
>>
File: o_00156_.png (1.8 MB)
1.8 MB PNG
>>
File: 17515452458612353.jpg (796.1 KB)
796.1 KB JPG
>>108582733
Who's this bird?
>>
>>
>>108582786
"a woman with blonde hair", yada yada, nostalgic 1970s American vibe, etc. via https://civitai.com/models/2289127/zit-nsfw-randommergetime-v3
>>
File: 1775788632131111.jpg (1.2 MB)
1.2 MB JPG
>>108582827
She looks a little rough around the edges but I like her.
>>
File: pixel-0001-1218524535.png (633.3 KB)
633.3 KB PNG
>>
>>
>>108582946
nice
>>
>>
>>108583092
standing in front of window, backlit, natural light, sunlight
"Bright, natural light streams in from the large window, creating a luminous backlighting effect and highlighting the subtle textures of her attire." yada.
>>
Is there a way to use a local LLM or some kind of img2img system in order to create prompts faster in SD? I don't know if they offer something like the old interrogate deepbooru function separately for NeoForge or ComfyUI, because I have a lot of images that I could really use with that function.
>>
File: Video_00001.mp4 (2.3 MB)
2.3 MB MP4
>>108581802
Using claude to help me with latent upscaling. Helped me figure out a few things. So much motion, it's amazing.
>>
>>
File: o_00157_.png (1.1 MB)
1.1 MB PNG
>>
File: scary.png (176.1 KB)
176.1 KB PNG
>>108583153
a number of ways to do that. ollama & comfyui ollama (see img, sys prompt stolen from 'nerdy rodent'), comfyui-gemini (need a google account for that), all with their pros and cons. the gemma 3:4b model is like 3.5gb, leaves enough headroom to do stuff. it doesn't have vision capability, so just for prompting
>>
>>108581783
>>108581815
Reminder that imagegen found itself in a similar dilemma when Dalle 3 came out, being stuck with SD/Stability seemed so hopeless, then just 2 years later BFL appeared out of nowhere and dropped Flux, and we're here now with a mix of both uncensored and censored models that have fully deprecated Dalle 3. Similar tale with audio and Suno/Udio, now we have ACEStep XL. As demand grows, someone always will drop something. Someone will realize how important something like Seedream is for producers and small creators, and many will grow sick and tired of the corporate restrictions, so they (investors etc...) will realize how big it would be to have it locally. The more demand there is, the more likely the chance we see it sooner rather than later.
>>
>>
File: Stablediffusion4.png (2.2 MB)
2.2 MB PNG
are you ready for stable diffusion 4?
>>
>>108583291
Also, by not releasing the model, they are delaying the inevitable. They are just giving themselves competition. Sentiment around their failed model releases isn't going to change once a big open model drops that matches API (and therefore surpasses by its open nature), yet the company behind the open model will get all the glory.
>>
>>
>>
File: o_00163_.png (974.1 KB)
974.1 KB PNG
>>
>>
>>108583410
Wan 2.2, latent upscaling.
No one ever seems to know that wan is trained on 480p. So by upscaling into low noise, you maintain the maximum motion, with enough steps.
I thought I had figured latent upscaling out, but it's volatile as fuck when I introduce loras. Currently using claude to figure it out.
>>
File: o_00168_.png (1.1 MB)
1.1 MB PNG
>>
File: o_00170_.png (993.8 KB)
993.8 KB PNG
>>
File: 191022095059251.png (74.5 KB)
74.5 KB PNG
>>
>>
>>
File: image.png (17.3 KB)
17.3 KB PNG
>>108583706
how is this supposed to work
>>
File: 1757159618552659.png (189.9 KB)
189.9 KB PNG
I should have been more careful using the install from URL function willy nilly, I have clearly broken my installation somehow even after removing all the extensions folder from their respective folder.
>>
File: file.png (695.7 KB)
695.7 KB PNG
>>108583118
>>108583224
>>108583424
I was thinking more like a proper photography term for this "dreamy / blurry" effect, is very distinct. I see it often on japanese photoshoots or old playboy magazines. I guess "Dreamy Look" would work. But I've found how they are made: it's called a "mist filter".
>>
>>108583792
mist filter, checks out I guess. 'proper photography terms' and how your SD model interprets those, long winded topic. something somewhere must've been tagged with that expression, otherwise you get just a vague interpretation by the model. it'll pick up the word 'mist' obviously in that context. and fuck she looks like she burns through 4 very rich husbands every year.
>>
File: comfy__516.jpg (1.3 MB)
1.3 MB JPG
>>108583792
fun fact: they used vaseline on the camera lens to get that effect
>>
>>
File: o_00173_.png (2.1 MB)
2.1 MB PNG
>>
File: 430127098681774.png (1.1 MB)
1.1 MB PNG
>>
>>
>>
File: 644906155708829.jpg (356.1 KB)
356.1 KB JPG
>>108584033
so it is
>>
>>
>>108583250
>>108583153
Is there an up to date version of WD14 tagger for WebUI as opposed to just ComfyUI?
>>
File: pixel-0002-4103295849.png (709.5 KB)
709.5 KB PNG
>>
>>108581691
>>108582946
>>108584178
Catbox or tips for any of these? Really like the style.
>>
>>
File: 171542359913040.png (3.4 MB)
3.4 MB PNG
>>
>>
>>
kino
>>
>>
File: 247336236653023.jpg (379.5 KB)
379.5 KB JPG
>>
File: 1744921366856832.png (2.5 MB)
2.5 MB PNG
https://www.ft.com/content/b39da303-3188-447b-8b65-3dd8dad8b59a?syn-25 a6b1a6=1
it's so over...
>>
>>
>>
File: pixel-0003-2436757515.png (1005.9 KB)
1005.9 KB PNG
>>108584314
prompts were:
[subject/setting], traditional media, colored pencil \(medium\), pixel art, texture, 1970s \(style\), blurry, soft focus,
masterpiece,best quality,amazing quality
negative:
bad quality,worst quality,worst detail,sketch,censor, signature, watermark, shiny skin, flat color, simple background
using pixel art extension to limit color palette
>>
File: 544740533811128.jpg (335 KB)
335 KB JPG
>>
>>
File: deJA_zi_00017_.png (2.2 MB)
2.2 MB PNG
>>
>>
File: degunk.gif (3.2 MB)
3.2 MB GIF
Here is latest diffusion degunker for comfyui, related to https://rentry.org/RemovingDiffusionGunk
https://github.com/L33chKing/ComfyUI_LatentResidueCleaner
>>
File: pixel-0004-296154482.png (952.6 KB)
952.6 KB PNG
>>108584587
no loras. an illustrious mix of mine, unreleased. the mix isn't particularly good.
>>
File: deJA_zi_00019_.png (2.4 MB)
2.4 MB PNG
>>108584615
>>
>>
File: deJA_zi_00025_.png (2.8 MB)
2.8 MB PNG
>>108584629
>the mix isn't particularly good.
disagree
>>
>>
File: pixel-0005-4221785105.png (656.6 KB)
656.6 KB PNG
>>
File: deJA_zi_00026_.png (2.6 MB)
2.6 MB PNG
>>108584715
yea its a mess
been experimenting with llm prompt processing but its not very successful
>>
File: chinasaas winning.png (18.8 KB)
18.8 KB PNG
>>108584504
those of us with a brain called this months ago. saas is the future of AI, only unironically deranged freetards think otherwise. local models are nothing more than a party trick. there is always the connotation of "for local" at the end
>LTX is pretty good, for local
>Chroma is amazing considering it's a local model
even localkeks prefaced their happyhorse hype with "this is insane, SOTA and local too?!" because they too know local is on its deathbed and will never receive top-10 models again.
>>
>>
File: pixel-0006-666030637.png (197.2 KB)
197.2 KB PNG
>>
>>
File: pixel-0007-201280990.png (249.8 KB)
249.8 KB PNG
>>
>>
>>
>>
>>
File: _AnimaPreview3_00042_.jpg (314.7 KB)
314.7 KB JPG
>>
Fresh when ready
>>108585019
>>108585019
>>108585019
>>108585019
>>
File: ComfyUI_temp_pivpu_00011_.png (2.5 MB)
2.5 MB PNG
>>
File: example3.gif (1.7 MB)
1.7 MB GIF
>>108584942
I tried it on few flux2 klein images and yeah, there is cum on some parts of the image.
as far am aware all local models do this, but photorealistic images are less influence by it.
>>
>>108584620
Added to https://rentry.org/ldg-lazy-getting-started-guide#anon-guides-and-reso urces
>>
>>
>>