Thread #108563476
File: highlights_g_108558395_1775714841_1.jpg (2.7 MB)
2.7 MB JPG
Discussion and Development of Local Image and Video Models
Previous: >>108558395
https://rentry.org/ldg-lazy-getting-started-guide
>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows
>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe
>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo
>Anima
https://huggingface.co/circlestone-labs/Anima
https://tagexplorer.github.io/
>Qwen
https://huggingface.co/collections/Qwen/qwen-image
>Klein
https://huggingface.co/collections/black-forest-labs/flux2
>LTX-2
https://huggingface.co/Lightricks/LTX-2
>Wan
https://github.com/Wan-Video/Wan2.2
>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46
>Illustrious
https://rentry.org/comfyui_guide_1girl
>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Collage: https://rentry.org/ldgcollage
>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg
>Local Text
>>>/g/lmg
>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
312 RepliesView Thread
>>
>mfw Resource news
04/09/2026
>MAR-GRPO: Stabilized GRPO for AR-diffusion Hybrid Image Generation
https://github.com/AMAP-ML/mar-grpo
>HybridScorer: Score, sort, and cut large sets down fast with GPU-accelerated AI review
https://github.com/vangel76/HybridScorer
04/08/2026
>OrthoFuse: Training-free Riemannian Fusion of Orthogonal Style-Concept Adapters for Diffusion Models
https://github.com/ControlGenAI/OrthoFuse
>MIRAGE: Benchmarking and Aligning Multi-Instance Image Editing
https://github.com/ZiqianLiu666/MIRAGE
>Few-Shot Semantic Segmentation Meets SAM3
https://github.com/WongKinYiu/FSS-SAM3
>PoM: A Linear-Time Replacement for Attention with the Polynomial Mixer
https://github.com/davidpicard/pom
>RS Nodes for ComfyUI: Cmprehensive custom node pack focused on LTXV audio-video generation, LoRA training and post-processing
https://github.com/richservo/rs-nodes
>FLUX.2 Small Decoder: Distilled VAE decoder for faster decoding and lower VRAM usage
https://huggingface.co/black-forest-labs/FLUX.2-small-decoder
>Nvidia snaps up AI chip packaging capacity as TSMC expands in U.S.
https://www.cnbc.com/2026/04/08/tsmc-nvidia-advanced-packaging-intel.h tml
04/07/2026
>Anima preview3 released
https://huggingface.co/circlestone-labs/Anima#preview3
>FrameFusion Image Interpolation: Compact image interpolation model for generating in-between frames
https://github.com/BurguerJohn/FrameFusion-Model
>An Inside Look at OpenAI and Anthropic’s Finances Ahead of Their IPOs
https://www.wsj.com/tech/ai/openai-anthropic-ipo-finances-04b3cfb9
>PrismML debuts energy-sipping 1-bit LLM in bid to free AI from the cloud
https://www.theregister.com/2026/04/04/prismml_1bit_llm
>ComfyUI Hires Fix Ultra - All in One
https://github.com/ThetaCursed/ComfyUI-HiresFix-Ultra-AllInOne
>ATSS: Detecting AI-Generated Videos via Anomalous Temporal Self-Similarity
https://github.com/hwang-cs-ime/ATSS
>>
>mfw Research news
04/08/2026
>GenLCA: 3D Diffusion for Full-Body Avatars from In-the-Wild Videos
https://onethousandwu.com/GenLCA-Page
>Grounded Forcing: Bridging Time-Independent Semantics and Proximal Dynamics in Autoregressive Video Synthesis
https://arxiv.org/abs/2604.06939
>Evolution of Video Generative Foundations
https://arxiv.org/abs/2604.06339
>VersaVogue: Visual Expert Orchestration and Preference Alignment for Unified Fashion Synthesis
https://arxiv.org/abs/2604.07210
>Controllable Generative Video Compression
https://arxiv.org/abs/2604.06655
>Not all tokens contribute equally to diffusion learning
https://arxiv.org/abs/2604.07026
>FlowInOne:Unifying Multimodal Generation as Image-in, Image-out Flow Matching
https://arxiv.org/abs/2604.06757
>Holistic Optimal Label Selection for Robust Prompt Learning under Partial Labels
https://arxiv.org/abs/2604.06614
>Towards Robust Content Watermarking Against Removal and Forgery Attacks
https://arxiv.org/abs/2604.06662
>PhyEdit: Towards Real-World Object Manipulation via Physically-Grounded Image Editing
https://arxiv.org/abs/2604.07230
>Noise Constrained Diffusion (NC-Diffusion) Framework for High Fidelity Image Compression
https://arxiv.org/abs/2604.06568
>RefineAnything: Multimodal Region-Specific Refinement for Perfect Local Details
https://limuloo.github.io/RefineAnything
>Visual prompting reimagined: The power of the Activation Prompts
https://arxiv.org/abs/2604.06440
>MoRight: Motion Control Done Right
https://research.nvidia.com/labs/sil/projects/moright
>Fast-dVLM: Efficient Block-Diffusion VLM via Direct Conversion from Autoregressive VLM
https://arxiv.org/abs/2604.06832
>DesigNet: Learning to Draw Vector Graphics as Designers Do
https://arxiv.org/abs/2604.06494
>FP4 Explore, BF16 Train: Diffusion Reinforcement Learning via Efficient Rollout Scaling
https://arxiv.org/abs/2604.06916
>When to Call an Apple Red: Humans Follow Introspective Rules, VLMs Don't
https://arxiv.org/abs/2604.06422
>>
File: 1765189754590886.png (256 KB)
256 KB PNG
>>108562783
>even a ComfyUi employee fell for it
lol, lmao even
>>
>>
File: file.png (495.1 KB)
495.1 KB PNG
>>108563499
now I understand why it's called HappyHorse, they arr all the horse face kek
>>
>>
>>
>>108563505
>new model under the pseudonym 'happyhorse' gets teased on arenas, beats top API model seedance 2
>news spreads of this new model, people wondering if it's the new google VEO, others speculating it's china because of the name
>jeet vibe-codes a generic pop-up throwaway website for 'happy-horse ai' claiming it is SOTA, 15b parameters, and will be locally released
>the exact same grift that happened with the 'mogao' model that turned out to be seedream api
>localbrowns itt fall for it >>108555676
>it spreads, chinese anime man falls for it (picrel)
>gets reposted to reddit, redditoids fall for it
>kijai shuts him down and calls it fake, which it obviously is
>doesn't matter, news spreads and now everyone thinks a SOTA video model will be released locally within 48 hours
the backlash will be funny when it releases as API-only and the comments get flooded with outrage, even though whatever company behind it never even claimed it would be local. though it's still deserved, as every model should be local (even if it's fun laughing at localkeks)
>>
File: this.png (256.8 KB)
256.8 KB PNG
>>108563557
>the backlash will be funny when it releases as API-only and the comments get flooded with outrage
based, at least it had the effect to create some hate to an API model, can't wait to see the comments
>>
File: happyhorse05.png (92.4 KB)
92.4 KB PNG
>>108563567
it's kind of sad to see so many people get hyped for nothing, when the reality is we won't receive a model like this for at least another full year
>>
File: 1753632922542443.png (423.7 KB)
423.7 KB PNG
>>108563573
>we won't receive a model like this for at least another full year
bold to assume we'll get something better ever, it's obvious now that Alibaba has abandonned us, do you even realize that we've been waiting for Z-image edit for more than 4 months now? Pretty fair to say this is all over, chinese culture won
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1749992175112340.jpg (923.1 KB)
923.1 KB JPG
>muh fagollage
lmao
>>
>>108563557
the model looks like complete slop, i don't think anyone will care one way or the other.
i guess it might be ok if it is an API model and it can hold character likeness properly, bytedance kind of killed seedance 2.0 over that shit.
>>
>>
>>
>>
File: 1748109295030791.png (281.2 KB)
281.2 KB PNG
>>108563640
KJGod said it
>>
File: FAKE.png (98.8 KB)
98.8 KB PNG
>>108563640
>>108563651
https://github.com/brooks376/Happy-Horse-1.0/issues
kek
>>
File: take this my kind sir.png (89.5 KB)
89.5 KB PNG
>>108563655
https://github.com/brooks376/Happy-Horse-1.0/issues/3#issue-4225521889
>The author is using a deceptive title and README to exploit the open-source community's trust.
>The open-source community is a place for developers to share and collaborate, not a dumping ground for your vanity metrics or clickbait schemes.
>>
File: ComfyUI_Anima_00038_.png (1.7 MB)
1.7 MB PNG
ACEStep 1.5 XL Turbo. I am speechless, these are all first shots.
J-Core/Ballad
https://vocaroo.com/12JLSQwAuKIH
Electronic, Hatsune Miku included in prompt-
https://vocaroo.com/1l5RndzPCRbL
Country-
https://vocaroo.com/104rQ4A0Ux62
Gabber-
https://vocaroo.com/1b9C9ss8CTh9
Prompts are all enhanced with Gemini's help. Lyric alignment tends to be perfect now. We're now at Udio/Suno v5 territory
>>
File: __00037_.png (1.6 MB)
1.6 MB PNG
>>
>>
>>108563679
>Electronic, Hatsune Miku included in prompt-
it doesn't sound like miku at all, and the sound is still metalic as fuck, why did you use the turbo model though? the sft is supposed to have better quality right?
>>
>>
>Alibaba
>Trusting that this lab will release anything SOTA since not releasing Qwen Image 2 despite releasing its parameter count and top researchers from the team departing shortly after the CEO claimed that he is not happy about the state of open source
Unless BFL steps in and gives us a good video model, they have no reason to give us one themselves. There's no competition and no reason to release.
>>
>>108563693
>an open model on par with seedance would be the death of /ldg/
nothing is close to seedance, I've seen some videos from HappyHorse they are mid as fuck
https://youtu.be/mmk9C6bkV_c?t=161
>>
>>108563640
>saaars that is fake???????
this is the website. if you believe this is legit, you're brown: https://happyhorse-ai.com
the github repo with the 'source code' is full of made up bullshit as well. the most obvious tell is that they call it 'happy horse' when artificialanalysis lets companies use code-names for models (mogao = seedream, blueberry/strawberry were flux-2). they admit directly it's a pseudonym:
>We’ve added a new pseudonymous video model to our Text to Video and Image to Video Arenas.‘HappyHorse-1.0’
this is so that if the model turns out to be dogshit, the company behind it can just silently pull it without the whole world knowing that openai/grok's latest api model is a complete flop. saars are using this to their advantage to make fake websites surrounding these codenames to get people to enter their login information or pay crypto for 'credits' to use it
>>
>>
>>108563688
Definitely sounds like her, at least in style, the faded sounds in that particular song were prompted for (glitchy synths was included in many parts, so it glitched out her voice). That is roughly what you'd get out of a cloud model like Suno after prompting for her, I'd imagine it gets better with LoRAs.
>why did you use the turbo model though? the sft is supposed to have better quality right?
I always test Turbo first especially since on S it had better creativity, I'll try SFT/Base later. So far what this model has outputted puts it way above S SFT though.
>>
File: 1775573203286570.png (1 MB)
1 MB PNG
>>108563729
obviously, hollywood threatened them with lawsuits, the """based""" chinks bent the knee to the jews
>>
>>
>>
File: 9679858348383.jpg (1.1 MB)
1.1 MB JPG
>>
>>108563476
Is there anything that rivals nano-banana or similar in terms of being able to just describe what you want to see, rather than a whole "prompt engineer" slopfest? I want to connect SillyTavern to ComfyUI and gemma 4 likes to generate intricate descriptions of settings
>>
File: LMAOOOOO.gif (2.8 MB)
2.8 MB GIF
>>108563769
>Is there anything that rivals nano-banana or similar in terms of being able to just describe what you want to see, rather than a whole "prompt engineer" slopfest?
IS THIS NIGGA SERIOUS??
>>
>>108563769
no, not anything close really
>>108563779
yeah it's kinda grim
>>
File: michael-2354825030.jpg (43.6 KB)
43.6 KB JPG
>>108563779
>use the word banana
>a nigger chimps out at me
I don't know what I was expecting
>>
>>
File: 1763748844098575.png (694.4 KB)
694.4 KB PNG
>>108563769
Nano Banana isn't even the best image model anymore, I think GPT-Image 2 dethroned it
https://xcancel.com/WolfRiccardo/status/2041573176681918728
>>
>>
>>
>>
File: 34708937240872480804762.jpg (212.1 KB)
212.1 KB JPG
>>108563798
do you get paid to shill chatgpt here? you've been spamming this slop for days.
>>
>>
>>
>>108563769
Qwen Image 2, but it's closed and only has been rumored to open (doubt it). Joy AI Image Edit seems like a viable alternative to Qwen, it in fact is better at prompt following than everything else we have, but Comfy hasn't added support for it yet.
>>
>>108563809
>>108563814
look at those localkeks seething and jealous that we get quality toys while all you can have is cucked plastic slop keek
>>
>>
>>108563818
>only has been rumored to open (doubt it)
that chink insider used to be reliable, now he can't stop missing lol >>108563557
>>
>>
>>
>>
>>
File: remind me of this yeah.png (274.9 KB)
274.9 KB PNG
>>108563849
I'm not the one who made that image, I'm showing you the realism and you focus on something completly unrealted like the color of the skin, remind me of something...
>>
File: deepdream2-1346427197.jpg (152.7 KB)
152.7 KB JPG
>>108563847
I have hope. It wasn't too long ago I was thirsting over Deep Dream and imagining how cool it'd be to run that. If local is only as good as SOTA was 6 months ago (more like a year for image gen right now) it's worth it just for the freedom and FOSS spirit.
>>
>>
>>
>>
>>
>>
>>
>>108563923
sdwebui had its time (in the corpo I currently work for I actually coded a UI similar to it for internal usage for promo material creation), but I elected to use comfy as the backend because it's just too flexible to work with
>>
>>
File: kek.png (372.4 KB)
372.4 KB PNG
>>108564039
>i'd pay for it if I could
are you that broke ani? I thought you had secured some investments in japan? maybe you should come back to comfy and ask him the 1 million grant again, you're talking about him in such a nice way on /ldg/ I'm sure he'll consider giving some of those bits to such a good non-treasonus friend
>>
>>
>>
>>108564080
>>108564129
>move your mouse with your feet, now!
>what do you mean you don't want to deal with that because there's more elegant ways to manipulate a mouse, like hands for example...
>just say that you're too retarded to do it
ComfyAPI shills in a nutshell
>>
File: bb9b328e-3bc5-4a1a-b6ff-f6cbcdc126c9.jpg (49.3 KB)
49.3 KB JPG
Hey /ldg/,
I’ve been working on Spellcaster, an open-source plugin that seamlessly integrates 30+ AI tools directly into GIMP and Darktable. It uses ComfyUI as the backend engine, running entirely locally on your own GPU. It is essentially the GIMP version of what you guys are doing/using.
My goal was to bring Photoshop-level AI features to open-source editors, without the steep learning curve, cloud requirements, or subscription fees.
What it does right inside your editor:
Inpaint & Outpaint: Generative fill to change objects or extend your canvas.
Enhance & Fix: 1-click AI upscaling, background removal, face restoration, and object erasure.
Relighting & Video: Change lighting direction on portraits (IC-Light) or turn still layers into short video clips (Wan 2.2).
One-Click Install: The installer handles all the backend complexity (detects GPU, downloads models, sets up ComfyUI, and links to your editor).
https://github.com/laboratoiresonore/spellcaster/blob/main/README.md
I am looking for collaborators / feedback. There are a couple of advanced features that I'd like to implement next:
-full LTX2 support
-parsing any workflow method (at the moment, the script is designed with noobs in mind but comfyui veterans will likely want to use their own workflows and special sauces)
-making the "studio" system as advanced as possible
-clever refractoring and reorganizing the script
-better theming
The script is pretty recent but it's good enough that it is all I personally use, instead of comfy
>>
>>108562337
> >Nobody here talks about ACEStep 1.5 XL which just dropped
> https://ace-step.github.io/ace-step-v1.5.github.io/#XLDemos
> It's a different class of model bros, I'm not hearing any slop...
lol remember when locals have claiming that acestep is at suno/udio level of quality?
what level is now then?
>>
>>108564129
if you give the corpo drone a noodle gui... he's not going to understand, that's just how it is. You'd expect people here to be more tech savvy but even subhuman retards found their way here instead of staying in plebbit sooo whatever.
>>
>>
>>
>>
>>
>>
>>
File: 94685637372.jpg (1.4 MB)
1.4 MB JPG
>>
>>108563679
For the anon who mentioned instrumentals last thread, here's Spaghetti Western
https://vocaroo.com/1i8vvdmjVnDD
This is a genre the previous version could not do at all, and even its base model struggled with, that's a one shot from Turbo. Only issue I've noticed so far is it speaks some stuff in parentheses out loud in the midst of instrumental, which is not hard to fix in post processing and perhaps SFT does better here.
>>108564154
Well, with LoRAs the previous version absolutely was. However, LoRA training to improve everything is tedious. This one doesn't need LoRAs and has much better musical knowledge out of the box, which is now for the first time competitive with Udio/Suno at just 4B.
>>
>>108564260
1) add lipsync voice generation to an existing video.
2) Generate videos with audio or lipsync (if not possible the 1)
I'm curious about both options, and how they seen in the usual tiktok feed it seems pretty fast to made. Generate with sound is a new thing for me.
>>
>>
File: c304a3ad-1a56-47c8-b361-d44962850c49.png (1.5 MB)
1.5 MB PNG
>>108564262
>>108563755
Ummmmmmmmmm... prompt?
>>
>>
>tdrussel
>diffusion pipe: initial commit 2 years ago
But that didn't trigger FUD. It was Anima that put a target on him. If we follow the money, who could feel threatened by Anima but not diffusion-pipe? I think the answer is NovelAI. They're funding the troll farm.
>>
>>
>>108564310
>Only issue I've noticed so far is it speaks some stuff in parentheses out loud in the midst of instrumental
I see what the issue is kek, it should all be in brackets instead, parentheses are only for whispers and background noises.
>>
>>
File: 1753954094570315.png (1.3 MB)
1.3 MB PNG
the small gemma 4 models are so ass on vision task, it's a shame they went for a smaller mmproj relative to the 26 and 31b models
>>
File: 1762691525124100.jpg (770.9 KB)
770.9 KB JPG
ayo
>>108564685
yeah theyre cooked.
even the MOE is shit btw
I went back to qwen3vl
>>
>>
>>
>>108564727
Klein is not NBP tier with prompts or text because it's not autoregressive. It's very impressive for what it is, and it probably doesn't get any better than what Klein does with prompts for its particular architecture, but it's still not quite there yet.
>>
File: f087b952-d547-41e9-9235-adb45b929c45.jpg (404.8 KB)
404.8 KB JPG
https://civitai.com/articles/28368/chenkinnoob-xl-v05-is-coming-soon
We are thrilled to announce that ChenkinNoob-XL-V0.5, the direct successor to V0.2, has completed its training phase and will be officially released on April 10th (Beijing Time)!
After months of architectural refactoring and dataset expansion, V0.5 is no longer just a "gacha toy." We have pushed it to industrial-grade productivity standards.
What to Expect in V0.5:
Massive Dataset Leap: Built directly upon V0.2, we have added 2.17 million high-quality, open-source game-related images. The total training dataset now reaches ~12 million images, effortlessly capturing the latest anime art styles and popular characters.
Pro-Level Aesthetics: Built with industrial-grade standards, V0.5 fundamentally eliminates the cheap "AI-generated look," ensuring top-tier composition, lighting, and native anime aesthetics.
A Mysterious Ecosystem Addition: Alongside the V0.5 base model, we will also be releasing a highly capable new model within the ckn ecosystem. What exactly is it? We'll leave that as a surprise for you to guess until release day!
The wait is almost over. Get ready for the next evolution of anime AI generation.
Stay tuned for April 10th!
>>
>>
>>
>>
>>
File: REEEEE.png (52.4 KB)
52.4 KB PNG
>>108564801
>SDXL
WHY?? We now have Z-image base and Klein 4b, what is wrong with youuu??
>>
File: mogged.jpg (2.3 MB)
2.3 MB JPG
>>108564685
>>108564690
kek
>>
>>
File: 1772090458692516.jpg (1.1 MB)
1.1 MB JPG
>>108564862
moral of the story, stick with gemini if you want to caption images lol
>>
>>
File: 1771787291393912.jpg (40.9 KB)
40.9 KB JPG
what's the lora training sample aesthetic called?
>>
>>
>>
>>
File: 1747212654368284.png (1.1 MB)
1.1 MB PNG
>>108564989
>its not local
>>
>>
>>
>>
>>
>>
>>
>>
File: _AnimaPreview3_00105_.jpg (424.8 KB)
424.8 KB JPG
>>
>>
File: _AnimaPreview3_00119_.jpg (504.3 KB)
504.3 KB JPG
>>
>>
File: o_00053_.png (1.4 MB)
1.4 MB PNG
>>
>>
>>
>>
File: 1774081832750822.png (39.4 KB)
39.4 KB PNG
>I decide to take a look at what StabilityAI is doing in the year of ourd 2026
https://xcancel.com/StabilityAI/status/2021322296707908034#m
>safety, safety safety
I see that you never changed, after all those years
>>
>>
>>108565426
>q4 is nearly indistinguishable from q8
https://www.youtube.com/watch?v=H47ow4_Cmk0
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 1762080854657300.jpg (451.7 KB)
451.7 KB JPG
>tfw reze lost
>>
I don't browse these threads much but I have a question on hardware.
I currently have an ubuntu system I just run for shit projects I do, can I just slap my 3090 on it and would it work right away with local LLMs because I've heard nvidia drivers are a pain on linux?
It's currently on my main windows PC right now but I can't multitask whenever I have a model loaded on it.
>>
>>
>>108565806
>I just slap my 3090
I mean if it isn't a complete potato like 8g ddr3 system, yes.
>local LLMs
Not the llm thread but sure.
>I've heard nvidia drivers are a pain on linux?
They werk fine for the most part, it's just bunch of cultists seething.
Plus nvidia is a lot better for AI irrespective of your OS.
>>
>>
>>108564801
What's the fucking point of waiting a day to release / announcing a day before?
>>108565785
You in the nice-girl thread too?
>>
File: 1758671165167472.jpg (706.2 KB)
706.2 KB JPG
>>108565873
of course
>>
>>
>>
File: o_00059_.png (1.2 MB)
1.2 MB PNG
>>
File: o_00060_.png (1.8 MB)
1.8 MB PNG
>>
>>
>>108563679
This time around, I do not see advantages to using SFT. The gap between it and Turbo is much lower than on previous version, and I do not notice too much difference in sound quality, plus I think Turbo being more creative still stands.
This seems to be general consensus on Discord as well, everyone is using Turbo, though I'm admittedly a bit worse at prompting SFT and tuning its settings so maybe it's just bias (this is default settings on qinglong UI with steps changed to 50).
Here are some samples from XL SFT, these are prompted with Gemini's help and a tiny change will make something sound 20x better so take results with grain of salt
Keygen music- https://vocaroo.com/11B2ndXidclH
Very sensitive on that one and made everything really fast paced, will prob. need LoRA for that, but chiptunes sound more authentic
Denpa/hyperpop with romaji lyrics
https://vocaroo.com/19gaTvuK3VQg
Eurobeat
https://vocaroo.com/1mF8vI2ppaK6
>>
>>
>>
>>
>>108566325
>how good is it at actually singing all the lyrics in the prompt?
XL Turbo gets it right almost every seed, same with SFT. Not perfect, but both have extremely high pass rates, so much it's not a concern anymore.
>>
>>
>>108566352
Though, as before, what you put in the duration matters. Small duration/slow bpm but too long lyrics can speed up lyrics or increase errors, but generally much more forgiving and seems to adapt really well even if you mess up duration.
>>
>>108566356
fp8scaled, mixed models, the whole purpose of gguf models was to save vram not enhance quality, since vram management and speed has improved by a ton there is no incentive on using gguf models anymore, they are slower to load and clunkier to run, especially on video models
>>
File: o_00062_.png (1.6 MB)
1.6 MB PNG
>>
>>
File: deWA_zi_00005_.png (3.1 MB)
3.1 MB PNG
this is two days old but I just saw it
>Over 1,000 Exposed ComfyUI Instances Targeted in Cryptomining Botnet Campaign
https://thehackernews.com/2026/04/over-1000-exposed-comfyui-instances. html
>>
File: 1775738237.jpg (125.3 KB)
125.3 KB JPG
>She is being embraced from behind by a large, muscular man in plate armor with his head mare
Was supposed to write bare.
>>
File: Flux2-Klein_01432_.png (711 KB)
711 KB PNG
>>108566455
>internet-exposed instances running ComfyUI
couldn't be me
>>
>>
File: 1746425087826946.jpg (819.2 KB)
819.2 KB JPG
Anima 3 is incredible at following artist styles.
>>
>>
>>
>>
>>
>>
>>
>>108566689
>>108566703
we're getting another video model besides LTX and Wan? i thought that was exposed as a fake a few generals ago due to the github looking sketchy.
>>
>>
File: nothing burger chudbob.jpg (70.6 KB)
70.6 KB JPG
Open weights 15b miracle model that beats API thingy was Chinese guy getting baited by some fake website.
We ain't getting jackshit.
>>
>>
>>
>>
>>
>>
File: 1751631366795490.png (105 KB)
105 KB PNG
>>108566757
nothing ever happens, we're stuck with Z-image turbo and Wan 2.2 as the best models until the end of time
>>
>>
>>
>>
>>108566741
Thinking it's not now is just delusional though. One thing is maybe Udio has more catchy songs out the box, because they did insane RLHF, but obviously an ACE Step LoRA or good prompt/seed surpasses that. One thing I have already noticed consistently from testing XL is lyric alignment surpasses the Udio 1.0/1.5 models.
>>
File: deWA_zi_00010_.png (2.3 MB)
2.3 MB PNG
>>
File: 1752197612404711.png (243.2 KB)
243.2 KB PNG
https://huggingface.co/happyhorseai/happyhorse-ai-video-generator
lmao
>>
>>
>>
File: o_00068_.png (1.3 MB)
1.3 MB PNG
>>
I'm attempting to generate multiple different prompts in one run and that all of the interacting characters are in the same prompt, there's a ton of bleeding.
How can I set this up so that it works like the BREAK prompt in forge, but in comfy?
>>
>>
File: 4687474.png (465 KB)
465 KB PNG
Holy based, bluvoll made two anima finetunes in one day, 200% more than any anima defender shills here.
So you can see, tdrusell, /ldg/ is a sham, stop visiting it and check out the anime generals instead, it’s basically an empty shilling general here, no value, no real users of Anima.
>>
>>
>>
>>
>>
>>
>>108567281
Until Tdrusell posts in real anime generals, because Anima was made for anime, I’ll stay here exposing how fake you all are, how you don’t care about anime aesthtetics or Anima, and only use it as a pretext to brag that important people show up here.
Last message for today, I have things to do.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
File: 2d90106ed9dd28e419716802fd01d154.png (516 KB)
516 KB PNG
>>108567221
Tried alternatives to it, but none works. In order for my multiple prompt workflow to work, it needs to be in a single prompt in a single line, a BREAK in that prompt would save it.
>>
>>
>>
>>
>>
>>
>>108567395
only makes sense if you don't care for local at all
using wasteful languages like python makes it easier for the dev, but the consumer needs a lot more hardware and everything takes unreasonably longer than needed
only makes sense if you want to sell something really
>>
File: _AnimaPreview3_00164_.jpg (413.9 KB)
413.9 KB JPG
>>
File: _AnimaPreview3_00179_.jpg (570.8 KB)
570.8 KB JPG
>>
>>108567323
Catjack, you have to understand that anime is like a religion for some weebs, they take it way more seriously than your shallow and cheap view about diffusion. While you were relaxing all this time not genning, there are weebs who keep genning, perfecting their artist tags, loras, etc. making anime diffusion things.
>>
>>
File: _AnimaPreview3_00183_.jpg (528.7 KB)
528.7 KB JPG
>>
>>
File: _AnimaPreview3_00194_.jpg (438.3 KB)
438.3 KB JPG
>>
>>
File: deWA_zi_00014__c.jpg (520 KB)
520 KB JPG
>>
>>
>>
File: _AnimaPreview3_00210_.jpg (687.9 KB)
687.9 KB JPG
>>
>>
File: deWA_zi_00015_.png (2.2 MB)
2.2 MB PNG
>>108567666
is this real?
>>
File: _AnimaPreview3_00255_.jpg (303.3 KB)
303.3 KB JPG
>>
File: 1761592497095071.png (149.2 KB)
149.2 KB PNG
i tested anima. not bad at all, and hot bodies already added. oh yeah
>>
File: rename.png (147.7 KB)
147.7 KB PNG
order is important anons. without order, everything goes to shit.
>>
File: lmao.png (662.2 KB)
662.2 KB PNG
babe wake up, another mid image model got released
https://huggingface.co/CSU-JPG/FlowInOne
>>
File: 1755826836342975.jpg (766.8 KB)
766.8 KB JPG
>>
>>
>>
File: _AnimaPreview3_00282_.jpg (330.3 KB)
330.3 KB JPG
>>
>>
>>
>>
File: shoot.jpg (723 KB)
723 KB JPG
>>108568473
Is pretty much explained there. Just read.
>>
File: 1770503989350880.png (112 KB)
112 KB PNG
>>108568473
when you go on civitai.red you get this shit lool
>>
>>108568473
they're rebranding to cope yet again after getting raped by payment processors and will continue to increasingly moderate characters/concepts that they don't deem 'appropriate'. it's just as censored as always. meanwhile NovelAI is uncensored and doesn't have to cope like this. local remains the censored kek option without any celebs or nono-concepts while API allows complete freedom
>>
>>
File: ComfyUI_20369.png (2.2 MB)
2.2 MB PNG
>>108568229
Dual sampling is the way to go.
I use Z-Image -> Latent Upscale (point resize) -> Z-Image Turbo -> RTX Upscale (and then I downscale before posting).
>>
>>108568575
>NovelAI is uncensored and doesn't have to cope like this.
seriously though how do they manage to get away with it, look at OpenAI they're getting an investigation from the government
https://xcancel.com/AGJamesUthmeier/status/2042258048115265541
>>
>>
>>
>>
>>108568575
and seedance 2.0 is now available for every country (including the US) now, local lost
https://xcancel.com/nickvnturi/status/2042345384299892828#m
https://files.catbox.moe/3csh8s.mp4
>>
>>
>>108568632
>#m
i can see why you like it.
and the reality of seedance
https://www.reddit.com/r/Seedance_AI/comments/1sgh0az/how_to_bypass_se edance_20_face_detection_method_2/
>>
>>
>>
>>
File: look at this dood.png (1.6 MB)
1.6 MB PNG
>>108568645
>the reality of seedance
https://xcancel.com/JSFILMZ0412/status/2042347708250292333#m
>Topview the only one that:
>lets you upload faces without any hacks
YOU LOST LOCALKEK
>>
>>108568593
I find it tricky to dial in z-image turbo, the upscaling part. I made a tiled upscale workflow with the TTP nodes, it works I guess. that rtx upscaler is fast as fuck tho, damn
>>
>>
File: deWA_zi_00019_.png (2.4 MB)
2.4 MB PNG
>>
>>
>>
>>
>>108568637
it was fake >>108563651 :( (seriously though, they made a bullshit video model that was wan 2.7 and people believe they actually made something better and will open source it??)
>>
>>108568685
we've had anons sneak api gens in here for ages
>>
>>
>>
>>
>>
File: ComfyUI_19435.png (2.4 MB)
2.4 MB PNG
>>108568682
>that rtx upscaler is fast as fuck tho, damn
Yeah, it's basically free. Quality-wise, I probably wouldn't keep the final output, but for supersampling back down it's nice to have the resolution available (3072x4096 in my case).
>>
>>
File: _AnimaPreview3_00376_.jpg (366 KB)
366 KB JPG
>>
>>
>>
>>
>>
>>108568815
>>108568593
based jenner, do you think a machine with a 3090 24gb of vram + 64gb ram could generate jennies of this magnitude?
>>
File: tung-tung-sahur.png (106.8 KB)
106.8 KB PNG
>>108568882
SaaS doesn't need a thread because SaaS outputs are now naturally integrated into the cultural zeitgeist. Dall-E 3 generated characters are now featured in Fortnite https://fortnite.fandom.com/wiki/Tung_Tung_Tung_Sahur
API-animated series 'fruit love island' became the world's fastest growing TikTok channel, reaching over 3 million followers in 9 days. Superbowl ads are being animated with API video models, Elevenlabs is being used in Hollywood productions. SaaS is being used for actual revenue-generating work and doesn't need a dedicated tinkertranny helpdesk because SaaS models actually output what you ask without needing to fiddle with 50 knobs (and still get lackluster results)
>>
File: kek.png (107 KB)
107 KB PNG
>>108568943
AIEEE DON'T RELEVATE THAT WE HAVE 0 RELEVANCE IN THE ZEITGEIST
>>
>>
>>
>>
File: ComfyUI_20265.png (2.6 MB)
2.6 MB PNG
>>108568925
Sure. I'm on the same 24/64 memory config. I waited too long on a $440 128GB memory kit because I was saving for a Threadripper build. Now that same kit is like $2k... if you can find it in stock.
Anyway, my workflow spits out an image in ~38s (and ~45s before caching).
>>
>>
>>
File: 635872472572.jpg (2.1 MB)
2.1 MB JPG
>>
File: file.jpg (119.4 KB)
119.4 KB JPG
>>108568473
So now they don't have payment processors to work with anymore?
>>
>>
>>108569190
I think she sneaked a gang sign in there, sorry. vatos locos forever, ese
>>
File: deWA_zi_00023_.png (2.2 MB)
2.2 MB PNG
>>
>>
>>108568766
Video? Sure, but not for long. API has never caught up to image in raw creative capability since Chroma. Funnily, Dalle 3 was closest thing to Chroma, but then censored. New GPT Image may be able to sneak a few things here and there, but still lacks some sovl from Dalle. API starts backwards, it starts ahead then regresses. Local starts behind, but progresses.
>>
File: 1747990745703870.png (214.8 KB)
214.8 KB PNG
>>108569456
>Video? Sure, but not for long.
I hope you're right anon
>>
Fresh when ready
>>108569503
>>108569503
>>108569503
>>
>>108569470
Remember anon, SaaS always regresses. Sora 2 and Seedream 2 quickly became a shell of what they used to be due to censorship. Sure, technically nothing local matches them yet. But something will soon that can do more like NSFW, and can be finetuned to do more based on edge cases, which is where local shines.
On the music gen side of things, Udio quickly removed its ability to be useful as a tool for musicians who are looking to do remixes, covers etc... after they took ownership of every song its users made. Suno could face similar levels of censorship, with that possibility hanging over their head, ACEStep is leading the way with XL (which now has covers very close to Suno's quality).
There's never any guarantee that one will own API made assets, it's just impossible.
>>
>>108568186
>instead of just typing a prompt, you can now make a shitty jpeg of your prompt to make slopped SD1.5 gens!
Why? I mean, I guess being able to circle the designated spot to place objects or drawing an arrow to point at something would be helpful but why did they train it with text overlaid on the input image?
>>
Is there any way to regional prompt in comfyui that isn't completely fucking insane?, i've been at it for 2 hours now and i get dogshit and it doesn't work, i tried Invokeai but the "regional prompting" there is more like regional suggesting, not nearly as good as forge or a1111 but those 2 are slow and clunky, do i have to pick the lesser evil? thoughts?
>>
>>
>>
>>
>>
>>
>>
>>
File: 612320840_122160219734892895_5837496806532520797_n.jpg (167.6 KB)
167.6 KB JPG
is WAN still king for local i2i? any workflows people can link me to? specifically for photo-realistic...