Thread #108585019
File: highlights_g_108575392_1775939293_1.jpg (1.5 MB)
1.5 MB JPG
AIDS and Vaseline Edition
Discussion and Development of Local Image and Video Models
Previous: >>108575392
https://rentry.org/ldg-lazy-getting-started-guide
>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows
>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe
>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo
>Anima
https://huggingface.co/circlestone-labs/Anima
https://tagexplorer.github.io/
>Qwen
https://huggingface.co/collections/Qwen/qwen-image
>Klein
https://huggingface.co/collections/black-forest-labs/flux2
>LTX-2
https://huggingface.co/Lightricks/LTX-2
>Wan
https://github.com/Wan-Video/Wan2.2
>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46
>Illustrious
https://rentry.org/comfyui_guide_1girl
>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Collage: https://rentry.org/ldgcollage
>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg
>Local Text
>>>/g/lmg
>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
339 RepliesView Thread
>>
>>
File: _AnimaPreview3_00044_.jpg (350 KB)
350 KB JPG
>>
>>
File: Flux2-Klein_00669_.png (3.2 MB)
3.2 MB PNG
can we tone the seethe down a bit
>>
File: ComfyUI_temp_pivpu_00014_.png (2.6 MB)
2.6 MB PNG
>>
>>
>>
>mfw Resource news
04/11/2026
>ComfyUI-RookieUI: The ultimate A1111-style sidebar
https://github.com/rookiestar28/ComfyUI-RookieUI
>Qwen3.5-4B-Base-ZitGen-V1: Image captioning fine-tune of Qwen 3.5 4B optimized for Z-Image Turbo
https://huggingface.co/lolzinventor/Qwen3.5-4B-Base-ZitGen-V1
>ComfyUI Memory Visualization
https://github.com/kijai/ComfyUI-MemoryVisualization
04/10/2026
>JoyAI-Image-Edit now supports ComfyUI
https://github.com/jd-opensource/JoyAI-Image#-news
>Two Front Doors: Civitai.com, Civitai.red, and What's Next
https://civitai.com/articles/28369/two-front-doors-civitaicom-civitair ed-and-whats-next
>Uni-ViGU: Towards Unified Video Generation and Understanding via A Diffusion-Based Video Generator
https://fr0zencrane.github.io/uni-vigu-page
>PrivFedTalk: Privacy-Aware Federated Diffusion with Identity-Stable Adapters for Personalized Talking-Head Generation
https://github.com/mazumdarsoumya/PrivFedTalk
>AVGen-Bench: A Task-Driven Benchmark for Multi-Granular Evaluation of Text-to-Audio-Video Generation
http://aka.ms/avgenbench
>Cross-Modal Emotion Transfer for Emotion Editing in Talking Face Video
https://chanhyeok-choi.github.io/C-MET
>ChenkinNoob-XL-V0.5
https://modelscope.ai/models/ChenkinNoob/ChenkinNoob-XL-V0.5
>Control Order & Free Memory: Controls the order of node execution with device-agnostic memory management
https://github.com/mkim87404/ComfyUI-ControlOrder-FreeMemory
>DMax: Aggressive Parallel Decoding for dLLMs
https://github.com/czg1225/DMax
04/09/2026
>MAR-GRPO: Stabilized GRPO for AR-diffusion Hybrid Image Generation
https://github.com/AMAP-ML/mar-grpo
>HybridScorer: Score, sort, and cut large sets down fast with GPU-accelerated AI review
https://github.com/vangel76/HybridScorer
04/08/2026
>OrthoFuse: Training-free Riemannian Fusion of Orthogonal Style-Concept Adapters
https://github.com/ControlGenAI/OrthoFuse
>MIRAGE: Benchmarking and Aligning Multi-Instance Image Editing
https://github.com/ZiqianLiu666/MIRAGE
>>
>>
>mfw Research news
04/11/2026
>M2StyleGS: Multi-Modality 3D Style Transfer with Gaussian Splatting
https://arxiv.org/abs/2604.03773
>SafeCtrl: Region-Aware Safety Control for Text-to-Image Diffusion via Detect-Then-Suppress
https://arxiv.org/abs/2604.03941
>SymphoMotion: Joint Control of Camera Motion and Object Dynamics for Coherent Video Generation
https://grenoble-zhang.github.io/SymphoMotion
>NavCrafter: Exploring 3D Scenes from a Single Image
https://arxiv.org/abs/2604.02828
>Reinforcement-Guided Synthetic Data Generation for Privacy-Sensitive Identity Recognition
https://arxiv.org/abs/2604.07884
>Collaborative Multi-Mode Pruning for Vision-Language Models
https://arxiv.org/abs/2604.02956
>GENFIG1: Visual Summaries of Scholarly Work as a Challenge for Vision-Language Models
https://arxiv.org/abs/2604.04172
>Stochastic Generative Plug-and-Play Priors
https://arxiv.org/abs/2604.03603
>Symbiotic-MoE: Unlocking the Synergy between Generation and Understanding
https://arxiv.org/abs/2604.07753
>Differentiable Stroke Planning with Dual Parameterization for Efficient and High-Fidelity Painting Creation
https://arxiv.org/abs/2604.02752
>Graphic-Design-Bench: A Comprehensive Benchmark for Evaluating AI on Graphic Design Tasks
https://arxiv.org/abs/2604.04192
>Token-Efficient Multimodal Reasoning via Image Prompt Packaging
https://arxiv.org/abs/2604.02492
>DINO-QPM: Adapting Visual Foundation Models for Globally Interpretable Image Classification
https://arxiv.org/abs/2604.07166
>Do Audio-Visual Large Language Models Really See and Hear?
https://arxiv.org/abs/2604.02605
>AutoSOTA: An End-to-End Automated Research System for State-of-the-Art AI Model Discovery
https://arxiv.org/abs/2604.05550
>Neural Network Pruning via QUBO Optimization
https://arxiv.org/abs/2604.05856
>Beyond Fixed Inference: Quantitative Flow Matching for Adaptive Image Denoising
https://arxiv.org/abs/2604.02392
>>
File: pixel-0008-1554111285.png (1.2 MB)
1.2 MB PNG
who remembers disco diffusion
>>
>>
>>
File: _AnimaPreview3_00052_.jpg (318.4 KB)
318.4 KB JPG
>>
>>
File: _AnimaPreview3_00075_.jpg (313.5 KB)
313.5 KB JPG
>>
File: ComfyUI_temp_pivpu_00018_.png (2.5 MB)
2.5 MB PNG
>>
File: ComfyUI_temp_pivpu_00019_.png (2.4 MB)
2.4 MB PNG
>>
File: _AnimaPreview3_00082_.jpg (306.8 KB)
306.8 KB JPG
>>
File: ComfyUI_temp_pivpu_00020_.png (2.4 MB)
2.4 MB PNG
>>
File: _AnimaPreview3_00086_.jpg (243.6 KB)
243.6 KB JPG
>>
File: _AnimaPreview3_00091_.jpg (418.7 KB)
418.7 KB JPG
>>
File: ComfyUI_temp_pivpu_00023_.png (1.9 MB)
1.9 MB PNG
>>
>>108585079
>>108585221
>>108585236
>>108585258
chroma sucks
>>
>>
>>
>>
>>
File: _AnimaPreview3_00104_.jpg (276 KB)
276 KB JPG
>>108585317
bruhs @ u
>>108585378
the bold and the beautiful
>>
>>
File: ComfyUI_temp_pivpu_00033_.png (2.9 MB)
2.9 MB PNG
>>
File: _AnimaPreview3_00119_.jpg (364.3 KB)
364.3 KB JPG
>>
File: images.jpg (10.9 KB)
10.9 KB JPG
Which anime model is most used on 4chan threads right now Anima, SDXL or NAI?
Vote!
https://strawpoll.com/B2ZB9rDajgJ
>>
>>
File: ComfyUI_temp_iuvrh_00001_.png (2.1 MB)
2.1 MB PNG
>>
>>
>>
>>108585500
>>108585512
>/^ComfyUI_temp_/i;type:filename;
you'll thank me later
>>
>>
>>
>>
File: _AnimaPreview3_00121_.png (1.9 MB)
1.9 MB PNG
>>
File: pixel-0009-2084854573.png (1.1 MB)
1.1 MB PNG
>>
>>
File: 00013-3076487574.jpg (1.7 MB)
1.7 MB JPG
>>
File: ComfyUI_temp_lheqp_00001_.png (3.6 MB)
3.6 MB PNG
>>
>>
>>
>>
File: ComfyUI_temp_lheqp_00006_.png (2.4 MB)
2.4 MB PNG
>>
>>
>>
File: ComfyUI_temp_lheqp_00008_.png (1.6 MB)
1.6 MB PNG
>>
File: 00018-1158535645.jpg (1.5 MB)
1.5 MB JPG
>>
File: 1762160217564214.png (3.1 MB)
3.1 MB PNG
>>
>>
File: 00020-4042355821.jpg (2.2 MB)
2.2 MB JPG
>>
>>108585891
Come on, ugly male feet, ugly male toes, ugly male ankles, ugly male thighs, ugly male buttocks.
>but it looks like it was drawn by hand!
Fuck off. Artists are learning too, and many of the artists on Danbooru are amateurs and have slop eyes.
>>
File: 1744668127760498.jpg (977.9 KB)
977.9 KB JPG
can someone fix this
>>
File: 00022-2281910583.jpg (2.2 MB)
2.2 MB JPG
>>
>>108585891
If I covered her from the waist up, would I still be able to tell it was Frieren just from the waist down? What would her legs and feet look like, considering she is a small elf, and what about her toes and ankles?
None of this would happen if you used NoobAI based models, but it is Saturday, it is your free day, it is your "casual animu genning day" you turned on your PC and chose the lowest effort model of all and posted this troon crossdressing as Frieren in the least anime general of all:
ANIMA and /LDG/
>>
>>
File: deJA_zi_00039_.png (2.6 MB)
2.6 MB PNG
>>108586032
thanks i puked
>>
File: 1771032806653727.webm (2.9 MB)
2.9 MB WEBM
what upscaler do you guys use?
seedvr2 changes the face too much and looks like slop.
using z image as a 2 pass completely messes up the skin and adds a weird white haze like someone smeared cum all over the pic.
im really at a loss for good upscalers
>>
File: 00024-2218325491.jpg (1.5 MB)
1.5 MB JPG
>>
File: deJA_zi_00046_.png (2.6 MB)
2.6 MB PNG
>>108586182
this upscaler is basically instant
https://github.com/Comfy-Org/Nvidia_RTX_Nodes_ComfyUI
>>
>>
File: Z-image_00132.png (1.2 MB)
1.2 MB PNG
what is there left to look forward to now that china has officially abandoned us?
>>
>>
>>
>>
https://civitai.com/models/2536147?modelVersionId=2850290
Style lora example for Anima, full captioned dataset and all config files are shared. The model trains extremely well I don't know why some people say otherwise.
>>
is there any LTX 2.3 workflow that doesn't have a hundred random custom nodes? Why do these faggots feel the need to install every random piece of shit node set rather than making things work with the most popular nodes?
>>
>>
File: 00026-1292305904.jpg (1.7 MB)
1.7 MB JPG
>>
>>108586348
wait for the next company to do the same thing.
>here are a bunch of great open source models
they build up a userbase and then try to monetize a new model and another company comes in and fills the void.
or some rich neckbeard like notch or kim dotcom throw a bunch of money into a new model just because they can.
>>
>>
>>108586348
we just accept that we're no longer a part of the cutting edge of tech, we're retro tinkertroons now who enjoy fiddling with outdated hardware. like the people who try to push the limits of the nintendo 64. we will be seeing if we can push out models to get 1/10 as good as seedance 2.0, or if loras can get local models to properly fill a wine glass to the brim.
>>
>>
https://huggingface.co/Lightricks/LTX-2.3/blob/main/ltx-2.3-spatial-up scaler-x2-1.1.safetensors
use the updated upscaler with 2.3, helps a lot it seems.
https://files.catbox.moe/gpmk06.mp4
>>
>>108586485
https://huggingface.co/RuneXX/LTX-2-Workflows/tree/main
I use these, they work well with ltx 2.3 distilled
>>
>>
>>108586573
I just want a good API that isn't completely gimped after the first week.
Seedance2 looked good when it was first showcased, what we have now is a joke.
Bad physics, plastic skin, inconsistent generations, stiff animations.
I guess we wait for Happyhorse, but it will probably get hit with a cease and desist on day one.
Fucking bleak.
>>
>>
>>
>>108586032
This reminded me of when I went to the ENT doctor and he had a colossal scar on his throat, it shocked me because I thought the scar was from him slicing his throat but it was from a thyroid surgery lmao
>>
>>
>>
>>
>>
>>
>>
>>108586449
I HAVE TO POST MY ANIME NEWS TO THIS 3DPG SLOP GENERAL!!!
MUH CATJACK MUST READ MY ANIME NEWS OR I WILL LOSE MY MIND!!!
MUH CATJACK!!! MUH MEAT!!!
THEY ARE VERY VERY IMPORTANT!!! NOT THE 200 ANIME POSTERS OF ALL 4CHAN, NO NO NO, THEY ARE TRASH, WORTHLESS, BENEATH CONTEMPT!!! ONLY MUH CATJACK AND THE ZIT AND THE CHROMA SLOPPERS MATTER TO ME!!!
/LDG/ MUST STOP EVERYTHING AND READ MY ANIME NEWS RIGHT NOW THIS INSTANT!!
>>
File: tenshi hakase by greg rutkowski.png (943.6 KB)
943.6 KB PNG
>>108586449
Onegai, realism lora kudasai!
>>
>>
>>
File: image-4.jpg (239.5 KB)
239.5 KB JPG
What's the easiest local hardware I can use to make slop like this where I'm just going to take pictures and say "Give her a silver dress" or "Give her blue eye shadow" like you can do with cloud tools like Gemini and Grok
>>
>>108586841
>>108585019
>>Klein
>https://huggingface.co/collections/black-forest-labs/flux2
>>
>>
>>108586851
I've got a 5080 and this is the most exciting use case of it now that I've beaten RE9 on Nightmare
>>108586844
thank you will give it a shot
>>
>>
>>
>>
>>108586880
>>108586869
Is there a way to get Qwen edit running locally on Linux instead of the hugging face version?
>>
>>
>>
>>
>>108586902
Three versions of Qwen are listed here https://comfyanonymous.github.io/ComfyUI_examples/qwen_image/ godspeed my celeb gooner
>>
>>108586449
I spent 8 hours today of my saturday using your model and sharing artist tag and comparisons on /h/, /e/ and /adt/ with other anons who use Anima. Watching you ignore us makes me want to never use your model again.
>>
>>
>>
File: 1758948263662488.png (1.5 MB)
1.5 MB PNG
>>
>>108586449
>I don't know why some people say otherwise.
If you spend 5 minutes on any place where people discuss lora training you discover why, most people train to overfit because they caption poorly and prompt poorly, so for the loras to work for them they have to imprint the DNA of the image in the model
>>
>>
>>
>>
>>
>>
File: projecting.png (471.2 KB)
471.2 KB PNG
>>108586862
>>
File: Chroma_Generated_00054_.png (695.1 KB)
695.1 KB PNG
>>108585744
Most of the discord seems to have moved on from this thread. I had to stop chroma gooning but if I remember correctly all three are still in training epochs. Kaleidoscope is fairing better than Zeta, to the surprise of nobody.
>>
>>
>>
>>
>>
>>
>>
>>
https://higgsfield.ai/original-series/zephyr/episode-1
>Traditional directors flimmaxxxing using Seedance 2.0 on Higgsfield. Watch “Zephyr” FULL Ep.1 – this is what happens when filmmakers face ZERO gatekeeping. With Unlimited Seedance 2.0 now LIVE everywhere for anyone with up to 70% OFF* - YOU can build your next viral AI movie. 2 minute intro got MILLIONS in a day. Now see how full Zephyr takes over your feed.
>Dir. by ILYA KARCHIN & the team.
>Zephyr (2026)
>>
>>108586609
see, quality is better vs the 1.0 one (both use the new one):
https://files.catbox.moe/5pid7f.mp4
>>
File: big eyed freak boy.png (1.6 MB)
1.6 MB PNG
>>108586931
>>
File: ComfyUI_temp_jutls_00002_.png (2.6 MB)
2.6 MB PNG
>>
>>
File: ComfyUI_09207_.png (1.9 MB)
1.9 MB PNG
guess ill try out the greg lora...
>>
File: ComfyUI_temp_jutls_00003_.png (2.4 MB)
2.4 MB PNG
>>
>>
File: ComfyUI_temp_jutls_00005_.png (2.5 MB)
2.5 MB PNG
>>
>>
>>
>>
File: promptmaxing.png (118 KB)
118 KB PNG
who else is promptmaxxxing?
>>
>>
>>108587318
>>108587278
Very, very impressively realistic
>>
>>
File: ComfyUI_temp_jutls_00015_.png (2.9 MB)
2.9 MB PNG
>>
>>
File: ComfyUI_temp_jutls_00022_.png (2.2 MB)
2.2 MB PNG
>>
>>108585160
>>108585250
>>108585250
>>108585268
>>108585394
>>108585480
I like the 90s anime filter, but you gotta train your model to stop making Frieren look so unhinged/retarded
>>
>>
>>
>>
File: Screenshot_20260412_005017.png (462.3 KB)
462.3 KB PNG
Can anyone make some recommendations for why I'm not getting the results I'm expecting to get? I know that if I were doing this with something that had natural language processing it wouldn't change anything about the image other than her outfit
Is Qwen 2512 not a good fir for what I'm trying to do?
>>
File: ComfyUI_temp_jutls_00024_.png (3.2 MB)
3.2 MB PNG
>>108587646
its better to inpaint if you wanna add specific stuff
>>
>>
>>
>>108587646
>>108587728 this kek
>>
File: Screenshot_20260412_011933.png (147 KB)
147 KB PNG
>>108587728
Sorry I'm super new to this- Where do I find that under templates? Is it just called Qwen Image Edit or is it this one?
>>
If you're doing SFW stuff just stick to API. It's not worth using Qwen Edit for basic shit that API models can do 100x faster and better. You can get Qwen Edit here, but the model itself is outdated https://huggingface.co/Qwen/Qwen-Image-Edit-2511
Flux Klein 9b is the best edit model available locally, and it can gen/edit without needing separate models. Qwen was working on a model that could gen/edit in one, but they decided to abandon local for API like everyone else.
>>
>>
>>
>>
>>108587775
I've never upscaled anything that didn't feel like a sidegrade. You always lose something you liked about the original gen, unless you keep the denoise so low that you wonder "what am I even wasting my GPU cycles on? It looks the same."
>>
>>108587784
API = 'the online stuff'. If your task is just putting a fur scarf on a girl, then there's no reason not to just use google AI or whatever. The benefit of local is nsfw stuff and niche use-cases that API cannot achieve (loras which are trained on specific concepts or styles).
>>
File: Screenshot_20260412-014235.png (920.3 KB)
920.3 KB PNG
>>108587801
The main thing is I just don't want to get constrained by stuff when I am doing something NSFW or if it's just some random guideline it doesn't agree with. I also despise paying for cloud software in general
I feel like what I'm asking for isn't necessarily outlandish, I'm sure it would be slower on local hardware even with a 5080, basically just something that can interpret prompts and then apply them to images.
I mean look at this, why would this be moderated? It's something I'm sure local hardware is capable of, I'm trying to find the best tool for the job
>>
File: Screenshot_20260412-015115.png (1.5 MB)
1.5 MB PNG
>>108587834
Another example, this stuff seems like it's what people have been working on for years, I'm surprised there isn't a consensus best tool for something like this. Granted there are a million different directions that people are working on
>>
>>
>>108587856
Okay will be trying klein in the morning, thanks Anon
>>108587860
Sadly after playing for a couple hours I think Qwen might not be what I'm looking for, what I'm looking for is a really simple tool that's kinda hard to fuck up
>>
>>108587855
There isn't a single best tool because all companies are competing. Here is a list of pretty much all the relevant edit models:
https://artificialanalysis.ai/image/leaderboard/editing
>>
>>
>>
File: 1765336873951395.png (3.9 MB)
3.9 MB PNG
my wife seira
>>
File: 1772319093603289.jpg (831.7 KB)
831.7 KB JPG
>>
File: Why.png (8.2 KB)
8.2 KB PNG
Sorry if this is a dumb question, I'm new to this. Why does this keep popping up? I already downloaded and selected a VAE and put it in the VAE folder. (vaelsem). What else do I need to do?
>>
>>
>>
File: What.png (31.7 KB)
31.7 KB PNG
>>108588068
I downloaded Stability Matrix and just downloaded the first package available there (WebUI Forge NEO). And these are the other stuff
>>
>>
>>108588117
more action:
https://litter.catbox.moe/n1p1j6e4ns8bxn83.mp4
>>
File: comfy__520.jpg (990.2 KB)
990.2 KB JPG
>>108588117
>>108588132
instant classics
>>
File: 1761981410156904.png (131.1 KB)
131.1 KB PNG
https://github.com/Comfy-Org/ComfyUI/pull/13369
What is this model? Are we saved?
>>
File: 1770027281971472.png (350 KB)
350 KB PNG
>>108588222
looks like it's a 8b model and uses ministral 3d as a text encoder, but so far it doesn't seem there's anything else about that model on the internet
>>
>>
>>
File: 1750705985290236.png (3.2 MB)
3.2 MB PNG
>>
File: 85673838227.jpg (2.3 MB)
2.3 MB JPG
>>
>>
>>
>>
>>
>>
>>
>>108588341
>>108588331
the seethe is delicious
>>
>>
File: 1774752637044401.png (347.6 KB)
347.6 KB PNG
>>108588349
baidu's site is one of the most visited in the world
>>
>>
>>
>>
File: 1756511195721900.jpg (822.4 KB)
822.4 KB JPG
so many new kinos to watch, so little time
>>
>>
File: 1753423010278987.jpg (652.6 KB)
652.6 KB JPG
>>
File: 1751022907777741.jpg (558.2 KB)
558.2 KB JPG
>>
>>
>>108588473
You don't train loras for Anima. Bluvoll already made 2 finetunes this week. Stop pretending to be interested in anime.
>>108588320
It's the only anime model successor of Illustrious, and you don't even post in anime generals. You're trash. I know it's you because you're the only one who cares about your model in this 3DPG general.
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>108588733
Change sd.next settings to FP16, read the wiki (Compute setting) https://github.com/vladmandic/sdnext/wiki/Performance-Tuning
>>
>>
File: 8653786373.jpg (1.9 MB)
1.9 MB JPG
>>
File: 554557516881795.png (460 KB)
460 KB PNG
>>108586449
It do work.
>>
>>
>>
>>
>>
>>108588922
gemma 4 + that system prompt >>108588368
>>
>>
>>108588974
it can with that jailbreak prompt >>108588960
>>
File: images.jpg (28.2 KB)
28.2 KB JPG
>>108588985
>>
>>
>>
>>
File: 777.jpg (2.3 MB)
2.3 MB JPG
>>108589090
sir pls understand
>>
File: r7qbi2yz8qqg1.png (1.8 MB)
1.8 MB PNG
saw this guy on reddit using generated pics for a fake onlyfans. any idea what model he could be using?
>>
>>
>>
>>
>>
File: 1758174372714758.png (69.5 KB)
69.5 KB PNG
>>108588222
https://github.com/huggingface/diffusers/pull/13432
based, there will be a base model and its turbo variant
>>
File: 1685272533435.png (27.4 KB)
27.4 KB PNG
>>108589177
Lustify maybe.
>picrel
HATE
>>
>>108589268
https://github.com/HsiaWinter/diffusers/blob/3aec976fc30347e4ea70e5f97 c1bb4123cc218fd/docs/source/en/api/ pipelines/ernie_image.md
>ERNIE-Image is designed with a relatively compact architecture and solid instruction-following capability, emphasizing parameter efficiency. Based on an 8B DiT backbone, it provides performance that is comparable in some scenarios to larger (20B+) models, while maintaining reasonable parameter efficiency.
big if true
>>
>>
>>
>>
>>
>>108589307
>Big Bird Image
When bird game 3 image??
https://www.tiktok.com/@ancient_meme_archive/video/7557971057102114079
>>
>>108589318
I trained ltx23 lora in AITmostly with videos. It worked decently but I think the images were not close enough in style and slowed the learning and made the model stiffer. I think I'll try videos entirely next time.
>>
>>108588222
So Comfy implemented that baidu model one but not this one?
https://huggingface.co/jdopensource/JoyAI-Image-Edit
why?
>>
>>
File: 1751917912441562.mp4 (1.9 MB)
1.9 MB MP4
how did wan 2.7 fuck up this much?
https://xcancel.com/ChrisGwinnLA/status/2039960196458680366
https://www.youtube.com/watch?v=RERsGjQrQ6E
wan 2.5/6 was marginal improvement if even that, and now this is just trash.
>>
>>108589402
I'm not talking about the starting frame. It was a character lora. I had two datasets, one video dataset and one image dataset. The image dataset had images from photoshoots etc. which didn't match the style of the videos.
>>
>>108589423
anon, Alibaba actually has a good video model, it's called HappyHorse
https://xcancel.com/AlibabaGroup/status/2042530517799887326#m
https://xcancel.com/lovart_ai/status/2043282414605332813#m
>>
File: budgetpixel-image-1400866.png (1.7 MB)
1.7 MB PNG
>>108587834
nigga look for site that has flux klein, qwen image 2.0 or wan2.7. Budget pixel is my favorite it because of it variety of models and mentions the various levels on strictness a of a model. I find local image generation at the moment to very stale and boring at the moment.
https://budgetpixel.com/
>>
>>108589423
I've used wan2.7 image and video generation. its absolute censored dogshit that even makes wan2.5 look a lot better. The shit model has a filter that re-writes your prompt to be sfw pg13. Basically making de-clothing and nudity prompts difficult to get right. Many people are disappointed with wan2.7 and it's basically DOA saas model.
>>
>>
>>108589506
>Image
>Lightyears behind SAAS
maybe baidu will save us >>108588222
>>
>>
>>
>>
File: local definitely lost.gif (14.6 KB)
14.6 KB GIF
https://xcancel.com/obscaries/status/2043304041053397437
>>
File: woof.png (222.8 KB)
222.8 KB PNG
>>108589523
>now we're talking
*yawn*, pixel space or gtfo
>>
>>
>>108589585
obviously
https://xcancel.com/AzeAlter/status/2043027227374436827
>>
>>
File: Untitled.png (355.9 KB)
355.9 KB PNG
Trellis 2
>>
>>
>>108589179
He is literally generating not literal children.
>108589265
I make videos of sexy kids, not images, because while still images work for big tits hags since you can appreciate them as meat, little girls are more of a vibe so video works better for that. It's why I'm waiting for video+audio so eagerly to add more dimensionality to that vibe.
Unfortunately it's looking like the odds of the evasi@n website being shut down due to costs before a local video+audio model comes out go up every month
>>
>>
>>
>>
File: Gemma 4 31b caption.png (1.4 MB)
1.4 MB PNG
>>108589637
>it'll get better along with hardware.
I used to believe that, but Z-image turbo and Gemma 4 proved to me that you can get insane quality with a relatively small model, the future is bright
>>
File: Untitled.png (376 KB)
376 KB PNG
>>108589670
If you look up the latest repos, they use something called DINO lock that helps a lot. But it's still meh.
>>
>>108589689
>you can pay $10
That's the whole concern, his stripe got banned and most of his potential market are retarded sooner nocoders who don't understand crypto so I'm worried about the financial health. I've gotten many times the worth of my gold donation and have shared hundreds of videos of youthful beauty with the world and have done my small part in displacing and substituting demand for the real thing.
No seriously, a couple of times on some pedo-adjacent forum I see some guy post an old Gen and it makes me happy knowing that it is inarguable that this person has consumed something AI generated instead of the real thing
The monthly begging prompt is back though and unlike previous years there's less momentum and the project is much more private, and 4chan is getting less and less popular.
Video+audio doesn't share well on /g/ anyways but I'm just excited for the extra world knowledge the audio dimension brings. I will FINALLY be able to actually prompt for something like a home family vlog where the dad is holding the camera because 1000% that training data is in current models but there's no way to express that knowledge given the relationships you make when captioning videos with just text and not learning the audio information
>>
File: 00001-1260334831.jpg (2.2 MB)
2.2 MB JPG
>>
File: PLEASE.png (33.9 KB)
33.9 KB PNG
>>108588222
please be good, we haven't gotten anything decent this year so far (except klein I guess)
>>
File: 00004-1077369241.jpg (2.1 MB)
2.1 MB JPG
>>
File: 1738142509041313.jpg (79.9 KB)
79.9 KB JPG
>>108589423
>>108589439
i'm really fed up with chinks. i wouldn't have said anything if they'd released something, but now, fuck them. i don't even touch wan anymore for sfw content. ltx is much better
>>
>>
>>
>>108589985
Depending on how many steps you need and which model (I am presuming something like SDXL with that res + batch size combo) it can cost less than a dollar to train a lora with online compute. (vast, runpod, etc.)
>>
File: 1773673613343768.jpg (686.4 KB)
686.4 KB JPG
OWO
>>
>>108590008
uwu, what's this?
https://www.youtube.com/watch?v=7mBqm8uO4Cg
>>
>>
>>
File: 1774861054051915.jpg (514.8 KB)
514.8 KB JPG
>>
File: 00009-3934370304.jpg (2.5 MB)
2.5 MB JPG
>>
File: o_00179_.png (1.1 MB)
1.1 MB PNG
>>
>>
File: o_00181_.png (1.7 MB)
1.7 MB PNG
>>108590150
"anons might find this image amusing.... oh wait an anon is seething about it, even better, a successful post"
>>
>>
File: o_00184_.png (909 KB)
909 KB PNG
>>108590187
ah, i thought they are always uniformly bad. maybe on occasion less bad.
i am just bored.
>>
>>
>>
File: 00033-2433693979.png (3.6 MB)
3.6 MB PNG
>>108590008
prompt and settings please? you used z image base right?
>>
>>
File: Anima_01713_.png (1.4 MB)
1.4 MB PNG
>>
File: 1774838182679087.png (128.6 KB)
128.6 KB PNG
>be me reusing ltx2
>be me testing a medieval prompt that gives good results everywhere.
>the gen is full of happy modern indians...
>>
>>
>>
>>108590356
>the regular one just refuses
even with this jailbreak? >>108588960
>>
File: Chroma1-HD-Flash.safetensors_00031_.png (1.4 MB)
1.4 MB PNG
>>
>>
File: Anima_01717_.png (2.1 MB)
2.1 MB PNG
Gotta post this one to >>>/aco/9173986
>>
File: 1652614388904.png (203.8 KB)
203.8 KB PNG
>>108590357
I have white characters in version 2.3, but with medieval carnival costumes...they have serious problems with medieval content.
>>
Wow, ComfyUI updated the front end, but Nodes 2.0 still don’t work with text autocompletion, and I can’t align the nodes 2.0 horizontally or vertically or redistribute them.
Thanks for doing nothing, Fennecfaggot.
>>
>>
File: Anima_01719_.png (2.1 MB)
2.1 MB PNG
>>108590437
Just disable Nodes 2.0? They are shit anyways.
>>
>>
File: 1759740297338035.jpg (748.5 KB)
748.5 KB JPG
>>
File: o_00190_.png (719.9 KB)
719.9 KB PNG
>>
File: Anima_01722_.png (1.9 MB)
1.9 MB PNG
I love the kind of LCD 1girls anima can spit out.
Preview 2 performs better than 3 for this, among other things.
>>
comfy bred
>>
>>
>>
File: 03405-animaOfficial_preview3Base.png (3.7 MB)
3.7 MB PNG
There are quite a lot of really nice 'western style' loras for Anima now and they blend pretty well with anime styles as well if you screw around with step scheduling. I'm having lots of fun.
>>
File: Anima_01726_.png (2 MB)
2 MB PNG
>>108590493
More buggy to use than the original node design which is what I care about most. When (if) it stops being a buggy piece of shit I might care about leanness.
>>108590498
I am curious where do you typically switch and from which to which?
>>
>>
>>
File: Anima_01730_.png (1.9 MB)
1.9 MB PNG
>>108590539
thx
>>
File: 03392-animaOfficial_preview3Base.png (2.9 MB)
2.9 MB PNG
>>108590510
I haven't done much systematic testing regarding order but the timing serves as a pretty robust way of dialing in the amount of influence you want a style to have. In the attached image the relevant section is pasted below. I'm only using mikoyan's style half the time. In my previous image I only let it in with the last 10% steps. Super fun to play around with.
<lora:Sanjulian:0.7> sanjulian style,painting \(medium\)
[@mikoyan::0.5]
>>
>>
>>
I tried this training for the first time in three years. On Klein 9B Snofs 1.3 Checkpoint, I completed a Lokr with 6 images in 15 minutes.
A very good result. Times have changed. Setting up the trainers is just as annoying as it used to be. Kek
>>
>>108590587
In this particular case, the lora does have a trigger word "sanjulian style" and its effect is constant. I am using this bit: [@mikoyan::0.5] to add ON TOP of the lora with an artist style that is already known by anima but only for the first 50% of steps.
>>
File: Anima_01737_.png (1.7 MB)
1.7 MB PNG
>>108590615
Ok thanks for the clarification anon.
>>
File: 858734683568258.jpg (2.8 MB)
2.8 MB JPG
>>
>>
File: Z-Image_00026_.png (1.8 MB)
1.8 MB PNG
most of my nodes are vibe coded, at this point i might as well vibe code my own CLI and avoid using cumfart entirely
i assume thats what most of you guys are doing
>>
>>
File: Anima_01742_.png (1.6 MB)
1.6 MB PNG
>>108590669
I think you are underestimating the difference in scope and complexity.
It's not a "one and done" deal neither as you will need to slop in support for every new model, method, etc. that you want to use.
Not the biggest fan of the state Comfy is in, and it's bloated as hell, but it's not all for nothing.
>>
File: 00010-3141267921.png (1.2 MB)
1.2 MB PNG
>>
File: o_00194_.png (959.6 KB)
959.6 KB PNG
>>
File: Anima_01745_.png (1.9 MB)
1.9 MB PNG
>>
File: ComfyUI_temp_rdmcr_00049_.png (915.5 KB)
915.5 KB PNG
>>
>>108590669
I vibe coded a more asset centric node slop app. It uses comfy UI as a backend and you dump in an API workflow and tell it which inputs you actually care about (prompt, resolution, loras). Assets are connected on a graph to indicate their origin and relationships. This is for a video game with branching paths and shit and I wanted something to keep better track of assets like animations because doing it manually would occasionally lead to catastrophic fuckups (like say I put the wrong resolution for an early generation in a sequence and didn't realize it).
>>
>>