Doing some simple basic testing with a character turnaround. Melita's design is really simple so I'm not exactly looking for high consistency with her and wanted to have some sort of baseline for loopback testing. Unfortunately, I wasn't very successful so it's not really worthwhile writing and article for. I'll probably need to design a character with more complex details later for more stress testing. It seems for Illustrious knowing the proper prompts/artist keywords seem to help with the posing. The dataset is only a limited turnaround so the posing is very stiff. The anytest CN can help with adding dimensionality to the character which I think is only possible as her design is really straightforward.
A complex character design. Images are very cherrypicked on the basis of the outfit but the LoRA performs very well in terms of inpainting. More overtrained compared to the other LoRAs due to loopback usage.
A rather simple character design this time to test out a couple of things regarding accessory training, and clustered objects along with experimenting with loopback.
3/4/24 Update:
Blind Test with SDXL training. I'll write an article about this later. Read 'About this Model' for more details.
Kie recommended weight:1
I am probably going to need to change the title of the page now since I am using txt2img ControlNets to augment the dataset for style. (I did use a LoRA to loopback but I used weight of .1~.2 and outd to weaken the effect. I did another round without the LoRA and the difference wasn't very noticable). Read below if you want see the specifics of the augmentation.
Enna LoRA focuses on a simpler design to see what mistakes in the turnaround make an impact on the LoRA design
This is a place where I keep my tests with single image OC LoRAs. The goal is to try and create a consistent character LoRA with a minimum amount of effort. If you're willing to spend time, you can create a consistent character in this tutorial here (not mine). Keep in mind that these experiments are more of a learning effort rather than producing an accurate LoRA. Due to the limited dataset, these LoRA are highly likely overfit in weird scenarios that I haven't tested yet. I'll probably write an article about the process in the future.
Images are of-course cherry picked with adetailer and hi-res fix with (AnimeSharp4x\UltraSharp4x) at 0.2 ~ 0.5 denoising strengths.
Isabella's original image is the second image in V1. (Base prompt might be useful)
Recommended Weight for V4: (0.5~0.8) Block Weight can help but isn't completely necessary.
Recommended Weight: 0.6 ~ 0.8 + (LoRA Block Weight with OUTD preset recommended; For version 3, the block weight extension is highly recommended).
For Enna: Recommended Weight: 0.8 ~ 1 (Lower if encounter concept bleeding)
Use terms such as: flat_chest,petite to get character proportions correct.
You can get away with 1 weight with the block weight extension but will need to lower the weight if you want do custom outfits. Requires long prompts that describe that character and the environment to work well.
It should work well enough with anime-based models, I haven't tried anything realistic or semi-realistic. Probably best used with an inpainting workflow.
Inconsistency with coloring and outfit
facial expression is overfit
Hairline is occasionally flipped
V2 is probably better than V3; I ran into a roadblock but decided to publish my results for now
V1
1 Full body shot +1 crop image of face (same image)
Trained on NovelAi
V2
Reused dataset from V1
Added upperbody,lowerbody, and leggings crop
lowered repeats
more captioning
V3
removed flip augmentation
reused data set from v1
added 3/4 images of the head and skirt, boots to neck,close-up of bangs
more captioning
Others
I likely hit a wall with Isabella so I might not be able to improve the LoRA any further. There are a few more things I could try with kohyaSS trainer but I'm not expecting much. The main limitation is that the base image isn't very high quality and has image artifacts on zoom up.
V4
reused data from v3
moved regularization over into the main dataset and added trigger words for style
added style images using txt2img with canny and lineart using regional prompter to augment the dataset for style, tagged as 'alternativecostume'
selected images with closest colors
added style and outfit triggerwords
n tokens increased to 4
removed block weight during training
V1
Based on a character turnaround
Uses same concepts from V4
V1.1
minor update with more captioning and lower training rates by 1/10th
V2
Added full turnaround of swimsuit and underwear
Blind attempt at training a character with medium-level details with AnimagineXL
-Observations
Trigger word is a hard requirement otherwise the LoRA will not work
Recommended Strength:1
Clothing is fully removable and easily swappable but prefers long and descriptive prompts
use best_quality,masterpiece,modern for ease of nsfw according to AnimagineXL documentation
Style keywords are usable but I'm not familiar with all of the related keywords with AnimagineXL
Overfitting
Anime Screencap artstyle is bleeding over into the environment details causing generation images to have overly flat backgrounds
may require words such as 'head_out_of_frame' and 'from_behind' in negative
There is still some level of style bleeding but I'm not completely familiar with how Animagine handles style keywords
Strong preference towards cowboy shots
some overfitting on face coloring
Undertrained
Waist ribbon may or may not appear
Doesn't seem to understand asymmetric so details can be flipped
Belt color is undertrained but is occasionally correct
Base outfit seems to lose detail when using other artstyles
Augmentations:
Img2Img ControlNet augmentation with Flat style lora and Add Detail LoRA
Swimsuit and underwear images
Notable Differences from SD1.5
Small rank compared to SD1.5 but SDXL has larger size overall
removed Style captioning to reduce overfitting of style
removed Manual Pose regularization data
Single Repeats folder for most of the data
Go ahead and upload yours!
Your query returned no results – please try removing some filters or trying a different term.