Training Images

#1
by ftopal - opened

Hi,
Is it possible to share your images that you used to train?

Hi there @ftopal ,

Thank you for your interest! The images I used for training were sourced from an Instagram page that focuses on AI-generated content. The owner of that page has kept their methods quite secretive, and I believe they may have used Leonardo AI to create those images. While I’d love to share more details, the page itself is a core part of my approach, so I can’t disclose its exact name.

The training was primarily an experiment to see if I could replicate the style of that page. In the future, I plan to fine-tune the model further by selectively retraining it on its best outputs, which I’ll choose manually.

Thanks again for reaching out, and feel free to ask if you have any more questions!

Hi @aleksa-codes

Thanks for your reply! I was wondering if you are planning to provide this SD 3.5 as well. That was my reason for asking. I do like image styles and would love to see it in other architectures as well.

Hi @ftopal ,

Yes, I'm considering it now that I see people having interest in this LoRA/style. I haven't tested SD 3.5 myself yet, so I’m curious about its performance compared to Flux. Have you had the chance to experiment with SD 3.5? I’d love to hear if you think it’s noticeably better.

Also, I noticed that a new LoRA trainer for SD 3.5 was recently released on Replicate, which could streamline this process: SD 3.5 fine-tuner by lucataco.

Really depends on the use case, i know some people are into exceptional realistic images and having better human anatomy and there FLUX seems to be better but for this image style, it's hard to say. That's why i wanted to like compare and see how both of them look.

Hi @aleksa-codes ,

Did you have person images in your dataset as well? I am asking because sometimes faces are not coming out that great. I was wondering if that has something to do with the lora itself.

Hi @ftopal ,

There were no person images in the dataset. I was getting decent results of person faces during my tests, the only problem I was noticing with people was that skin usually had wounds and was bruised up, which I think comes from the style of images it was trained on. Here are some of my results with people:

a8050b00-3841-4350-84d0-de5bf9a8d0bd.jpg

3ff89ebb-00c5-47c5-b1e3-f785560f9131.jpg

0de2b2a6-4a3f-4970-9d0f-8b3fc734f924.jpg

291d61f9-0d13-41ad-bd1b-bef79d2c9541.jpg

Yes, close up shots/portraits work fine for faces but natural environments ends up very less detailed faces and mostly eyes are not detectable. Let me share some results:

idx1__scene_4_1__g3.5_base0.1_mask5_steps50_d1s1.png
chap1_scene6_1_46578686.jpg
chap1_scene1_2_46578686.jpg
chap2_scene4_1_987454698.jpg
chap2_scene2_1_987454698.jpg
chap1_scene6_1_987454698.jpg

In most cases, when the people are bit away, their face quality shrink significantly. I thought maybe it's FLUX related but trying out other LoRas showed me it's not.
One example of what I mean used with the same prompt/same seed:

With this LoRa:
idx0__scene_5_1__g3.5_base0.1_mask5_steps50_d1s1.png

with another LoRa:

scene_5_1__g9_base0.1_mask5_steps50_d1s1.png

I think this is tied to having little to no human in the training dataset. This also affects the quality of hands/fingers. Again mostly for medium to long shots. I sort of think it would be improved if the dataset contains more human samples. Do you plan to have like v2 version of this by any chance? :)

Sign up or log in to comment