16G VRAM; 30 trains per images, 4 epochs, 30 images => 3600 training steps; took 4 hours
Tested the model through ComfyUI with text prompt.
Here are some interesting outputs:
Result: Great with environments, not necessarily with characters.
So I looked for a free online Arcane style LoRa on civitai in case that I need it for style transferring the character in my video.
Moving onto style transferring the video, I exported jpg sequence from UE and selected images for keys so that I can use them on EbSynth.
In the process of style transferring these key frames, I fixed the seed and also use ChatGPT to create text prompt to ensure that the style stay consistent.
Ebsynth doesn't work great with the character when I tried out to do the whole thing at once, so I have to rotoscope out the dancer through Runway AI. After doing so, I generated an Arcane version of the dancer with the Civitai model.
Then I use this image and the rotoscoped footage of the dancer on Viggle to create a greenscreen footage of this Arcane dancer.
After having everything prepared, I used After Effects to edit the video. When the composition is mostly done, I found a copyright free audio online edited it with Reaper to match the composition.