--- license: apache-2.0 base_model: Wan-AI/Wan2.1-T2V-14B tags: - wan - video - text-to-video - diffusion-pipe - lora - template:sd-lora - standard library_name: diffusers pipeline_tag: text-to-video instance_prompt: Jodorowsky psychedelic montage 1970s film by jodorowsky widget: - text: >- Wan2.1 LoRA:Run3@2950 output: url: samples/1755569403645__000002950_0.webp - text: >- Wan2.1 LoRA:Run3@3450 output: url: 3450_Run3_ii.mov - text: >- Wan2.1 LoRA:Run3@2750 output: url: samples/1755566258585__000002750_0.webp - text: >- Wan2.1 LoRA:Run2@1600 output: url: samples/1755545990540__000001600_0.webp - text: >- Wan2.1 LoRA:Run1@1000 output: url: 1000_Wan21i.mov - text: >- Wan2.1 LoRA:Run3@2450 output: url: samples/1755561576004__000002450_0.webp - text: >- Wan2.1 LoRA:Run3@3450 output: url: 3450_Run3_i.mov - text: >- Wan2.1 LoRA:Run3@2350 output: url: samples/1755560004870__000002350_0.webp - text: >- Wan2.1 LoRA:Run2_pEMA_SigmaRel0.19 output: url: 2000pEMA019_Run2_i.mov - text: >- Wan2.2 LoRA:Run1@1000 output: url: Wan22_St1000_jodorowsky1.mov --- ## ALEJANDRO JODOROWSKY's CINE-SURREELS ## A Low(ish) Rank Adapter (LoRA) ## For Wan2.* 14B Text to Video Models ## ____||| By SilverAgePoets.com |||____ Artistically-specialized text to video generative fine-tuned low-rank adapter (Rank 16 LoRA) for the 14billion-parameter Wan2.1, Wan2.2, and derived base models.
This LoRA was trained on a custom dataset of video clips from classic films by **Alejandro Jodorowsky**: the great filmmaker, artist, author, psychoanalyst, sage, & occultist/psyche-mage...
## To reinforce the adapter, pre-phrase/amend prompts with: `[Jodorowsky] psychedelic montage 1970s film by Alejandro Jodorowsky`, etc...
Other suggested prompt-charms: `surrealist occult cinema, eclectically collaged scene, dynamic motion, kodachrome, classic countercultural movie, experimental arthouse analog footage`, etc...
## Training/Usage Notes: The training, orchestrated using Ostris' [ai-toolkit trainer](https://github.com/ostris/ai-toolkit/tree/main), was conducted in several stages/runs, with each pause/re-start involving a partial changing-out of trained-on clips and substantial modifications of hyperparameters:
**Run 1**: Steps 0 thru 1000. With lr: 1e-4, `content_or_style: content` (high noise stage emphasis), and medium resolution samples.
**Run 2**: Steps 1001 thru 1800. With `content_or_style: balanced` (balanced noise schedule), lr: 9e-5, lower resolution and changed out/more numerous samples.
**Run 3**: Steps 1801 thru 3450. With higher resolution samples than previous runs, plus an additional (to linear) training of `conv` & `conv_alpha` networks at rank 16, with `force_consistent_noise: True`, and `content_or_style: content` (high noise stage emphasis), and lr: 1e-4.
This adapter works with both Wan2.1 and Wan2.2. Euler schedulers work best in our tests. Lower "shift" values typically yield more realism/analog quality, depending on other factors.
## With accelerated Inference LoRAs: All checkpoints are confirmed to work with the Wan2.1 Self-Forcing (T2V), FastWan, and CauseVid accelerater adapter LoRAs.
Checkpoints from Run 3 (but not Runs 1 or 2) are confirmed to work with the Wan2.2 T2V Lightning/4-step Adapter (for the Low Noise Expert transformer).
All checkpoints are also likely to work with the Wan2.2 High Noise T2V Lightning/4-step Adapter, but at worse quality/detailing/chromatic range.