Update README.md
This commit is contained in:
parent
26177ccf40
commit
b14022ea92
@ -1,6 +1,6 @@
|
||||
# ddpo-pytorch
|
||||
|
||||
This is an implementation of [Denoising Diffusion Policy Optimization (DDPO)](https://rl-diffusion.github.io/) in PyTorch with support for [low-rank adaptation (LoRA)](https://huggingface.co/docs/diffusers/training/lora). Unlike our original research code (which you can find [here](https://github.com/jannerm/ddpo)), this implementation runs on GPUs, and if LoRA is enabled, requires less than 10GB of GPU memory to finetune a Stable Diffusion-sized model!
|
||||
This is an implementation of [Denoising Diffusion Policy Optimization (DDPO)](https://rl-diffusion.github.io/) in PyTorch with support for [low-rank adaptation (LoRA)](https://huggingface.co/docs/diffusers/training/lora). Unlike our original research code (which you can find [here](https://github.com/jannerm/ddpo)), this implementation runs on GPUs, and if LoRA is enabled, requires less than 10GB of GPU memory to finetune Stable Diffusion!
|
||||
|
||||

|
||||
|
||||
@ -42,4 +42,4 @@ The image at the top of this README was generated using LoRA! However, I used a
|
||||
You can find the exact configs I used for the 4 experiments in `config/dgx.py`. For example, to run the aesthetic quality experiment:
|
||||
```bash
|
||||
accelerate launch scripts/train.py --config config/dgx.py:aesthetic
|
||||
```
|
||||
```
|
||||
|
Loading…
Reference in New Issue
Block a user