From 895795d9e434d264d65be2b43c31f3f86915bc53 Mon Sep 17 00:00:00 2001 From: YaYaB Date: Thu, 6 Oct 2022 12:00:37 +0200 Subject: [PATCH 1/2] Fix push_to_hub for dreambooth and textual_inversion --- examples/dreambooth/train_dreambooth.py | 5 ++--- examples/textual_inversion/textual_inversion.py | 5 ++--- 2 files changed, 4 insertions(+), 6 deletions(-) diff --git a/examples/dreambooth/train_dreambooth.py b/examples/dreambooth/train_dreambooth.py index 322c5506156a..5a4b5dd2e313 100644 --- a/examples/dreambooth/train_dreambooth.py +++ b/examples/dreambooth/train_dreambooth.py @@ -13,6 +13,7 @@ from accelerate.logging import get_logger from accelerate.utils import set_seed from diffusers import AutoencoderKL, DDPMScheduler, StableDiffusionPipeline, UNet2DConditionModel +from diffusers.hub_utils import push_to_hub from diffusers.optimization import get_scheduler from huggingface_hub import HfFolder, Repository, whoami from PIL import Image @@ -575,9 +576,7 @@ def collate_fn(examples): pipeline.save_pretrained(args.output_dir) if args.push_to_hub: - repo.push_to_hub( - args, pipeline, repo, commit_message="End of training", blocking=False, auto_lfs_prune=True - ) + push_to_hub(args, pipeline, repo, commit_message="End of training", blocking=False, auto_lfs_prune=True) accelerator.end_training() diff --git a/examples/textual_inversion/textual_inversion.py b/examples/textual_inversion/textual_inversion.py index 5b5ba9a2078e..3cbd38c6566a 100644 --- a/examples/textual_inversion/textual_inversion.py +++ b/examples/textual_inversion/textual_inversion.py @@ -17,6 +17,7 @@ from accelerate.logging import get_logger from accelerate.utils import set_seed from diffusers import AutoencoderKL, DDPMScheduler, PNDMScheduler, StableDiffusionPipeline, UNet2DConditionModel +from diffusers.hub_utils import push_to_hub from diffusers.optimization import get_scheduler from diffusers.pipelines.stable_diffusion import StableDiffusionSafetyChecker from huggingface_hub import HfFolder, Repository, whoami @@ -569,9 +570,7 @@ def main(): save_progress(text_encoder, placeholder_token_id, accelerator, args) if args.push_to_hub: - repo.push_to_hub( - args, pipeline, repo, commit_message="End of training", blocking=False, auto_lfs_prune=True - ) + push_to_hub(args, pipeline, repo, commit_message="End of training", blocking=False, auto_lfs_prune=True) accelerator.end_training() From c34a76601474f1526e15651f88287b92b71ea963 Mon Sep 17 00:00:00 2001 From: YaYaB Date: Fri, 7 Oct 2022 10:54:03 +0200 Subject: [PATCH 2/2] Use repo.push_to_hub instead of push_to_hub --- examples/dreambooth/train_dreambooth.py | 3 +-- examples/textual_inversion/textual_inversion.py | 3 +-- .../unconditional_image_generation/train_unconditional.py | 4 ++-- 3 files changed, 4 insertions(+), 6 deletions(-) diff --git a/examples/dreambooth/train_dreambooth.py b/examples/dreambooth/train_dreambooth.py index 5a4b5dd2e313..56deb4a720d8 100644 --- a/examples/dreambooth/train_dreambooth.py +++ b/examples/dreambooth/train_dreambooth.py @@ -13,7 +13,6 @@ from accelerate.logging import get_logger from accelerate.utils import set_seed from diffusers import AutoencoderKL, DDPMScheduler, StableDiffusionPipeline, UNet2DConditionModel -from diffusers.hub_utils import push_to_hub from diffusers.optimization import get_scheduler from huggingface_hub import HfFolder, Repository, whoami from PIL import Image @@ -576,7 +575,7 @@ def collate_fn(examples): pipeline.save_pretrained(args.output_dir) if args.push_to_hub: - push_to_hub(args, pipeline, repo, commit_message="End of training", blocking=False, auto_lfs_prune=True) + repo.push_to_hub(commit_message="End of training", blocking=False, auto_lfs_prune=True) accelerator.end_training() diff --git a/examples/textual_inversion/textual_inversion.py b/examples/textual_inversion/textual_inversion.py index 3cbd38c6566a..18469af3d4ae 100644 --- a/examples/textual_inversion/textual_inversion.py +++ b/examples/textual_inversion/textual_inversion.py @@ -17,7 +17,6 @@ from accelerate.logging import get_logger from accelerate.utils import set_seed from diffusers import AutoencoderKL, DDPMScheduler, PNDMScheduler, StableDiffusionPipeline, UNet2DConditionModel -from diffusers.hub_utils import push_to_hub from diffusers.optimization import get_scheduler from diffusers.pipelines.stable_diffusion import StableDiffusionSafetyChecker from huggingface_hub import HfFolder, Repository, whoami @@ -570,7 +569,7 @@ def main(): save_progress(text_encoder, placeholder_token_id, accelerator, args) if args.push_to_hub: - push_to_hub(args, pipeline, repo, commit_message="End of training", blocking=False, auto_lfs_prune=True) + repo.push_to_hub(commit_message="End of training", blocking=False, auto_lfs_prune=True) accelerator.end_training() diff --git a/examples/unconditional_image_generation/train_unconditional.py b/examples/unconditional_image_generation/train_unconditional.py index d3b4e17743f4..8ddbdd77ba42 100644 --- a/examples/unconditional_image_generation/train_unconditional.py +++ b/examples/unconditional_image_generation/train_unconditional.py @@ -9,7 +9,7 @@ from accelerate.logging import get_logger from datasets import load_dataset from diffusers import DDPMPipeline, DDPMScheduler, UNet2DModel -from diffusers.hub_utils import init_git_repo, push_to_hub +from diffusers.hub_utils import init_git_repo from diffusers.optimization import get_scheduler from diffusers.training_utils import EMAModel from torchvision.transforms import ( @@ -185,7 +185,7 @@ def transforms(examples): if epoch % args.save_model_epochs == 0 or epoch == args.num_epochs - 1: # save the model if args.push_to_hub: - push_to_hub(args, pipeline, repo, commit_message=f"Epoch {epoch}", blocking=False) + repo.push_to_hub(commit_message=f"Epoch {epoch}", blocking=False) else: pipeline.save_pretrained(args.output_dir) accelerator.wait_for_everyone()