InvokeAI 2.3.1rc4__tar.gz → 2.3.2.post1__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (155) hide show
  1. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/InvokeAI.egg-info/PKG-INFO +13 -33
  2. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/InvokeAI.egg-info/SOURCES.txt +7 -3
  3. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/InvokeAI.egg-info/entry_points.txt +1 -0
  4. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/InvokeAI.egg-info/requires.txt +17 -9
  5. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/PKG-INFO +13 -33
  6. InvokeAI-2.3.2.post1/invokeai/configs/INITIAL_MODELS.yaml +88 -0
  7. InvokeAI-2.3.2.post1/invokeai/configs/stable-diffusion/v2-inference.yaml +67 -0
  8. InvokeAI-2.3.1rc4/invokeai/frontend/dist/assets/index-0e39fbc4.js → InvokeAI-2.3.2.post1/invokeai/frontend/dist/assets/index-c09cf9ca.js +46 -46
  9. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/index.html +1 -1
  10. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/en.json +7 -3
  11. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/es.json +102 -13
  12. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/pt_BR.json +70 -6
  13. InvokeAI-2.3.2.post1/invokeai/frontend/dist/locales/ro.json +1 -0
  14. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/generate.py +4 -0
  15. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/CLI.py +92 -31
  16. InvokeAI-2.3.2.post1/ldm/invoke/_version.py +2 -0
  17. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/args.py +5 -4
  18. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/ckpt_to_diffuser.py +23 -12
  19. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/conditioning.py +8 -6
  20. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/config/invokeai_configure.py +13 -8
  21. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/config/invokeai_update.py +1 -1
  22. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/config/model_install.py +45 -37
  23. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/config/model_install_backend.py +41 -2
  24. InvokeAI-2.3.2.post1/ldm/invoke/dynamic_prompts.py +535 -0
  25. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/generator/base.py +3 -11
  26. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/globals.py +6 -9
  27. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/merge_diffusers.py +2 -2
  28. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/model_manager.py +119 -78
  29. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/pngwriter.py +20 -5
  30. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/training/textual_inversion_training.py +1 -1
  31. InvokeAI-2.3.2.post1/pyproject.toml +187 -0
  32. InvokeAI-2.3.2.post1/tests/test_path.py +48 -0
  33. InvokeAI-2.3.2.post1/tests/test_textual_inversion.py +301 -0
  34. InvokeAI-2.3.1rc4/LICENSE-ModelWeights.txt +0 -294
  35. InvokeAI-2.3.1rc4/invokeai/configs/INITIAL_MODELS.yaml +0 -58
  36. InvokeAI-2.3.1rc4/ldm/invoke/_version.py +0 -1
  37. InvokeAI-2.3.1rc4/pyproject.toml +0 -136
  38. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/InvokeAI.egg-info/dependency_links.txt +0 -0
  39. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/InvokeAI.egg-info/top_level.txt +0 -0
  40. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/LICENSE +0 -0
  41. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/README.md +0 -0
  42. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/assets/web/caution.png +0 -0
  43. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/backend/__init__.py +0 -0
  44. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/backend/invoke_ai_web_server.py +0 -0
  45. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/backend/modules/__init__.py +0 -0
  46. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/backend/modules/create_cmd_parser.py +0 -0
  47. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/backend/modules/get_canvas_generation_mode.py +0 -0
  48. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/backend/modules/parameters.py +0 -0
  49. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/backend/modules/parse_seed_weights.py +0 -0
  50. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/configs/models.yaml.example +0 -0
  51. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/configs/stable-diffusion/v1-finetune.yaml +0 -0
  52. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/configs/stable-diffusion/v1-finetune_style.yaml +0 -0
  53. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/configs/stable-diffusion/v1-inference.yaml +0 -0
  54. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/configs/stable-diffusion/v1-inpainting-inference.yaml +0 -0
  55. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/configs/stable-diffusion/v1-m1-finetune.yaml +0 -0
  56. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/configs/stable-diffusion/v2-inference-v.yaml +0 -0
  57. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/assets/Inter-Bold-790c108b.ttf +0 -0
  58. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/assets/Inter-b9a8e5e2.ttf +0 -0
  59. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/assets/favicon-0d253ced.ico +0 -0
  60. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/assets/index-14cb2922.css +0 -0
  61. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/assets/logo-13003d72.png +0 -0
  62. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/ar.json +0 -0
  63. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/de.json +0 -0
  64. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/fr.json +0 -0
  65. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/it.json +0 -0
  66. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/ja.json +0 -0
  67. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/nl.json +0 -0
  68. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/pl.json +0 -0
  69. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/ru.json +0 -0
  70. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/uk.json +0 -0
  71. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/invokeai/frontend/dist/locales/zh_CN.json +0 -0
  72. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/__init__.py +0 -0
  73. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/data/__init__.py +0 -0
  74. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/data/base.py +0 -0
  75. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/data/imagenet.py +0 -0
  76. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/data/lsun.py +0 -0
  77. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/data/personalized.py +0 -0
  78. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/data/personalized_style.py +0 -0
  79. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/__init__.py +0 -0
  80. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/ckpt_generator/__init__.py +0 -0
  81. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/ckpt_generator/base.py +0 -0
  82. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/ckpt_generator/embiggen.py +0 -0
  83. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/ckpt_generator/img2img.py +0 -0
  84. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/ckpt_generator/inpaint.py +0 -0
  85. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/ckpt_generator/omnibus.py +0 -0
  86. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/ckpt_generator/txt2img.py +0 -0
  87. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/ckpt_generator/txt2img2img.py +0 -0
  88. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/concepts_lib.py +0 -0
  89. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/config/__init__.py +0 -0
  90. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/config/widgets.py +0 -0
  91. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/devices.py +0 -0
  92. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/generator/__init__.py +0 -0
  93. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/generator/diffusers_pipeline.py +0 -0
  94. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/generator/embiggen.py +0 -0
  95. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/generator/img2img.py +0 -0
  96. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/generator/inpaint.py +0 -0
  97. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/generator/omnibus.py +0 -0
  98. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/generator/txt2img.py +0 -0
  99. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/generator/txt2img2img.py +0 -0
  100. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/image_util.py +0 -0
  101. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/log.py +0 -0
  102. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/offloading.py +0 -0
  103. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/patchmatch.py +0 -0
  104. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/readline.py +0 -0
  105. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/restoration/__init__.py +0 -0
  106. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/restoration/base.py +0 -0
  107. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/restoration/codeformer.py +0 -0
  108. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/restoration/codeformer_arch.py +0 -0
  109. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/restoration/gfpgan.py +0 -0
  110. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/restoration/outcrop.py +0 -0
  111. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/restoration/outpaint.py +0 -0
  112. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/restoration/realesrgan.py +0 -0
  113. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/restoration/vqgan_arch.py +0 -0
  114. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/seamless.py +0 -0
  115. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/server.py +0 -0
  116. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/server_legacy.py +0 -0
  117. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/training/textual_inversion.py +0 -0
  118. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/invoke/txt2mask.py +0 -0
  119. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/lr_scheduler.py +0 -0
  120. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/__init__.py +0 -0
  121. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/autoencoder.py +0 -0
  122. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/__init__.py +0 -0
  123. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/classifier.py +0 -0
  124. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/cross_attention_control.py +0 -0
  125. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/cross_attention_map_saving.py +0 -0
  126. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/ddim.py +0 -0
  127. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/ddpm.py +0 -0
  128. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/ksampler.py +0 -0
  129. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/plms.py +0 -0
  130. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/sampler.py +0 -0
  131. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/models/diffusion/shared_invokeai_diffusion.py +0 -0
  132. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/__init__.py +0 -0
  133. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/attention.py +0 -0
  134. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/diffusionmodules/__init__.py +0 -0
  135. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/diffusionmodules/model.py +0 -0
  136. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/diffusionmodules/openaimodel.py +0 -0
  137. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/diffusionmodules/util.py +0 -0
  138. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/distributions/__init__.py +0 -0
  139. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/distributions/distributions.py +0 -0
  140. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/ema.py +0 -0
  141. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/embedding_manager.py +0 -0
  142. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/encoders/__init__.py +0 -0
  143. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/encoders/modules.py +0 -0
  144. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/image_degradation/__init__.py +0 -0
  145. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/image_degradation/bsrgan.py +0 -0
  146. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/image_degradation/bsrgan_light.py +0 -0
  147. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/image_degradation/utils_image.py +0 -0
  148. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/losses/__init__.py +0 -0
  149. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/losses/contperceptual.py +0 -0
  150. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/losses/vqperceptual.py +0 -0
  151. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/textual_inversion_manager.py +0 -0
  152. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/modules/x_transformer.py +0 -0
  153. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/simplet2i.py +0 -0
  154. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/ldm/util.py +0 -0
  155. {InvokeAI-2.3.1rc4 → InvokeAI-2.3.2.post1}/setup.cfg +0 -0
@@ -1,50 +1,30 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: InvokeAI
3
- Version: 2.3.1rc4
3
+ Version: 2.3.2.post1
4
4
  Summary: An implementation of Stable Diffusion which provides various new features and options to aid the image generation process
5
5
  Author-email: The InvokeAI Project <lincoln.stein@gmail.com>
6
- License: MIT License
7
-
8
- Copyright (c) 2022 InvokeAI Team
9
-
10
- Permission is hereby granted, free of charge, to any person obtaining a copy
11
- of this software and associated documentation files (the "Software"), to deal
12
- in the Software without restriction, including without limitation the rights
13
- to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
14
- copies of the Software, and to permit persons to whom the Software is
15
- furnished to do so, subject to the following conditions:
16
-
17
- The above copyright notice and this permission notice shall be included in all
18
- copies or substantial portions of the Software.
19
-
20
- THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
21
- IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
22
- FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
23
- AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
24
- LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
25
- OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
26
- SOFTWARE.
27
-
28
- Project-URL: Homepage, https://invoke-ai.github.io/InvokeAI/
29
- Project-URL: Documentation, https://invoke-ai.github.io/InvokeAI/
30
- Project-URL: Source, https://github.com/invoke-ai/InvokeAI/
6
+ License: MIT
31
7
  Project-URL: Bug Reports, https://github.com/invoke-ai/InvokeAI/issues
32
8
  Project-URL: Discord, https://discord.gg/ZmtBAhwWhy
33
- Keywords: stable-diffusion,AI
9
+ Project-URL: Documentation, https://invoke-ai.github.io/InvokeAI/
10
+ Project-URL: Homepage, https://invoke-ai.github.io/InvokeAI/
11
+ Project-URL: Source, https://github.com/invoke-ai/InvokeAI/
12
+ Keywords: AI,stable-diffusion
34
13
  Classifier: Development Status :: 4 - Beta
35
- Classifier: Environment :: GPU
36
14
  Classifier: Environment :: GPU :: NVIDIA CUDA
15
+ Classifier: Environment :: GPU
37
16
  Classifier: Environment :: MacOS X
38
- Classifier: Intended Audience :: End Users/Desktop
39
17
  Classifier: Intended Audience :: Developers
18
+ Classifier: Intended Audience :: End Users/Desktop
40
19
  Classifier: License :: OSI Approved :: MIT License
41
- Classifier: Operating System :: POSIX :: Linux
42
20
  Classifier: Operating System :: MacOS
43
21
  Classifier: Operating System :: Microsoft :: Windows
22
+ Classifier: Operating System :: POSIX :: Linux
44
23
  Classifier: Programming Language :: Python :: 3 :: Only
45
- Classifier: Programming Language :: Python :: 3.8
46
- Classifier: Programming Language :: Python :: 3.9
24
+ Classifier: Programming Language :: Python :: 3
47
25
  Classifier: Programming Language :: Python :: 3.10
26
+ Classifier: Programming Language :: Python :: 3.9
27
+ Classifier: Programming Language :: Python
48
28
  Classifier: Topic :: Artistic Software
49
29
  Classifier: Topic :: Internet :: WWW/HTTP :: WSGI :: Application
50
30
  Classifier: Topic :: Internet :: WWW/HTTP :: WSGI :: Server
@@ -53,12 +33,12 @@ Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
53
33
  Classifier: Topic :: Scientific/Engineering :: Image Processing
54
34
  Requires-Python: <3.11,>=3.9
55
35
  Description-Content-Type: text/markdown
36
+ Provides-Extra: dev
56
37
  Provides-Extra: dist
57
38
  Provides-Extra: docs
58
39
  Provides-Extra: test
59
40
  Provides-Extra: xformers
60
41
  License-File: LICENSE
61
- License-File: LICENSE-ModelWeights.txt
62
42
 
63
43
  <div align="center">
64
44
 
@@ -1,5 +1,4 @@
1
1
  LICENSE
2
- LICENSE-ModelWeights.txt
3
2
  README.md
4
3
  pyproject.toml
5
4
  InvokeAI.egg-info/PKG-INFO
@@ -24,12 +23,13 @@ invokeai/configs/stable-diffusion/v1-inference.yaml
24
23
  invokeai/configs/stable-diffusion/v1-inpainting-inference.yaml
25
24
  invokeai/configs/stable-diffusion/v1-m1-finetune.yaml
26
25
  invokeai/configs/stable-diffusion/v2-inference-v.yaml
26
+ invokeai/configs/stable-diffusion/v2-inference.yaml
27
27
  invokeai/frontend/dist/index.html
28
28
  invokeai/frontend/dist/assets/Inter-Bold-790c108b.ttf
29
29
  invokeai/frontend/dist/assets/Inter-b9a8e5e2.ttf
30
30
  invokeai/frontend/dist/assets/favicon-0d253ced.ico
31
- invokeai/frontend/dist/assets/index-0e39fbc4.js
32
31
  invokeai/frontend/dist/assets/index-14cb2922.css
32
+ invokeai/frontend/dist/assets/index-c09cf9ca.js
33
33
  invokeai/frontend/dist/assets/logo-13003d72.png
34
34
  invokeai/frontend/dist/locales/ar.json
35
35
  invokeai/frontend/dist/locales/de.json
@@ -41,6 +41,7 @@ invokeai/frontend/dist/locales/ja.json
41
41
  invokeai/frontend/dist/locales/nl.json
42
42
  invokeai/frontend/dist/locales/pl.json
43
43
  invokeai/frontend/dist/locales/pt_BR.json
44
+ invokeai/frontend/dist/locales/ro.json
44
45
  invokeai/frontend/dist/locales/ru.json
45
46
  invokeai/frontend/dist/locales/uk.json
46
47
  invokeai/frontend/dist/locales/zh_CN.json
@@ -63,6 +64,7 @@ ldm/invoke/ckpt_to_diffuser.py
63
64
  ldm/invoke/concepts_lib.py
64
65
  ldm/invoke/conditioning.py
65
66
  ldm/invoke/devices.py
67
+ ldm/invoke/dynamic_prompts.py
66
68
  ldm/invoke/globals.py
67
69
  ldm/invoke/image_util.py
68
70
  ldm/invoke/log.py
@@ -142,4 +144,6 @@ ldm/modules/image_degradation/bsrgan_light.py
142
144
  ldm/modules/image_degradation/utils_image.py
143
145
  ldm/modules/losses/__init__.py
144
146
  ldm/modules/losses/contperceptual.py
145
- ldm/modules/losses/vqperceptual.py
147
+ ldm/modules/losses/vqperceptual.py
148
+ tests/test_path.py
149
+ tests/test_textual_inversion.py
@@ -2,6 +2,7 @@
2
2
  configure_invokeai.py = ldm.invoke.config.invokeai_configure:main
3
3
  invoke.py = ldm.invoke.CLI:main
4
4
  invokeai = ldm.invoke.CLI:main
5
+ invokeai-batch = ldm.invoke.dynamic_prompts:main
5
6
  invokeai-configure = ldm.invoke.config.invokeai_configure:main
6
7
  invokeai-merge = ldm.invoke.merge_diffusers:main
7
8
  invokeai-model-install = ldm.invoke.config.model_install:main
@@ -1,10 +1,10 @@
1
- accelerate
1
+ accelerate~=0.16
2
2
  albumentations
3
3
  click
4
4
  clip_anytorch
5
5
  compel==0.1.7
6
6
  datasets
7
- diffusers[torch]~=0.13
7
+ diffusers[torch]~=0.14
8
8
  dnspython==2.2.1
9
9
  einops
10
10
  eventlet
@@ -26,41 +26,49 @@ omegaconf
26
26
  opencv-python
27
27
  picklescan
28
28
  pillow
29
- pudb
30
29
  prompt-toolkit
30
+ pudb
31
31
  pypatchmatch
32
32
  pyreadline3
33
33
  pytorch-lightning==1.7.7
34
34
  realesrgan
35
35
  requests==2.28.2
36
- safetensors
36
+ safetensors~=0.3.0
37
37
  scikit-image>=0.19
38
38
  send2trash
39
39
  streamlit
40
40
  taming-transformers-rom1504
41
41
  test-tube>=0.7.5
42
- torch>=1.13.1
43
42
  torch-fidelity
44
- torchvision>=0.14.1
43
+ torch>=1.13.1
45
44
  torchmetrics
46
- transformers~=4.25
45
+ torchvision>=0.14.1
46
+ transformers~=4.26
47
47
 
48
48
  [:sys_platform == "win32"]
49
49
  windows-curses
50
50
 
51
+ [dev]
52
+ black[jupyter]
53
+ flake8
54
+ flake8-black
55
+ flake8-bugbear
56
+ isort
57
+ pre-commit
58
+
51
59
  [dist]
52
60
  pip-tools
53
61
  pipdeptree
54
62
  twine
55
63
 
56
64
  [docs]
57
- mkdocs-material<9.0
58
65
  mkdocs-git-revision-date-localized-plugin
66
+ mkdocs-material==9.*
59
67
  mkdocs-redirects==1.2.0
60
68
 
61
69
  [test]
62
- pytest>6.0.0
63
70
  pytest-cov
71
+ pytest>6.0.0
64
72
 
65
73
  [xformers]
66
74
 
@@ -1,50 +1,30 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: InvokeAI
3
- Version: 2.3.1rc4
3
+ Version: 2.3.2.post1
4
4
  Summary: An implementation of Stable Diffusion which provides various new features and options to aid the image generation process
5
5
  Author-email: The InvokeAI Project <lincoln.stein@gmail.com>
6
- License: MIT License
7
-
8
- Copyright (c) 2022 InvokeAI Team
9
-
10
- Permission is hereby granted, free of charge, to any person obtaining a copy
11
- of this software and associated documentation files (the "Software"), to deal
12
- in the Software without restriction, including without limitation the rights
13
- to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
14
- copies of the Software, and to permit persons to whom the Software is
15
- furnished to do so, subject to the following conditions:
16
-
17
- The above copyright notice and this permission notice shall be included in all
18
- copies or substantial portions of the Software.
19
-
20
- THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
21
- IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
22
- FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
23
- AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
24
- LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
25
- OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
26
- SOFTWARE.
27
-
28
- Project-URL: Homepage, https://invoke-ai.github.io/InvokeAI/
29
- Project-URL: Documentation, https://invoke-ai.github.io/InvokeAI/
30
- Project-URL: Source, https://github.com/invoke-ai/InvokeAI/
6
+ License: MIT
31
7
  Project-URL: Bug Reports, https://github.com/invoke-ai/InvokeAI/issues
32
8
  Project-URL: Discord, https://discord.gg/ZmtBAhwWhy
33
- Keywords: stable-diffusion,AI
9
+ Project-URL: Documentation, https://invoke-ai.github.io/InvokeAI/
10
+ Project-URL: Homepage, https://invoke-ai.github.io/InvokeAI/
11
+ Project-URL: Source, https://github.com/invoke-ai/InvokeAI/
12
+ Keywords: AI,stable-diffusion
34
13
  Classifier: Development Status :: 4 - Beta
35
- Classifier: Environment :: GPU
36
14
  Classifier: Environment :: GPU :: NVIDIA CUDA
15
+ Classifier: Environment :: GPU
37
16
  Classifier: Environment :: MacOS X
38
- Classifier: Intended Audience :: End Users/Desktop
39
17
  Classifier: Intended Audience :: Developers
18
+ Classifier: Intended Audience :: End Users/Desktop
40
19
  Classifier: License :: OSI Approved :: MIT License
41
- Classifier: Operating System :: POSIX :: Linux
42
20
  Classifier: Operating System :: MacOS
43
21
  Classifier: Operating System :: Microsoft :: Windows
22
+ Classifier: Operating System :: POSIX :: Linux
44
23
  Classifier: Programming Language :: Python :: 3 :: Only
45
- Classifier: Programming Language :: Python :: 3.8
46
- Classifier: Programming Language :: Python :: 3.9
24
+ Classifier: Programming Language :: Python :: 3
47
25
  Classifier: Programming Language :: Python :: 3.10
26
+ Classifier: Programming Language :: Python :: 3.9
27
+ Classifier: Programming Language :: Python
48
28
  Classifier: Topic :: Artistic Software
49
29
  Classifier: Topic :: Internet :: WWW/HTTP :: WSGI :: Application
50
30
  Classifier: Topic :: Internet :: WWW/HTTP :: WSGI :: Server
@@ -53,12 +33,12 @@ Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
53
33
  Classifier: Topic :: Scientific/Engineering :: Image Processing
54
34
  Requires-Python: <3.11,>=3.9
55
35
  Description-Content-Type: text/markdown
36
+ Provides-Extra: dev
56
37
  Provides-Extra: dist
57
38
  Provides-Extra: docs
58
39
  Provides-Extra: test
59
40
  Provides-Extra: xformers
60
41
  License-File: LICENSE
61
- License-File: LICENSE-ModelWeights.txt
62
42
 
63
43
  <div align="center">
64
44
 
@@ -0,0 +1,88 @@
1
+ stable-diffusion-1.5:
2
+ description: Stable Diffusion version 1.5 diffusers model (4.27 GB)
3
+ repo_id: runwayml/stable-diffusion-v1-5
4
+ format: diffusers
5
+ vae:
6
+ repo_id: stabilityai/sd-vae-ft-mse
7
+ recommended: True
8
+ default: True
9
+ sd-inpainting-1.5:
10
+ description: RunwayML SD 1.5 model optimized for inpainting, diffusers version (4.27 GB)
11
+ repo_id: runwayml/stable-diffusion-inpainting
12
+ format: diffusers
13
+ vae:
14
+ repo_id: stabilityai/sd-vae-ft-mse
15
+ recommended: True
16
+ stable-diffusion-2.1-768:
17
+ description: Stable Diffusion version 2.1 diffusers model, trained on 768 pixel images (5.21 GB)
18
+ repo_id: stabilityai/stable-diffusion-2-1
19
+ format: diffusers
20
+ recommended: True
21
+ stable-diffusion-2.1-base:
22
+ description: Stable Diffusion version 2.1 diffusers model, trained on 512 pixel images (5.21 GB)
23
+ repo_id: stabilityai/stable-diffusion-2-1-base
24
+ format: diffusers
25
+ recommended: False
26
+ sd-inpainting-2.0:
27
+ description: Stable Diffusion version 2.0 inpainting model (5.21 GB)
28
+ repo_id: stabilityai/stable-diffusion-2-inpainting
29
+ format: diffusers
30
+ recommended: False
31
+ analog-diffusion-1.0:
32
+ description: An SD-1.5 model trained on diverse analog photographs (2.13 GB)
33
+ repo_id: wavymulder/Analog-Diffusion
34
+ format: diffusers
35
+ recommended: false
36
+ deliberate-1.0:
37
+ description: Versatile model that produces detailed images up to 768px (4.27 GB)
38
+ format: diffusers
39
+ repo_id: XpucT/Deliberate
40
+ recommended: False
41
+ d&d-diffusion-1.0:
42
+ description: Dungeons & Dragons characters (2.13 GB)
43
+ format: diffusers
44
+ repo_id: 0xJustin/Dungeons-and-Diffusion
45
+ recommended: False
46
+ dreamlike-photoreal-2.0:
47
+ description: A photorealistic model trained on 768 pixel images based on SD 1.5 (2.13 GB)
48
+ format: diffusers
49
+ repo_id: dreamlike-art/dreamlike-photoreal-2.0
50
+ recommended: False
51
+ inkpunk-1.0:
52
+ description: Stylized illustrations inspired by Gorillaz, FLCL and Shinkawa; prompt with "nvinkpunk" (4.27 GB)
53
+ format: diffusers
54
+ repo_id: Envvi/Inkpunk-Diffusion
55
+ recommended: False
56
+ openjourney-4.0:
57
+ description: An SD 1.5 model fine tuned on Midjourney; prompt with "mdjrny-v4 style" (2.13 GB)
58
+ format: diffusers
59
+ repo_id: prompthero/openjourney
60
+ vae:
61
+ repo_id: stabilityai/sd-vae-ft-mse
62
+ recommended: False
63
+ portrait-plus-1.0:
64
+ description: An SD-1.5 model trained on close range portraits of people; prompt with "portrait+" (2.13 GB)
65
+ format: diffusers
66
+ repo_id: wavymulder/portraitplus
67
+ recommended: False
68
+ seek-art-mega-1.0:
69
+ description: A general use SD-1.5 "anything" model that supports multiple styles (2.1 GB)
70
+ repo_id: coreco/seek.art_MEGA
71
+ format: diffusers
72
+ vae:
73
+ repo_id: stabilityai/sd-vae-ft-mse
74
+ recommended: False
75
+ trinart-2.0:
76
+ description: An SD-1.5 model finetuned with ~40K assorted high resolution manga/anime-style images (2.13 GB)
77
+ repo_id: naclbit/trinart_stable_diffusion_v2
78
+ format: diffusers
79
+ vae:
80
+ repo_id: stabilityai/sd-vae-ft-mse
81
+ recommended: False
82
+ waifu-diffusion-1.4:
83
+ description: An SD-1.5 model trained on 680k anime/manga-style images (2.13 GB)
84
+ repo_id: hakurei/waifu-diffusion
85
+ format: diffusers
86
+ vae:
87
+ repo_id: stabilityai/sd-vae-ft-mse
88
+ recommended: False
@@ -0,0 +1,67 @@
1
+ model:
2
+ base_learning_rate: 1.0e-4
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ linear_start: 0.00085
6
+ linear_end: 0.0120
7
+ num_timesteps_cond: 1
8
+ log_every_t: 200
9
+ timesteps: 1000
10
+ first_stage_key: "jpg"
11
+ cond_stage_key: "txt"
12
+ image_size: 64
13
+ channels: 4
14
+ cond_stage_trainable: false
15
+ conditioning_key: crossattn
16
+ monitor: val/loss_simple_ema
17
+ scale_factor: 0.18215
18
+ use_ema: False # we set this to false because this is an inference only config
19
+
20
+ unet_config:
21
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
22
+ params:
23
+ use_checkpoint: True
24
+ use_fp16: True
25
+ image_size: 32 # unused
26
+ in_channels: 4
27
+ out_channels: 4
28
+ model_channels: 320
29
+ attention_resolutions: [ 4, 2, 1 ]
30
+ num_res_blocks: 2
31
+ channel_mult: [ 1, 2, 4, 4 ]
32
+ num_head_channels: 64 # need to fix for flash-attn
33
+ use_spatial_transformer: True
34
+ use_linear_in_transformer: True
35
+ transformer_depth: 1
36
+ context_dim: 1024
37
+ legacy: False
38
+
39
+ first_stage_config:
40
+ target: ldm.models.autoencoder.AutoencoderKL
41
+ params:
42
+ embed_dim: 4
43
+ monitor: val/rec_loss
44
+ ddconfig:
45
+ #attn_type: "vanilla-xformers"
46
+ double_z: true
47
+ z_channels: 4
48
+ resolution: 256
49
+ in_channels: 3
50
+ out_ch: 3
51
+ ch: 128
52
+ ch_mult:
53
+ - 1
54
+ - 2
55
+ - 4
56
+ - 4
57
+ num_res_blocks: 2
58
+ attn_resolutions: []
59
+ dropout: 0.0
60
+ lossconfig:
61
+ target: torch.nn.Identity
62
+
63
+ cond_stage_config:
64
+ target: ldm.modules.encoders.modules.FrozenOpenCLIPEmbedder
65
+ params:
66
+ freeze: True
67
+ layer: "penultimate"