2410
Browse files- README.md +10 -6
- budget.jpg β media/budget.jpg +0 -0
- promo.png β media/promo.png +0 -0
- result_grid.jpg β media/result_grid.jpg +0 -0
- result_grid.png β media/result_grid.png +0 -0
- pipeline_sdxs.py +6 -6
- samples/unet_320x640_0.jpg +2 -2
- samples/unet_384x640_0.jpg +2 -2
- samples/unet_448x640_0.jpg +2 -2
- samples/unet_512x640_0.jpg +2 -2
- samples/unet_576x640_0.jpg +2 -2
- samples/unet_640x320_0.jpg +2 -2
- samples/unet_640x384_0.jpg +2 -2
- samples/unet_640x448_0.jpg +2 -2
- samples/unet_640x512_0.jpg +2 -2
- samples/unet_640x576_0.jpg +2 -2
- samples/unet_640x640_0.jpg +2 -2
- TRAIN.md β src/TRAIN.md +0 -0
- train.old.py β src/train.old.py +0 -0
- test.ipynb +2 -2
- train.py +1 -1
- unet/diffusion_pytorch_model.safetensors +1 -1
README.md
CHANGED
|
@@ -7,14 +7,18 @@ pipeline_tag: text-to-image
|
|
| 7 |
|
| 8 |
*XS Size, Excess Quality*
|
| 9 |
|
| 10 |
-

|
| 11 |
|
| 12 |
At AiArtLab, we strive to create a free, compact and fast model that can be trained on consumer graphics cards.
|
| 13 |
|
| 14 |
-
- 1.5b
|
| 15 |
-
- Qwen3
|
| 16 |
-
-
|
| 17 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
### Example
|
| 20 |
|
|
@@ -69,4 +73,4 @@ BTC: 3JHv9Hb8kEW8zMAccdgCdZGfrHeMhH1rpN
|
|
| 69 |
|
| 70 |
## More examples
|
| 71 |
|
| 72 |
-

|
| 11 |
|
| 12 |
At AiArtLab, we strive to create a free, compact and fast model that can be trained on consumer graphics cards.
|
| 13 |
|
| 14 |
+
- Model: 1.5b parameters
|
| 15 |
+
- Text encoder: Qwen3 (0.6B parameters)
|
| 16 |
+
- VAE: 8x16ch, [Simple VAE](https://huggingface.co/AiArtLab/simplevae)
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
The model was trained for ~3 months on 4xRTX 5090 GPUs using over 1 million images of various resolutions and styles, primarily anime and illustrations.
|
| 20 |
+
|
| 21 |
+
[Gradio](https://sdxs.aiartlab.org/)
|
| 22 |
|
| 23 |
### Example
|
| 24 |
|
|
|
|
| 73 |
|
| 74 |
## More examples
|
| 75 |
|
| 76 |
+

|
budget.jpg β media/budget.jpg
RENAMED
|
File without changes
|
promo.png β media/promo.png
RENAMED
|
File without changes
|
result_grid.jpg β media/result_grid.jpg
RENAMED
|
File without changes
|
result_grid.png β media/result_grid.png
RENAMED
|
File without changes
|
pipeline_sdxs.py
CHANGED
|
@@ -102,9 +102,9 @@ class SdxsPipeline(DiffusionPipeline):
|
|
| 102 |
def generate_latents(
|
| 103 |
self,
|
| 104 |
text_embeddings,
|
| 105 |
-
height: int =
|
| 106 |
-
width: int =
|
| 107 |
-
num_inference_steps: int =
|
| 108 |
guidance_scale: float = 5.0,
|
| 109 |
latent_channels: int = 16,
|
| 110 |
batch_size: int = 1,
|
|
@@ -192,9 +192,9 @@ class SdxsPipeline(DiffusionPipeline):
|
|
| 192 |
def __call__(
|
| 193 |
self,
|
| 194 |
prompt: Optional[Union[str, List[str]]] = None,
|
| 195 |
-
height: int =
|
| 196 |
-
width: int =
|
| 197 |
-
num_inference_steps: int =
|
| 198 |
guidance_scale: float = 5.0,
|
| 199 |
latent_channels: int = 16,
|
| 200 |
output_type: str = "pil",
|
|
|
|
| 102 |
def generate_latents(
|
| 103 |
self,
|
| 104 |
text_embeddings,
|
| 105 |
+
height: int = 640,
|
| 106 |
+
width: int = 640,
|
| 107 |
+
num_inference_steps: int = 50,
|
| 108 |
guidance_scale: float = 5.0,
|
| 109 |
latent_channels: int = 16,
|
| 110 |
batch_size: int = 1,
|
|
|
|
| 192 |
def __call__(
|
| 193 |
self,
|
| 194 |
prompt: Optional[Union[str, List[str]]] = None,
|
| 195 |
+
height: int = 640,
|
| 196 |
+
width: int = 640,
|
| 197 |
+
num_inference_steps: int = 50,
|
| 198 |
guidance_scale: float = 5.0,
|
| 199 |
latent_channels: int = 16,
|
| 200 |
output_type: str = "pil",
|
samples/unet_320x640_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_384x640_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_448x640_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_512x640_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_576x640_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_640x320_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_640x384_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_640x448_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_640x512_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_640x576_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
samples/unet_640x640_0.jpg
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
TRAIN.md β src/TRAIN.md
RENAMED
|
File without changes
|
train.old.py β src/train.old.py
RENAMED
|
File without changes
|
test.ipynb
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3da8c81f59608e0732fe9708b4cc2becf77ea6fa215601d0f30c057110e85b3c
|
| 3 |
+
size 5563222
|
train.py
CHANGED
|
@@ -31,7 +31,7 @@ project = "unet"
|
|
| 31 |
batch_size = 48
|
| 32 |
base_learning_rate = 5e-5
|
| 33 |
min_learning_rate = 1e-5
|
| 34 |
-
num_epochs =
|
| 35 |
# samples/save per epoch
|
| 36 |
sample_interval_share = 3
|
| 37 |
use_wandb = True
|
|
|
|
| 31 |
batch_size = 48
|
| 32 |
base_learning_rate = 5e-5
|
| 33 |
min_learning_rate = 1e-5
|
| 34 |
+
num_epochs = 40
|
| 35 |
# samples/save per epoch
|
| 36 |
sample_interval_share = 3
|
| 37 |
use_wandb = True
|
unet/diffusion_pytorch_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 6184944280
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0fd75e6dcdd3c39a84c5927b5b77f6cb8591b8feb0a9996aeca9f336b84f9d89
|
| 3 |
size 6184944280
|