itsadarshms commited on
Commit
a8fb76a
Β·
1 Parent(s): 3eb7e84

Update README.md

Browse files
Files changed (2) hide show
  1. .gitignore +2 -0
  2. README.md +137 -2
.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ assets/
2
+ .tests/
README.md CHANGED
@@ -1,3 +1,138 @@
1
  ---
2
- {}
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: openrail++
3
+ tags:
4
+ - text-to-image
5
+ - stable-diffusion
6
+ - logos
7
+ ---
8
+
9
+ ---
10
+
11
+ <div align="center"><h1 align="center">~ Chhavi ~</h1><img src="./cool.png" alt="Vector anime chibi style logo featuring a single piece of flaming piece of popcorn with a smiling face, with mirrorshades sunglasses, popcorn as morpheus, clean composition, symmetrical" width=150 style="border-radius: 150px;"></div>
12
+
13
+
14
+ <p align="center"><i>Democratizing access to generative-ai-models for the open-source community.<br>Let's advance AI, together. </i></p>
15
+
16
+ ---
17
+
18
+ ## Capability ~~Showcase~~ Show-off 😎
19
+
20
+ ![results](./demo.png)
21
+
22
+ ---
23
+
24
+ <h2> Introduction πŸŽ‰ </h2>
25
+
26
+ Welcome to **Chhavi**, a Latent Diffusion Model (LDM) fine-tuned on the foundation of StablitiyAI's open-source SDXL model. At Bud Ecosystem, we believe in the power of open-source collaboration to drive the advancement of technology at an accelerated pace. Our vision is to democratize access to fine-tuned generative-ai-models (LLMs, LDMs etc...), and to that end, we will be releasing a series of models for the open-source community to use, enhance, and build upon.
27
+
28
+ Chhavi is finetuned on a subset of our internal dataset and is first in the release cycle of our finetune series of Diffusion Models. The model still has room for improvements and is in active development.
29
+
30
+ And this isn't the end. It's just the beginning of a journey towards creating more advanced, more efficient, and more accessible diffusion models. We invite you to join us on this exciting journey. πŸš€
31
+
32
+ ---
33
+
34
+ <h2>Getting Started on Hugging Face πŸ€—</h2>
35
+
36
+ Getting up and running with our models on Hugging Face is a breeze. Follow these steps:
37
+
38
+ <h3>1️⃣ : Install dependencies</h3>
39
+
40
+ Make sure to upgrade diffusers to >= 0.19.0:
41
+ ```
42
+ pip install diffusers --upgrade
43
+ ```
44
+
45
+ In addition make sure to install `transformers`, `safetensors`, `accelerate` as well as the invisible watermark:
46
+ ```
47
+ pip install invisible_watermark transformers accelerate safetensors
48
+ ```
49
+
50
+ <h3> 2️⃣ : Import necessary modules</h3>
51
+
52
+ Start by importing the necessary modules from the 'diffusers' library and 'torch'.
53
+
54
+ ```py
55
+ from diffusers import DiffusionPipeline
56
+ import torch
57
+
58
+ pipe = DiffusionPipeline.from_pretrained("budecosystem/Chhavi", torch_dtype=torch.float16, use_safetensors=True)
59
+ pipe.to("cuda")
60
+
61
+ # if using torch < 2.0
62
+ # pipe.enable_xformers_memory_efficient_attention()
63
+
64
+ prompt = "Vector anime chibi style logo featuring a single piece of flaming piece of popcorn with a smiling face, with mirrorshades sunglasses, popcorn as morpheus, clean composition, symmetrical"
65
+
66
+ negative_prompt = "tiling, out of frame, watermark, signature, cut off, low contrast, underexposed, overexposed, bad art, beginner, amateur"
67
+
68
+ images = pipe(prompt=prompt, negative_prompt=negative_prompt).images[0]
69
+ ```
70
+
71
+ When using `torch >= 2.0`, you can improve the inference speed by 20-30% with torch.compile. Simple wrap the unet with torch compile before running the pipeline:
72
+ ```py
73
+ pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True)
74
+ ```
75
+
76
+ If you are limited by GPU VRAM, you can enable *cpu offloading* by calling `pipe.enable_model_cpu_offload`
77
+ instead of `.to("cuda")`:
78
+
79
+ ```diff
80
+ - pipe.to("cuda")
81
+ + pipe.enable_model_cpu_offload()
82
+ ```
83
+
84
+ <h3> 3️⃣ : Promp Guidelines</h3>
85
+
86
+
87
+ Chhavi doesn't follow any specific prompt style or template but we've observed the use of the keyword "logo" in the prompt yield better results. It's worth noting that the use of negative prompt will help in reducing any unwanted artifacts in the generation, the negative prompt mentioned in the above code snippet should serve as a good kick-start.
88
+
89
+ ---
90
+
91
+ <h2>Model Card for Chhavi πŸ“„</h2>
92
+
93
+ Here's a quick overview of everything you need to know about Chhavi.
94
+
95
+ <h3>Model Details:</h3>
96
+
97
+
98
+ - **Developed by**: Bud Ecosystem
99
+ - **Base pretrained model type**: SDXL Base 1.0
100
+ - **Model Description:** This is a model that can be used to generate and modify images (specifically finetuned for logos) based on text prompts. It is a [Latent Diffusion Model](https://arxiv.org/abs/2112.10752) that uses two fixed, pretrained text encoders ([OpenCLIP-ViT/G](https://github.com/mlfoundations/open_clip) and [CLIP-ViT/L](https://github.com/openai/CLIP/tree/main)).
101
+ - **License**: [CreativeML Open RAIL++-M License](https://huggingface.co/budecosystem/stable-diffusion-xl-logos/blob/main/LICENSE.md)
102
+
103
+ ---
104
+
105
+ <h2>Intended Use πŸ’Ό</h2>
106
+
107
+ When we created Chhavi, we had a clear vision of how it could be used to push the boundaries of what's possible with diffusion models. We also understand the importance of using such models responsibly. Here's a brief overview of the intended and out-of-scope uses for Chhavi.
108
+
109
+ <h3>Direct Use</h3>
110
+
111
+ Chhavi is designed to generate logo images, which can be utilized for the following:
112
+
113
+ - Generation of artworks and use in design and other artistic processes.
114
+ - Applications in educational or creative tools.
115
+
116
+
117
+ Excluded uses are described below.
118
+
119
+ <h3>Out-of-Scope Use 🚩</h3>
120
+
121
+ The model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.
122
+
123
+ <h3>Recommendations 🧠</h3>
124
+
125
+ We recommend users of Chhavi to take appropriate precautions and guardrails for any production use. Using Chhavi responsibly is key to unlocking its full potential while maintaining a safe and respectful environment.
126
+
127
+ ---
128
+
129
+ <h2>Looking Ahead πŸ‘€</h2>
130
+
131
+ We're excited about the journey ahead with Chhavi. We're committed to continuously improving and enhancing our models, and we're excited to see what the open-source community will build with them. We believe in the power of collaboration, and we can't wait to see what we can achieve together.
132
+
133
+ Remember, we're just getting started. This is just the beginning of a journey that we believe will revolutionize the world of Generative AI models. We invite you to join us on this exciting journey. Together, we can push the boundaries of what's possible with AI. πŸš€
134
+
135
+ ---
136
+
137
+
138
+ Check the GitHub for the code -> [Chhavi](https://github.com/BudEcosystem/Chhavi)