|
--- |
|
tags: |
|
- text-to-image |
|
- stable-diffusion |
|
- lora |
|
- diffusers |
|
- template:sd-lora |
|
widget: |
|
- text: '-' |
|
output: |
|
url: images/card2.jpg |
|
base_model: runwayml/stable-diffusion-v1-5 |
|
instance_prompt: dreambooth, text to image, bagan |
|
license: mit |
|
--- |
|
# bagan-text-to-image |
|
### Text-To-Image (Bagan Ai Generated) |
|
|
|
### Results |
|
|
|
We display the results using a range of training samples and images from different image categories, such as pagodas and Buddha statues. |
|
|
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6598b82502c4796342239a35/TpLTtrQBFLFQmbIvzdF5V.png) |
|
|
|
### Ai Generated Bagan Images: |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6598b82502c4796342239a35/MwR8pZ8xd6IXrNrvNL5ru.png) |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6598b82502c4796342239a35/w-7_MOhc0dMt6uEcdPoay.png) |
|
|
|
### Problem Statement: |
|
When we prompted the stable diffusion model to generate an image of Bagan, it produced an image depicting a pagoda from Thailand. Hence, our decision was to fine-tune the current stable diffusion model using a multitude of Bagan photos in order to attain a clearer outcome. |
|
|
|
### How to use: |
|
prompt = "fantasy bagan,hypper detailed , peaceful mood ,The central theme could revolve around a fantastical journey through a magical realm, featuring characters with ethereal and surreal qualities, set against a backdrop of vibrant and enchanting landscapes, The color palette would be a harmonious combination of Jean's bold and surreal hues, by yukisakura sunset." |
|
|
|
negative_prompt = "" |
|
|
|
num_samples = 5 |
|
|
|
guidance_scale = 9 |
|
|
|
num_inference_steps = 100 |
|
|
|
height = 512 |
|
|
|
width = 512 |
|
|
|
with autocast("cuda"), torch.inference_mode(): |
|
images = pipe( |
|
prompt, |
|
height=height, |
|
width=width, |
|
negative_prompt=negative_prompt, |
|
num_images_per_prompt=num_samples, |
|
num_inference_steps=num_inference_steps, |
|
guidance_scale=guidance_scale, |
|
generator=g_cuda |
|
).images |
|
|
|
for img in images: |
|
display(img) |
|
|
|
### Contributors: |
|
Main Contributor: [Ye Bhone Lin](https://github.com/Ye-Bhone-Lin) |
|
|
|
Supervisor: Sa Phyo Thu Htet |
|
|
|
Contributors: Thant Htoo San, Min Phone Thit |
|
|
|
### Limitation: |
|
We can't generate a photo of a human. |
|
|
|
### References: |
|
Wikipedia (2022). Stable Diffusion. Retrieved From: https://en.wikipedia.org/wiki/Stable_Diffusion |
|
|
|
Rombach, R., Blattmann, A., Lorenz, D., Esser, P., & Ommer, B. (2022). High-Resolution Image Synthesis with Latent Diffusion Models. Retrieved From: https://arxiv.org/abs/2112.10752 |
|
|
|
Naomi Brown (2022). What is Stable Diffusion and How to Use it. Retrieved From: https://www.fotor.com/blog/what-is-stable-diffusion |
|
|
|
Mishra, O. (June, 9). Stable Diffusion Explained. Medium. https://medium.com/@onkarmishra/stable-diffusion-explained-1f101284484d |
|
|
|
|
|
|