UI-Diffuser-V2

UI-Diffuser-V2 is fine tuned from "stabilityai/stable-diffusion-2-base" with the SCapRepo dataset for mobile UI generation.

A demo using diffusion model and large language model for UI generation is available at https://github.com/Jl-wei/ai-gen-ui

Using with Diffusers

import torch
from diffusers import StableDiffusionPipeline, EulerDiscreteScheduler

model_id = "stabilityai/stable-diffusion-2-base"
scheduler = EulerDiscreteScheduler.from_pretrained(model_id, subfolder="scheduler")
pipe = StableDiffusionPipeline.from_pretrained(model_id, scheduler=scheduler, torch_dtype=torch.float16)

lora_path = "Jl-wei/ui-diffuser-v2"
pipe.load_lora_weights(lora_path)
pipe.to("cuda")

prompt = "Mobile app: health monitoring report"
images = pipe(prompt, num_inference_steps=30, guidance_scale=7.5, height=512, width=288, num_images_per_prompt=10).images

columns = 5
fig = plt.figure(figsize=(20,10))
for i, image in enumerate(images):
    plt.subplot(int(len(images) / columns), columns, i + 1)
    plt.imshow(image)
for ax in fig.axes:
    ax.axis("off")

Citation

If you find our work useful, please cite our paper:

@misc{wei2024aiinspired,
      title={On AI-Inspired UI-Design}, 
      author={Jialiang Wei and Anne-Lise Courbis and Thomas Lambolais and Gérard Dray and Walid Maalej},
      year={2024},
      eprint={2406.13631},
      archivePrefix={arXiv}
}

Please note that the code and model can only be used for academic purpose.

UI-Diffuser-V1

This model, UI-Diffuser-V2, represents the second version of the UI-Diffuser model.

The initial version, UI-Diffuser-V1, was introduced in our paper titled Boosting GUI Prototyping with Diffusion Models

Downloads last month
132
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for Jl-wei/ui-diffuser-v2

Adapter
(5)
this model

Space using Jl-wei/ui-diffuser-v2 1