Artificial Intelligence 7 min read

AIGC Poster Generation Project: Methods and Optimizations

The AIGC Poster Generation Project employs Stable Diffusion enhanced with VAE, ControlNet, LoRA and other extensions to create product posters in four visual styles, exploring outpainting, inpainting, reference‑based diffusion and DreamBooth prototypes, and optimizes detail preservation, super‑resolution text, and masking to achieve over 90% detail fidelity, 95% success rate, and 3–5 second inference per image.

DaTaobao Tech
DaTaobao Tech
DaTaobao Tech
AIGC Poster Generation Project: Methods and Optimizations

The Inspiration Artist project leverages AIGC drawing capabilities to create a low‑threshold, high‑fun poster design competition for new product promotion.

Background: generate product posters (e.g., Chanel No.5) using large language models (Tongyi Qianwen) for prompts and focus on image generation.

Four visual styles are supported: product poster, Pixar, anime, realistic. The team selected Stable Diffusion (SD) as the core algorithm due to its open‑source nature and extensibility, implementing VAE, ControlNet, LoRA, embeddings, warm‑up and auto‑predict features via the diffusers library.

Four solution prototypes were explored:

SD + Outpainting – fixes background while keeping product fixed.

SD Inpainting + Reference Only – injects product image into UNet attention.

Reference‑based Diffusion (PBE, IP‑Adapter, AnyDoor) – generates similar products from a reference.

SD + LoRA/DreamBooth – fine‑tunes the model to embed product appearance.

Optimization directions include:

Enhancing VAE to preserve fine details.

Applying super‑resolution to improve text clarity.

Copy‑and‑paste of extracted text regions onto generated images.

The final offline‑online pipeline combines a background library, multi‑angle product images, ControlNet‑guided generation, SAM/LAMA masking, and light‑shadow projection to produce high‑quality posters.

Results: detail restoration above 90 %, generation success rate >95 %, and inference speed of 3–5 seconds per image on an A10 GPU.

Future work explores better occlusion handling, proportion harmony (e.g., GLIGEN), and VAE‑free consistency decoders.

Stable Diffusionimage generationAIGCdiffusion modelsControlNetPoster DesignVAE
DaTaobao Tech
Written by

DaTaobao Tech

Official account of DaTaobao Technology

0 followers
Reader feedback

How this landed with the community

login Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.