WAFFLE: Multi-Modal Model for Automated Front-End Development
Abstract
Web development involves turning UI designs into functional webpages, which can be difficult for both beginners and experienced developers due to the complexity of HTML's hierarchical structures and styles. While Large Language Models (LLMs) have shown promise in generating source code, two major challenges persist in UI-to-HTML code generation: (1) effectively representing HTML's hierarchical structure for LLMs, and (2) bridging the gap between the visual nature of UI designs and the text-based format of HTML code. To tackle these challenges, we introduce Waffle, a new fine-tuning strategy that uses a structure-aware attention mechanism to improve LLMs' understanding of HTML's structure and a contrastive fine-tuning approach to align LLMs' understanding of UI images and HTML code. Models fine-tuned with Waffle show up to 9.00 pp (percentage point) higher HTML match, 0.0982 higher CW-SSIM, 32.99 higher CLIP, and 27.12 pp higher LLEM on our new benchmark WebSight-Test and an existing benchmark Design2Code, outperforming current fine-tuning methods.
Community
If you are interested in letting LLM help you write HTML/CSS code from UI images, try our Waffle_VLM_WebSight here: https://huggingface.co/lt-asset/Waffle_VLM_WebSight
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Harnessing Webpage UIs for Text-Rich Visual Understanding (2024)
- OMCAT: Omni Context Aware Transformer (2024)
- Inferring Alt-text For UI Icons With Large Language Models During App Development (2024)
- LATTE: Improving Latex Recognition for Tables and Formulae with Iterative Refinement (2024)
- TextLap: Customizing Language Models for Text-to-Layout Planning (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper