Portable Agent Skills that raise the bar for AI-built interfaces: premium layout, typography, motion, and spacing instead of generic templates. The repo also includes image-only skills for strong reference frames (web, mobile apps, brand kits). They pair well with ChatGPT Images and other high-quality image models: generate a screen or board, then hand it to Codex, Cursor, or Claude Code to ship real code.
Taste Skill has no official token, coin, or crypto project. Any token using my name, image, or project is unaffiliated and not endorsed by me.
Disclaimer · Install · Skills · Settings · Examples · Sponsor · Research · FAQ · License
We would love your feedback. Suggestions and bug reports:
- Open a Pull Request or Issue on GitHub
- DM @lexnlin or @blueemi99
- Email us at hello@tasteskill.dev
The npx skills add CLI scans the skills/ folder in this repo, so all skills below (code and image-generation) install the same way.
npx skills add https://github.com/Leonxlnx/taste-skillInstall a single skill by name (example):
npx skills add https://github.com/Leonxlnx/taste-skill --skill "imagegen-frontend-mobile"You can also copy any SKILL.md into your project or paste it into ChatGPT / Codex conversations.
Each skill does one job; you do not need all of them at once. Implementation skills output code. Image-generation skills output reference images only.
| Skill | Description |
|---|---|
| taste-skill | Default all-rounder for premium frontend output without locking one narrow visual style. |
| gpt-taste | Stricter variant for GPT/Codex: higher layout variance, stronger GSAP direction, aggressive anti-slop. |
| image-to-code-skill | Image-first pipeline: generate site references, analyze them, then implement the frontend to match. |
| redesign-skill | Existing projects: audit the UI first, then fix layout, spacing, hierarchy, styling. |
| soft-skill | Polished, calm, expensive UI with softer contrast, whitespace, premium fonts, spring motion. |
| output-skill | When the model ships half-finished work: full output, no placeholder comments. |
| minimalist-skill | Editorial product UI (Notion/Linear vibes), restrained palette, crisp structure. |
| brutalist-skill | BETA Hard mechanical language: Swiss type, sharp contrast, experimental layout. |
| stitch-skill | Google Stitch-compatible rules, including optional DESIGN.md export format. |
These produce design images only (no code). Use with ChatGPT Images, Codex image mode, or any agent that generates images.
| Skill | Description |
|---|---|
| imagegen-frontend-web | Website comps: hero, landing, multi-section with strong typography, spacing, anti-slop art direction. |
| imagegen-frontend-mobile | Mobile screens and flows: iOS/Android/cross-platform, mockups, readable type, coherent sets. |
| brandkit | Brand-kit boards: logo directions, palettes, type, identity applications across categories. |
- Start with taste-skill for the safest general default.
- Use gpt-taste when you want the stricter GPT/Codex-oriented rules and motion/layout enforcement.
- Use image-to-code-skill for image → analyze → code website workflows.
- Use redesign-skill to improve an existing codebase instead of greenfield styling.
- Add soft-skill, minimalist-skill, or brutalist-skill when the visual direction is already chosen.
- Add output-skill if the agent keeps truncating output.
- Use imagegen-frontend-web, imagegen-frontend-mobile, or brandkit when the deliverable is images (comps, flows, identity boards), then pass results to your coding agent.
For image-to-code-skill, state the pipeline in the prompt, e.g.: follow the skill: generate images, then analyze, then code.
Attach or paste imagegen-frontend-web, imagegen-frontend-mobile, or brandkit and ask for the frames you need, then feed the renders to Codex, Cursor, or Claude Code. Use image-to-code-skill when you want one workflow that both generates references and implements the site in code.
Numbers at the top of the file are 1-10 dials:
- DESIGN_VARIANCE: Layout experimentation (lower: centered/clean · higher: asymmetric/modern).
- MOTION_INTENSITY: Animation depth (lower: hover · higher: scroll/magnetic).
- VISUAL_DENSITY: Information per viewport (lower: spacious · higher: dense dashboards).
Created with taste-skill:
If Taste Skill helps you, consider sponsoring:
Background writing that shaped these skills lives in research/.
How is this different from other AI design skills?
Multiple specialized variants, adjustable dials in key skills, anti-repetition rules informed by dedicated research. All are framework agnostic across major coding agents.
Does it work with React, Vue, Svelte?
Yes. Rules target design intent, not a single framework API.
What is SKILL.md?
A portable instruction file agents can load automatically; install via npx skills add or by copying into a repo or conversation.
Do image-generation skills install with npx skills add?
Yes. They live under skills/ alongside the code skills so the same CLI discovers them.
MIT License · Copyright (c) 2026 Leonxlnx

