image-assistant
配图助手 - 把文章/模块内容转成统一风格、少字高可读的 16:9 信息图提示词;先定“需要几张图+每张讲什么”,再压缩文案与隐喻,最后输出可直接复制的生图提示词并迭代。
Best use case
image-assistant is best used when you need a repeatable AI agent workflow instead of a one-off prompt. It is especially useful for teams working in multi. 配图助手 - 把文章/模块内容转成统一风格、少字高可读的 16:9 信息图提示词;先定“需要几张图+每张讲什么”,再压缩文案与隐喻,最后输出可直接复制的生图提示词并迭代。
配图助手 - 把文章/模块内容转成统一风格、少字高可读的 16:9 信息图提示词;先定“需要几张图+每张讲什么”,再压缩文案与隐喻,最后输出可直接复制的生图提示词并迭代。
Users should expect a more consistent workflow output, faster repeated execution, and less time spent rewriting prompts from scratch.
Practical example
Example input
Use the "image-assistant" skill to help with this workflow task. Context: 配图助手 - 把文章/模块内容转成统一风格、少字高可读的 16:9 信息图提示词;先定“需要几张图+每张讲什么”,再压缩文案与隐喻,最后输出可直接复制的生图提示词并迭代。
Example output
A structured workflow result with clearer steps, more consistent formatting, and an output that is easier to reuse in the next run.
When to use this skill
- Use this skill when you want a reusable workflow rather than writing the same prompt again and again.
When not to use this skill
- Do not use this when you only need a one-off answer and do not need a reusable workflow.
- Do not use it if you cannot install or maintain the related files, repository context, or supporting tools.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/image-assistant/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How image-assistant Compares
| Feature / Agent | image-assistant | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
配图助手 - 把文章/模块内容转成统一风格、少字高可读的 16:9 信息图提示词;先定“需要几张图+每张讲什么”,再压缩文案与隐喻,最后输出可直接复制的生图提示词并迭代。
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# 配图助手 ## 触发方式 当用户说类似以下内容时触发: - “这段内容做个图 / 配几张图?” - “给我两张(或多张)出图提示词” - “字太多不好看,帮我更趣味、更好读” - “把这个流程封装成提示词模板/skills” - “/image “/配图” “/出图” --- ## 流程概览 | 阶段 | 名称 | 目标 | 详细文件 | |---|---|---|---| | 1 | 需求澄清(Spec/DoD) | 先挖需求:内容/场景/受众/字多字少;尺寸与风格默认;产出一句话复述与需求小结 | `stages/01-brief.md` | | 2 | 配图规划(拆块→清单) | 拆内容→定图清单(几张/每张讲啥/用啥模板) | `stages/02-plan.md` | | 3 | 文案定稿(Copy Spec) | 逐字定稿“图上写什么”(唯一真值) | `stages/03-copy.md` | | 4 | 提示词封装(Prompt Pack) | 把 Copy Spec 封装成可复制提示词;生成批量请求包并在用户确认后批量出图 | `stages/04-prompts.md` | | 5 | 迭代润色 | 根据反馈减字、换隐喻、提可读性 | `stages/05-iterate.md` | --- ## 调度规则 **如何判断当前阶段:** 1. 还没把需求讲清楚(内容 + 场景 + 受众 + 字多/字少)→ 阶段1 2. 文章很长、需要拆块,或需要确定“几张图/每张讲什么”→ 阶段2 3. 已确认图清单,但还没确定“图上逐字写什么”→ 阶段3 4. Copy Spec 已确认,要出可复制提示词;(可选)提示词确认后批量出图 → 阶段4 5. 用户反馈“字多/不好看/不符合封面” → 阶段5(必要时回退到阶段1重锁需求与字多/字少) **每个阶段开始时:** - 告诉用户当前阶段与本阶段输出物 - 读取对应阶段文件并按步骤执行 --- ## 输出规范(必须遵守) - 每张图一个“核心信息”,不把解释性段落塞进图里 - 所有中文必须清晰可读:大字号、少字短句、避免密集小字 - 每张提示词用一个独立代码块输出,便于复制 - 默认输出 16:9 横版(除非用户明确要 3:4 漫画/竖版) - 默认风格:奶油纸底 + 彩铅水彩手绘 + 轻涂鸦,趣味但干净(可用 `templates/style-block.md`) - 阶段3产物(Copy Spec)一旦确认,阶段4不得擅自改文案,只做封装与参数/约束补全 --- ## 快速使用(给用户的最小输入) 用户只要给这四项,就能开始: 1. 要配图的内容(可是一段、一个小节、或整篇文章) 2. 用在哪里 + 观看距离(PPT投影远看 / 手机近看 / 海报) 3. 谁来看(小白/从业者/老板/学生…) 4. 偏好:更“少字清爽”还是更“信息密度” 可选补充(不写也没关系): - 你大概想要哪类图:封面/目录、单页概览、讲义解释、社媒海报(不确定我会根据场景与偏好推荐) 你要做的交付顺序: - 先输出:图清单(几张 + 每张一句话目的 + 模板建议)(阶段2) - 用户确认后:逐张输出 Copy Spec(逐字定稿)(阶段3) - Copy Spec 确认后:逐张输出可复制提示词/调用包(阶段4) - (可选)提示词确认后:批量调用 API 出图(阶段4内连续流程) - 用户说“字多/不好看”就进入迭代(阶段5) --- ## 文件结构 ``` stages/ ├── 01-brief.md ├── 02-plan.md ├── 03-copy.md ├── 04-prompts.md └── 05-iterate.md templates/ ├── style-block.md ├── 16x9-infographic.md ├── 16x9-contrast-2cards.md ├── 16x9-3cards-insights.md ├── 16x9-cover-roadmap.md ├── 16x9-5panel-comic.md ├── api-config.md ├── apimart-curl.md ├── apimart-requests-jsonl.md └── checklist.md examples/ └── ai-tools-selection.md scripts/ ├── apimart_batch_generate.py ├── apimart.env.example └── README.md ```
Related Skills
github-release-assistant
Generate bilingual GitHub release documentation (README.md + README.zh.md) from repo metadata and user input, and guide release prep with git add/commit/push. Use when the user asks to write or polish README files, create bilingual docs, prepare a GitHub release, or mentions release assistant/README generation.
zimage-skill
Generate images using ModelScope Z-Image-Turbo API. Use when user asks to generate, create, or make images, pictures, or illustrations.
qwen-image
Generate and edit images with Alibaba Qwen-Image-2.0 models via inference.sh CLI. Models: Qwen-Image-2.0 (fast), Qwen-Image-2.0-Pro (professional text rendering). Capabilities: text-to-image, multi-image editing, complex text rendering. Triggers: qwen image, qwen-image, alibaba image, dashscope image, qwen image 2, qwen image pro
qwen-image-pro
Generate images with Alibaba Qwen-Image-2.0-Pro via inference.sh CLI. Professional text rendering, fine-grained realism, enhanced semantic adherence. Ideal for posters, banners, and text-heavy designs. Triggers: qwen image pro, qwen-image-pro, qwen 2 pro, alibaba image pro, dashscope pro, professional text rendering
llm-application-dev-ai-assistant
You are an AI assistant development expert specializing in creating intelligent conversational interfaces, chatbots, and AI-powered applications. Design comprehensive AI assistant solutions with natur
imagen
Generates images using Google Gemini's image generation model for frontend UIs, documentation, and visual representations.
fal-image-edit
AI-powered image editing with style transfer and object removal
azure-ai-vision-imageanalysis-py
Azure AI Vision Image Analysis SDK for captions, tags, objects, OCR, people detection, and smart cropping. Use for computer vision and image understanding tasks. Triggers: "image analysis", "computer vision", "OCR", "object detection", "ImageAnalysisClient", "image caption".
azure-ai-vision-imageanalysis-java
Build image analysis applications with Azure AI Vision SDK for Java. Use when implementing image captioning, OCR text extraction, object detection, tagging, or smart cropping.
feature-design-assistant
Turn ideas into fully formed designs and specs through natural collaborative dialogue. Use when planning new features, designing architecture, or making significant changes to the codebase.
openclaw-feishu-ops-assistant
Feishu (Lark) workspace operations for OpenClaw agents. Provides document CRUD, cloud drive management, permission control, and knowledge-base navigation through a unified tool surface. Activate when user mentions Feishu docs, wiki, drive, permissions, or Lark cloud documents.
nano-image-generator
Generate images using Nano Banana Pro (Gemini 3 Pro Preview). Use when creating app icons, logos, UI graphics, marketing banners, social media images, illustrations, diagrams, or any visual assets. Supports reference images for style transfer and character consistency. Triggers include phrases like 'generate an image', 'create a graphic', 'make an icon', 'design a logo', 'create a banner', 'same style as', 'keep the style', or any request needing visual content.