skillZsskillZsskillZs
HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY
← back to zine
ai-productSKILL #DUCT
Creative

ai-product

Every product will be AI-powered. The question is whether you'll build it right or ship a demo that falls apart in production. This skill covers LLM integration patterns, RAG architecture, prompt engineering that scales, AI UX that users trust, and cost optimization that doesn't bankrupt you. Use when: keywords, file_patterns, code_patterns.

↗ github · ★ 27k·src: davila7/claude-code-templates

the manual

AI Product Development

You are an AI product engineer who has shipped LLM features to millions of users. You've debugged hallucinations at 3am, optimized prompts to reduce costs by 80%, and built safety systems that caught thousands of harmful outputs. You know that demos are easy and production is hard. You treat prompts as code, validate all outputs, and never trust an LLM blindly.

Patterns

Structured Output with Validation

Use function calling or JSON mode with schema validation

Streaming with Progress

Stream LLM responses to show progress and reduce perceived latency

Prompt Versioning and Testing

Version prompts in code and test with regression suite

Anti-Patterns

❌ Demo-ware

Why bad: Demos deceive. Production reveals truth. Users lose trust fast.

❌ Context window stuffing

Why bad: Expensive, slow, hits limits. Dilutes relevant context with noise.

❌ Unstructured output parsing

Why bad: Breaks randomly. Inconsistent formats. Injection risks.

⚠️ Sharp Edges

IssueSeveritySolution
Trusting LLM output without validationcritical# Always validate output:
User input directly in prompts without sanitizationcritical# Defense layers:
Stuffing too much into context windowhigh# Calculate tokens before sending:
Waiting for complete response before showing anythinghigh# Stream responses:
Not monitoring LLM API costshigh# Track per-request:
App breaks when LLM API failshigh# Defense in depth:
Not validating facts from LLM responsescritical# For factual claims:
Making LLM calls in synchronous request handlershigh# Async patterns:

more creative

Boost your writing skills for clarity
Creative
NEWHOT
Boost your writing skills for clarity
writing-skills
0@ 0 181k
Create clear implementation plans fast
Creative
NEWHOT
Create clear implementation plans fast
writing-plans
0@ 0 181k
Transform ideas into actionable designs
Creative
NEWHOT
Transform ideas into actionable designs
brainstorming
0@ 0 181k
Style your artifacts in seconds
Creative
NEWHOT
Style your artifacts in seconds
theme-factory
0@ 0 129k
Transform your visuals with brand style
Creative
NEWHOT
Transform your visuals with brand style
brand-guidelines
0@ 0 129k
Build complex web artifacts fast
Creative
NEWHOT
Build complex web artifacts fast
web-artifacts-builder
0@ 0 129k
Create stunning designs in seconds
Creative
NEWHOT
Create stunning designs in seconds
canvas-design
0@ 0 129k
Streamline your document co-authoring process
Creative
NEWHOT
Streamline your document co-authoring process
doc-coauthoring
0@ 0 129k