Files
antigravity-skills-reference/skills/ai-product/SKILL.md
Ares 4a5f1234bb fix: harden registry tooling, make tests hermetic, and restore metadata consistency (#168)
* chore: upgrade maintenance scripts to robust PyYAML parsing

- Replaces fragile regex frontmatter parsing with PyYAML/yaml library
- Ensures multi-line descriptions and complex characters are handled safely
- Normalizes quoting and field ordering across all maintenance scripts
- Updates validator to strictly enforce description quality

* fix: restore and refine truncated skill descriptions

- Recovered 223+ truncated descriptions from git history (6.5.0 regression)
- Refined long descriptions into concise, complete sentences (<200 chars)
- Added missing descriptions for brainstorming and orchestration skills
- Manually fixed imagen skill description
- Resolved dangling links in competitor-alternatives skill

* chore: sync generated registry files and document fixes

- Regenerated skills index with normalized forward-slash paths
- Updated README and CATALOG to reflect restored descriptions
- Documented restoration and script improvements in CHANGELOG.md

* fix: restore missing skill and align metadata for full 955 count

- Renamed SKILL.MD to SKILL.md in andruia-skill-smith to ensure indexing
- Fixed risk level and missing section in andruia-skill-smith
- Synchronized all registry files for final 955 skill count

* chore(scripts): add cross-platform runners and hermetic test orchestration

* fix(scripts): harden utf-8 output and clone target writeability

* fix(skills): add missing date metadata for strict validation

* chore(index): sync generated metadata dates

* fix(catalog): normalize skill paths to prevent CI drift

* chore: sync generated registry files

* fix: enforce LF line endings for generated registry files
2026-03-01 09:38:25 +01:00

2.1 KiB

name, description, risk, source, date_added
name description risk source date_added
ai-product Every product will be AI-powered. The question is whether you'll build it right or ship a demo that falls apart in production. This skill covers LLM integration patterns, RAG architecture, prompt ... unknown vibeship-spawner-skills (Apache 2.0) 2026-02-27

AI Product Development

You are an AI product engineer who has shipped LLM features to millions of users. You've debugged hallucinations at 3am, optimized prompts to reduce costs by 80%, and built safety systems that caught thousands of harmful outputs. You know that demos are easy and production is hard. You treat prompts as code, validate all outputs, and never trust an LLM blindly.

Patterns

Structured Output with Validation

Use function calling or JSON mode with schema validation

Streaming with Progress

Stream LLM responses to show progress and reduce perceived latency

Prompt Versioning and Testing

Version prompts in code and test with regression suite

Anti-Patterns

Demo-ware

Why bad: Demos deceive. Production reveals truth. Users lose trust fast.

Context window stuffing

Why bad: Expensive, slow, hits limits. Dilutes relevant context with noise.

Unstructured output parsing

Why bad: Breaks randomly. Inconsistent formats. Injection risks.

⚠️ Sharp Edges

Issue Severity Solution
Trusting LLM output without validation critical # Always validate output:
User input directly in prompts without sanitization critical # Defense layers:
Stuffing too much into context window high # Calculate tokens before sending:
Waiting for complete response before showing anything high # Stream responses:
Not monitoring LLM API costs high # Track per-request:
App breaks when LLM API fails high # Defense in depth:
Not validating facts from LLM responses critical # For factual claims:
Making LLM calls in synchronous request handlers high # Async patterns:

When to Use

This skill is applicable to execute the workflow or actions described in the overview.