AI: PromptVault - Prompt Library Manager

Model: google/gemini-3-pro-preview
Status: Completed
Cost: $2.09
Tokens: 286,814
Started: 2026-01-02 23:25

03. User Stories & Problem Scenarios

Deep dive into user personas, pain points, and the "Git for Prompts" workflow transformation.

1. Primary User Personas

πŸ‘¨β€πŸ’»

System Architect Sam

Lead AI Engineer | High Tech Savviness

Background: Sam leads a team of 5 engineers building a customer support bot. He treats prompts as code but struggles because they live in codebases, making them hard for non-engineers to edit.

Core Pain Points
  • Regression Chaos: A prompt change broke the bot, and he can't easily "git revert" just the prompt string.
  • Opaque Costs: Has no idea which prompt version is driving up the OpenAI bill.
  • Testing Fatigue: Manually running scripts to test prompts against 100 edge cases.

Goals: Automate regression testing for prompts; integrate prompt management into CI/CD pipeline.

Buying Trigger: A major production incident caused by a "tweaked" prompt that wasn't tested.

πŸ‘©β€πŸ”¬

Prompt Engineer Priya

Product Manager / Specialist | Med-High Tech

Background: Priya isn't a coder but is the "LLM whisperer" at her company. She spends her day in ChatGPT and Claude Playground tweaking language to get perfect outputs.

Core Pain Points
  • The "Notion Graveyard": Has a Notion page with 500+ prompts, but can never find the "one that worked for legal docs."
  • Copy-Paste Hell: Manually copying prompts between OpenAI, Anthropic, and Gemini to compare results.
  • Dependency: Needs to ask Sam (Persona 1) to deploy her changes to production.

Goals: Organize prompts by use-case; A/B test models side-by-side without coding.

Buying Trigger: Spending 4 hours manually testing a prompt across 3 models.

πŸ’Ό

Agency Alex

Founder, AI Consultancy | High Tech

Background: Alex builds AI wrappers for non-tech clients. He needs to manage prompt libraries for 10 different clients simultaneously.

Core Pain Points
  • IP Management: Accidentally leaking Client A's prompt strategy to Client B.
  • Client Handoff: Delivering prompts in a PDF or Google Doc looks unprofessional.
  • Version Drift: Clients change prompts themselves and break the app, then blame Alex.

Goals: Centralized dashboard for all client assets; professional delivery mechanism.

Buying Trigger: A client requesting a "full audit" of all prompts used in their product.

2. Problem Scenarios (Current State)

Scenario A: The "It Worked Yesterday" Mystery

Context: Tuesday, 10:00 AM. Production Incident.

Sam gets a Slack alert: The customer support bot has started answering in pirate speak again. He checks the codebaseβ€”the prompt string looks fine. He asks Priya if she changed anything. She says she "tweaked it slightly" in the admin panel to be friendlier. Because the change was made directly in the database/admin panel without version control, the previous working version is overwritten. Sam spends 2 hours digging through Slack history trying to find the exact phrasing of the old prompt. The team is paralyzed, afraid to touch the prompt again.

Pain: Lack of Version Control & Revert Capability.

Scenario B: The Multi-Model Manual Labor

Context: Thursday, 2:00 PM. Optimization Sprint.

Priya wants to see if the new Claude 3.5 Sonnet model is better/cheaper than GPT-4 for their summarization task. She opens two browser windows: OpenAI Playground and Anthropic Console. She copies the system prompt, the user prompt, and the test data into OpenAI. Runs it. Pastes the result into a spreadsheet. Switches tabs. Pastes the same three things into Anthropic. Runs it. Pastes result into spreadsheet. She repeats this for 20 different test cases. By 5:00 PM, she has a headache, a messy spreadsheet, and 30 open tabs. She realizes halfway through she forgot to set the temperature to 0 in one of the windows, invalidating half the data.

Pain: Manual Testing & fragmented tooling.

3. Prioritized User Stories

Priority As a... I want to... So that... Effort
πŸ”΄ P0 AI Engineer Version prompts automatically so that I can instantly revert to a working state if a new iteration fails in production. M
πŸ”΄ P0 Prompt Engineer Organize prompts with tags/folders so that I stop losing valuable prompts in my notes app. S
πŸ”΄ P0 Developer Access prompts via API so that my application always fetches the latest "approved" prompt without code deploys. L
πŸ”΄ P0 Prompt Engineer Define variables {{like_this}} so that I can test the same prompt structure with different dynamic inputs easily. S
🟑 P1 Prompt Engineer Run multi-model tests side-by-side so that I can compare GPT-4 vs Claude 3 output quality and latency in one view. XL
🟑 P1 Team Lead See a "Diff" view of changes so that I can understand exactly what word changes caused performance to drop. M
🟑 P1 Developer Use a VS Code Extension so that I can manage prompts without leaving my IDE. L
🟒 P2 Finance Lead View cost-per-prompt analytics so that we can identify which complex prompts are draining the budget. M
🟒 P2 Agency Owner Create read-only shared links so that I can show clients their prompts without giving them edit access. S

4. Jobs-to-be-Done Framework

Job 1: Optimize Performance

"When I am preparing a feature for production, I want to mathematically prove which prompt/model combo is best, so I can deploy with confidence regarding cost and quality."


Current Alternative: Excel spreadsheets + manual copy/paste.

Job 2: Ensure Governance

"When a team member changes a prompt, I want to track exactly who changed what and when, so I can maintain system stability and accountability."


Current Alternative: Trust, Slack messages, or Git commit logs (if hardcoded).

Job 3: Democratize AI

"When non-technical domain experts need to improve the AI, I want them to edit prompts safely without touching code, so engineering isn't a bottleneck."


Current Alternative: Sending text files to engineers to copy-paste into the repo.

5. Problem Validation Evidence

Problem Area Evidence Type Data Point / Source
Prompt Fragility Community Sentiment Top Reddit threads in r/OpenAI: "My prompt stopped working after the update" (Frequency: Weekly)
Testing Inefficiency Market Gap LangChain's popularity (200k+ stars) proves developers need abstraction, but LangChain is too complex for non-coders.
Cost Management Industry Survey Retool State of AI Report: "Cost visibility" cited as top 3 challenge for enterprise adoption.

6. Scenarios Transformed (With PromptVault)

Scenario A Resolved: The "One-Click Revert"

Context: Tuesday, 10:05 AM. Incident Resolution.

Sam gets the alert about the pirate-speak bot. He logs into PromptVault. He sees the "Customer Support Bot" prompt was updated 2 hours ago by Priya. He clicks the "History" tab, selects the version from yesterday (v14), and clicks "Rollback." The API endpoint instantly begins serving v14. Total time elapsed: 45 seconds. Crisis averted. Later, he creates a new "Test Branch" for Priya to experiment safely without affecting production.

Resolution Time
2 hours βž” 45 sec
Confidence
Low βž” High

Scenario B Resolved: The "Matrix Test"

Context: Thursday, 2:15 PM. Optimization Sprint.

Priya opens PromptVault's "Lab" view. She selects her prompt and chooses 3 models: GPT-4, Claude 3.5, and Gemini Pro. She uploads a CSV of 20 test cases. She clicks "Run All." PromptVault executes all 60 combinations in parallel. The results appear in a comparison grid with latency and cost calculated for each. She sees Claude 3.5 is 30% cheaper and faster with equal quality. She tags that version as "Production Candidate" for Sam to review.

Effort
3 hours βž” 5 mins
Data Quality
Error-prone βž” Exact