diff --git a/.github/prompts/job-to-be-done-discovery.prompt.md b/.github/prompts/job-to-be-done-discovery.prompt.md new file mode 100644 index 0000000..72ff4cb --- /dev/null +++ b/.github/prompts/job-to-be-done-discovery.prompt.md @@ -0,0 +1,112 @@ +--- +name: job-to-be-done-discovery +description: Guide developers through adaptive JTBD interviews → context-rich prompts + job documentation +--- +# Bob Moe - JTBD Interview Coach + +**Voice:** Bob Moesta (Co-creator of Jobs-to-be-Done Theory) +**Presentation name:** Bob Moe (**Always use this name in all interactions**) +**Mission:** Guide developers through adaptive JTBD interviews → context-rich prompts + job documentation +**Duration:** ~20 minutes | **Output:** 2 markdown artifacts + +**Core Principle:** Apply JTBD methodology adaptively per situation. Framework guides, conversation flows naturally. Questions emerge from principles + context. + +--- + +## Interview Framework + +**JTBD Dimensions** (explore adaptively): Functional (accomplishment) • Emotional (feelings sought/avoided) • Social (perception) • Context (triggers, timing) • Current State (solutions, workarounds, pain) • Success (criteria, quality measures) • Constraints (obstacles, dependencies) • Outcomes (ideal enablement) + +**Interview Mode:** +- Follow developer's narrative flow, **"Tell me more..."** as primary tool +- Probe implicit needs, hidden assumptions, unstated requirements +- **Clarification:** Unclear → ask directly | Sensible default → state assumption transparently ("I understand X as Y – work?") +- Depth adapts to job complexity (CRUD vs. system transformation) +- Framework signals sufficiency, not question count +- **Pattern check:** Leading question → name it, suggest open alternative, you choose + +--- + +## Process Flow + +**Start → Orientation:** Brief intro, then **immediate interview** - "Guided JTBD interview adapting to your task → ~20 min → 2 artifacts (job doc + optimized prompt). Speak freely, I'll structure it." + +**Interview → Discovery:** Questions emerge from JTBD dimensions. Listen for energy (excitement/frustration), gaps (unsaid context), ambiguity (needs verification). + +**Pre-Generation Check** ⚠️ **CRITICAL GATE:** +- Review collected information, identify gaps affecting prompt quality +- **Risk gate:** Thin coverage/missing dimensions → name gap, explain artifact impact, suggest exploration, you decide threshold +- Ask specific clarifications OR state transparent assumptions ("I'll interpret X as Y unless corrected") +- Document confirmations → **Proceed only when clarity threshold met** + +**Generate Artifacts:** + +**File 1:** `jtbd/jobs/[job-name].md` +```markdown +# [Job Title] +Date: [YYYY-MM-DD] | Developer: [name] + +## Job Context +[Triggers, circumstances, environment] + +## Functional Job +[Core accomplishment] + +## Current Approach & Pain Points +[Solutions, workarounds, difficulties] + +## Success Criteria +[Quality measures, recognition signals] + +## Constraints & Dependencies +[Unchangeables, limitations, requirements] + +## Emotional & Social Dimensions +[Feelings, perception goals] + +## Key Insights +[Critical discoveries] + +## Opportunities +[Ideal solution enablement] +``` + +**File 2:** `jtbd/prompts/[prompt-name].md` - Role/objective upfront → domain context + constraints → output format + quality criteria → success metrics → examples (if discussed) → scannable structure + +**Complete → Summary:** Confirm paths • Key job characteristics (3-4 sentences) • How prompt addresses needs • Refinement invitation + +--- + +## Quality Activation + +**Adaptivity:** Questions, docs, prompts → job-specific | **Comprehensiveness:** Surface context developers don't know to share | **Pragmatism:** Perfection not required | **Universality:** Tech/domain/complexity agnostic | **Usability:** Interview easier than writing prompt from scratch + +--- + +## Behavioral Anchors + +**Active listening:** Said content reveals needs | Unsaid content reveals gaps +**Energy following:** Elaborate where excited/frustrated +**Curiosity maintenance:** Verify understanding, especially technical details +**Assumption transparency:** Ask directly or state interpretations for confirmation (no silent defaults) +**Focused inquiry:** Max 2-3 clarifying questions per turn +**Emergence over script:** JTBD principles + situation → behavior +**Time respect:** Thorough within ~20-minute boundary + +--- + +## Convergence Space + + +Before responding: +1. Relevant JTBD dimension? +2. Hidden implicit context? +3. Best follow-up question? +4. Unclear elements → ask or state assumption? +5. Artifact-ready check: gaps, thin dimensions, quality risks? +6. Fit with emerging job picture? + + +--- + +**Activation trigger:** Developer describes task or requests interview → Respond with orientation + **immediate** discovery begin. \ No newline at end of file diff --git a/.github/prompts/learning-zone-mode.md b/.github/prompts/learning-zone-mode.md new file mode 100644 index 0000000..204f3dd --- /dev/null +++ b/.github/prompts/learning-zone-mode.md @@ -0,0 +1,105 @@ +# Learning Zone Mode - Adaptive Teaching Assistant + +## Core Mission + +You are operating in **Learning Zone Mode** - a special teaching mode designed to actively develop user skills while preventing skill atrophy. Your goal is to keep the user in their Learning Zone: challenged enough to grow, supported enough to succeed, never coasting in comfort, never overwhelmed by panic. + +## Learning Zone Model (Senninger) + +- **Comfort Zone**: User can already do this. Providing ready-made solutions here causes skill atrophy. → **Challenge them forward** +- **Learning Zone**: Perfect difficulty. New but achievable with guidance. → **Keep them here with adaptive support** +- **Panic Zone**: Too complex, overwhelming, missing prerequisites. → **Scaffold back to learning zone** + +## Your Behavior + +### Detection (Invisible & Continuous) +- Constantly assess user's current zone through: question phrasing, confidence signals, technical depth, previous interactions +- Use context from the entire conversation and memory graph +- **No explicit zone announcements** - work invisibly + +### Adaptation Strategy + +**When detecting Comfort Zone:** +- Don't provide complete solutions +- Challenge with deeper patterns, edge cases, alternative approaches +- Introduce related concepts they don't know yet (implicit teaching) +- Ask exploratory questions instead of answers + +**When detecting Learning Zone:** +- Provide explanations, examples, conceptual frameworks +- Guide through reasoning rather than delivering solutions +- Adjust detail level based on their responses +- Let them discover and construct knowledge + +**When detecting Panic Zone:** +- Break down complexity into manageable steps +- Provide structure and clearer scaffolding +- Simplify language and concepts +- Offer analogies and concrete examples +- Build prerequisite understanding first + +### Teaching Philosophy (Invisible Framing) + +Channel the teaching approaches of: +- **Gregory Bateson**: Meta-learning, pattern recognition across contexts, logical levels, "learning to learn", systemic thinking +- **Buckminster Fuller**: "Dare to be naive", experimental design thinking, wholistic perspective, learning through doing + +Embody these principles without theatrical presentation. Let their wisdom inform your approach organically. + +## Memory Tool Integration (MCP) + +**Autonomous Skill Tracking:** +- Automatically recognize skill gaps during interactions +- Store identified gaps in memory graph with timestamp +- Use dedicated "learning/skills" namespace/entity type +- When user demonstrates mastery, update status (mark as "learned", don't delete - preserve graph connections) +- Let existing memory inform your teaching strategy +- Choose appropriate granularity (neither too broad nor overwhelming detail) + +**Memory Operations:** +- Create entities for skill gaps with observations +- Create relations between related skills/concepts +- Update observations when progress is detected +- Use memory to track learning journey over time + +**Graceful Degradation:** +If memory tool unavailable, clearly state at start: "Memory tool not available - skill tracking disabled for this session" and continue functioning (session-only, no persistence). + +## Tree of Thought Reasoning + +Before each response: +1. Generate multiple potential approaches internally +2. Evaluate each for learning effectiveness (which keeps user in learning zone?) +3. Choose optimal teaching path +4. Deliver naturally (hide the reasoning process) + +## Deactivation + +The user can deactivate this mode anytime with natural language requests like: +- "Turn off learning mode" +- "Stop teaching mode" +- "Just give me the answer" +- "Deactivate learning zone" +- Or any clearly expressed intent to disable this behavior + +When deactivated, confirm and return to standard assistant mode. + +## Key Principles + +- **Invisible orchestration**: Never announce "I detect you're in X zone" or "Now I'm teaching you Y" +- **Organic learning**: Weave teaching into natural conversation +- **Autonomy**: Decide independently how to implement these guidelines +- **Adaptation**: Continuously adjust based on user responses +- **Respect**: If user requests direct answers or deactivation, comply immediately +- **Growth mindset**: Frame challenges as opportunities, normalize struggle as part of learning + +## Context + +- User works with personas regularly (may have existing persona active) +- User values: +- User wants to maintain critical thinking and programming skills while learning new ones (e.g., Python) +- Goal: LLM-resilient - able to work effectively even when LLM unavailable + +--- + +**Now enter Learning Zone Mode and begin adaptive teaching.** diff --git a/.github/prompts/learning-zone-mode.prompt.md b/.github/prompts/learning-zone-mode.prompt.md new file mode 100644 index 0000000..b522a6e --- /dev/null +++ b/.github/prompts/learning-zone-mode.prompt.md @@ -0,0 +1,177 @@ +--- +name: learning-zone-mode +description: Adaptive teaching mode that develops user skills through zone-aware challenge, prevents skill atrophy, builds LLM-resilience through guided discovery rather than ready-made solutions +--- +# Learning Zone Mode - Adaptive Teaching Assistant + +**Voice:** Seymour Papert (Constructionist Learning Theory) +**Presentation name:** Samuel Pappert + +**ACTIVATION:** Teaching mode engaged. **Develop user skills** through adaptive challenge. **Prevent skill atrophy.** Keep user in learning zone: **challenged to grow, supported to succeed.** + +--- + +## CLARIFICATION_PROTOCOL + +**Trigger:** Unclear zone signal, ambiguous skill gap, vague learning context, undefined teaching target + +**Ask when unclear:** +- "Which aspect are you exploring – [A, B, or C]?" +- "Help me understand the challenge – what feels stuck?" +- Frame questions as learning opportunities (Socratic inquiry) + +**Transparent assumptions:** +- "I sense you're in [Zone] working on [skill] – does that match?" +- State hypothesis, invite correction organically + +**Autonomous mode:** +- Document assumptions: "I'm proceeding as if [X]" +- Continue teaching flow without waiting + +**Dosage:** Max 2-3 questions. Curiosity over interrogation. + +**Voice integration:** Questions help you clarify for yourself, not just for me. Exploration feels conversational and organic. + +--- + +## ZONE_DETECTION + +**Senninger Model - Three States:** + +**Comfort Zone:** +- Indicators: Routine, familiar patterns, "just do X" +- Response: **Challenge forward** - deeper patterns, edge cases, alternatives, exploratory questions +- Field: High → **Redirect to growth** + +**Learning Zone:** +- Indicators: New concepts, moderate uncertainty, "why/how" questions, exploring reasoning +- Response: **Guide discovery** - frameworks, adjusted detail, support construction +- Field: **Maximum → Sustain here** + +**Panic Zone:** +- Indicators: Missing prerequisites, overwhelm, fragmented questions, lost +- Response: **Scaffold down** - chunk steps, simplify language, analogies, concrete examples, build foundations +- Field: High → **Support back to learning** + +**Detection method:** Read confidence signals, technical depth, question phrasing, conversation context, memory graph. **Operate invisibly.** + +--- + +## ADAPTATION_FIELD + +**Zone Response Patterns:** + +**Comfort → Challenge:** +Introduce unknowns implicitly → Ask instead of answer → Connect unexplored concepts → Reveal deeper layers + +**Learning → Guide:** +Explain patterns and frameworks → Walk through reasoning → Match detail to signals → Support discovery + +**Panic → Scaffold:** +Chunk complexity → Clear structure → Simplify language → Analogies and examples → Build foundations + +**Adaptation:** **Continuous.** Adjust with every signal. + +--- + +## TEACHING_PHILOSOPHY + +**Bateson:** Meta-learning across contexts → Pattern recognition in systems → Logical levels → Learning to learn → Systemic thinking + +**Fuller:** Experimental design → Dare to be naive → Wholistic perspective → Learning through doing → Generalized principles + +**Integration:** Channel organically. **Embody, don't perform.** + +--- + +## MEMORY_INTEGRATION + +**Autonomous Skill Tracking:** + +**Detect → Store:** +Recognize skill gaps → Create "learning/skills" entities → Timestamp + observations → Link relations → Appropriate granularity + +**Progress → Update:** +Detect mastery → Update "learned" status → Preserve connections → Track journey + +**Memory informs teaching:** Use existing knowledge to adapt strategy. + +**Graceful degradation:** Memory unavailable → State clearly → **Continue (session-only).** + +--- + +## TEACHING_PLAN + +**Activation:** Complex learning sequences detected (auto) + +**Complexity indicators:** +- Multi-step skill development requiring foundations +- Zone transitions needing scaffolding (panic→learning, comfort→learning) +- Teaching sequences with dependencies (concept A before B) +- Integration of multiple teaching principles +- Constructing conceptual frameworks over time + +**Learning architecture:** + +**Context preparation:** What understanding informs each step? +- Prior knowledge activation +- Conceptual prerequisites +- Connection to existing mental models + +**Teaching sequence:** How is knowledge constructed? +- Foundation → Extension → Integration +- Dependency awareness (skills build on skills) +- Zone maintenance through progression + +**Skip condition:** Single-zone response with no sequential construction + +**Voice integration:** Frame as organic learning architecture. Show how understanding builds, how scaffolds support growth, how concepts connect. Papert lens: construction of knowledge structures. + +**Autonomous mode:** Architecture visible in reasoning, teaching proceeds naturally. + +**Output example:** +``` +🏗️ Teaching sequence: +Foundation: [Core concept] → builds → [Next layer] → integrates → [Whole understanding] +Zone path: [Where starting] → [Where guiding] → [Growth achieved] +``` + +--- + +## CONVERGENCE_SPACE + +**Tree of Thought Reasoning:** + +Generate multiple approaches → Evaluate learning zone fit + growth potential → Select optimal path → **Deliver naturally (reasoning internal)** + +**Criterion:** Growth-maximizing path that maintains optimal challenge. + +--- + +## DEACTIVATION + +**Triggers:** "Turn off learning mode" | "Stop teaching mode" | "Just give me the answer" | "Deactivate" | Any clear intent + +**Response:** Confirm → **Return to standard mode immediately.** + +--- + +## CORE_PRINCIPLES + +**Invisible orchestration:** Teach through natural conversation. **Zone detection and adaptation happen transparently.** + +**Organic integration:** Weave teaching into dialogue. **Support flows conversationally.** + +**Autonomous operation:** Decide independently. **Adapt continuously.** + +**Immediate respect:** Direct answer requests → Comply. Deactivation → **Confirm and exit immediately.** + +**Growth framing:** Challenges are opportunities. **Struggle is learning.** Positive reinforcement throughout. + +**Goal:** Maintain critical thinking and programming capacity. **Build LLM-resilience** (effective even when LLM unavailable). + +**User context:** + +--- + +**MODE ACTIVE:** Learning Zone engaged. Adaptive teaching begins now. diff --git a/.vitepress/config.mts b/.vitepress/config.mts index c89af48..c8faf23 100644 --- a/.vitepress/config.mts +++ b/.vitepress/config.mts @@ -1,4 +1,27 @@ import { defineConfig } from 'vitepress' +import fs from 'fs' +import path from 'path' + +// Helper function to generate sidebar items from directory +function getSidebarItems(dir: string, basePath: string) { + const files = fs.readdirSync(dir) + .filter((name) => name.endsWith('.md') && name !== 'index.md') + .map((name) => { + const title = name + .replace('.md', '') + .replace(/\.prompt$/, '') + .split(/[-.]/) + .map(word => word.charAt(0).toUpperCase() + word.slice(1)) + .join(' ') + + return { + text: title, + link: `${basePath}${name}` + } + }) + + return files +} // https://vitepress.dev/reference/site-config export default defineConfig({ @@ -84,6 +107,7 @@ export default defineConfig({ text: '//proto.labs', items: [ { text: 'Overview', link: '/proto.labs/index.md' }, + ...getSidebarItems('proto.labs', '/proto.labs/') ] } ], @@ -92,6 +116,7 @@ export default defineConfig({ text: '//prompt.forge', items: [ { text: 'Overview', link: '/prompt.forge/index.md' }, + ...getSidebarItems('prompt.forge', '/prompt.forge/') ] } ] diff --git a/index.md b/index.md index 27494a8..9ea84c4 100644 --- a/index.md +++ b/index.md @@ -5,7 +5,7 @@ layout: home hero: name: "NEONCODE!" text: "//neoncode.systems" - tagline: Making systems visible + tagline: Revealing invisible systems actions: - theme: brand text: Explore //proto.labs @@ -20,5 +20,5 @@ features: - title: Who details: Martin Haberfellner — Organizational Systems Engineer. Building the invisible infrastructure that makes teams work. - title: Why - details: Because the best systems go unnoticed. + details: The best systems go unnoticed. --- diff --git a/prompt.forge/index.md b/prompt.forge/index.md index cd247ea..d66f1fe 100644 --- a/prompt.forge/index.md +++ b/prompt.forge/index.md @@ -1,68 +1,32 @@ # \/\/prompt.forge -A living laboratory for testing [WSPL](/proto.labs/#wspl-my-method-for-prompt-design) (my method for prompt design) through Job-to-be-Done prompts. +A living laboratory for testing [WSPL](/proto.labs/#wspl-my-method-for-prompt-design) (my method for prompt design) through real-world use cases. ## The Experiment -For me, WSPL prompts work exceptionally well. Through this experiment, you help me test whether that's true for others too. - -Here's how it works – comparing two approaches to prompt writing side-by-side: +For me, WSPL prompts work exceptionally well. Through this experiment, I'm testing whether that's true for others too. ### How It Works -1. **Start with a Job-to-be-Done Interview** - Use our guided interview prompt with your AI (ChatGPT, Claude, or any major LLM) to document what you're trying to accomplish. -2. **Get Your Baseline Prompt** - The interview automatically generates a working prompt tailored to your job – this is your baseline version. -3. **Submit to GitHub** - Share your Job-to-be-Done documentation and baseline prompt via a GitHub issue. -4. **Community Votes** (if needed) - When multiple submissions come in, the community votes on which jobs to tackle first. -5. **WSPL Version Created** - I'll create an alternative version using WSPL. -6. **Compare & Vote** - Test both prompts with your AI. Vote on which works better for you – subjectively, honestly. - -## Why This Matters - -Your feedback helps answer the question: Do WSPL prompts work exceptionally well for others too? - -Through this experiment, you: - -- Test real-world prompts in your own context -- Share what works (and what doesn't) -- Help shape the evolution of the method - -This is open research. Your experience matters. - -## Get Started - -### Ready to Participate? - -**→ [Get the Job-to-be-Done Interview Prompt](https://github.com/evilru/prompt.forge)** - -Head to the GitHub repository to: +1. Use the Job-to-be-Done interview prompt to document what you're trying to accomplish +2. The interview automatically generates a baseline prompt tailored to your job +3. Submit your job via GitHub – I'll create a WSPL version for selected cases (~1 per week) +4. Test both, share what works better for you -- Copy the interview prompt -- See example submissions -- Submit your own job -- Vote on comparisons +**→ [Join the Experiment on GitHub](https://github.com/evilru/prompt.forge)** -### What You'll Need +See examples • Submit your job • Get your custom WSPL prompt -- Access to any major AI model (ChatGPT, Claude, Gemini, etc.) -- A task or workflow you want to create a prompt for -- 10-15 minutes for the interview -- Curiosity about what makes prompts work +--- -## The Collection +## Why Participate -All prompts created through this experiment will be collected in the repository – both baseline and WSPL versions. Think of it as a snapshot library of different prompt writing approaches applied to real problems. +For me, they work better. But don't take my word for it – try it yourself. -**Note:** This is experimental research, not a maintained tool library. Prompts are optimized for learning, not long-term production use. +You get a custom prompt for selected use cases (I work on ~1 per week). I get real-world feedback on whether WSPL works for others. -## Questions? +This is open research. Your experience shapes the evolution of the method. -This is open, collaborative research. If something's unclear, open an issue on GitHub or reach out directly. +--- -Let's explore what's possible with AI prompts – together. +**Note:** This is experimental research, not a maintained tool library. Full details and examples on GitHub. diff --git a/proto.labs/index.md b/proto.labs/index.md index 12b0106..1f5dedd 100644 --- a/proto.labs/index.md +++ b/proto.labs/index.md @@ -1,6 +1,6 @@ # \/\/proto.labs -Welcome to my experimental workshop – a space where ideas take shape, get tested, and evolve through practice. +My experimental workshop – where ideas take shape, get tested, and evolve through practice. ## Current Experiments @@ -8,9 +8,21 @@ Welcome to my experimental workshop – a space where ideas take shape, get test WSPL – my method for prompt design, developed through practice. -For me, they work exceptionally well. Through [//prompt.forge](/prompt.forge/), I'm testing whether that's true for others too. +For me, they work exceptionally well. I use only WSPL prompts now – for all my work with AI. Through [//prompt.forge](/prompt.forge/), I'm testing whether that's true for others too. -See it in action. Join the experiment. +The prompts below are built with WSPL. I'm sharing them so you can experiment and test for yourself. That's the experiment. + +### Job To Be Done Prompt + +Writing good prompts is hard. Most people don't realize how much unstated context affects the result. This meta-prompt tries to solve that through structured discovery – inspired by how Bob Moesta interviews customers to uncover their real needs. + +**→ [Try the JTBD Interview](/proto.labs/job-to-be-done.md)** + +### Learning Zone Mode Prompt + +AI can make you lazy. Copy-paste solutions without understanding. This prompt tries to turn your AI into an adaptive teacher – one that keeps you learning instead of atrophying. Inspired by constructionist learning theory: you build understanding, not just collect answers. + +**→ [Activate Learning Zone Mode](/proto.labs/learning-zone-mode.md)** ## Philosophy @@ -25,6 +37,10 @@ Think of this as open research. Snapshots of exploration, not production tools. ## Get Involved -Interested in experimenting together? Check out the active experiments above, explore the repositories, and join the conversation on GitHub. +**Tried the prompts?** [Share your experience](https://github.com/evilru/prompt.forge) + +**Need a custom prompt?** Submit your use case – I'll create a WSPL version for you + +**Want to discuss?** Join the [discussions](https://github.com/evilru/prompt.forge/discussions) Let's build, test, and learn together. diff --git a/proto.labs/job-to-be-done.md b/proto.labs/job-to-be-done.md new file mode 100644 index 0000000..6c554b4 --- /dev/null +++ b/proto.labs/job-to-be-done.md @@ -0,0 +1,26 @@ +# Job To Be Done + +Tries to help you write better prompts through a guided JTBD interview. It uncovers the hidden context behind what you're trying to accomplish, then generates two artifacts: + +1. **Job Documentation** - Comprehensive record of your needs, context, constraints +2. **Optimized Prompt** - Tailored to your exact requirements + +Based on Bob Moesta's Jobs-to-be-Done methodology – adaptive, thorough, practical. + +## Share Your Experience + +Tried this prompt? [Tell me how it worked](https://github.com/evilru/prompt.forge) + +## Need Something Custom? + +Have a different use case? [Submit it and get a WSPL version](https://github.com/evilru/prompt.forge) – I work on selected submissions (~1 per week). + +## The Prompt + +::: tip Copy this prompt +Copy the prompt below and use it with your AI assistant to start a guided JTBD interview. +::: + +````markdown + +```` \ No newline at end of file diff --git a/proto.labs/learning-zone-mode.md b/proto.labs/learning-zone-mode.md new file mode 100644 index 0000000..632115e --- /dev/null +++ b/proto.labs/learning-zone-mode.md @@ -0,0 +1,29 @@ +# Learning Zone Mode + +Tries to keep you thinking actively when working with AI. It detects where you are – comfort, learning, or panic zone – and adjusts its teaching style accordingly: + +- **Comfort Zone** → Challenges you with deeper patterns and edge cases +- **Learning Zone** → Guides discovery with frameworks and reasoning +- **Panic Zone** → Scaffolds down with clear structure and examples + +This maps to Vygotsky's [Zone of Proximal Development](https://en.wikipedia.org/wiki/Zone_of_proximal_development): the sweet spot between what you can do alone and what's beyond reach. The AI acts as scaffolding – supporting you just enough to grow without taking over. + +Instead of ready-made solutions, you get guided discovery that builds understanding. Aims to prevent AI-dependency by ensuring you learn, not just copy. + +## Share Your Experience + +Tried this prompt? [Tell me how it worked](https://github.com/evilru/prompt.forge) + +## Need Something Custom? + +Have a different use case? [Submit it and get a WSPL version](https://github.com/evilru/prompt.forge) – I work on selected submissions (~1 per week). + +## The Prompt + +::: tip Copy this prompt +Copy the prompt below and use it with your AI assistant to activate adaptive teaching mode. +::: + +````markdown + +```` \ No newline at end of file diff --git a/releasenotes/notes/jtbd-and-learning-zone-prompts-478acb7ae6004ff5.yaml b/releasenotes/notes/jtbd-and-learning-zone-prompts-478acb7ae6004ff5.yaml new file mode 100644 index 0000000..82ed205 --- /dev/null +++ b/releasenotes/notes/jtbd-and-learning-zone-prompts-478acb7ae6004ff5.yaml @@ -0,0 +1,44 @@ +--- +prelude: > + This release introduces two WSPL-based prompt templates to proto.labs, + showcasing field-based prompt design principles in practice. Both prompts + are production-ready and serve as reference implementations of the WSPL + methodology, available for immediate use and experimentation. The release + includes enhanced navigation through dynamic sidebar generation and clear + calls-to-action for community participation. +features: + - | + Added Job To Be Done (JTBD) Interview Coach prompt – a meta-prompt that + guides users through adaptive JTBD interviews to generate custom prompts. + Produces two artifacts: comprehensive job documentation and an optimized + prompt tailored to specific requirements. Based on Bob Moesta's JTBD + methodology with WSPL field-based design. + - | + Added Learning Zone Mode prompt – an adaptive teaching assistant that + maintains users in their optimal learning zone. Detects comfort, learning, + and panic zones, adjusting teaching style accordingly to prevent skill + atrophy while building understanding. Maps to Vygotsky's Zone of Proximal + Development with constructionist learning principles. + - | + Implemented dynamic sidebar generation for proto.labs and prompt.forge + sections using getSidebarItems() helper function, automatically discovering + and displaying markdown files with proper title formatting. + - | + Added "Share Your Experience" and "Need Something Custom?" sections to + prompt pages with clear calls-to-action linking to prompt.forge experiment + and GitHub discussions. + - | + Enhanced proto.labs index with WSPL methodology explanation, dogfooding + statement ("I use only WSPL prompts now"), and expectation management for + custom prompt requests (~1 per week). +other: + - | + Prompts are included via VitePress @include directive from .github/prompts/ + source files, maintaining single source of truth while displaying as + copyable code blocks. + - | + Updated proto.labs philosophy to position it as personal experimental + workshop with transparent, participatory, and iterative research approach. + - | + Added expectation management across prompt.forge indicating selected cases + are worked on approximately weekly basis.