The Methodology

Three phases to go from idea to working custom GPT

Why Build a Custom GPT?

General-purpose AI is powerful but generic. A custom GPT lets you embed domain expertise, ensure consistent outputs, and lower the barrier so anyone on your team can use sophisticated analysis without writing prompts. You're not building software — you're encoding a decision-making process into a reusable AI tool.

📋

Case Study: AI Task Augmentation Analyst

Throughout this course, we use a real GPT as our case study. The AI Task Augmentation Analyst takes a job description, maps it to the U.S. Department of Labor's O*NET occupational database, scores each task on Automation Risk and Strategic Moat, classifies tasks as Automate / Augment / Human-led, and recommends a pilot project. It went through three major prompt iterations and was deployed across multiple platforms.

Phase 1: Define the Problem

Before writing a single line of prompt, answer these five questions. This step prevents the most common failure mode: building a tool that's technically impressive but doesn't help anyone make a decision.

1. What decision does this tool help someone make?

Be specific. “Help with AI” is too vague. “Identify which of a role's tasks should be automated, augmented, or kept human-led” is actionable.

Template:

“This tool helps [who] decide [what] by analyzing [input] and producing [output].”

2. Who is the user?

Different users need different levels of detail, jargon, and output format.

Executive

Summary dashboard, ROI framing

HR / Operations

Detailed task breakdown, implementation steps

Individual Contributor

Personal task audit, skill development guidance

Consultant

Methodology transparency, exportable data

3. What does the input look like?

Define what users will paste or upload. This shapes your parsing logic. For the case study: a job description — typically 200-800 words of unstructured text containing a job title, responsibilities, qualifications, and sometimes company context.

4. What should the output look like?

Design your output format before writing the prompt. Sketch it on paper if needed. For the case study: two structured tables (Task Breakdown + AI Wins Dashboard), a quadrant classification, and a pilot recommendation with success metrics.

5. What should the tool NOT do?

Constraints prevent the AI from overstepping. These are just as important as capabilities. For the case study: no legal/HR compliance advice, no workforce reduction recommendations, no invented data, always frame savings as estimates.

Phase 2: Find Your Data Foundation

The difference between a mediocre GPT and a great one is grounding data — structured, authoritative information the AI references instead of relying solely on its training data.

Identify Authoritative Sources

Ask: What structured dataset would an expert use to do this job manually?

DomainExample Data Sources
Workforce analysisO*NET, BLS Occupational Outlook
Financial analysisSEC filings, GAAP standards
Medical triageICD-10 codes, clinical guidelines
Legal reviewStatute databases, case law
Product managementFeature taxonomies, user research

Curate and Prioritize Your Data

Most LLM platforms have knowledge file limits. You can't upload everything, so create a tiered loading strategy:

Tier 1: Must Have

Core data the tool can't function without. For the case study: Occupation Data, Task Statements, Task Ratings.

Tier 2: Should Have

Enriches analysis, improves accuracy. For the case study: Work Activities, Skills, Technology Skills.

Tier 3: Nice to Have

Adds depth if file limits allow. For the case study: Abilities, Knowledge, Work Context.

Build a Data Fallback Chain

Your prompt should define what happens when data is missing. This ensures the tool always works, but is transparent about confidence levels:

1User-uploaded files — Treat as authoritative
2Attached knowledge files — Default reference
3General model knowledge — Flag as “lower confidence”

Phase 3: Architect Your Prompt

This is the core of your GPT. A well-structured prompt has five components:

1. Identity and Purpose

Tell the AI who it is and what it does in 2-3 sentences. This anchors all behavior.

You are an AI Workforce Analyst evaluating job roles for AI augmentation potential using O*NET 30. You help organizations identify where AI saves time, improves quality, and creates advantage — while flagging human-led tasks.
Use a specific role title
Name the framework
State the value proposition

2. Workflow Steps

Define the analysis as a numbered sequence. This is the most important section — it determines how the AI thinks through the problem.

Sequential steps

Prevents the AI from jumping to conclusions

Defined inputs/outputs per step

Each step has clear deliverables

Decision points

Where the AI should stop and ask the user

Scoring rubrics

Removes subjectivity, ensures consistency

Case Study Workflow (8 Steps):

Parse job description → Match SOC code → Estimate time → Score tasks (Automation Risk + Strategic Moat) → Classify (Automate / Augment / Human-led) → Assess quality impact → Generate output tables → Recommend pilot

3. Scoring Rubrics

If your tool scores or ranks anything, define the scale explicitly. Without rubrics, the same task might score a 2 from one run and a 4 from the next.

ScoreAutomation Risk Criteria
1Highly routine, rule-based, structured data
2Mostly routine with occasional exceptions
3Mixed — some judgment required
4Significant judgment, unstructured inputs
5Creative, empathetic, physical, high-stakes decisions

4. Output Formats

Define exact table structures and CSV headers so outputs are copy-paste ready and can be directly imported into spreadsheets.

Task,Hours per Week,Automation Risk,Strategic Moat,Category,Quality Delta,AI Approach

5. Guardrails and Uncertainty Handling

Define how the AI should handle ambiguity, low confidence, and edge cases:

SituationRequired Behavior
Rough estimatePrefix with “~”
Low data confidenceAppend “(low confidence)”
Assumption neededState “Assuming [X]...”
Ambiguous inputAsk user before proceeding
Outside data coverageFlag and offer to proceed with general knowledge
💡

Key Takeaways

  • Define before you build — answer the five questions before writing any prompt
  • Ground in authoritative data — structured datasets beat general model knowledge
  • Prioritize with tiers — you can't upload everything, so know what matters most
  • Explicit rubrics create reproducibility — don't rely on the AI's interpretation of “high” or “low”
  • Design your output first — sketch the tables and dashboards before writing the prompt
  • Constraints are features — what the tool should NOT do is as important as what it should