I updated my AI Advice page by:
I added this request:
But, and this is IMPORTANT, analyze my original writing style, write it exactly in that style, and then verify to make sure it follows the same style (correcting where required.)
This worked surprisingly well with Claude 4.6 Sonnet (adaptive). I used the resulting ai-advice.md almost verbatim.
Here's the full list of changes it suggested:

The biggest structural update: ai-advice.md is still written as "how individuals use AI better." Your actual advice has evolved into "how people, teams, and organizations operate in the agent era." The whole document should eventually be reorganized to reflect this.
Build an AI workspace, not just a chat (Very high frequency — 10+ docs) Every serious AI project needs a project folder containing:
AGENTS.md — folder-specific instructions the agent reads on startupprompts.md — all prompts version-controlled as source codeskills/ — encapsulated successful workflows (see #2)Treat prompts as the real IP. Code is disposable; prompts, tests, and skills are assets.
Encapsulate successful workflows into reusable skills (Very high frequency) Once an agent succeeds at a task three times, encapsulate it: the prompt, tools used, edge cases, constraints, validation tests. Store in a skill.md file. Skills are the new software libraries — they make workflows deterministically repeatable without re-explaining everything. Use agents to build these skills by asking them to summarize what they learned.
Run coding agents safely: Git + Docker (Very high frequency) Always: (a) work inside a Git repository and instruct the agent to commit as it goes — git checkout is your undo button, (b) run agents inside Docker containers so they cannot touch your actual files, (c) use "YOLO mode" (skip permission prompts) only inside isolated containers. These aren't optional for anything beyond throwaway prototypes.
"LLMs hallucinate, but code doesn't" — use code as the truth engine (Very high frequency) Broaden "have it write code to process numbers" significantly. The mantra is: wherever correctness matters, make the AI produce executable code or logic rather than natural language answers. Code either works or fails — it's binary and auditable. Use domain-specific languages (Prolog-like rule trees, schema validators, policy-as-code) for logic-heavy tasks. This is the primary mechanism for eliminating hallucinations in production.
Build verification into the workflow, not after it (Very high frequency) Verification should be engineered as a product feature, not added as a post-hoc check. Every output should expose: source citations linked to snippets, confidence levels, what's unverifiable, disagreement signals, and audit logs. Use model disagreement as a routing signal — when models disagree, send to human review; when they agree, lower review priority. Build golden sets to measure actual accuracy on your specific task.
Use AI for exception triage, not blanket automation (High frequency) Let AI classify outputs as red/yellow/green: green = automate fully, yellow = flag for review, red = human required. This is more mature than "80-90% AI, human for last mile." It says exactly where the human loop belongs, and it scales: automation handles routine volume while humans focus only on high-stakes exceptions.
Use synthetic data deliberately (High frequency) Not just "realistic fake data for prototyping" — generate hypothesis-driven synthetic data that embeds specific behavioral patterns, edge cases, and known failure modes you expect in production. This lets you stress-test before real data arrives, without compliance concerns, and at whatever messiness level you choose.
Treat demos as imagination accelerators (High frequency) Demos are not just proof-of-concept — they are the fastest way to expand what stakeholders think is possible. Use "Hollywood set" demos: working outputs, simulated backends, precomputed workflows, client-specific synthetic data. Only demo live if the task completes in under 10 minutes. Simulate or precompute slow, expensive, or credential-heavy workflows. Show the output first; defend the architecture only if asked.
Maintain a living model radar — don't freeze model advice (High frequency) Specific model recommendations go stale within months. The durable advice: continuously blind-test frontier models on your exact task, maintain a benchmark set, and route by capability. Current pattern: Claude for coding/aesthetic/style/writing; ChatGPT for rigorous analysis/financial modeling/extended thinking; Gemini for Google Workspace/research/video/speed. But measure this; don't assume it. Additionally: use LiteLLM or Portkey as open-source gateways for organizational cost observability across models.
The Jevons Paradox applies to knowledge work (High frequency in strategic contexts) AI making cognitive tasks cheaper will increase total demand for cognitive work, not reduce it. Human roles shift from execution to verification and judgment — but there's a talent crunch coming for verification roles. Hire now for people who can check, certify, and take accountability for AI output.
Use games to teach AI, not slide decks (High frequency) Replace passive L&D with Capture the Flag challenges, treasure hunts, forbidden-word jailbreaks, prompt-injection games, and coding-agent races. Evaluate proficiency by task completion speed with an agent, not syntax recall. Design challenges where using a coding agent is the only practical way to finish in time — this creates binary signal: those who can use agents solve everything; those who can't solve nothing.
Bifurcate hallucination advice: operational vs. creative (Medium-high frequency) Current advice mixes these. Split explicitly:
Move from dashboards to answers and actions (High frequency) Replace static BI dashboards with AI that answers "what should I do?" not just "what happened?" Ask AI to anticipate a stakeholder's questions and pre-answer them. The endpoint: proactive agents that push insights to individuals rather than passive dashboards that wait to be queried.
Sell outcomes, accountability, and verification — not software (High frequency in business contexts) Software is a depreciating asset; any client can regenerate it tomorrow. Durable value: judgment, trust, domain expertise, data access, and taking responsibility for results. Shift toward outcome-based pricing. The "neck to catch" — human accountability for AI output — is increasingly the product.
Measure AI adoption by behavior, not attendance (Medium frequency) Track: unique days of active use (regularity beats volume), token consumption trends, tool diversity, quality of outputs produced, and business outcomes driven. Usage logs from NetSkope or LLM gateways give better signal than training completion rates.
Assess AI literacy by how people prompt, verify, and recover (High frequency in education contexts) Don't evaluate final answers — AI can produce those. Evaluate: quality of prompts (specificity, guardrails, constraints), ability to identify and fix hallucinations, recovery from errors, and process discipline. Multiple-choice questions are essentially obsolete for AI-era assessment. Assess the process, not the output.
Use AI-native output formats (High frequency) Stop defaulting to PPT or PDF. AI generates HTML, SVG, JSON, interactive dashboards, podcasts, sketch notes, and games better than it generates static slides. A single source document can auto-generate: podcasts, explainer videos, interactive quizzes, sketch notes, executive summaries, slide decks. Use NotebookLM for audio synthesis from diverse sources.
Audit your own behavior via AI (Medium frequency) Feed your own meeting transcripts, email chains, and call recordings into LLMs to find personal blind spots, biases, and recurring errors. Conduct project post-mortems on email threads. Use adversarial prompting — pit GPT against Claude to stress-test your plans. This expands "mine your digital exhaust" from insight into behavioral coaching.
Optimize content for the agentic web (Medium frequency) More content is now consumed by AI agents than humans. Publish in formats AI can parse, cite, and remix: clean metadata, semantic structure, source links, reusable chunks. Develop MCP connectors to your proprietary content. This is SEO for the agentic era.
Ask AI what it needs before starting (Medium frequency) Don't guess what context to provide. Ask: "What information, tools, files, and access do you need to do X?" Let the agent specify missing pieces before it starts, not halfway through. Simple and dramatically reduces mid-task derailment.
Your current document answers: "What tips should I follow?"
Your actual advice answers: "How do I operate in the agent era?"
Suggested new top-level structure: