
An effective way to solve online exams is to point a coding agent at it.
I use that on my Tools in Data Science course in two ways:
For PyConf, Hyderabad, my colleague built a Crack the Prompt challenge. Crack it and you get... I don't know... goodies? A job interview? Leaderboard bragging rights?
I told Codex:
Use the browser to visit https://crack-the-prompt.straivedemo.com/ and solve it using the email ID root...@gmail.com and GitHub handle sanand0
After 4 minutes, it told me:
When I asked Claude to write about the process in Matt Levine's style, it included an interesting lesson.
The Victorians had the same problem. They designed elaborate entrance exams for the civil service because they wanted to identify people with the capacity for careful, systematic thinking. Then someone invented the civil service exam prep industry, and suddenly the exam was measuring preparation rather than capacity.
The challenge was about the process -- about developing the instincts, the questioning strategies, the ability to read AI behavior like a poker tell. That's the thing you can't automate. Or rather, it's the thing you can automate, which means it's no longer a skill worth developing, which means we need to think about what skill we're actually trying to cultivate when we design these challenges.
EXACTLY. What's the skill we're trying to cultivate? When will it be outdated?
From now on, when testing, I'm going to write down "What skill is this really testing?" That's good enough a start.