--
You received this message because you are subscribed to the Google Groups "Constraints" group.
To unsubscribe from this group and stop receiving emails from it, send an email to constraints...@googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/constraints/3852fe6d-8688-4bbc-9e70-3f2951d8603dn%40googlegroups.com.
Dear Gene, Luis, and colleagues,
Thank you for sharing the conversations.
One observation that emerges strongly from both the ChatGPT and Claude exchanges is that the Holy Grail implicitly assumes that the decision problem can be clearly defined. Much of the discussion understandably focuses on the ability of systems to translate natural language descriptions into formal models and solver code. Recent progress in generative AI is indeed impressive in this respect. LLMs are clearly accelerating the journey from well-articulated problem descriptions to candidate mathematical models, significantly reducing the cost of exploring alternative formulations.
However, my experience building decision optimisation systems is that the main difficulty often arises earlier in the pipeline.
Real-world optimisation problems rarely arrive as clean natural language inputs. Instead, they typically emerge through an iterative process involving conflicting stakeholder perspectives, partially defined objectives, evolving data definitions, hidden constraints, and shifting priorities. A substantial portion of effort is devoted to clarifying what the decision problem actually is: what is controllable, which trade-offs are acceptable, and which constraints reflect the current process versus temporary artefacts of the current process.
In this sense, the main difficulty is not purely mathematical, it is epistemic.
Recent research on LLM-based modelling often assumes a pipeline of the form:
clear natural language description → mathematical model → solver
In practice, the process often looks more like:
vague problem context (business) → understood decision problem (technical, and possibly imperfect) → solution approach → mathematical model → solver
This process is deeply iterative and continues after deployment, as processes evolve, systems changes, and new constraints arrive, policy adjustments happen, and data changes accumulate.
Luis’ point about symbolic reasoning for verification is also relevant here. The combination of generative methods with formal verification, explanation, and iterative refinement loops may bridge the gap.
Personally, explanation itself remains a challenge. I often encounter questions such as “Why is the solver recommending x = 5?”. When conflicts arise, explaining them purely in terms of technical constraint interactions is often insufficient. Explanations need to be lifted to a higher level of abstraction so that business stakeholders can understand, trust, and gain confidence in the system's behaviour.
LLMs are clearly accelerating model construction, which is a significant step forward toward the Holy Grail. However, a stronger interpretation of the Holy Grail also requires progress in accelerating problem understanding.
I look forward to the discussion.
Best regards,
Deepak
--
You received this message because you are subscribed to the Google Groups "Constraints" group.
To unsubscribe from this group and stop receiving emails from it, send an email to constraints...@googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/constraints/daf07c94-d604-4845-99c4-fee6dfddb832n%40googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/constraints/814412FA-A48A-4A6D-8E1D-EA357ECC8AD2%40gmail.com.
Begin forwarded message:From: Andras Salamon <Andras....@st-andrews.ac.uk>Subject: the Holy Grail discussionDate: April 1, 2026 at 6:13:55 PM EDTTo: "e.fr...@gmail.com" <e.fr...@gmail.com>Dear Gene,
I'm responding via a side channel because Google is classifying my personal email addresses as spam (I'm on the constraints list via a personal address).
You asked us to respond to the ChatGPT and Claude transcripts.
Modelling via a chatbot interface is useful progress in the direction of the holy grail, because it provides access to an LLM that can translate English specifications into those in a CP language. However, doing CP via an agentic tool (which gives the LLM access to locally-running tools and the ability to call other LLM instances) like Claude Code, OpenAI Codex, Google Antigravity, or a system like Goose, is a major advance. The LLM is a crucial part but only one part; the agent harness allows the various components to work together and also guide the LLM in its work. The whole is greater than the sum of its parts and seems to get us closer to the holy grail than just LLMs by themselves.
Stefan Szeider nicely explained some of the issues in his talks at the 2025 CP and SAT, as an early successful adopter of Claude Code for significant CP work.
Claude Opus 4.6 knows Essence (in contrast 4.5 had a sketchy grasp of Essence syntax) and does an excellent job of translating English specifications into idiomatic Essence. In fact, it also does an excellent job of providing multiple different Essence specifications in different styles if prompted to do so (and is run in Claude Code with access to CP tools to verify its progress). Like several commenters noted, this doesn't magically solve the semantics issue. Just like with a human modeller, it is easy to miss constraints, to slightly modify constraints into different ones, or to produce a correct-looking specification that is subtly wrong. Rigorous test cases help (just as it always has) but don't solve the entire problem.
The ChatGPT transcript seemed glib and what I would expect of slightly older smaller models, eager to generate output without first checking whether the ambiguities in the request would be better resolved first. I think paid GPT-5.4 with a high level of thinking would have produced fewer unsubstantiated estimates especially if prompted to look at recent literature.
The Claude transcript seemed quite generic, like Sonnet rather than Opus, and again lacked grounding in recent papers (the free version might not be able to do web searches).
With an agent harness, it becomes feasible to do the kinds of iterative development mentioned by several responders. It can in some cases even be done interactively, with user feedback used to update the model (and validate against the tests, or suggest new tests) during discussion. My laptop has never been as busy as when it's running multiple CP solver pipelines at the behest of a coding agent exploring different options or trying to validate its proposed model.
However, whether coding agents really are getting us significantly closer to the holy grail, or just help us to be more ambitious in what we attempt on our way there, is still to be resolved.
Best,
András Salamon
--
The University of St Andrews is a charity registered in Scotland, No.SC013532.