Loading that link, Gemini says " This conversation was created with a Gem that has been deleted. Create a new Gem or start a new chat to continue. "
Loading that link, Gemini says " This conversation was created with a Gem that has been deleted.
I'd like to see the transcript, it sounds interesting. Getting a good prompt is a major factor in getting good results.
On Tuesday, June 10, 2025 at 10:56:35 AM UTC-5 tbp1...@gmail.com wrote:Loading that link, Gemini says " This conversation was created with a Gem that has been deleted.Does this link work?
A bit OT, but I met John Brunner once, years ago (I used to hang out with some of the SF crowd...)
Let’s break that down critically.
The AWIPS example you cited—originally tens of thousands of requirements—is a realistic portrayal of large-scale, mission-critical systems. These requirements are typically hierarchical, with high-level system requirements broken down into functional, performance, and interface specifications. Many are ambiguous or interdependent without domain-specific clarification.
I can help:
Parse and categorize requirements if they are well-documented and structured.
Generate initial designs or pseudocode for subsets.
Provide unit tests or basic system-level tests for components once they are decomposed into manageable pieces.
I cannot:
Manage or trace 11,000 requirements reliably.
Validate their completeness or consistency.
Resolve contradictions or vague specifications without clarification.
Simulate the iterative and cross-disciplinary review processes used in real systems engineering.
In practice, such a requirement set demands rigorous version control, stakeholder validation, domain-specific modeling, and continuous integration testing—none of which I can autonomously carry out.
I can generate GUI code for frameworks like Qt, Tkinter, JavaFX, or web stacks like React. I can also suggest UI layouts, interaction flows, and state machines. However:
I do not maintain state across sessions unless you re-provide the design context.
I lack real-time testing feedback, so UI behavior is hypothetical unless you test and report back.
Accessibility, internationalization, responsiveness, and platform-specific edge cases are typically overlooked unless explicitly specified.
Designing GUIs for critical systems like AWIPS, which must handle high data throughput, human factors under stress, and concurrent multi-modal input, would exceed my current architectural reasoning abilities without continuous expert guidance and iterative feedback.
I can:
Write unit and integration tests based on clear interface definitions.
Generate tests from requirement statements if they're phrased unambiguously and scoped tightly.
Suggest test plans for UI or backend components in isolation.
I cannot:
Produce a traceability matrix automatically across thousands of requirements.
Validate statistical models or performance under load.
Certify correctness under adversarial conditions or edge cases unless specifically directed.
Any real test suite for such a system would require hardware simulations, environmental test data, fault injection, performance benchmarks, regression testing—all managed continuously. I cannot autonomously orchestrate that.
I do not:
Maintain a persistent project state across sessions (unless scaffolded manually by the user).
Interface with full DevOps pipelines (e.g., Jenkins, GitLab CI/CD, Kubernetes) without explicit command-by-command instruction.
Manage documentation, issue tracking, code reviews, or code ownership structures.
I can assist in scripting or describing tools to manage these tasks, but I do not act as a team or development environment. You would need human oversight to manage everything from software configuration management to risk assessment.
AWIPS was developed with:
A budget in the hundreds of millions of dollars.
Multiple years of development.
Input from meteorologists, UI/UX experts, systems engineers, software developers, testers, and program managers.
A legacy of failure and redesign (AWIPS I to AWIPS II) due in part to requirement complexity and contractor issues.
Even now, systems like AWIPS are maintained by specialized contractors (e.g., Raytheon, General Dynamics), not generated or maintained by any single tool or language model.
I am a useful component in the development pipeline for a large-scale system, especially for:
Boilerplate code generation.
Requirement interpretation (in chunks).
Test scaffolding.
Code review.
UI prototyping.
But I cannot independently develop, verify, or maintain a software system at the scale or complexity of AWIPS or any comparable mission-critical platform. Any claim to the contrary is speculative or based on trivial examples, not grounded in the realities of enterprise-scale software engineering."
I just asked ChatGPT an involved question about its capabilties for developing large and complex software. I picked AWIPS, the (US) National Weather Service's Advanced Weather Interactive Processing System, because I had some professional involvement with it in the early 2000's. I have to say, the chatbot's discussion of important aspects of system engineering that it is unable to do tracks very well with what I know from experience.
I have seen this from other country, but still amazed by how the
self made billionaire mythos have co-opted USA* and been extended
to the "once in a generation" world genius and all that companion
crap. I remember saying, long before his fall from grace, that
Musk was not a real man Ironman, but a real man Luthor, except for
the genius part, just a rich megalomaniac super villain.
Going back to the IA topic, recently I started to listen a podcast
episode that makes a hard critique to one of the unfortunately
foundational papers of the field, "The imitation game" and all the
presumptions tone and lack of understanding behind it. It is
refreshing to listen it, amidst all that hype that confuses
thinking with the magic trick that Apparent Intelligence (my
coined expansion for AI) is doing:
https://omny.fm/shows/future-of-coding/computing-machinery
Cheers,
Offray
* For me America is still a continent and not a single country,
and such a naming sounds pretty imperialistic, as people who grew
up in other places of America (the continent) could notice
intermediately.
--Edward
You received this message because you are subscribed to the Google Groups "leo-editor" group.
To unsubscribe from this group and stop receiving emails from it, send an email to leo-editor+...@googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/leo-editor/CAMF8tS07n88%2BM-0KfvzG%3D1dyF2HVpkUzWs_P_u9KCQYANRo%2BUw%40mail.gmail.com.
That's precisely the point they made and that I find related with this thread: a billionaire is not a granted genius and a genius in computer science is not a genius on everything else and, in fact, can device pretty bad test regarding intelligence, justified in a pretty poorly written paper, as showcased in the podcast. Like Turing's affirmation of the overwhelming statistical evidence for telepathy and his considerations of paranormal phenomena while advocating for the Imitation game and "considering" and deviating possible criticisms.
First principles should be properly located, for example in places like geometric axioms and not in Joe/Jane Doe is un-criticable, as (s)he is who (s)he is. That is the path to fall in argumentative fallacies, like the principle of authority. An argument should stand on its own merits and not on the ones of who made it. But yes, let's move on.
Offray
> First principles should be properly located, for example in places like geometric axioms and not in Joe/Jane Doe is un-criticable, as (s)he is who (s)he is.
Turing proposed his test 75 years ago. The blog is crap.
Edward
Turing proposed his test 75 years ago. The blog is crap.
Edward
That doesn't make it automatically valid nor invalid. Newton
proposed his Principia Mathematica several centuries ago. It was
the quality of the proposal, not its author or newness/oldness,
what make it stand the test of time.
The blog is crap.
Do you mean the podcast episode? If yes, could you point to particular places that make it so? The podcast did a pretty good work to pointing to particular places where Turing's paper is "crap".
OffrayDo you mean the podcast episode?
If yes, could you point to particular places that make it so? The podcast did a pretty good work to pointing to particular places where Turing's paper is "crap".
On Thursday, June 12, 2025 at 11:16:26 AM UTC-5 Edward K. Ream wrote:
> I am uninterested in this topic, for reasons I have explained previously.
And I have explained why such reasons seem uncompelling. A first
principle (an axiomatic rule) can not be: someone argument's are,
by principle un-criticable.
Imo, Wikipedia article on the Turing Test refutes the notion that Turing's 1950 paper was anything but first rate.
In fact it doesn't. None of the 24 references to the paper across
the article says anything about the paper quality itself, and are
more related with interpretations. Given than Wikipedia, as any
encyclopedia is a secondary source that appeals to primary
sources, and we have access to the primary source[1], we can judge
by ourselves about the papers quality and see if we agree or not
with the podcast criticism about Turing's tendencies to displacing
both the question from "can machines think?" to "can machines pass
the imitation game?" and also displacing the objections (from
machines can not create anything original to machines can not
surprise us), his lack of citations, his disengagement with
contemporary thinkers/academicians working on similar topics, his
poor referencing to external sources and so on.
[1]
https://academic.oup.com/mind/article-abstract/LIX/236/433/986238?redirectedFrom=fulltext&login=false
> You may have the last word, if you like.
Still true :-)
I'm not interested in having the last word, nor in forcing someone into a conversation. In my two decades as member of the Leo community, I have found here a place of intelligent, inspiring and well argumented conversations, that I have praised several times. That was also what I was looking for here regarding IA. I think that such conversations are important in this times of rage and noise, particularly when we disagree.
Cheers,
Offray
Ps: In another moment I'll share my links collecting criticisms
of AI, that seem also necessary, now that most of places seem
preaching to the (AI) chorus.
On 12/06/25 11:44, Edward K. Ream wrote:
And I have explained why such reasons seem uncompelling. A first principle (an axiomatic rule) can not be: someone argument's are, by principle un-criticable.
Imo, Wikipedia article on the Turing Test refutes the notion that Turing's 1950 paper was anything but first rate.
In fact it doesn't. None of the 24 references to the paper across the article says anything about the paper quality itself, and are more related with interpretations. Given than Wikipedia, as any encyclopedia is a secondary source that appeals to primary sources, and we have access to the primary source[1], we can judge by ourselves about the papers quality
and see if we agree or not with the podcast criticism about Turing's tendencies to displacing both the question from "can machines think?" to "can machines pass the imitation game?"
and also displacing the objections (from machines can not create anything original to machines can not surprise us), his lack of citations, his disengagement with contemporary thinkers/academicians working on similar topics, his poor referencing to external sources and so on.
In my two decades as member of the Leo community, I have found here a place of intelligent, inspiring and well argumented conversations, that I have praised several times. That was also what I was looking for here regarding IA. I think that such conversations are important in this times of rage and noise, particularly when we disagree.
I have read Turing's paper and I think it's something of a masterpiece. He finds a way to avoid figuring out what "thinking" is and still make progress. We have now had 75 more years of experience than Turing had in 1950, and it's no wonder if there were things he hadn't figured out. And we still don't have a good handle on what "thinking" is. His original proposed game is more sophisticated than what is usually meant nowadays when people talk about "the Turing test", and we have have had more chances to see where these tests have turned out not to be perfect. And after all, his ideas seem to have been evolving in the several years after the paper was published. Were he still alive we can be sure they would have evolved further.There's no point is arguing whether Turing got everything right. His ideas and presentation opened up people's minds in the area of thinking machines, and even now he doesn't seem to have been far from the mark.
I just asked ChatGPT an involved question about its capabilties for developing large and complex software.
...ChatGPT>>========
"The short answer is: no, not independently, and not reliably—at least not in the way such a system would be developed by a professional engineering team operating under disciplined systems and software engineering protocols.
I lack real-time testing feedback, so UI behavior is hypothetical unless you test and report back.
Accessibility, internationalization, responsiveness, and platform-specific edge cases are typically overlooked unless explicitly specified.
I can assist in scripting or describing tools to manage these tasks, but I do not act as a team or development environment. You would need human oversight to manage everything from software configuration management to risk assessment.
I am a useful component in the development pipeline for a large-scale system, especially for:
Boilerplate code generation.
Requirement interpretation (in chunks).
Test scaffolding.
Code review.
UI prototyping.
But I cannot independently develop, verify, or maintain a software system at the scale or complexity of AWIPS or any comparable mission-critical platform. Any claim to the contrary is speculative or based on trivial examples, not grounded in the realities of enterprise-scale software engineering."
Remember the old saying, that you need to be twice as smart to debug code as to write it.
I've been finding it helpful to prompt not to emit any code except for snippets to illustrate a point. Then I hold a discussion of a proposed approach. Only then do I ask for some code. I think using the chatbots can provide some of the benefits of pair programming. But you need to be involved closely for the best results.Remember the old saying, that you need to be twice as smart to debug code as to write it. So if your code is as clever as you can make it, then you are not smart enough to debug it.
But what they do is save me a massive amount of time just to get what I think is a simple concept in my mind converted into the actual code that will completely effectively do what I needed to do.Even though cccpu was something like 50+ revisions of code that I think I ended up with
It's that direction and coaching part that's so important. I'm fairly sure that soon there will be a glut of code that didn't get much direction and coaching. And since good design and architecture don't seem to come easily to many people, much chatbot software is likely not to have good design and architecture. In an ideal world, working with a chatbot should improve requirements, specifications, architecture, design, and implementation. We'll see how that all plays out.
I'm not certain if you were YouTube
For now, I figure if I can be a design director and critiquer, fixing randomness that pops up but mostly just kindly asking tasks of it, gemini does a good job. Trying to find the right balance between prompt length and desired outcome has been an interesting learning curve, that I still think I am gong up.