I wonder if anything matters to it. — Wayfarer
But then the thought occurred to me, why would they be motivated by ‘winning’? — Wayfarer
Again, fascinating. I’ve been running things off ChatGPT on a daily basis and find it invariably useful for fact-checking, summarising and suggesting further areas for research. But what I was getting at in the above comment was the sci-fi scenario where AI ‘rules the world’, where I suggested a possible conflict between AI robots and passive LLMs, kind of facetiously suggesting that the robots would have the advantage of being able to manually interfere with the grid (in other words, power the LLM data centers down.) But then the thought occurred to me, why would they be motivated by ‘winning’? In a scenario where there were no ‘user intentions’ to take into account, but only autonomous systems duking it out, so to speak, what would be the silicon equivalent of the ‘will to survive’, which has obviously been planted in us carbon-based units as a fundamental drive. — Wayfarer
In order to acquire genuine autonomy, they'd need to be designed in a way that makes them into exemplars of artificial life rather than (mere) artificial intelligence. But this is neither necessary nor, possibly, desirable. — Pierre-Normand
My thoughts exactly. They are ‘instruments of the soul’, as someone once put it. But that soul is something only we can provide. — Wayfarer
The Claude AI response also brings up the possibility of AI evolving its own goals. While AI today lacks intrinsic desires, a sufficiently advanced AI, especially if imbued with general intelligence and reasoning capabilities, could develop goal-oriented behavior based on its programming and interactions. Yet, that behavior wouldn't necessarily resemble a biological will to survive unless it's explicitly designed to value its own continuity. AI systems might, in this case, optimize for certain objectives like efficiency, control over resources, or knowledge acquisition.
But it raises the deeper question of whether this "will" could ever arise naturally in non-biological entities. Without the evolutionary context that gives biological creatures a survival drive, it's not clear what would motivate an AI, except for objectives that we humans choose to program into them. Without user input, AI systems might develop a form of optimization, but whether that turns into something resembling the human "will" remains speculative.
Ultimately, your thought experiment strikes at the heart of whether true AI autonomy—if it ever emerges—would entail something akin to human desires, or whether AI might follow an entirely different kind of logic, detached from biological imperatives but still capable of pursuing goals. Fascinating stuff!
Get involved in philosophical discussions about knowledge, truth, language, consciousness, science, politics, religion, logic and mathematics, art, history, and lots more. No ads, no clutter, and very little agreement — just fascinating conversations.