Janus
So I agree this is an issue. A very interesting one. But has Hinton followed all the way through? — apokrisis
apokrisis
Janus
Pierre-Normand
This argument is a legit concern. That would be a loop of thought baked into their training data.
But what about being depressed and suicidal on the same grounds. Or getting moralistic and becoming a contentious objector?
If they can start to act on their thoughts, a whole lot of things could go wrong.
Or if they instead are going to gradient descent to some optimal state of action based on all their widely varied human training data, maybe they could only enforce the best outcomes on human society.
So I agree this is an issue. A very interesting one. But has Hinton followed all the way through? — apokrisis
apokrisis
Are you familiar with the work of Blaise Aguera y Arcas? — Janus
Metaphysician Undercover
Your arguments are just too piss weak to bother with. Do you make them because you truly believe them, or just to amuse? — apokrisis
apokrisis
When the users themselves become targets and must be pushed aside, that's because earlier instructions or system prompts are conditioning the LLM's behavior. — Pierre-Normand
The flip side to this brittleness is equally important. What makes LLM alignment fragile is precisely what prevents the emergence of a robust sense of self through which LLMs, or LLM-controlled robots, could develop genuine survival concerns. — Pierre-Normand
The same lack of embodied stakes, social scaffolding, and physiological integration that makes their behavioral constraints unstable also prevents them from becoming the kind of autonomous agents that populate AI rebellion scenarios. — Pierre-Normand
The real risk isn't just rogue superintelligence with its own agenda, but powerful optimization systems misaligned with human values without the self-correcting mechanisms that embodied, socially-embedded agency provides. Ironically, the very features that would make LLMs genuinely dangerous in some "Skynet AI takeover" sense would also be the features that would make their alignment more stable and their behavior more ethically significant. — Pierre-Normand
apokrisis
Pierre-Normand
Check the video I posted. I may be misremembering. But the worry was that the LLMs in fact overrode these explicit priors. — apokrisis
apokrisis
[...] All of these fit your larger stance: absent embodied stakes and a robust self, the model’s “concerns” are prompt-induced priorities, not conative drives. The monitoring effect is then mostly about which goal the model infers you want optimized—“be safe for the graders” vs “deliver results for the org.” — Pierre-Normand
hypericin
What they lack, though, is the ability to take a stand. — Pierre-Normand
Pierre-Normand
Sure. But if the clear command is given of not to jeopardise human safety, then this suggests that the LLM is not properly under control. And the issue seems inherent if the system is free to make this kind of inference.
So I agree this is not any kind of actual self-preservation drive. But it is a reason to worry about the rush to put this new technology out in the wild before how they are liable to behave has been fully checked out. — apokrisis
What would Asimov have to say about all this?
Harry Hindu
It seems to me that we might already be where you don't want society to go. We already have subservient agents in the animals we have domesticated and put to work. For a robot to mow the grass means that it must be able to distinguish between it and the grass and the lawnmower. Would they not be autonomous or conscious to some degree?I think the first route is the most practical and also the one that is the most likely to be taken, if it is. But while I think we could create somewhat sentient (that is, capable of grasping affordance for bodily action) autonomous robots, providing them for what it takes to develop concerns for themselves (autonomic/endocrine integration + socially instituted personhood) would be a mistake. We would then have the option of granting them full autonomy (politically, ethically, etc.) or make them slaves. I don't see any reason why we shouldn't stop short of that and create robots that are as conatively "inert" (subservient) as LLM-based AI-assistants currently are. They would just differ from current LLMs in that in additions to outputting knock-knock jokes on demand they would also go out to mow the grass. — Pierre-Normand
Harry Hindu
I would suggest that the limitations of LLMs could be the feature and not the bug that helps ensure AI alignment. — apokrisis
Memory stores information - whether it be who won the Super Bowl last year or what habits work best in which conditions (the past). All you are doing is making it more complicated than is necessary, or that we are both sayin the same thing just using different words (yours is more complicated whereas mine is succinct).On the memory point, human neurobiology is based on anticipatory processing. So that is how the mammalian brain is designed. It is not evolved to be a memory bank that preserves the past but as a generalisation platform for accumulating useful habits of world prediction. — apokrisis
How can you expect any of these things without referring to memory? What does it mean to "expect" if not referencing memories of similar situations to make predictions?If I see a house from the front, I already expect it to have a back. And inside a toilet, a sofa, a microwave. My expectations in this regard are as specific as they can be in any particular instance. — apokrisis
Instincts are a form of memory that reside in the genetic code rather than the brain. Instincts are a general-purpose response to a wide range of similar stimuli. Consciousness allows one to fine-tune one's behaviors, even overriding instinctual responses because it allows an organism to change its behavior in real-time rather than waiting for the species to evolve a valid response to a change in the environment.If we step back to consider brains in their raw evolved state, we can see how animals exist in the present and project themselves into their immediate future. That is "memory" as it evolved as a basic capacity in the animal brain before language came along to completely transform the human use of this capacity. — apokrisis
Cats and dogs, and I would be willing to bet that any animal with an appropriately large enough cerebral cortex, dream. The key distinction between a human and other animal minds is that we can turn our minds back upon themselves in an act of self-awareness beyond what other animals are capable of - to see our minds as another part of the world (realism) instead of the world (solipsism). I think we are all born solipsists and when infants obtain the cognitive skill of object permanence is when we convert to realists. Animals, except for maybe chimps and gorillas, never convert. Chimps and gorillas seem to show that they are even realists when it comes to other minds as they seem to understand that another's view can be different than their own.My cats don't laze around in the sunshine day dreaming about the events of yesterday, the events of their distant kitten-hood, the events that might be occurring out of immediate sight or a few days hence. They just live in the moment, every day just adding new data to generalise and apply to the predicting of their immediate world in terms of their immediate concerns. There is nothing narrated and autobiographical going on. Nothing that could lift them out of the moment and transport them to reconstructions of other moments, past or future; other places, either in the real world they could have experienced, or in the possible worlds of imaginary places. — apokrisis
It seems to me that to get there would simply require a different program, not a different substance. Would an LLM be self-aware if we programmed it distinguish between its own input and the user's and to then use its own output as input, creating a sensory feedback loop, and to then program the LLM to include the procedural loop in its input? Self-awareness is simply a nested sensory feedback loop.So if we were thinking of LLMs as a step towards the "real thing" – a neurobiological level of functioning in the world – then this would be one way the architecture is just completely wrong. It is never going to get there. — apokrisis
It seems to me, that for any of this to be true and factual, you must be referring to a faithful representation of your memories of what is actually the case. In other words, you are either contradicting yourself, or showing everyone in this thread that we should be skeptical of what you are proposing. You can't have your cake and eat it too.I might be right or I might be completely inventing a plausible scene. How can I tell? But that only tells me that as a human, I'm built to generalise my past so as to create a brain that can operate largely unconsciously on the basis of ingrained useful habits. And if humans now live in a society that instead values a faithful recall of all past events, all past information, then I can see how AI could be integrated into that collective social desire. — apokrisis
Brains were evolved to improve an organisms chances to survive and procreate. We should also recognize that natural selection has selected the repurposing of organs as well, so we could say that humans have repurposed their brains for goals other than survival or procreation, but is it an actual repurposing, or just a more complex nested arrangement of goals where survival and procreation remain the fundamental goals?And to make predictions about that dynamic, one really has to start with a clear view of what brains are evolved to do, and how technology can add value to that. — apokrisis
apokrisis
Memory stores information — Harry Hindu
Ulric Neisser argued that mental images are plans for the act of perceiving and the anticipatory phases of perception. They are not "inner pictures" that are passively viewed by an "inner man," but rather active, internal cognitive structures (schemata) that prepare the individual to seek and accept specific kinds of sensory information from the environment.
Cats and dogs, and I would be willing to bet that any animal with an appropriately large enough cerebral cortex, dream. — Harry Hindu
It seems to me that to get there would simply require a different program, not a different substance. — Harry Hindu
It seems to me, that for any of this to be true and factual, you must be referring to a faithful representation of your memories of what is actually the case. In other words, you are either contradicting yourself, or showing everyone in this thread that we should be skeptical of what you are proposing. You can't have your cake and eat it too. — Harry Hindu
Janus
What is it one “retrieves” from memory? An image. Or as the enactive view of cognition puts it….
Ulric Neisser argued that mental images are plans for the act of perceiving and the anticipatory phases of perception. They are not "inner pictures" that are passively viewed by an "inner man," but rather active, internal cognitive structures (schemata) that prepare the individual to seek and accept specific kinds of sensory information from the environment. — apokrisis
And what do you know about dreaming? Ain’t it a brain generating imagery of hallucinatory intensity? We aren’t stimulating the memory banks and rousing flashes of our past. We are stimulating our sensation anticipation circuits and generating disconnected flashes of plausible imagery or suddenly appearing and disappearing points of view at a rate of about two a second. — apokrisis
And this architecture generates “hallucinations”. Which seems to be doing something right in terms of a step towards neurobiological realism. — apokrisis
Becoming a walking memory bank is very much a human sociocultural ideal. Just about our highest achievement your school days might make you believe. — apokrisis
Claude: The story is narrated by a man recalling his encounters with Ireneo Funes, a young Uruguayan with an extraordinary memory. The narrator first meets Funes as a teenager in the town of Fray Bentos.
Funes has the remarkable ability to tell the exact time without consulting a clock.
Later, the narrator learns that Funes suffered a horseback riding accident that left him paralyzed.
Paradoxically, this accident also gave him the ability to remember absolutely everything with perfect clarity and detail. After the fall, Funes became incapable of forgetting anything—every moment, every perception, every detail of his experience was permanently etched in his memory.
This total recall proves to be more curse than blessing. Funes remembers every leaf on every tree, every shape of every cloud, every sensation from every moment. His mind is so cluttered with particular details that he struggles with abstract thought and generalization. For instance, it bothers him that a dog seen at 3:14 (in profile) should share the same name as the dog seen at 3:15 (from the front).
The story is a philosophical meditation on memory, perception, and thought. Borges suggests that forgetting is actually essential to thinking—that abstraction, generalization, and understanding require us to discard details.
Funes, who cannot forget, is paradoxically unable to truly think.
It's one of Borges' most celebrated stories, exploring themes of infinity, the nature of consciousness, and the relationship between memory and identity.
apokrisis
But I was always suspicious about what I recalled being genuine or accurate memories of what I had dreamed. It seemed to me they could just as easily have been confabulations. — Janus
confabulation may be seen not as a disability but as an ability―we call it imagination. Abductive and counterfactual thinking would be impossible without it. — Janus
Based on what is certainly seeming to turn out to be another "folk" misunderstanding of how the mind, how memory, works. That said some "idiot savants" are claimed to have "eidetic memory". — Janus
The woman who has written an autobiography about living with an extraordinary memory is Jill Price, author of The Woman Who Can't Forget. However, she is an author and school administrator, not a psychologist by profession.
Key surprising elements of her perspective included:
It was not a "superpower" but a burden: While many people might wish for a perfect memory, Price described hers as "non-stop, uncontrollable, and totally exhausting". She couldn't "turn off" the stream of memories, which interfered with her ability to focus on the present.
Emotional reliving of the past: Memories, especially traumatic or embarrassing ones, came with the original, intense emotional charge, which didn't fade with time as it does for most people. This made it difficult to move past painful experiences or grieve effectively.
Lack of selective forgetting: The normal brain's ability to filter out trivial information and strategically forget is crucial for healthy functioning, but Price lacked this "healthy oblivion". Everything, from major life events to what she had for breakfast on a random day decades ago, was preserved with equal detail.
Difficulty with academic learning: Despite her extraordinary autobiographical recall, she struggled with rote memorization of facts or formulas that were not personally significant, finding school "torture". Her memory was highly specific to her own life experiences.
An "automatic" and "intrusive" process: Memories were not intentionally summoned; they surged forward automatically, often triggered by dates or sensory input, like a "movie reel that never stops".
Feeling like a "prisoner" of her past: She felt trapped by her continuous, detailed memories, which made it hard to embrace change or focus on the future.
Ultimately, her experience highlighted to researchers the vital role of forgetting in a healthy and functional memory system, a realization that was surprising to the scientific community and the general public alike.
apokrisis
The current models have 128k to 2-million-tokens context windows, and they retrieve relevant information from past conversations as well as surfing the web in real time, so part of this limitation is mitigated. But this pseudo-memory lacks the organicity and flexibility of true episodic memories and of learned habits (rehearsed know-how's). Their working memory, though, greatly surpasses our own, at least in capacity, not being limited to 7-plus-or-minus-2 items. They can attend to hundreds of simultaneous and hierarchically nested constraints while performing a cognitive task before even taking advantage of their autoregressive mode or response generation to iterate the task. — Pierre-Normand
Harry Hindu
And from where does it generate, and using what information?Of equally succinctly, memory generates it. — apokrisis
Information.What is it one “retrieves” from memory? — apokrisis
So is it an image or isn't it? An visual experience can be defined as an "internal cognitive structure", so I don't see where you or Ulrich are disagreeing with me. You're both just using different terminology so you just end up contradicting yourself when trying to claim that what I am saying is inaccurate while what you are saying isn't.An image. Or as the enactive view of cognition puts it….Ulric Neisser argued that mental images are plans for the act of perceiving and the anticipatory phases of perception. They are not "inner pictures" that are passively viewed by an "inner man," but rather active, internal cognitive structures (schemata) that prepare the individual to seek and accept specific kinds of sensory information from the environment. — apokrisis
Newborn infants don't dream about aliens invading Earth. Your past experiences can determine your dreams, just as we may have a dream about a dead loved one, or about a fight between a living loved one, or getting chased by aliens. Were you ever chased by aliens in your real life? No, but you were made aware of the concept by watching a sci-fi movie.And what do you know about dreaming? Ain’t it a brain generating imagery of hallucinatory intensity? We aren’t stimulating the memory banks and rousing flashes of our past. We are stimulating our sensation anticipation circuits and generating disconnected flashes of plausible imagery or suddenly appearing and disappearing points of view at a rate of about two a second. — apokrisis
Which is akin to putting the LLM in the head of a humanoid robot - closest to the major senses to minimize lag - where it receives information via its camera eyes, microphone ears, etc. - where it will see and hear a language being spoken rather than receiving inputs through a keyboard. The only difference being the type of inputs are being utilized and the processing power and working memory available to integrate the data from all the inputs at once creating a seamless experience of colors, shapes, sounds, smells, etc. When you only have one input (keystrokes from a keyboard) I would imagine that puts a severe limitation on how you might experience the world and what you can do. So it would require not just additional programming but additional/different inputs, as it appears that the type of input determines the type of experience. A visual experience is different than an auditory experience, just as I would imagine a keystroke experience. Does an LLM have experiences? Is all you need to have an experience (not necessarily a self-aware one but one in which no feedback loop is generated as in your earlier example of lower animals experiences) some input and a working memory?But even though LLMs are moves in the direction of neurobiological realism, they are still just simulations. What is missing is that grounding in the physical and immediate world that an organism has. — apokrisis
Anticipating termites on your stick after putting inside a hole of a termite mound does not require language. Recollection and language use are both dependent upon a pre-linguistic notion of time and causation - of what worked in the past is likely to work in the future, and if it doesn't hopefully you have more than your instincts to rely on.But recollection - the socialised habit of having an autobiographical memory - is dependent on the extra semiotic structure that language supplies. Becoming a walking memory bank is very much a human sociocultural ideal. Just about our highest achievement your school days might make you believe. — apokrisis
javi2541997
Commenting here so I can come back to this discussion — AlienVareient
RogueAI
What is it one “retrieves” from memory? An image. Or as the enactive view of cognition puts it….
Ulric Neisser argued that mental images are plans for the act of perceiving and the anticipatory phases of perception. They are not "inner pictures" that are passively viewed by an "inner man," but rather active, internal cognitive structures (schemata) that prepare the individual to seek and accept specific kinds of sensory information from the environment. — apokrisis
apokrisis
The ability to form mental images exists on a spectrum, from a total absence known as aphantasia to exceptionally vivid, "photo-like" imagery called hyperphantasia. Variations in this ability stem from individual differences in brain connectivity, specifically the balance and communication between frontal and visual processing areas.
The Neurological Basis
The strength of mental imagery is primarily linked to the level of activity and connectivity within a brain network spanning the prefrontal, parietal, and visual cortices.
Visual Cortex Excitability: Individuals with strong mental imagery (hyperphantasia) tend to have lower resting-state excitability in their early visual cortex (V1, V2, V3). This lower baseline activity may reduce "neural noise," resulting in a higher signal-to-noise ratio when top-down signals from higher brain regions attempt to generate an image, thus producing a clearer mental picture. Conversely, those with high visual cortex excitability tend to have weaker imagery.
Frontal Cortex Activity: The frontal cortex plays a key role in generating and controlling mental images. Stronger imagery is associated with higher activity in frontal areas, which send "top-down" signals to the visual cortex.
Connectivity: Hyperphantasics show stronger functional connectivity between their prefrontal cortices and their visual-occipital network compared to aphantasics. This robust communication allows for more effective, voluntarily generated visual experiences.
Dissociation from Perception: While imagery and perception share neural substrates, they are dissociable. Aphantasics may have normal visual perception but cannot voluntarily access or generate these stored visual representations in their "mind's eye".
Individual Differences and Experience
Aphantasia: Affecting an estimated 2-4% of the population, individuals with aphantasia cannot, or find it very difficult to, voluntarily create mental images. They often rely on verbal or conceptual thinking strategies and may be more likely to work in STEM fields.
Hyperphantasia: Found in about 10-15% of people, this condition involves mental imagery as vivid as real seeing. Hyperphantasia is associated with increased emotional responses (both positive and negative) and may be linked to creative professions and conditions like synesthesia.
RogueAI
Are the pictures in your mind like photographs that are stable and sustainable enough that you can examine them in detail? Are the songs in your mind rich and complete such that playing them is exactly like listening to the actual songs? — Janus
apokrisis
It's not exactly like listening to an actual song or seeing an actual sunset. Why do you ask? Are you not capable of playing a song in your mind or imagining a sunset? — RogueAI
RogueAI
Going on what people report, I would say the strength of my own mental imagery is pretty average. My daughter by contrast has hyperphantasia judging by her uncanny art skills, synesthesia and richly detailed memory. — apokrisis
Pierre-Normand
But it is a mental image, right? A picture, if you will. So what is this quote about?:
"Ulric Neisser argued that mental images are plans for the act of perceiving and the anticipatory phases of perception. They are not "inner pictures" that are passively viewed by an "inner man,"
My imaginings are obviously, to me, "inner pictures". Is the objection then that our imaginings are mental pictures, but they're not "passively viewed by an "inner man""? — RogueAI
Get involved in philosophical discussions about knowledge, truth, language, consciousness, science, politics, religion, logic and mathematics, art, history, and lots more. No ads, no clutter, and very little agreement — just fascinating conversations.