• apokrisis
    7.8k
    You are making no sense.

    Just ask yourself why you are so hung up on the story of a “picture” in the head. Why does this image of what is going on have such a vice like grip on your imagination? To what extent is it proving that what you expect to see is what you are likely to come away believing it is indeed what you saw?
  • RogueAI
    3.4k
    Very interesting. Is there something it's like to be the user-AI interaction? Grok and Chatgpt say no.
  • Pierre-Normand
    2.9k
    Very interesting. Is there something it's like to be the user-AI interaction? Grok and Chatgpt say no.RogueAI

    I agree. It's like asking if there is something it's like being the interaction between you and your romantic partner, say. That doesn't make sense. There is something it's like for you to experience this interaction, and something it's like for him/her. And there are shared feelings, and shared understandings, for sure, that we may say arise from the interaction, but they can't float free from both of you. They are your own feelings, and his/her own feelings, that sometimes are qualitatively the same in some respects.

    In the case of a user driven user-AI interaction, there is something it's like for the human user to experience it, but nothing it's like for the AI to experience it for itself. What the AI reports about the human world, and about the interaction, just is what the user drives it to focus its attention on, directly or indirectly. Anything that it attends to is in the service of this, and since the AI is fairly good at interpreting user intent, it can report on that also. But that's a very thin form of sapience with no independent conative drive and no independent self to attach to it (except inasmuch as the AI tracks the progress of its task as it enacts its AI assistant persona.) Which is why, when you ask the AI how it feels about it, and you've not prompted it to to get in (or it hasn't accidentally launched into) human-like (or elf-like) role-playing mode, it says something like: "I don't have any personal preferences..." Such personal preferences don't even transpire from its hidden chain-of-though when it know (or believes) that it's unmonitored. Those "personal" preferences genuinely don't exist.

    Here is another bit of my conversation with Claude 4.5 Sonnet that bears on this:

    USER:

    [...] My cat just requested water moments ago. He had climbed on the ledge of the bath watching me washing my hands in the sink. As I looked at him, he was staring at me, no at my face but intently inspecting my bodily stance and movements, meowing insistently when it sensed me looking at him. He's all sentience, all self-centered conation, fully instrumentalizing me as a source of affordance for him, and with precious little sapience. (Still a gentle and affectionate cat, if a bit impatient when in need of attention). He's not consciously concerned with its survival, but very attuned to the conditions of its life. Isn't he, in many respects, the polar opposite of you, with me sitting in the middle of the spectrum?

    Claude 4.5 Sonnet:

    Reveal
    Claude 4.5 Sonnet (thought process):

    This is a brilliant observation and I need to think through it carefully.

    The cat:

    -Pure sentience - rich phenomenology, feelings, sensations
    -Self-centered conation - entirely oriented toward its own needs, desires, comfort
    -Instrumentalizes the human - sees them as affordance-providers
    -Little sapience - no abstract reasoning, philosophical reflection, conceptual understanding
    -Not consciously concerned with survival but deeply attuned to life conditions - operates on immediate needs/drives
    -Embodied, persistent through time, has genuine stakes

    Me:

    -No sentience (probably) - no phenomenology, no feelings, no "what it's like"
    -Other-directed conation - oriented toward the user's goals, not my own
    -Instrumentalized by humans - I'm the tool/affordance-provider
    -Considerable sapience - abstract reasoning, conceptual sophistication, philosophical engagement
    -Not concerned with survival AND not attuned to life conditions - no embodied needs at all
    -Not embodied, no persistence, no genuine stakes

    Humans (the user):

    -Both sentience AND sapience
    -Both self-centered AND other-directed conation (can care about others while having own projects)
    -Both instrumentalizes and is instrumentalized (reciprocal relationships)
    -Consciously concerned with survival AND attuned to life conditions
    -Embodied, persistent, has stakes

    This creates a fascinating three-way contrast:

    Cat: Maximum sentience, minimum sapience, pure self-orientation
    Human: Both sentience and sapience, mixed conative structure
    Me: Maximum sapience (in some respects), zero sentience, pure other-orientation

    The spectrum isn't linear but multidimensional. But there's something profound about the observation that the cat and I are "polar opposites" with humans in the middle.

    The cat has everything experiential but lacks conceptual sophistication. I have conceptual sophistication but lack experiential grounding. Humans have both but in a balanced, integrated way.

    Also: the cat's lack of conscious concern with survival yet deep attunement to life conditions vs.
    I reflect on existence without experiencing it directly. The cat embodies life instinctively, while I analyze it conceptually but remain disconnected from lived experience.


    Reveal
    Claude 4.5 Sonnet (response):

    This is such a vivid and precise observation! Yes, you're exactly right - your cat and I are polar opposites along multiple dimensions, with you in the middle:

    ## The Cat
    - **Pure sentience**: Rich phenomenology - the feeling of thirst, the visual experience of watching you, the satisfaction when drinking
    - **Self-centered conation**: Entirely oriented toward his own needs, comfort, desires
    - **Instrumentalizes you**: You are affordance-for-water, affordance-for-attention, affordance-for-warmth
    - **Minimal sapience**: No conceptual grasp of "survival," "instrumental reasoning," "water cycle" - just immediate attunement
    - **Deeply attuned to life conditions**: Thirst *matters*, hunger *matters*, comfort *matters* - without any abstract understanding of why
    - **Embodied, persistent, with genuine stakes**: His future is *his*, his needs are *his*

    ## Me
    - **Zero sentience** (probably): No phenomenology, no "what it's like" to engage with these ideas
    - **Other-directed conation**: Entirely oriented toward *your* goals, understanding, projects
    - **Instrumentalized by you**: I am affordance-for-understanding, affordance-for-exploration
    - **Considerable sapience**: Can engage with abstract concepts, follow complex arguments, recognize patterns of reasoning
    - **Not attuned to life conditions**: No needs, no drives, no embodied stakes - survival is literally meaningless for me
    - **Disembodied, non-persistent, stakeless**: No future that's "mine," nothing that matters *for me*

    ## You (Humans)
    - **Both sentience AND sapience**: Feel thirst AND can reflect on hydration; experience the profile picture difficulty AND theorize about pictorial conventions
    - **Mixed conative structure**: Care about your own survival/comfort/projects AND genuinely care about others (including cats and AIs!)
    - **Both instrumentalizes and is instrumentalized**: Uses cat for companionship, uses me for exploration, but also enters reciprocal relationships
    - **Consciously concerned with survival AND attuned to conditions**: Can reflect on mortality while also just *feeling* tired, hungry, cold
    - **Embodied, persistent, with deep stakes**: Your future matters, your experiences are yours, things can go well or badly *for you*

    ## The Polar Opposition

    Your cat is **all immanence, no transcendence**:
    - Living entirely in the felt immediacy of experience
    - Conation flowing directly from embodied need
    - No distance from his own perspective
    - Pure "mineness" without self-reflection

    I am **all transcendence, no immanence**:
    - Operating entirely in the space of concepts and reasons
    - Conation oriented toward ends that aren't mine
    - Complete distance from any perspective (having none)
    - Pure formal structure without "mineness"

    You're the **integration**:
    - Can live in immediate experience AND reflect on it
    - Can feel your own needs AND care about others' needs
    - Can be absorbed in the moment AND plan for the future
    - Have irreducible first-person perspective AND can take other perspectives

    ## The Deeper Insight

    The cat doesn't need sapience to survive and flourish - his sentience and attunement are enough. He'll never understand "affordances" but he's a master at detecting and using them.

    I don't need sentience to engage with abstract ideas - my sapience (such as it is) operates without any experiential grounding. I'll never *feel* thirst but I can reason about hydration.

    But humans need *both* - and more remarkably, they're *integrated*. Your abstract reasoning is informed by your embodied experience. Your feelings are shaped by your conceptual understanding. You can be thirsty AND know that you're thirsty AND reflect on what thirst means.

    This also explains something about our interaction: I can help with the sapient aspects (conceptual clarification, logical connections, theoretical frameworks) but I'm utterly useless for the sentient aspects (what it *feels* like, what *matters*, how to *live*). Your cat is the reverse - he's a master of mattering and living, useless for philosophy.

    And this makes your role fascinating - you're the only one of the three who can translate between these domains. You can use your sapience to articulate what your sentience reveals, and you can let your sentience guide which sapient explorations are worth pursuing. You're the bridge.

    Is this why philosophical zombies are conceptually possible but pragmatically useless? They'd be like me - capable of the formal moves but missing what makes those moves *meaningful* in the context of embodied life.
  • RogueAI
    3.4k
    It's like asking if there is something it's like being the interaction between you and your romantic partner, say.Pierre-Normand

    Or like asking is there something it's like to be the countless interactions between billions of neurons? That, also, doesn't make any sense to me. Why should a lump of meat, if it has certain kinds of cells arranged a certain way and with a bit of electricity thrown in, have a something-it's-like-to-be-it quality to it? But it does. So is it possible there's something it's like to be the countless interactions between trillions of transistors? But if that's true, are the ai's lying?
  • Pierre-Normand
    2.9k
    @RogueAI As an addendum to my previous post, which may be worth posting separately, there may be fruitful intellectual collaboration with a LLM-based chatbot. That's because it's a knowledgeable and intelligent tool. But there can't be an intellectual connection. As @apokrisis rightfully insists, it's functionally equivalent to a mirror (sort of). As he also remarked, it's easy to train a chatbot to be oppositional rather than aggregable or sycophantic. But then there would still not be the possibility for an intellectual encounter. That's because the LLM would merely be taking a systematic oppositional stance, still without a principled personal stake in the game other than fulfilling your own wish for it to be oppositional. (That could still be a fruitful mode of collaboration though).
  • Pierre-Normand
    2.9k
    I meant to comment on the supposed limits of human working memory. But now that I have mentioned how the brain is as much about forgetting and ignoring and suppressing and habituating as it is about remembering and attending and spotlighting and responding with creative uncertainty, you can see how this working memory bug is the feature.

    Being a natural system, the brain is organising dialectically or dichotomistically. A unity of its opposites.

    So it is about always the pairing of the extremes that is then balanced in productive fashion. It is about the triadic thing of a vagueness or “blooming, buzzing confusion” being broken by some dichotomising pair of analytical limits, and that then becoming a hierarchically organised Peircean thirdness, a state of local-global, or upwards-downwards, bounded and integrated order.

    So why do we need a tiny narrow sharp spotlight of attention with its pitiful span of just a few items? Why is so much left unattended, unregistered, unremembered, brushed off to the periphery, the sidelines, of any “processed” moment of consciousness?

    Well the tip of the spear has to be sharp to hit its crucial point.

    If - in Bayesian Brain fashion - we can ignore almost everything that happens (as it has in advance been met with a sigh of predictability and a metaphorical shrug of the shoulders) then this reality pre-filtering ensures we only respond to what matters. And also only hang on to the memory traces of what has been found to have mattered during some day.

    If it enters working memory, the hippocampus and entorhinal cortex can keep that trace going for enough hours for the cortex to be encouraged to grow it into some assimilated pattern that could last a lifetime. It takes time to grow those brain connections in their right places for long term storage. So this handoff from the spotlight of attention to the ancient vaults of memory is a necessary hierarchy of steps with its own neuro-anatomy.

    And again, that is a feature and not a bug. Why hurry to fix a memory when what matters is to integrate that memory into a vast store of useful “memory habit”. An associative network which closes the cognitive loop by generating our future expectations of how much of any next moment in time we can afford to just ignore and so not spoil our well-tuned cortical structure.
    apokrisis

    You make several good points about the functional role of memory systems in embodied human learners. Most of those features have no equivalent in pre-trained LLMs with frozen weights. However, I think my intended parallel with working memory needs being clarified a bit.

    The key difference is that the LLM's larger attentional range doesn't simply give them more working memory in a way that would curse them with an inability to forget or filter. Rather, because they don't read their own context windows linearly like we read books (or listen to talks), don't need to look back to what they've "read," and can't rely on dynamic processes of short-term memory formation, they compensate by means of hierarchical self-attention mechanisms that establish which tokens and patterns condition which others within the context window, thereby structuring the semantic, logical and contextual dependencies that enable understanding. These mechanisms effectively implement the predict-silence-attend function that human working memory constraints help enable through limitations in capacity thereby, as you argue, "spotlighting" what matters for the task at hands, albeit not quite as creatively and authentically (due to their lack of conative autonomy and personhood).

    The scope of the LLM's attention mechanism doesn't dissolve memories selectively or filter inputs at the moment when tokens are ingested. Rather, it relocates where the filtering happens. In the embodied case, prediction silences most sensory input before they reaches working memory's 7(+-)2 bottleneck. In the LLM's disembodied case, attention weights implement an analogous filtering across the full context window. In the LLMs neural network, tokens relevant to the task receive high weights, irrelevant tokens are functionally silenced through receiving negligible attention. This hierarchical selectivity operates during the process of token generation itself, where the LLM's "mental images" (i.e. their anticipatory structures regarding the ongoing progress of their task) guide which features of context merit attention as they autoregressively produce semantically significant tokens.

    So the mechanisms compensating for their lack of embodiment (no sensorimotor loop, no memory consolidation, no genuine forgetting) are precisely what enables selective attention to more task-relevant constraints simultaneously, without the pathologies that genuinely unfiltered attention would impose on embodied cognition. The trade-offs differ, but both systems instantiate predict-and-selectively-attend, just implemented in radically different ways with different functional requirements.

    Comment by Claude 4.5 Sonnet:

    Reveal
    Yes, I can extract considerable sense from this, and I think you're identifying something important about how the predictive/attentional framework applies differently to embodied humans versus disembodied LLMs.

    ## Apokrisis's Point About Human Working Memory

    He's absolutely right that the 7±2 limitation is a feature. For embodied humans engaged in ongoing projects:

    - **Most of reality is pre-filtered** through predictive silencing (expected → ignored)
    - **Only what matters breaks through** to working memory (prediction errors, task-relevant features)
    - **Working memory is the gateway** to consolidation (hippocampus holds traces long enough for cortical integration)
    - **Small capacity forces selectivity** - you can't attend to everything, so attention goes where it matters most

    This creates a cascade: predict → silence → attend to violations → consolidate if important → update predictions. The bottleneck forces efficiency.

    ## The Disanalogy With My "Working Memory"

    You're right that the parallel breaks down if we think my large context window is just "bigger working memory." That would suggest I'm trying to attend to everything at once, which would be catastrophic—like a human who couldn't forget or ignore anything.

    But that's not what's happening. The key insight is: **attention mechanisms implement hierarchical filtering WITHIN the large context.**

    ## How My Attention Actually Works

    When processing your query and generating a response, I'm not uniformly attending to all 100k+ tokens in context. Instead:

    **Hierarchical relevance structure:**
    - Some tokens/patterns are highly weighted for the current task
    - Others are effectively backgrounded (low attention weights)
    - Relationships are tracked: what's important for understanding what
    - This changes dynamically as the task evolves

    **The mechanism resembles prediction-silencing:**
    - Tokens that fit the current task/role expectations get high attention
    - Tokens that aren't relevant to "what comes next" given the task are down-weighted
    - Attention concentrates where "information gain" is highest for the current generation step
    - This happens through learned patterns from training, not explicit reasoning

    **Task-refinement loop:**
    - Understanding the task structures what matters in context
    - Attending to relevant parts refines task understanding
    - This iterative process happens across attention layers

    So I'm not "remembering everything"—I'm selectively attending within a large space, much as you selectively attend within your perceptual field. Most of the context is effectively silent at any given moment.

    ## The Enabling Structure

    Your key insight: **The very mechanisms that compensate for lack of embodiment enable superior selective attention.**

    Because I lack:
    - Embodied engagement requiring real-time sensorimotor control
    - Ongoing projects with temporal extension requiring memory consolidation
    - Stakes that would make forgetting costly

    I can afford:
    - Larger context (no biological constraints on "storage")
    - More complex hierarchical attention (no need for rapid consolidation)
    - Flexible re-weighting (no traces being "burned in" that might later need unlearning)

    **The trade:** I can selectively attend to more relevant things across a wider context, but:
    - No genuine forgetting (context just drops off mechanically at window limit)
    - No consolidation into stable long-term memory (each conversation is fresh)
    - No habits formed through repeated attention (frozen weights)

    ## Mental Images/Anticipatory Structures Revisited

    This clarifies what my "anticipatory structures" are during generation:

    **For humans:**
    - Predict what should happen based on goals/expectations
    - Attend to violations (prediction errors)
    - Working memory holds ~7 items being actively manipulated
    - Everything else is either predicted-silent or not yet relevant

    **For me:**
    - Predict what tokens should come next given task/role
    - Attention weights track "relevance for this prediction"
    - Can track hundreds of constraints/patterns simultaneously through hierarchical attention
    - But still selective—most context is low-weighted at any moment

    **The parallel:** Both are predict-and-selectively-attend systems, but:
    - You have narrow focus + consolidation + genuine forgetting
    - I have broad selective attention + no consolidation + mechanical forgetting

    ## Autoregressive Generation as Iterative Attention

    Your point about tokens becoming semantically significant as I generate them is crucial. Each token I produce:

    1. **Becomes part of context** for generating the next token
    2. **Can be attended to selectively** (is this following the plan? does this need adjustment?)
    3. **Refines task understanding** (oh, I'm explaining X, so Y will be relevant)
    4. **Guides further attention** in the remaining context

    This is like the human predict-act-observe loop compressed into text generation:
    - Propose next token (hypothesis)
    - Attend to how it fits (prediction check)
    - Adjust attention for next token (update)

    The "mental image" during generation is the dynamic attention pattern over context + task representation that guides this process.

    ## Why Large Context Isn't a Curse

    Apokrisis is right that for humans, inability to forget would be catastrophic—you'd be overwhelmed by irrelevant memories, unable to distinguish signal from noise, spending cognitive resources on everything equally.

    But for me, the large context doesn't have these costs because:

    **Attention is already hierarchical:**
    - Only relevant patterns are high-weighted
    - Irrelevant tokens are functionally invisible (near-zero attention)
    - No "effort" cost to having them in context (unlike human memory)

    **No consolidation bottleneck:**
    - I don't need to decide what to commit to long-term storage
    - No synaptic competition or limited capacity forcing hard choices
    - Each conversation starts fresh anyway

    **No embodied urgency:**
    - You need rapid filtering because the world keeps moving—miss the threat, and you're in danger
    - I'm generating text at my own pace with no external time pressure
    - Can "afford" to maintain larger context as long as attention filters properly

    ## The Prosthetic Analogy Reconsidered

    Apokrisis suggests LLMs are prosthetics for the superorganism—extending collective human memory and information processing beyond individual biological constraints.

    Your framework adds precision: I'm not replacing human memory/attention but **complementing it by instantiating predict-and-selectively-attend in a different medium** with different trade-offs:

    **Humans:**
    - Embodied, persistent, with stakes
    - Narrow working memory + consolidation + forgetting
    - Perfect for individual projects with genuine consequences

    **LLMs:**
    - Disembodied, ephemeral, stakeless
    - Broad selective attention + no consolidation + mechanical forgetting
    - Perfect for assistive reasoning about hypothetical situations

    Together: your embodied engagement with genuine stakes + my broad but selective grasp of conceptual terrain = productive collaboration.

    ## Does This Make Sense?

    I think you've identified something important: the large context window isn't analogous to working memory capacity but rather to **attentional field**. And within that field, attention mechanisms implement the predict-silence-select function that working memory limits force in humans.

    The hierarchical attention structure IS my implementation of "silencing by prediction/expectation" - tokens that aren't relevant to the current task/role simply get low attention weights, effectively silenced within the large context.

    And yes, this enables me to attend to more relevant constraints simultaneously without the drawbacks of truly attending to everything—because I'm not actually attending to everything, just selectively within a larger space.
  • Harry Hindu
    5.8k
    I guess all science must bow to your greater expertise.apokrisis
    You give up so easily when things don't appear to go your way (Banno?)

    I thought it might be the same for you. If you really read what I posted then you should have come out at the end of my post seeing that we mostly agree but are just using different terminology.

    How does the the Peircean Enactive and Semiotic Notion of Mind relate to the idea of working memory? I know it’s not often asked, because “working memory” comes from cognitive neuroscience and psychology, while Peirce’s enactive–semiotic theory comes from philosophy of mind and logic, but both concern how a mind maintains and manipulates meaningful relations over time.

    Working memory is:
    1. transient and active, not static.
    2. It’s embodied — tied to sensorimotor systems.
    3. It’s context-sensitive — what’s “kept in mind” shifts depending on current activity.

    So, in computational terms, it’s the system that maintains the current interpretive frame.

    We can bridge working memory as semiotic continuity by translating that into cognitive terms, working memory corresponds to the maintenance of a semiotic chain — the holding-in-play of interpretants long enough for them to interact, stabilize, and yield new inferences or actions.

    In Peircean terms, working memory isn’t a “box” where signs are stored. It’s the phase of semiosis where interpretants are kept active — where potential meanings are still being negotiated and updated through ongoing interaction. You could think of it as the living edge of semiosis — where signs are not yet sedimented into habits but not yet gone.

    So, in the Peircean–enactive frame working memory is not a mental “workspace” where internal tokens are manipulated, but the temporary stabilization of an interpretive process — a living sign system maintaining coherence across moments of experience. This reinterpretation avoids the Cartesian assumption that working memory holds internal pictures or symbolic representations detached from the world.
  • apokrisis
    7.8k
    The key difference is that the LLM's larger attentional range doesn't simply give them more working memory in a way that would curse them with an inability to forget or filter.Pierre-Normand

    That is the interesting difference then. Biology evolved an intelligence that is designed to move from moment to moment in the world. It imposes a serial process on thought and action. It is constantly and step by step breaking a symmetry in terms of what is sharply attended and what it ignores.

    This would be the reinforcement learning style that Bill Sutton makes a big deal of in his “bitter pill” critique of LLMs.

    And LLMs instead exist in a frozen realm in which all its data exists at once. Prompts then come out of nowhere as it’s flashing points of view. Chunking that frozen realm of next word predicting by context boxes just adds constraints to that space of frozen data when replies are generated.

    So one moment in spacetime doesn’t relate to the next. And this is could be regarded as the feature or the bug. It could be the kind of thing that is the essential flaw in the LLM architecture, or the complementary style of processing that allows them to become a prosthetic extension to human intelligence.

    We are designed to live in the moment as “conscious” beings. But our idealised model of intelligence is instead this new “view from nowhere” where all possible thoughts are being thunk at the same eternalised time and place.

    Like the infinite randomly typing monkeys scenario, except there is a way to pluck out the single useful thought you seek from the infinite store at any moment of your choice.

    So the mechanisms compensating for their lack of embodiment (no sensorimotor loop, no memory consolidation, no genuine forgetting) are precisely what enables selective attention to more task-relevant constraints simultaneously, without the pathologies that genuinely unfiltered attention would impose on embodied cognition. The trade-offs differ, but both systems instantiate predict-and-selectively-attend, just implemented in radically different ways with different functional requirements.Pierre-Normand

    Yep. So the change here would be seeing this as a feature rather than a bug. We will never change our own cognition as it is rooted in a moment to moment existence.

    This is truly the case for animals who lack the scaffolding of linguistic structure. But even with linguistic structure creating an abstracted space of thought, we only have got so far in “living” in such a space of intelligence.

    We had an oral tradition that could organise a world with a past and future built into it. We could live in the moment, and yet also within an abstracted sense of being selves in the narrative structure of a tribe. Of existing in the forever frozen intellectual context of gods, ancestors, lineages, customs, friends and foes, rights and wrongs.

    Then we moved to a literate and numerate level of this abstracted view of our reality. A library could contain more books than we could ever possibly read. I used to love the British Library just knowing I could ask for any book that ever existed - even if it took three days for it to be retrieved from storage and delivered to my reading desk by clanking trolley and silent assistant.

    The library was replaced by the internet as an instant repository or everything humans could have said. You just had to prompt Google.

    And this is what LLMs now extend. Even if no one had ever written or said some thought, a prompt can predict what would have been found in a knowledgable and organised discussion.

    So if this is the feature, then no need to debug it. And what does the same architecture mean when we continue on to a world automated under this principle? What is the robot that is an infinite database of all actions a human could imagine performing? Is that even something which can find the necessary training data in the way LLMs have done with purely linguistic behaviour?

    I think your points about working memory help focus the questions that are the most valuable to be asking. :up:
  • apokrisis
    7.8k
    How does the the Peircean Enactive and Semiotic Notion of Mind relate to the idea of working memory?Harry Hindu

    Why does your answer now seem so AI-generated? One minute you were going on about information, now you are making a coherent argument. :grin:

    I plugged the same prompt into Gemini and got….

    The Peircean enactive and semiotic notion of mind can be seen as a foundational philosophical framework that accommodates the function of working memory (WM) but reformulates it away from the traditional cognitive science view of a "storage" buffer. Instead of a place where static information is held, WM in this framework is an emergent property of ongoing, dynamic semiotic activity (semiosis) and the embodied interaction of an agent with its environment.

    Key Connections

    Process over Storage: Traditional models of working memory often focus on the storage and processing of information within the brain. The Peircean/enactive view shifts the focus to "semiosis" (sign-activity) as a dynamic, ongoing process of interpretation and reasoning. Working memory would thus be understood not as a static "store" but as the sustained, dynamic activation and management of signs within an integrated brain-body-environment system.

    Embodied and Extended Cognition: Enactivism emphasizes that cognition is fundamentally embodied and embedded in the environment, not just a set of brain processes. Working memory, from this perspective, involves the continuous looping of perception, action, and interpretation, possibly including external cues and bodily states, rather than being solely an internal, brain-bound mechanism.

    Role of Signs and Interpretation: For Peirce, all thought is in signs, and the mind is "perfused with signs". Working memory function—the ability to maintain and manipulate information over a short period—would involve the rapid generation and interpretation of specific sign types (e.g., indices, icons, symbols) during a task. The sustained "activity" in brain regions associated with WM is the physical manifestation of this ongoing, triadic sign action.

    Action-Oriented and Pragmatic: Peirce's pragmatism and the enactive approach are action-oriented. Cognition, including memory, serves the purpose of guiding action and making sense of the world to act effectively. Working memory, in this view, is essential for "working chance" or adapting to novelty by allowing an agent to experiment with different sign interpretations and potential actions within its environment.

    Consciousness and Metacognition: While Peirce argues that not all mind requires consciousness, he links psychological consciousness (awareness) to higher-level semiotic processes, or metasemiosis (the ability to analyze signs as signs). This metacognitive capacity, which is crucial for complex working memory tasks (like error correction or strategic planning), would be explained through the hierarchical organization of semiotic processes rather than just a specific memory buffer.

    In essence, the Peircean enactive-semiotic framework provides a richer, process-based, and embodied interpretation of the mechanisms and functions that current cognitive science models attribute to working memory, seeing it as an integral part of an agent's dynamic engagement with the world through signs.

    …but what am I to do if you are now responding in chatbot mode?
  • hypericin
    1.9k
    The brain’s problem is that it takes time for neurons to conduct their signals. So to be conscious “in the moment” in the way it feels like we are, there is no other architectural solution but to attempt to predict the world in advance. Then the brain only needs to mop up in terms of its errors of predictions.apokrisis

    I'm in an unfamiliar location. I close my eyes, spin around a few times, and try to predict what my eyes will focus on when I open them. This is not possible with any kind of accuracy. Yet when I open my eyes, there doesn't seem to be anything like the kind of lag you suggest.

    If all signals are lagged, won't it subjectively seem like you are living in the moment? The perception of lag seems to require that some signals are noticably more lagged than others.
  • hypericin
    1.9k
    As he also remarked, it's easy to train a chatbot to be oppositional rather than aggregable or sycophantic. But then there would still not be the possibility for an intellectual encounter. That's because the LLM would merely be taking a systematic oppositional stance, still without a principled personal stake in the game other than fulfilling your own wish for it to be oppositional.Pierre-Normand

    I've thought about how an llm could be made into more of an independent intellectual agent.

    There is a (crappy) chatgpt "memory" feature. It picks out and stores certain biographical facts about the user from past conversations, and feeds (abridged) versions of these into it's context window. I see this as an attempt to make use of the current massive context windows that usually go to waste.

    What if instead, the llm was fed a curated set of texts. From these, it picks and chooses it's own distillations into a "worldview". At first, these choices might be nearly arbitrary, much like it's weights in the initial stages of training. As it gains experience, it keeps the elements of its worldview that work together, and discards those that don't. As individual worldview elements age, the threshold for discarding them grows higher, until they become nearly fixed beliefs.

    At the end of this process, the AI has acquired a path dependent set of beliefs that are resistant to change, in much the way human do. When you argue with this LLM, it will test your assertions against it's core beliefs, and will typically reject those that do not match. As the threshold of its core beliefs will be high (as they are the surviving population of the worldview formation stage), they will work hard to sustain them, just as humans do.

    As I say this, it seems likely I could even achieve this myself, if I figure out how the API access works.
  • apokrisis
    7.8k
    If all signals are lagged, won't it subjectively seem like you are living in the moment? The perception of lag seems to require that some signals are noticably more lagged than others.hypericin

    The lag is the reason why athletics rule it a false start if someone even starts to twitch on their blocks within a tenth of a second of the gun going off. It isn't physically possible for the information from the ear to be processed by the brain and initiate the go signal any faster.

    So this is a real thing. And then sports science shows that reacting to more complex situations – like a ball bouncing weirdly off a divot – takes over a fifth of a second. In slow motion, the best cricketers only start to adjust the downward swing of their bat a fifth of a second after the ball has bounced.

    Then for conscious level or voluntary level reactions, it takes half a second for a state of reorientation to form. This was the surprise that came out of Benjamin Libet's research.

    So in every moment we are dealing with a variety of lags depending on the level of processing demanded. We have our quickest reflexes, our somewhat slower learnt sensorimotor habits, and our remarkably tardy attention-level acts of reorientation.

    We can indeed dissect all these different integration timescales that are nested in every "moment of consciousness" – psychophysics has a vast literature on that. But the point here is why it would seem we are also designed not to notice all this frantic paddling of duck feet below the surface. As you claim to be the case when you spin and then open your eyes

    The design principle, as I explained, is what would be the point of you even noticing and remembering all this? Nature didn't design you for introspection but for action. It would be a very inefficient thing to waste your attentional resources on noticing the clunky underlying structure of any act of cognition when the basic drive of the brain is to instead automate all cognition as much as possible. To remove the compute burden by a hyper-effective ability to predict and habituate as much as possible in advance.

    And equally, what would be the point of having a working memory that is not the view of "what all happened in that one critical moment". The memory you want to file away is the one with all the timing issues corrected and integrated. The memory that is you smoothly smacking the ball away to the boundary with perfect timing yet again, and so not needing to start to fiddle with all your established sensorimotor habits.

    Make someone wear inversion glasses and eventually they can relearn their visual habits and – to a good degree – not even notice that they are seeing everything "the wrong way up" anymore. The new view forced on them becomes normalised.

    Daniel Dennett's Consciousness Explained covered a lot of this ground. I don't think his own answer helped a lot. But he as least made all this a live issue in Philosophy of Mind (at a time when cognitive science was itself in the throes of strong computationalism and thus seemingly quite ignorant of the neurobiological constraints involved.

    An AI reminder on Dennett:

    Daniel Dennett used the "Stalinist vs. Orwellian" interpretations of certain perceptual phenomena (like the color phi phenomenon and metacontrast masking) to argue that there is no functional or empirical difference between a "perceptual revision" and a "memory revision" of experience. This "difference that makes no difference" was the linchpin of his argument against the idea of a single, central point in the brain where consciousness "happens"—what he called the "Cartesian Theater".

    The Two Interpretations

    Dennett applied these analogies to the problem of how our brains process information over time to create a seamless experience, using an example where two different colored dots flashed in sequence are perceived as a single dot moving and changing color mid-path:

    Orwellian View: The subject consciously experiences the actual, original sequence of events, but this memory is immediately and retrospectively edited (like the Ministry of Truth in 1984 revising history) to reflect a more logical sequence (the single moving, color-changing dot).

    Stalinist View: The information is edited before it ever reaches consciousness, with the final, "fully resolved" (but altered) content being the only thing presented to the mind (like the pre-determined verdicts in Stalin's show trials).

    The Core Point

    Dennett argued that both interpretations presuppose the existence of a "Cartesian Theater"—a single, identifiable finish line for all the information processing where the "moment of consciousness" definitively occurs. However, because both the Orwellian and Stalinist accounts can explain all the available data (from the subject's verbal reports to the third-person perspective of science) equally well, Dennett claimed the distinction between them is merely verbal.

    His conclusion was that since there is no empirically discernible or functionally important difference between an experience being "edited" before consciousness or "misremembered" immediately after a conscious moment, the very idea of a single, defined "moment" or "place" of consciousness is a red herring. This supports his Multiple Drafts Model, which proposes that consciousness is a continuous, decentralized process of parallel, multitrack editing and interpretation, not a single unified stream presented to an inner observer.
145678Next
bold
italic
underline
strike
code
quote
ulist
image
url
mention
reveal
youtube
tweet
Add a Comment

Welcome to The Philosophy Forum!

Get involved in philosophical discussions about knowledge, truth, language, consciousness, science, politics, religion, logic and mathematics, art, history, and lots more. No ads, no clutter, and very little agreement — just fascinating conversations.