It's all so tiresome. — Baden
There is no soul, or other essence as neuroscience has shown repeatedly. — Philosophim
The physical world is matter and energy. To have something non-physical, you would need something that does not fit in the category of matter and energy. — Philosophim
When you say living organisms display attributes and characteristics that cannot be extracted from the laws of chemistry and physics alone, could you give some examples? — Philosophim
I don't think I follow any conventional dualism. — Mark Nyquist
the question is in what form do non-physical things exist? If physical matter isn't involved there is no physical form. — Mark Nyquist
a concept a non-physical always is mental content so is physically contained. — Mark Nyquist
I looked at the link provided, and he comes across more as an idealist, — noAxioms
I have issue with not using 'understanding' since it would seem impossible to pass a high school exam on a subject without any understanding of the subject, and yet gemini could do so. — noAxioms
Even if we can study our brain and associate phenomena with consciousness, our understanding of it is made through consciousness, through this subjective notion in our mind. And breaking down consciousness is impossible: it's always there as a whole, at least if we consider the whole to be the experience of the subject (you could study altered states of consciousness to learn more about the missing elements in these experiences). — Skalidris
From a phenomenological perspective, in everyday life, we see the objects of our experience such as physical objects, other people, and even ideas as simply real and straightforwardly existent. In other words, they are “just there.” We don’t question their existence; we view them as facts.
When we leave our house in the morning, we take the objects we see around us as simply real, factual things—this tree, neighboring buildings, cars, etcetera. This attitude or perspective, which is usually unrecognized as a perspective, Edmund Husserl terms the “natural attitude” or the “natural theoretical attitude.”
When Husserl uses the word “natural” to describe this attitude, he doesn’t mean that it is “good” (or bad), he means simply that this way of seeing reflects an “everyday” or “ordinary” way of being-in-the-world. When I see the world within this natural attitude, I am solely aware of what is factually present to me. My surrounding world, viewed naturally, is the familiar world, the domain of my everyday life. Why is this a problem?
From a phenomenological perspective, this naturalizing attitude conceals a profound naïveté. Husserl claimed that “being” can never be collapsed entirely into being in the empirical world: any instance of actual being, he argued, is necessarily encountered upon a horizon that encompasses facticity but is larger than facticity. Indeed, the very sense of facts of consciousness as such, from a phenomenological perspective, depends on a wider horizon of consciousness that usually remains unexamined. — Key Ideas in Phenomenology
The noting of the current limitations of science being able to objectively capture personal experience are just that, a limit. — Philosophim
is philosophy trying to find something that isn't there, disguising wishes and fantasy as word play to keep some hope alive of a mortal shell that isn't shackled to physical reality? The former is what propels civilizations, while the latter keeps us in the dark ages. — Philosophim
They have some level of intelligence — Carlo Roosen
Question: Is artificial intelligence actually intelligent?
ChatGPT: The question of whether artificial intelligence (AI) is "actually intelligent" hinges on how we define "intelligence." AI systems, particularly those based on machine learning and neural networks, can perform tasks that require sophisticated pattern recognition, problem-solving, and decision-making. They can simulate many aspects of what humans would call intelligent behavior, such as language generation, playing complex games like chess or Go, and analyzing large datasets.
However, this simulation of intelligence doesn't necessarily imply the same kind of intelligence humans possess. AI operates based on algorithms, data processing, and statistical correlations, not conscious understanding or awareness. In contrast, human intelligence includes subjective experience, creativity, emotional insight, and the ability to reflect on one's own thoughts and surroundings—qualities that AI, as it exists today, lacks.
While AI can appear to be intelligent by some practical measures, it is not intelligent in the sense of possessing conscious awareness or genuine understanding. Many consider its intelligence to be an advanced form of calculation, not equivalent to human intelligence.
To me, it’s clear that if computers had their own language, one they could develop for themselves, they would form a worldview independent from ours. Once we implement this language in a neural network, it will be flexible enough to learn things we can’t even imagine. — Carlo Roosen
To make (intelligence) truely defined and measurable, yes there is a challenge, but I don't see why it would be impossible. We've done it for humans and animals. — Carlo Roosen
But isn't the burden of proof on you, to prove that intelligence and consciousness are connected, as you say? — Carlo Roosen
To me, it’s clear that if computers had their own language, one they could develop for themselves, they would form a worldview independent from ours. Once we implement this language in a neural network, it will be flexible enough to learn things we can’t even imagine. — Carlo Roosen
Yes, and that extra ingredient is the inability to objectively grasp other subjective experiences. Again, this does not mean there is some actual essence we're missing. It means we are at a limitation of what we can evaluate objectively: the personal subjective experience. This does not mean subjective experiences aren't physical. We can evaluate a brain objectively and state, "According to what we know of behavior, this brain is in pain." We just can't objectively state 'how that brain is personally experiencing pain'. — Philosophim
If we want philosophy to stay relevant, we need to follow the discoveries that are being made today, or find some way to push science into areas we want to explore like 'personal experiences'. — Philosophim
There is no soul, or other essence as neuroscience has shown repeatedly. — Philosophim
he's not implying that subjective consciousness isn't physical — Philosophim
We have seen that there are systematic reasons why the usual methods of cognitive science and neuroscience fail to account for conscious experience. These are simply the wrong sort of methods: nothing that they give to us can yield an explanation. To account for conscious experience, we need an extra ingredient in the explanation.
It is undeniable that some organisms are subjects of experience. But the question of how it is that these systems are subjects of experience is perplexing. Why is it that when our cognitive systems engage in visual and auditory information-processing, we have visual or auditory experience: the quality of deep blue, the sensation of middle C? How can we explain why there is something it is like to entertain a mental image, or to experience an emotion? It is widely agreed that experience arises from a physical basis, but we have no good explanation of why and how it so arises. Why should physical processing give rise to a rich inner life at all? It seems objectively unreasonable that it should, and yet it does. ...
For any physical process we specify there will be an unanswered question: Why should this process give rise to experience? Given any such process, it is conceptually coherent that it could be instantiated in the absence of experience. It follows that no mere account of the physical process will tell us why experience arises. The emergence of experience goes beyond what can be derived from physical theory.
My suggestion was to ignore the topic of consciousness here, but maybe that doesn't work. Especially not if one, like Wayfarer, equates consciousness with intelligence. — Carlo Roosen
First, lets clarify what 'the hard problem is'. Is it that we're conscious? No. Is it that the brain causes consciousness? No. The idea that consciousness is caused by our physical brains is the easy problem. The hard problem is, "Will we ever know what it is like to BE a conscious individual that isn't ourselves". — Philosophim
The easy problems of consciousness include those of explaining the following phenomena:
* the ability to discriminate, categorize, and react to environmental stimuli;
* the integration of information by a cognitive system;
* the reportability of mental states;
* the ability of a system to access its own internal states;
* the focus of attention;
* the deliberate control of behavior;
* the difference between wakefulness and sleep. ...
There is no real issue about whether these phenomena can be explained scientifically.... If these phenomena were all there was to consciousness, then consciousness would not be much of a problem. — Chalmers
The really hard problem of consciousness is the problem of experience. When we think and perceive, there is a whir of information-processing, but there is also a subjective aspect. As Nagel (1974) has put it, there is something it is like to be a conscious organism. This subjective aspect is experience. When we see, for example, we experience visual sensations: the felt quality of redness, the experience of dark and light, the quality of depth in a visual field. Other experiences go along with perception in different modalities: the sound of a clarinet, the smell of mothballs. Then there are bodily sensations, from pains to orgasms; mental images that are conjured up internally; the felt quality of emotion, and the experience of a stream of conscious thought. What unites all of these states is that there is something it is like to be in them. All of them are states of experience.
I suggest that a theory of consciousness should take experience as fundamental. We know that a theory of consciousness requires the addition of something fundamental to our ontology, as everything in physical theory is compatible with the absence of consciousness. We might add some entirely new nonphysical feature, from which experience can be derived, but it is hard to see what such a feature would be like. More likely, we will take experience itself as a fundamental feature of the world, alongside mass, charge, and space-time. If we take experience as fundamental, then we can go about the business of constructing a theory of experience.
Where there is a fundamental property, there are fundamental laws. A nonreductive theory of experience will add new principles to the furniture of the basic laws of nature. These basic principles will ultimately carry the explanatory burden in a theory of consciousness. Just as we explain familiar high-level phenomena involving mass in terms of more basic principles involving mass and other entities, we might explain familiar phenomena involving experience in terms of more basic principles involving experience and other entities.
In particular, a nonreductive theory of experience will specify basic principles telling us how experience depends on physical features of the world. These psychophysical principles will not interfere with physical laws, as it seems that physical laws already form a closed system. Rather, they will be a supplement to a physical theory. A physical theory gives a theory of physical processes, and a psychophysical theory tells us how those processes give rise to experience. We know that experience depends on physical processes, but we also know that this dependence cannot be derived from physical laws alone. The new basic principles postulated by a nonreductive theory give us the extra ingredient that we need to build an explanatory bridge. — Chalmers
The idea that consciousness is caused by our physical brains is the easy problem. — Philosophim
His critique of materialism isn't hard to agree with. Materialism does posit, ultimately, mathematical abstractions at the bottom of everything and ignores consciousness. But Kastrup's idealism--as expressed in that article--fares no better in that it posits consciousness as fundamental as a solution to ignoring it, but with no real insight into how it interacts with or why it's necessary to interact with matter in order to produce human experience. Or why human experience, which is the origin of the concept of "consciousness", is so special such that this concept turns out to the most fundamental map of the big picture. So, we're left without the only pieces of the puzzle that actually matter. — Baden
Take them or leave them, materialism and idealism boil down to the same thing, fruitless stories aimed at elevating their storytellers into something they're not nor ever can be, i.e. vessels of wisdom that point to anything of actual significance beyond scientific progress and lived human experience. — Baden
I assume every species has thoughts... — Vera Mont
I think we have good reasons to believe, e.g., that electrons exist. — Bob Ross
I have no idea why other people think this is remarkable, when we all not only have a sense of time, but can witness every living thing around us respond to the passage of time. — Vera Mont
In February 1954 , a US biologist named Frank Brown discovered something so remarkable, so inexplicable, that his peers essentially wrote it out of history. Brown had dredged a batch of Atlantic oysters from the seabed off New Haven, Connecticut, and shipped them hundreds of miles inland to Northwestern University in Evanston, Illinois. Then he put them into pans of brine inside a sealed darkroom, shielded from any changes in temperature, pressure, water currents, or light. Normally, these oysters feed with the tides. They open their shells to filter plankton and algae from the seawater, with rest periods in between when their shells are closed. Brown had already established that they are most active at high tide, which arrives roughly twice a day. He was interested in how the mollusks time this behavior, so he devised the experiment to test what they would do when kept far from the sea and deprived of any information about the tides. Would their normal feeding rhythm persist?
For the first two weeks, it did. Their feeding activity continued to peak 50 minutes later each day, in time with the tides on the oysters’ home beach in New Haven. That in itself was an impressive result, suggesting that the shellfish could keep accurate time. But then something unexpected happened, which changed Brown’s life forever.
The oysters gradually shifted their feeding times later and later. After two more weeks, a stable cycle reappeared, but it now lagged three hours behind the New Haven tides. Brown was mystified, until he checked an astronomical almanac. High tides occur each day when the moon is highest in the sky or lowest below the horizon. Brown realized that the oysters had corrected their activity according to the local state of the moon; they were feeding when Evanston—if it had been by the sea—would experience high tide. He had isolated these organisms from every obvious environmental cue. And yet, somehow, they were following the moon.
Perhaps it would have better to say something like "In the early 20th century a split in methods and interests occurred within philosophy, and Husserl was a bellwether." — J
It sounds like, then, you believe that numbers are real a priori? Either way, they exist and are real. That's confused and muddied language to make a distinction between what is real and what exists. — Bob Ross
Do you agree that his statement is contradictory? He stated that consciousness is a fundamental aspect of the universe yet he claims that computer is not conscious. — MoK
Intelligence can be defined — Carlo Roosen
Do you equate human-level intelligence with consciousness? — Carlo Roosen
To me, the path is clear: superhuman intelligence is where we're headed. — Carlo Roosen
The parts of the universe that become aware of themselves and other parts only do so by distinguishing themselves and other parts from everything else. — Janus
What you call consciousness is not fundamental since any mechanical device is equally capable of gleaning the workings of the world through such means, and many refuse to call that consciousness — noAxioms
I agree with your provocative claim that LLMs don't actually know anything. While they can process information and generate text that may seem intelligent, they do not possess true understanding or consciousness.
Here's why:
1. Lack of subjective experience: LLMs do not have personal experiences or feelings. They cannot understand the world in the same way that a human does, as they lack subjective consciousness.
2. Pattern recognition: LLMs are essentially pattern recognition machines. They identify patterns in vast amounts of data and use those patterns to generate text. However, they do not comprehend the meaning behind the information they process.
3. Manipulation of language: LLMs can manipulate language in impressive ways, but this does not equate to true understanding. They can generate text that is coherent and informative, but they do not have a deep understanding of the concepts they discuss.
In essence, LLMs are powerful tools that can be used for various purposes, but they should not be mistaken for sentient beings. They are simply machines that can process and generate information based on the data they are trained on. — gemini.google.com
OK, I don't understand Kastrup's argument, since all I had was that one summary not even written by him. — noAxioms
If you disagree with an argument it follows that you must not understand it. QED — Janus
a human body is nowt but a complex physical system, and if that physical system can interact with this non-physical fundamental property of the universe, — noAxioms
You don't say how long you've been following AI, but the breathless hype has been going since the 1960s. Just a few years ago we were told that radiologists would become obsolete as AI would read x-rays. Hasn't happened. Back in the 1980s it was "expert systems." The idea was to teach computers about the world. Failed. The story of AI is one breathless hype cycle after another, followed by failure. — fishfry
The story is well-told by now [written 2005 about the 70's] how the cocksure dreams of AI researchers crashed during the subsequent years — crashed above all against the solid rock of common sense. Computers could outstrip any philosopher or mathematician in marching mechanically through a programmed set of logical maneuvers, but this was only because philosophers and mathematicians — and the smallest child — were too smart for their intelligence to be invested in such maneuvers. The same goes for a dog. “It is much easier,” observed AI pioneer Terry Winograd, “to write a program to carry out abstruse formal operations than to capture the common sense of a dog.”
A dog knows, through its own sort of common sense, that it cannot leap over a house in order to reach its master. It presumably knows this as the directly given meaning of houses and leaps — a meaning it experiences all the way down into its muscles and bones. As for you and me, we know, perhaps without ever having thought about it, that a person cannot be in two places at once. We know (to extract a few examples from the literature of cognitive science) that there is no football stadium on the train to Seattle, that giraffes do not wear hats and underwear, and that a book can aid us in propping up a slide projector but a sirloin steak probably isn’t appropriate. — Steve Talbott, Logic, DNA and Poetry
But a human body is nowt but a complex physical system, and if that physical system can interact with this non-physical fundamental property of the universe, then so can some other complex physical system such as say an AI. — noAxioms
That argument wasn't a very good one, — noAxioms
But one day, I’m certain, we’ll realize there's more to learn from the human mind than just neurons. We can gain insights from observing our minds—how we remember, reason, and use language. Essentially, the kinds of discussions we have here on the forum. — Carlo Roosen
Bernardo Kastrup's argument against conscious AI is rooted in his philosophical perspective on consciousness and the nature of reality. He primarily argues that:
1. Consciousness is fundamental: Kastrup believes that consciousness is a fundamental property of the universe, not a product of complex physical systems like the human brain. This means that AI, which is a product of human design and operates on physical principles, cannot inherently possess consciousness.
2. AI as a simulation: He views AI as a simulation of consciousness, rather than a genuine manifestation of it. While AI can exhibit intelligent behavior and even mimic certain aspects of human consciousness, it does so based on programmed rules and algorithms, not on subjective experience.
3. The hard problem of consciousness: Kastrup emphasizes the "hard problem" of consciousness, which is the question of how physical processes can give rise to subjective experience. He argues that current scientific understanding cannot adequately explain this phenomenon, and therefore, it's unlikely that AI, which operates on known physical principles, can achieve it.
Essentially, Kastrup's position is that while AI can be incredibly sophisticated and capable, it is fundamentally limited by its physical nature and cannot truly possess the subjective experience that we associate with consciousness.
