No that is not it at all.So every time I point to a fundamental difference, your reply is simply that differences can be minimised. And when I point out that minimising those differences might be physically impractical, you wave that constraint away as well. It doesn't seem as though you want to take a principled approach to your OP. — apokrisis
Anyway, another way of phrasing the same challenge to your presumption there is no great problem here: can you imagine an algorithm that could operate usefully on unstable hardware? How could an algorithm function in the way you require if it's next state of output was always irreducibly uncertain? In what sense would such a process still be algorithmic in your book if every time it computed some value, there would be no particular reason for the calculation to come out the same? — apokrisis
I was seeking to make a distinction between simulating a human being and simulating general intelligence....I was using the criterion of if a computer could learn any problem and or solution to a problem that a human could... — m-theory
Ok. But from my biophysical/biosemiotic perspective, a theory of general intelligence just is a theory of life, a theory of complex adaptive systems. You have to have the essence of that semiotic relation between symbols and matter built in from the smallest, simplest scales to have any "intelligence" at all.
So yes, you are doing the familiar thing of trying to abstract away the routinised, mechanics-imitating, syntactical organisation that people think of as rational thought or problem solving. If you input some statement into a Searlean Chinese room or Turing test passing automaton, all that matters is that you get the appropriate output statement. If it sounded as though the machine knew what you were talking about, then the machine passes as "intelligent". — apokrisis
So again, fine, its easy to imagine building technology that is syntactic in ways that map some structure of syntax that we give it on to some structure of syntax that we then find meaningful. But the burden is on you to show why any semantics might arise inside the machine. What is your theory of how syntax produces semantics? — apokrisis
Biology's theory is that of semiotics - the claim that an intimate relation between syntax and semantics is there from the get-go as symbol and matter, Pattee's epistemic cut between rate independent information and rate dependent dynamics. And this is a generic theory - one that explains life and mind in the same physicalist ontology. — apokrisis
But computer science just operates on the happy assumption that syntax working in isolation from material reality will "light up" in the way brains "light up". There has never been any actual theory to back up this sci fi notion.
Instead - given the dismal failure of AI for so long - the computer science tendency is simply to scale back the ambitions to the simplest stuff for machines to fake - those aspects of human thought which are the most abstractly syntactic as mental manipulations.
If you just have numbers or logical variables to deal with, then hey, suddenly everything messy and real world is put at as great a distance as it can possibly be. Any schoolkid can learn to imitate a calculating engine - and demonstrate their essential humanness by being pretty bad, slow and error-prone at it, not to mention terminally bored. — apokrisis
Then we humans invent an actual machine to behave like a machine and ... suddenly we are incredibly impressed at its potential. Already our pocket calculators exceed all but our most autistic of idiot savants in rapid, error-free, syntactical operation. We think if a pocket calculator can be this unnaturally robotic in its responses, then imagine how wonderfully conscious, creative, semantic, etc, a next generation quantum supercomputer is going to be. Or some such inherently self-contradicting shit. — apokrisis
Second, I think there is a strong tendency to underrate animal (wet) intelligence. It isn't learning how to recite Beowulf from memory that is the only impressive human achievement. It's also remembering the odor of the room where we learned Anglo-Saxon and now feel nostalgia for that faint musty odor when we recite Beowulf, that's distinctive. [SEE NOTE] Dry intelligence can replay Beowulf, but it can't connect odors and texts and feelings. It can't feel. It can't smell. — Bitter Crank
Dry intelligence can't connect with the feelings of a dog excited by the walk it's about to take. Dry intelligence can't lay on the floor and determine whether the guy walking around is getting ready to go to work (alone) or is going to take the dog for a walk. Dogs can do that. They can tell the difference between routine getting ready to go to work and getting ready to go out of town (which the dog will probably disapprove of, considering what happened the last time "they" left). So can cats. — Bitter Crank
Wet brains and wet intelligence have developed over an exceedingly long time. Wet brains aren't the only defense animals have, but they are remarkably effective. A rat's wet brain does, and will, out-performs Deep Blue and all of it's Blue successors, Screwed Blue, Dude Blue, Rude Blue, etc. because it has capabilities that can be reproduced by an algorithm.
It's not the algorithm, it's the structure of the body and its history.
[NOTE] I never learned Anglo Saxon and I can't recite Beowulf. I can pretend I did, and even feel like I did. Betcha Deep Blue can't do that. — Bitter Crank
I argue that because this algorithm has to learn from scratch it must discover it's own semantics within the problem and solution to that problem. — m-theory
Consider the task of creating robot hand that is deleterious as the human hand. — m-theory
So again...this algorithm, if it does have semantic understanding...it does not and never will have human semantic understanding. — m-theory
Pattee's epistemic cut was not very clear to me, and he seems to have coined this term. — m-theory
That is the question. Does it actually learn its own semantics or is there a human in the loop who is judging that the machine is performing within some acceptable range? Who is training the machine and deciding that yes, its got the routine down pat? — apokrisis
The thing is that all syntax has to have an element of frozen semantics in practice. Even a Turing Machine is semantic in that it must have a reading head that can tell what symbol it is looking at so it can follow its rules. So semantics gets baked in - by there being a human designer who can build the kind of hardware which ensures this happens in the way it needs to.
So you could look at a neural network as a syntactical device with a lot of baked-in semantics. You are starting to get some biological realism in that open learning of that kind takes place. And yet inside the black box of circuits, it is still all a clicking and whirring of syntax as no contact with any actual semantics - no regulative interactions with material instability - are taking place.
Of course my view relies on a rather unfamiliar notion of semantics perhaps. The usual view is based on matter~mind dualism. Meaning is held to be something "mental" or "experiential". But then that whole way of framing the issue is anti-physicalist and woo-making.
So instead, a biosemiotic view of meaning is about the ability of symbol systems - memory structures - to regulate material processes. The presumption is that materiality is unstable. The job of information is to constrain that instability to produce useful work. That is what mindfulness is - the adaptive constraint of material dynamics.
And algorithms are syntax with any semantics baked in. The mindful connection to materiality is severed by humans doing the job of underpinning the material stability of the hardware that the software runs on. There is no need for instability-stabilising semantics inside the black box. An actual dualism of computational patterns and hotly-switching transistor gates has been manufactured by humans for their own purpose. — apokrisis
Yes. But the robot hand is still a scaled-up set of digital switches. And a real hand is a scaled-up set of molecular machines. So the difference is philosophically foundational even if we can produce functional mimickry.
At the root of the biological hand is a world where molecular structures are falling apart almost as soon as they self-assemble. The half-life of even a sizeable cellular component like a microtubule is about 7 minutes. So the "hardware" of life is all about a material instability being controlled just enough to stay organised and directed overall.
You are talking about a clash of world views here. The computationalist likes to think biology is a wee bit messy - and its amazing wet machines can work at all really. A biologist knows that a self-organising semiotic stability is instrincally semantic and adaptive. Biology know itself, its material basis, all the way down to the molecules that compose it. And so it is no surprise that computers are so autistic and brittle - the tiniest physical bug can cause the whole machine to break-down utterly. The smallest mess is something a computer algorithm has no capacity to deal with.
(Thank goodness again for the error correction routines that human hardware designers can design in as the cotton wool buffering for these most fragile creations in all material existence). — apokrisis
But the question is how can an algorithm have semantic understanding in any foundational sense when the whole point is that it is bare isolated syntax?
Your argument is based on a woolly and dualistic notion of semantics. Or if you have some other scientific theory of meaning here, then you would need to present it. — apokrisis
Pattee has written a ton of papers which you can find yourself if you google his name and epistemic cut.
This is one with a bit more of the intellectual history.... http://www.informatics.indiana.edu/rocha/publications/pattee/pattee.html
But really, Pattee won't make much sense unless you do have a strong grounding in biological science. And much of the biology is very new. If you want to get a real understanding of how different biology is in its informational constraint of material instability, then this a good new pop sci book.... — apokrisis
I have read some more and you are right he is very technically laden. — m-theory
I was hoping for a more generalized statement of the problem of the epistemic cut because I believe that the Partially observable Markov decision process might be a very general solution to establishing an epistemic cut between the model and the reality in an A.I. agent. — m-theory
What I see as his main issue is that he believes there is something like the measurement problem when dealing with the origin of life.Right. Pattee requires you to understand physics as well as biology. ;) But that is what makes him the most rigorous thinker in this area for my money. — apokrisis
Good grief. Not Mr Palm Pilot and his attempt to reinvent Bayesian reasoning as a forward modelling architecture? — apokrisis
But I don't agree that we have to solve the origin of life and the measurement problem to solve the problem of general intelligence. — m-theory
I suppose if you want to argue that the mind ultimately takes place at a quantum scale in nature then Pattee may well be correct and we would have to contend with the issues surrounding the measurement problem. — m-theory
What is wrong with bayesian probability I don't get it either? — m-theory
Great. So in your view general intelligence is not wedded to biological underpinnings. You have drunk the Kool-Aid of 1970s cognitive functionalism. When faced with a hard philosophical rebuttal to the hand-waving promises that are currency of computer science as a discipline, suddenly you no longer want to care about the reasons AI is history's most over-hyped technological failure. — apokrisis
That is nothing like what I suggest. Instead I say "mind" arises out of that kind of lowest level beginning after an immense amount of subsequent complexification.
The question is whether computer hardware can ever have "the right stuff" to be a foundation for semantics. And I say it can't because of the things I have identified. And now biophysics is finding why the quasi-classical scale of being (organic chemistry in liquid water) is indeed a uniquely "right" stuff. — apokrisis
Well I think I get it...Pattee argues that life may be like a unique state of matter at the quantum scale and we just might not be able to tell because of the measurement problem (I know it is much more complicated then that I just could not think of a better analogy for breviaries sake).I explained this fairly carefully in a thread back on PF if you are interested....
http://forums.philosophyforums.com/threads/the-biophysics-of-substance-70736.html
So here you are just twisting what I say so you can avoid having to answer the fundamental challenges I've made to your cosy belief in computer science's self-hype. — apokrisis
I thought you were referring to the gaudy self-publicist, Jeff Hawkins, of hierarchical temporal memory fame - https://en.wikipedia.org/wiki/Hierarchical_temporal_memory
But Bayesian network approaches to biologically realistic brain processing models are of course what I think are exactly the right way to go, as they are implementations of the epistemic cut or anticipatory systems approach.
Look, it's clear that you are not even familiar with the history of neural networks and cybernetics within computer science, let alone the way the same foundational issues have played out more widely in science and philosophy.
Don't take that as in insult. It is hardly general knowledge. But all I can do is point you towards the arguments.
And I think they are interesting because they are right at the heart of everything - being the division between those who understand reality in terms of Platonic mechanism and those who understand it in terms of organically self-organising processes. — apokrisis
That is the bottom up approach.
We are reverse engineering from the top down as you pointed out.
And I believe that somewhere in the middle is where the mind breakthrough will happen. — m-theory
That is a good point maybe you are right.So you hope to discover the software by examining the hardware? The trouble is, since we don't know what we're looking for, how could we recognise it? — tom
Back to epistemology. If we want to create an AGI then the problem of how to create knowledge will have to be solved. You can't transfer knowledge from one mind to another. Instead one mind creates cultural artefacts, from which the other mind discerns something not contained within the artefact - its meaning. As Karl Popper said, "It is impossible to speak in such a way that you cannot be misunderstood. This by the way, dispenses with the Chinese Room. — tom
I am not so sure.t has been suggested that the human brain evolved the way it did in order to facilitate efficient knowledge transfer. Humans are unique (i.e. they are the last remaining species) in that they interpret meaning and intention - i.e. they create knowledge from artefacts and behaviours.
Now, here's the amazing thing if this account of our evolutionary history is true: once you can create knowledge, there is no stopping you. This is a leap to universality. Once you are an explainer you are automatically a universal explainer because the same mechanisms are involved.
Prior to the leap to universal explainer, there must have been another leap - the leap to computational universality in the human brain. This is a hardware problem, which we have long solved! — tom
If we had a thinking machine that interacted with humans there is no reason to assume it would not be able to communicate with the conventions humans use. — m-theory
I am not so sure.
It could be that the brains software became more efficient to and that it is not strictly a hardware leap. — m-theory
And I believe that somewhere in the middle is where the mind breakthrough will happen.
I believe this because a great deal of what the body and brain do is completely autonomous from the mind...or at least what we mean by the term mind. — m-theory
For this reason I think simulations of thought do not have to recreate the physics of biology at the nano scale before a mind can be modeled. — m-theory
I just don't agree that intelligence is necessarily dependent upon that state.
I don't see why computers can not be the "right stuff" as you put it.
Pattee does not provide conclusive evidence that such is the case.
And you haven't either. — m-theory
Nice example of misunderstanding a cultural aretfact. — tom
And again it seems. The leap to computational universality (the hardware problem) is fully understood. The leap to universal explainer (the software problem) is not understood. — tom
Do you mean a dualistic folk psychology notion of mind? I instead take the neurocognitive view that what you are talking about is simply the difference between attentive and habitual levels of brain processing. And these are hardly completely autonomous, but rather completely interdependent. — apokrisis
But the burden of proof is on you here. The only sure thing is that whatever you really mean by intelligence is a product of biology. And so biological stuff is already known to be the right stuff. — apokrisis
This misrepresents my argument again. My argument is that there is a fundamental known difference between hardware and wetware as BC puts it. So it is up to you to show that this difference does not matter here. — apokrisis
That would be why it seems easy to work from the top down. Computers are just mechanising what is already us behaving as if we were mechanical. But as soon as you actually dig into what it is to be a biological creature in an embodied relation with a complex world, mechanical programs almost immediately break down. They are the wrong stuff.
Neural networks buy you some extra biological realism. But then you have to understand the detail of that to make judgements about just how far that further exercise is going to get. — apokrisis
We might disembody a head and sustain the life of the brain without a body by employing machines.
Were we to do so we would not say that this person has lost a significant amount of their mind.
Would we? — m-theory
My notion was that we might hope to model something like the default mode network. — m-theory
If you state that the origins of life must be understood in order that we understand the mind that is claim that entails burdens of proof. — m-theory
The main issue at hand is whether or not computational theory of the mind is valid.
Not whether or inorganic matter can compute. — m-theory
That is irrelevant because you are talking about an already fully developed biology. The neural circuitry that was the result of having a hand would still be attempting to function. Check phantom limb syndrome.
Then imagine instead culturing a brain with no body, no sense organs, no material interaction with the world. That is what a meaningful state of disembodiment would be like. — apokrisis
I was talking about the biological basis of the epistemic cut - something we can examine in the lab today. — apokrisis
We also know that matter can compute...surely I am not expected to prove as much?Again, we know that biology is the right stuff for making minds. You are not expecting me to prove that? — apokrisis
And we know that biology is rooted in material instabilty, not material stabilty? I've given you the evidence of that. And indeed - biosemiotically - why it has to be the case. — apokrisis
And I've made the case that computation only employs syntax. It maps patterns of symbols onto patterns of symbols by looking up rules. There is nothing in that which constitutes an understanding of any meaning in the patterns or the rules? — apokrisis
So that leaves you having to argue that despite all this, computation has the right stuff in a way that makes it merely a question of some appropriate degree of algorithmic complication before it "must" come alive with thoughts and feelings, a sense of self and a sense of purpose, and so you are excused of the burden of saying just why that would be so given all the foregoing reasons to doubt. — apokrisis
Or it could be a nice example of a poorly constructed artifact.
But I will assume the fault lies with me...and hope you can forgive that. — m-theory
It could be that the brains software became more efficient to and that it is not strictly a hardware leap. — m-theory
Notice that subjectivity has already appeared! AlphaGo has no subjectivity. — tom
You are completely missing the point. It is impossible to transfer knowledge from one mind to another. Minds construct new knowledge from artefacts, problem-situations, background knowledge, by a fundamentally creative ability.
So, the creator of the artefact, and the interpreter of the artefact, are engaged in an inter-subjective dialogue. Each person is conjecturing theories about what each other means or interprets. Perfection and justification are impossible. — tom
AlphaGo can be as efficient as it likes. It will always fail the Chinese Room. It cannot create the knowledge that it is playing Go! — tom
Of course I disagree that the mind must necessarily always be biological...but that is a semantic debate surrounding how the term is defined.
You have decided that the term mind must be defined biologically to the exclusion of a computational model. — m-theory
Yes and as far as I could tell from your source material it was claimed that the origin of life contains a quantum measurement problem.
The term epistemic cut was used synonymously with the quantum measurement problem and the author continuously alluded to the origins of self replicating life. — m-theory
Imagine if the body and brain had a sudden interruption in the supply of electrons within its neurological system?
Biology is not without stability. — m-theory
I don't agree semantics can only occur in biology. — m-theory
Again I refer to the alternative of a undecidable mind.
We could not know if we had one if the mind is not algorithmic it is that simple.
If we can know without error that we have minds this is the result of some algorithm which means the mind is computational. — m-theory
2) The best theory of that what kind of stuff that actually is, is what you would expect biologists to produce. And the standard answer from biologists is biology is material dynamics regulated by semiotic code - unstable chemistry constrained by evolving memory. Agreed? — apokrisis
3) Then the question is whether computation is the same kind of stuff as that, or a fundamentally different kind of stuff. And as Pattee argues (not from quantum measurement, but his own 1960s work on biological automata), computation is physics-free modelling. It is the isolated play of syntax that builds in its presumption of being implementable on any computationally suited device. And in doing that, it explicitly rules out any external influences from the operation of physical laws or dissipative material processes. Sure there must be hardware to run the software, but it is axiomatic to universal computation that the nature of the hardware is irrelevant to the play of the symbols. Being physics-free is what makes the computation universal. Agreed? — apokrisis
from von Neumann (1955, p. 352). He calls the system being measured, S, and the measuring device, M, that must provide the initial conditions for the dynamic laws of S. Since the non-integrable constraint, M, is also a physical system obeying the same laws as S, we may try a unified description by considering the combined physical system (S + M). But then we will need a new measuring device, M', to provide the initial conditions for the larger system (S + M). This leads to an infinite regress; but the main point is that even though any constraint like a measuring device, M, can in principle be described by more detailed universal laws, the fact is that if you choose to do so you will lose the function of M as a measuring device. This demonstrates that laws cannot describe the pragmatic function of measurement even if they can correctly and completely describe the detailed dynamics of the measuring constraints.
Given the above - that biological stuff is fundamentally different from computational stuff in a completely defined fashion - the burden is then on the computationalist to show that computation could still be the right stuff in some way. — apokrisis
This is another unhelpful idee fixe you have developed. As said, Pattee's theoretical formulation of the epistemic cut arose from being a physicist working on the definition of life in the 1950s and 1960s as DNA was being discovered and the central mechanism of evolution becoming physically clear. From von Neumann - who also had an interest in self-reproducing automata - Pattee learnt that the epistemic cut was also the same kind of problem as had been identified in quantum mechanics as the measurement problem. — apokrisis
The epistemic cut or the distinction between subject and object is normally associated with highly evolved subjects with brains and their models of the outside world as in the case of measurement. As von Neumann states, where we place the cut appears to be arbitrary to a large extent. The cut itself is an epistemic necessity, not an ontological condition. That is, we must make a sharp cut, a disjunction, just in order to speak of knowledge as being "about" something or "standing for" whatever it refers to. What is going on ontologically at the cut (or what we see if we choose to look at the most detailed physics) is a very complex process. The apparent arbitrariness of the placement of the epistemic cut arises in part because the process cannot be completely or unambiguously described by the objective dynamical laws, since in order to perform a measurement the subject must have control of the construction of the measuring device. Only the subject side of the cut can measure or control.
Fine. Now present that evidence. — apokrisis
No idea what you are talking about here — apokrisis
No of course I don't agree that the best theory of the mind must be biological. — m-theory
I offered the that the pomdp could be a resolution.
You did not really bother to suggest any reason why that view was not correct. — m-theory
Mind is only found in living organic matter therefor only living organic matter can have a mind.
That is an unasailable argument in that it defines the term mind to the exclusion of inorganic matter.
But that this definition is by necessity the only valid theory of the mind is not simply a resolved matter in philosophy. — m-theory
It is not immediately clear to me how this general statement can be said to demonstrate necessarily that computation can not result in a mind. — m-theory
My argument is on the first page below Tom's post. — m-theory
Get involved in philosophical discussions about knowledge, truth, language, consciousness, science, politics, religion, logic and mathematics, art, history, and lots more. No ads, no clutter, and very little agreement — just fascinating conversations.