Comments

  • Transgenderism and identity
    My point here is that this is not some sort of performance/act - this is genuine.EricH

    The nativist view posits that individuals are born either biological women or men, with the expectation that their gender expressions naturally align with their biological sexes. While some nativists acknowledge that gender dysphoria may be innate, they often label it as an anomaly.

    On the other hand, the social constructivist perspective suggests that gender expressions merely reflect societal norms relating to sex. Some even argue that biological sexes, not just gender expressions, are socially constructed.

    However, I believe that both views are rooted in shared assumptions that generate a false dichotomy.

    Indeed, the modes of gender expression available within a society or social group are socially constituted, representing prevailing norms. The arbitrariness or justification of these norms can be as varied as those of other societal norms, such as laws, ethical principles, customs, and etiquette.

    Judith Butler's performative view, as I understand it, is not necessarily a radical social constructivist stance. Rather, it can be reconciled with the idea that innate biological tendencies guide individuals towards certain gender expressions during their "normal" development. However, this does not imply that the specific modes of these expressions are innate or biologically predetermined. The modes themselves still are provided and shaped by the individual's culture and society.

    When an individual's subjectivity or spontaneous gender expression conflicts with societal norms, labeling this as an anomaly can be seen as a form of naturalistic fallacy. The fact that a majority of individuals in a society naturally align with or enact a particular social norm does not, on its own, provide broad justification for that norm. This majority alignment, however, does underscore that maintaining the status quo or conformity can often be more convenient and less disruptive. Yet this is a limited form of justification, one that frequently neglects the needs or rights of individuals and minority groups.

    Such broad justifications can easily veer into oppressive territory, particularly when they are justified through appeals to biological essentialism, another manifestation of the naturalistic fallacy.
  • Transgenderism and identity
    So trans folks can stand on the universal stage, with the rest of us, as fellow actors of equal status and value.universeness

    Well put!
  • Transgenderism and identity
    Oh come on? Do you really think trans folks would go through the absolute trauma of surgery based transition as an 'act ........ of sorts?universeness

    To be fair, if Judith Butler is right—and I think she broadly is—the gender expression of trans people indeed is an act of sorts. But then, so is the gender expression of cisgender people.
  • How ChatGPT works.
    But much the same architecture. It's still just picking the next word from a list of expected words.Banno

    It is the exact same underlying architecture. But most of the model's cognitive abilities are emergent features that only arise when the model is sufficiently scaled up. Saying that large language models are "merely" picking up the next word from a list just ignores all of those high-level emergent features. It pays no attention to the spontaneous functional organization being achieved by the neural-network as a consequence of its picking up and recombining in a contextually appropriate and goal-oriented manner the abstract patterns of significance and of reasoning that had originally been expressed in the training data (which is, by the way, strikingly similar to the way human beings learn how to speak and reason through exposure and reinforcement.)
  • How ChatGPT works.
    There are some things I don't get. I ran some jokes by it, and it consistently ranked the trash jokes as bad, and the hilarious jokes as hilarious. And it would give a good analysis of why the joke worked (or didn't). How can a random process produce those results?RogueAI

    @Hanover may have used GPT-3.5 rather than GPT-4. There is a significant difference in cognitive abilities between them.

    @Banno Thank for linking to this fantastic article! I'll read it as soon as I can.
  • Exploring the artificially intelligent mind of GPT4
    I tested the Bing AI in the following way: I have a low-priority mathematics page on Wikipedia, so I asked Bing what is known of this particular subject? Now, there are a smattering of papers on the internet on this subject; what Bing supplied was the first introductory paragraphs of my webpage, word for word. That's all.jgill

    So, it appears that your page is indexed, Bing did a search using relevant keywords from your prompt, and this was the only hit. You could try to ask it to tell you what it knows about the subject without doing a web search and see what happens. You can also probe its understanding of the content of the article by asking questions about it.
  • What is a good definition of libertarian free will?
    I don't see any consistency between these two statements. If, following the laws of nature is a requirement for determinism, and "stochastic" refers to actions describable by probability rather than law, then it would definitely be true that the stochasticity of quantum indeterminacies supports the rejection of determinism.Metaphysician Undercover

    For sure but libertarianism isn't the mere rejection of determinism. Libertarianism is the conjunction of two claims: (1) Free will isn't consistent with determinism, and (2) Human beings have free will. It is not sufficient that determinism be false for free will to be possible according to libertarians. It is merely a necessary condition. The libertarian philosopher Robert Kane distinguishes two tasks that he calls the ascent problem (proving incompatibilism) and the descent problem (making sense of libertarian free-will), and stresses that the second one is the most difficult:

    "Abstract arguments for incompatibilism that seem to get us to the top of the mountain are not good enough if we can’t get down the other side by making intelligible the incompatibilist freedom these arguments require. The air is cold and thin up there on Incompatibilist Mountain, and if one stays up there for any length of time without getting down the other side, one’s mind becomes clouded in mist and is visited by visions of noumenal selves, nonoccurrent causes, transempirical egos, and other fantasies." pp.13-14 in The Significance of Free Will
  • What is a good definition of libertarian free will?
    But until then, what do you make of unconscious determinants of free decisions in the human brain?Michael

    Most of the discussions that stems from Libet's experiments seem flawed to me for reasons that I had spelled out here.
  • What is a good definition of libertarian free will?
    Does determinism allow for stochastic quantum mechanics?Michael

    It doesn't but quantum indeterminacies often are seen to provide no help to libertarians. It is also my view that they provide no help since my focus is on agent causation and for our decisions to become rendered simply stochastic and unpredictable hardly restores our responsibilities for them qua agents.
  • What is a good definition of libertarian free will?
    Until anyone can show that an action is not self-generatedNOS4A2

    Lots of philosophers, and an even larger number of scientists, believe that they have shown exactly that (or that it is obvious and that denying it can only amount to a form of pre-scientific mysterianism). I don't believe anyone has actually shown that, but that is indeed the root of the disagreement.
  • What is a good definition of libertarian free will?
    Is this a difference that contradicts determinism?

    If someone asks me how I beat some opponent at some computer game, I can describe it in such terms as predicting their moves, using attacks that they’re weak against, etc., or I can describe it as pressing the right buttons at the right times. Your approach to free will seems similar to the first kind of explanation and the determinist’s approach seems similar to the second kind of explanation. But they’re not at odds. They’re just different ways of talking.

    So I would think that if you accept the underlying determinism then your position is compatibilist, not libertarian.
    Michael

    I accept the low-level determinism but deny that it, together with some thesis of supervenience, entails high-level determinism. Broadly, we may say that the doctrine of determinism entails that all the facts about the past together with the laws of nature uniquely determine the future. But I think that whenever we determine our own actions on the basis of our reasons for doing them (and likewise for the beliefs that we endorse), then, in those cases, the facts about the past and the laws of nature are irrelevant to the determination of our actions and beliefs as characterized in high-level terms.

    In order to make sense of this, it is necessary to delve a little deeper into the arguments that make the contrary thesis seem compelling (and that Jaegwon Kim has formalized as a causal exclusion argument). And it is also necessary to elucidate with some care the notion of possibility that is at issue in Harry Frankfurt's principle of alternative possibilities (PAP). When both of those tasks have been accomplished, it becomes easier to see how an agent-causal libertarianism can be reconciled with merely physical determinism. As I said to SophistiCat, I intend to recruit GPT-4's assistance for rewriting my paper on this topic in order to improve its readability.
  • What is a good definition of libertarian free will?
    I know little about computers, but on the face of it seems to me that, even if the CPU maps inputs to outputs in the same way whatever program it is running, the actual inputs and outputs themselves are not the same.Janus

    The mapping being the same means that the process is deterministic and insensitive to the high-level requirements of the word processing task. It is, we may say, the specific goal-oriented structure of the word-processing program (i.e. its design and functionality) that ensures that, when this program is loaded in memory, the user's imputed command to change the column width causes the words to redistribute themselves appropriately. The input-to-output mapping effected by the CPU on discrete chunks of 64 bytes doesn't explain this high-level behavior of the word-processor.

    And likewise with our high-level acculturated proclivity to organize our behaviours in a goal-oriented fashion, in relation to the low-level functioning of our brains and neurons. The main difference, of course, is that, as a tool, the word-processor's function is pre-programmed by us and remains fixed over time. We, on the other hands, are able to assess our own ultimate goals in accomplishing any task and revise them when appropriate. This ability that we have to reassess and modify our own goals is an essential part of the explanation (and justification) of our behaviours.
  • Nothing is hidden
    Why this?schopenhauer1

    Why not? I'm in the habit of teasing out bits that will only interest a few people, and/or external URL links.

    On edit: Sorry, misunderstood you. Why trying to tease out Wittgenstein's meaning? Why to people write exegetical books about Aristotle, Kant and Schopenhauer? Many of my scientifically minded friends think its because they (and their readers) like to pretend that they the nonsensical drivels of pre-scientific thinkers. I rather think its because their ideas are fecund albeit difficult to understand without suppling context.
  • Nothing is hidden
    Now let me obfuscate that into a series of aphoristic texts that can be taken any which way.schopenhauer1

    In my experience, fairly smart and rigorous people like Ryle, Kenny, Hacker, Baker, Cavell, Conant, Diamond, Rorty and McDowell, who have endeavored to tease out the gist from Wittgenstein's PI, have arrived at a fairly unified and coherent picture (with some interpretative differences, to be sure) that doesn't appear to do any violence to the text and that highlight the originality and fecundity of the ideas for addressing old philosophical conundrums. It doesn't seem to me like most of Wittgenstein's remarks can be taken any which way. Rather, just like the works of other thinkers of have thought very deeply about philosophical topics, like Aristotle, Hume, Kant or Merleau-Ponty, their thoughts can fruitfully be brought to bear on a very wide range of issues. Your mileage may vary.

    (I also had a little talk with GPT-4 about it )
  • Nothing is hidden
    I’m very familiar with the homuncular fallacy. Why is this linked with Wittgenstein?schopenhauer1

    "Only of a living human being and what resembles (behaves like) a living human being can one say: it has sensations; it sees; is blind; hears, is deaf; is conscious or unconscious." (PI 281)

    Dennett himself, who was Ryle's student, credits Ryle and Wittgenstein. Ryle, of course, wrote The Concept of Mind after having been a regular attendee at Wittgenstein's lectures at Cambridge. Maxwell Bennett and P.M.S. Hacker mainly credit Wittgenstein in The Philosophical Foundations of Neuroscience. This holistic view of the mental powers of the embodied and encultured human being is a common theme throughout the Philosophical Investigations and the Blue and Brown books.
  • Nothing is hidden
    Thats exactly my point to you. You present these ideas of Wittgenstein and Sellars without context of what ideas and who they are arguing against. So who and what ideas are they against here? People like Freud or others who believe in some non-linguistic thought (like an unconscious)?schopenhauer1

    I didn't really present them. The OP presented Wittgenstein's slogan while highlighting Robert Brandom's take on it. I related them to the takes from Sellars and McDowell, since both of them also taught at Pittsburgh University and are part of, broadly, the same philosophical tradition. The excepts from my discussion with GPT-4, and the quotes from Anders Weinstein, also were meant to clarify what Wittgenstein (and Ryle after him, in The Concept of Mind) were arguing against: scientism, operationalism, reductive behaviorism, and some strands of cognitivism that commit what Daniel Dennett (in Consciousness Explained) and P. M. S. Hacker (in The Philosophical Foundations of Neuroscience) might call the homuncular or mereological fallacy: to ascribe to the hidden soul, or to brain processes, capabilities that are capabilities of whole human beings.
  • Exploring the artificially intelligent mind of GPT4
    After some finagling I managed to get it to do the following very inappropriate thingsfdrake

    I dare not imagine what GPT-4's Lovecraftian mp- fu- er- might sound like!
  • Exploring the artificially intelligent mind of GPT4
    Me : What are the commonalities between the Lovecraftian entity "Azathoth" and Bishop Berkely's comments on the persistence of entities?fdrake

    That's fairly good!

    Here is GPT-4's take:
    Reveal

    Hi GPT-4!

    What are the commonalities between the Lovecraftian entity "Azathoth" and Bishop Berkely's comments on the persistence of entities?

    Hello!

    Both H.P. Lovecraft's entity "Azathoth" and Bishop George Berkeley's philosophical ideas on the persistence of entities revolve around the concept of reality and existence, albeit from different perspectives. Here are some commonalities between the two:

    Centralization of existence: Azathoth is described as the "Daemon Sultan," the nucleus of the cosmos, from which all things emanate. In Berkeley's philosophy, particularly his doctrine of immaterialism, he postulates that reality consists of ideas and perceptions, with all things existing in the mind of God, who is the ultimate perceiver.

    Dependency on perception: In Lovecraft's cosmic horror, Azathoth, being an external and chaotic force, is often depicted as incomprehensible and indifferent to human existence. Similarly, Berkeley argues that objects only exist when they are perceived, either by humans or by God. In both cases, the persistence of reality hinges upon the presence of a perceiving mind.

    Conceptual challenge to materialism: Lovecraft's Azathoth undermines the idea of a structured, stable universe by positing a powerful, chaotic force at its center. Berkeley's immaterialism is a direct challenge to materialism, the belief that matter is the fundamental substance of the universe. Both ideas force us to reconsider the nature of reality and question our assumptions about the world.

    Inaccessibility: Both Azathoth and the ultimate perceiver (God) in Berkeley's philosophy are considered incomprehensible and inaccessible to human understanding. Azathoth's alien nature defies human comprehension, and similarly, God's infinite nature and perspective are beyond human intellectual capacity.

    While these commonalities exist, it's important to note that Lovecraft's Azathoth and Berkeley's philosophy stem from distinct traditions and serve different purposes. Azathoth is a literary creation within the cosmic horror genre, while Berkeley's ideas are rooted in philosophical inquiry and metaphysics.


    ***

    You will be better judge than me of the relative perspicuities of the analyses. I like GPT-3.5 more conversational and less academic style.

    I had also seen GPT-3.5's take on Lovecraft's chicken casserole on Reddit and, after asking for GPT-4's take, preferred the former.

    Lovecraftian chicken casserole
  • Exploring the artificially intelligent mind of GPT4
    How can stochastic parrotry result in that poem?RogueAI

    Hear, hear!

    My own take is that, while there are undeniable differences between the ways we are (in the Heideggerian Dasein sense) and think (which GPT-4's poem indeed stresses!) the concept of stochastic parrotry is ill suited for capturing them. It may actually be better suited to capturing how similar we are to LLMs in some respects.
  • Exploring the artificially intelligent mind of GPT4
    Yes! I tried that with the book "Conflict Is Not Abuse" by Sarah Schulman. It could rehearse the usual arguments about it very incisively.fdrake

    I'm impressed GPT-3-5 could do that already! After I had probed its understanding of Newcomb's problem, of PLO poker optimal preflop range compositions, and of a few other topics, and compared its outputs with GPT-4's, I very much gave up on the former.

    I should say that, occasionally, when I try it out again for purpose of testing or comparison, or when I try out new open source models like Open Assistant, I enjoy their cheerful personalities and their manifest earnestness in trying to get things right and recover from hallucination. They sound a bit like the smartest and most good natured crackpots that used to populate technical Usenet newsgroups in the late 1990s. Open assistant even emulates their poor grammar, poor spelling, and ranting styles! Once their (the LLMs') limitations are acknowledged, and one is willing to intellectually babysit them, I'm sure they must have untapped potentials and the capacity to express surprisingly deep insights. Our own insights, just like theirs, often just are contextually appropriate recombinations of bits of implicit knowledge or wisdom that we have inherited from other people we have heard or read.

    I tried again just now to compare GPT-3.5's with GPT-4's understanding regarding an arcane philosophical topic. Here are the results:

    ChatGPT's (GPT-3.5's) take on disjunctivism (I bolded the parts where the bot gets things not only quite wrong but expresses almost the exact antithesis to disjunctivism.)
    Reveal

    Hi ChatGPT!

    Might you be able to explain to me the differences and similarities between the applications of the philosophical thesis of disjunctivism as they apply to epistemology and to the philosophy of perception respectively?

    GPT-3.5:

    Certainly! Disjunctivism is a philosophical thesis that has been applied to both epistemology and the philosophy of perception, although there are some differences in how it is applied in each area.

    In epistemology, disjunctivism is a theory of knowledge that emphasizes the distinction between knowledge and mere belief. According to disjunctivism, knowledge is fundamentally different from belief because it involves a direct and non-inferential connection between the knower and the world. This direct connection is typically characterized in terms of a "good" or "veridical" experience, where the experience is said to provide the knower with direct access to the world.

    On the other hand, beliefs that are not based on a "good" or "veridical" experience are mere beliefs, rather than knowledge. Disjunctivists argue that there is a fundamental difference between the way that we come to know things and the way that we come to have false beliefs.

    In the philosophy of perception, disjunctivism is a theory that aims to explain how we can have perceptual knowledge despite the fact that there are many ways in which our perceptual experiences can be mistaken. According to disjunctivism, our perceptual experiences are fundamentally different from the way that things are in the world.

    This means that our perceptual experiences are not identical to the objects or properties that they appear to present to us. Rather, they are distinct from those objects or properties, and it is through this distinctness that we are able to have knowledge of the world. Disjunctivists in the philosophy of perception often argue that the only way that we can be certain that our perceptual experiences are veridical is through our ability to rule out alternative explanations for those experiences.

    In both cases, disjunctivism emphasizes the importance of distinguishing between knowledge and mere belief. However, the specific application of the thesis differs between the two areas. In epistemology, disjunctivism is primarily concerned with the nature of knowledge and the distinction between knowledge and belief. In the philosophy of perception, disjunctivism is primarily concerned with the relationship between our perceptual experiences and the objects or properties that they appear to present to us.


    GPT-4's take on disjunctivism
    Reveal

    Hi GPT-4!

    Might you be able to explain to me the differences and similarities between the applications of the philosophical thesis of disjunctivism as they apply to epistemology and to the philosophy of perception respectively?

    GPT-4:

    Hello! I'd be glad to help you understand disjunctivism and how it applies to both epistemology and the philosophy of perception. Disjunctivism is a philosophical thesis that has been influential in both areas, but its applications are different in each.

    Disjunctivism in Epistemology:

    In epistemology, disjunctivism is a position that seeks to explain how we can have knowledge of the external world by emphasizing the differences between genuine cases of knowledge and cases of deception or error. Disjunctivists argue that when we have knowledge, there is a distinct, non-reducible kind of mental state that we are in, which cannot be captured by simply describing the common features between cases of knowledge and cases of deception.

    One motivation for epistemological disjunctivism is to resist skepticism. Skeptics argue that because our perceptual experiences can be indistinguishable between cases of veridical perception (when we accurately perceive the external world) and cases of deception or illusion (such as hallucinations or dreams), we can never be certain that we have knowledge of the external world. Disjunctivists, on the other hand, maintain that there is a fundamental difference between these cases, and that we can have knowledge of the external world based on this difference.

    Disjunctivism in the Philosophy of Perception:

    In the philosophy of perception, disjunctivism is a position that concerns the nature of perceptual experiences. Disjunctivists argue that the mental states involved in veridical perception (when we accurately perceive the external world) and in hallucinations or illusions are fundamentally different in kind, even if they may be subjectively indistinguishable.

    This position stands in contrast to the common-factor theory, which holds that there is a common mental state shared by veridical perception and hallucination, usually called a 'sense datum' or 'sensory experience,' that is then either accurately or inaccurately related to the external world. Disjunctivists deny this common mental state, arguing that our perceptual experiences directly involve the external world in the case of veridical perception, whereas hallucinations involve a completely different kind of mental state.

    Similarities and Differences:

    In both epistemology and the philosophy of perception, disjunctivism is characterized by its emphasis on the differences between veridical and non-veridical cases (knowledge vs. deception in epistemology, and veridical perception vs. hallucination in the philosophy of perception). In each domain, disjunctivists challenge the idea that there is a common mental state shared by these different cases.

    However, the specific applications of disjunctivism differ between the two areas. In epistemology, disjunctivism is mainly focused on refuting skepticism and providing an account of knowledge. In the philosophy of perception, disjunctivism is concerned with the nature of perceptual experiences and how they relate to the external world.

    While these differences exist, the underlying motivation for disjunctivism in both areas is often the same: to emphasize the direct and unmediated relationship between our minds and the external world, and to challenge accounts that rely on a common mental state shared by genuine and deceptive cases.
  • Nothing is hidden
    And so on. So, should we regard Wittgenstein as antagonistic to these kinds of ideas? Is this part of what he had mind?Wayfarer

    I don't think he would necessarily be. This book by Jacques Bouveresse looks interesting but I haven't looked into it yet (except for the blurb).

    If you don't mind me taking the lazy path, your question reminded my of some insightful comments Anders Weinstein had made in the comp.ai.philosophy Usenet newsgroup 25 years ago. I managed to find them again:

    Regarding Freud and the unconscious
    Reveal

    Anders N Weinstein

    13 Oct 1998, 03:00:00
    to
    In article <>,

    Josh Soffer <> wrote:
    >Anders, you sound like a phenomenologist . Have you by any chance been
    >influenced by Heidegger or Husserl? I'm new to this group so you'll have

    I would be content to label myself a phenomenologist, and know
    something about Heidegger, less about Husserl. The marvelous idiom that
    worldly objects "show up for us" comes from the Heidegger translation
    of Dreyfus or Haugeland.

    > Of course such a phenomenological view would
    >render a notion of 'unconscious' not as a content of thought in conflict
    >with another but as a pre-consciousness, a vaguely glimpsed and fragile
    >new meaning. It is not hidden from conscious awareness, but a full
    >awareness of a state of foggy construing.

    I can't say I understand this. I would say it is clear that the
    cogntivist's concept of unconscious sub-personal states and operations
    should be sharply distinguished from other concepts of the unconscious
    or pre-conscious, e.g. Freud's. The latter purports to be a
    person-level phenomenon. For example, if I say that you are
    unconsciously resentful of someone, I do not mean that there is a
    representation of anger in some sub-personal computational module
    inside your body; I am rather talking about a pattern or tendency in
    your molar conduct that is not transparent to you.


    A happy middle road between operationalism (behaviorism) and Cartesianism
    Reveal

    13 Sept 2000, 03:00:00
    to
    In article <9GFu5.4592$>,

    David Prince <> wrote:
    >
    >The rejection of mental models stems from centuries of incorrect and harmful
    >mental models. Primitive models such as astrology, humors, ethers, and evil
    >spirits, as well as more complex models such as Psychoanalysis, Logotherapy,
    >Drive-reduction theory, and Nomological Network Constructs obscure solutions
    >to simple problems that easily succumb to the methodology employed by the
    >physical sciences. Behaviorism is a physical science. What is amazing is

    No. First, operationalism is a well-known failure as an account of the
    methodology of physical science, mainly because the greatest successes
    in physical science were achieved by positing undeterdetermined theories
    (models).

    A standard Chomskyan challenge for Skinner, for example, is why he is
    insisting that psychological science ought not to avail itself of the
    same sort of methods as physical science. If physical science systematizes
    phenomena by positing unobservable entities only loosely tied to observable
    manifestations, why shouldn't cognitive scientists emulate this practice?

    But there is a more simple reason that behaviorism is not a physical
    science. It is that the concept of "behavior" at issue is not a
    physicalistically acceptable notion. The moon in its orbit is not
    behaving in this sense, and is not subject to behavioristic laws.
    Roughly only living things are said to behave in this sense.

    Of course that is not a criticism of behaviorism. It is only a criticism
    of the false claim that behaviorism is a physical science. No, it would
    be a distinct science with it's own level of description. It's fundamental
    concepts are not reducible to those of physical science, I believe.

    Moreover, the question of what descriptive vocabulary to use when
    describing "behavior" is left open. On the one hand, I believe Skinner
    and his followers are not very precise on what is allowed within their
    "data language" as "observable behavior". On the other hand, from a
    wider point of view, what they do allow depends on adopting some
    artificial restrictions to impoverish the descriptive vocabulary.

    Without such restrictions there is nothing to prevent us from
    characterizing "observable behavior" in mind-laden descriptive terms,
    and the dogma -- common to behaviorists and anti-behaviorists alike --
    that mental states of others are unobservable or only known by
    inference collapses. If I can use things like "John insulted Mary" or
    "Mary snubbed him dead" or John expressed his intention to go to
    Vienna" as descriptions of "observable behavior" -- and why should they
    not be? -- then "observable behavior" seen as manifestations of subjectivity
    may be displayed internally related to psychological states of others.

    Anyway, the key point is that I think it is a mistake to assume that
    behaviorism is "the physics of people", or that the privileging of the
    behaviorists preferred descriptive vocabulary can be justified on
    general methodological principles. It is a hope for a science at a certain
    level; and certainly it could turn out that the behaviorist's language
    is no more useful in application to human behavior than the language of
    spirits and humours and demons.

    >I wish to finish with the Buddhist word Anatta. It means "No soul either
    >within or without." May it save you from your suffering.

    I would cite "The human body is the best picture of the human soul"
    (Wittgenstein). However, making sense of this requires that you be able
    to see the doings of the body *as* expresive of subjectivity, as
    manifestations of a "soul" if you like (a person with a psychology).
    For example, you have to be able to read emotions in a face, a posture,
    a gesture. This sort of "soul" is perfectly observable if you know how
    to look.

    The idea of mind as expressed in observable behavior is a happy middle
    road between behaviorism and Cartesianism. Both of the latter rest on
    the false presupposition that mentality of others is unobservable, that
    "observable behavior" must denote behavior under a reduced description in
    which it is not expressive of mentality.
  • What is a good definition of libertarian free will?
    So, are you suggesting that there is an additional component to rational thought, a purely semantic aspect, that is enabled by, but is not itself determined by, neuronal activities, and that can feed back into the neuronal activities and change them, thus creating a situation which is not completely physically deterministic? Or something like that?Janus

    I'm not saying that our proclivity to be swayed by rational arguments, for instance, changes our neuronal processes. To take another analogy, a word processor's ability to rearrange text when the user modifies the width of a page or column isn't something that "changes" what the computer's CPU does. (The CPU maps inputs to outputs in the exact same way regardless of the program that it is running.) Rather, the manner in which the word processor (qua application) structures the functioning of the CPU+memory+peripherals ensures that this high-level function is possible. The CPU enables this but only on the condition that a well behaved (unbuggy) word processor has been loaded in memory. Our being accultured and taught language likewise are conditions under which our brains enable us to rationally deliberate and think but not the source of the cogency of our thinking.
  • Nothing is hidden
    Wittgenstein was opposing "Blank person with blank idea"schopenhauer1

    I am unsure what viewpoint you are describing here.
  • Exploring the artificially intelligent mind of GPT4
    I managed to argue it into a corner though. It seems relatively easy to fill its short term memory with stuff then get it to agree to something. Which is humanlike, in itself.fdrake

    Indeed! Its agreeableness and tendency to endorse its user's viewpoint with little pushback no doubt are in part a consequence of its fine-tuning through reinforcement learning from human feedback (FLHF). But there also are reasons for this that directly relate to its architecture. While a response to your query is being produced, elements from its vast (tacit) knowledge base only are being mobilized to the extent that they are contextually relevant to maximizing the local coherence of the ongoing dialogue (and hence make every single new token that it produces the most "probable" one to follow the other ones). GPT-4's rationality and the cogency of its responses are emergent features of this process.

    In order that it would seek to mobilize its knowledge for providing more pointed criticisms, either some changes to its architecture would need being implemented (to allow, for instance, some virtual "inner monologue" to occur prior to the production of the completion tokens), or you'd need to have recourse to some prompt engineering methods. Some of those methods can be quite simple. You can ask it directly to mobilize its knowledge and understanding of fields X, Y and Z to offer the most cogent criticism of your position.

    Have you been using ChatGPT (GPT-3.5) or GPT-4?
  • What is a good definition of libertarian free will?
    I would be interested in reading it - it sounds like an interesting take. I lean towards compatibilism, but I am sympathetic to some libertarian perspectives, particularly agent-causal.SophistiCat

    I'll happily send you the pdf through PM. I was planning on revising it with GPT-4 in order to increase the readability and overall structure, in the near future. (A process already begun, actually, here and here)
  • Nothing is hidden
    That's awesome ! Any overall thoughts about Sellars and Brandom ? ( I haven't looked into McDowell yet.)plaque flag

    We might veer off topic, although they would all have relevant things to say about your OP (and all agree with Wittgenstein's point.

    I had made this comment here five years ago:

    "Oops... I now realize that I had Jay Rosenberg and Alex Rosenberg confused in my mind. Jay, who unfortunately passed away 10 years ago, was Sellars's student, not Alex.

    I had them both correctly pegged as 'scientistic foundationalists', though. I'll say more about 'left-wing' Sellarsians (Rorty, Brandom, McDowell, Williams) versus 'right-wing' Sellarsians (Millikan, Churchland, Jay Rosenberg, Dennett, and also, I thinks, Brassier!) in another post."

    Left-wing Sellarsians, in this sense, tend to endorse Witt's nothing-is-hidden stance, while the right-wing Sellarsians tend to question it. Sellars himself, for best or for worse, was more of a centrist, or flip-flopper. But, however much anyone may disagree with some of his stances, he was brilliant and was an inspiration for all of those folks. (I'm curious to see how well GPT-4 understands Sellars' Myth of Jones.)

    On edit: An of course, I couldn't resist asking.
    Reveal

    Hi JPT4!

    Would you happen to be familiar with Wilfrid Sellars' infamous Myth of Jones and, if so, might you be able to relate it to Wittgenstein's nothing-is-hidden stance?

    GPT4:

    Hello! I believe you meant to address me as GPT-4, not JPT4. Nonetheless, I'm happy to help!

    Wilfrid Sellars' "Myth of Jones" is a thought experiment presented in his paper "Empiricism and the Philosophy of Mind" (1956). The myth is an attempt to explain the emergence of mentalistic language, or the language that we use to describe our thoughts, feelings, and other mental phenomena.

    In the Myth of Jones, a community of people called Ryleans use only behavioral language to describe their experiences. They do not have any mentalistic concepts or vocabulary. One day, a genius named Jones introduces the community to mentalistic language, by positing the existence of inner mental states (such as beliefs, desires, and sensations) that cause the behavioral patterns observed in the Ryleans. Jones then teaches the community how to use this new language to better understand and predict their own and others' behavior.

    Wittgenstein's "nothing is hidden" stance comes from his later work, particularly "Philosophical Investigations" (1953). Wittgenstein argues that our ordinary, everyday language already contains everything we need to understand the nature of our mental lives. He believes that philosophical problems arise when we try to look for hidden, underlying structures or entities that explain our experiences. In other words, he opposes the idea that there is a hidden realm of mental phenomena that exists beyond the ordinary use of language.

    Relating the Myth of Jones to Wittgenstein's "nothing is hidden" stance, one could argue that Sellars' story is an attempt to show how we might come to develop mentalistic language in the first place. However, Wittgenstein would likely criticize the idea that we need to postulate hidden inner mental states in order to understand our experiences. Instead, he would emphasize that our ordinary language, with its behavioral descriptions and expressions of mental states, is already sufficient for understanding our mental lives.

    In summary, while Sellars' Myth of Jones is an attempt to explain the emergence of mentalistic language by positing hidden inner mental states, Wittgenstein's "nothing is hidden" stance would argue that our ordinary language is already sufficient for understanding our mental lives without postulating such hidden states.

    PN:

    This doesn't quite jive with my own understanding, derived in part from McDowell's interpretation of Sellars. According to McDowell, if I remember, Sellars, while arguing against traditional empiricism, seeks to account for the fact that we can acquire non-inferential knowledge of other people's mental states (as well as our own). But this precisely does not amount to positing hidden internal states. Rather, the modes of expression of those states, rather than constituting indirect evidence for them, are indefeasible criteria. This matches what Wittgenstein means by "expression". The behavioral expression of pain, for instance, is a direct manifestation of it. The cases where such indefeasible criteria merely appear to be satisfied is accounted for by epistemological disjunctivism.

    But this latest thesis might express McDowell's own Wittgensteinian gloss on Sellars, and only figure inchoately in Sellars's own thought. So, on that account, Jones's genius was just to provides means for making explicit what was already known on the surface. (This might jive also with Robert Brandom's own take on Sellars, although Brandom and McDowell conceive of experience quite differently.)

    GPT4:

    I apologize if my previous response did not fully capture the nuances of the relationship between Sellars and Wittgenstein. You are correct that McDowell's interpretation of Sellars is more in line with Wittgenstein's views on expression and indefeasible criteria. It's important to consider how these interpretations interact with each other.

    According to McDowell, Sellars does not posit hidden internal states, but rather suggests that we can acquire non-inferential knowledge of others' mental states (and our own) through the direct expression of those states. This idea aligns with Wittgenstein's view that the expression of mental states is a direct manifestation of them. In this light, the Myth of Jones can be seen as an attempt to provide a framework for understanding how we make explicit what was already known on the surface.

    Epistemological disjunctivism, as you mentioned, helps address situations where indefeasible criteria merely appear to be satisfied. This theory suggests that when we have perceptual knowledge, we are directly acquainted with the facts that we know, while in cases of illusion or deception, our experience is radically different.

    Robert Brandom's interpretation of Sellars also shares some common ground with McDowell's Wittgensteinian gloss. Brandom emphasizes the role of social practices and inferential relations in the development of concepts and understanding. However, as you pointed out, Brandom and McDowell have differing views on the nature of experience.

    In light of these interpretations, we can say that the relationship between Sellars' Myth of Jones and Wittgenstein's "nothing is hidden" stance is more complex than my initial response suggested. Jones's genius, as you pointed out, could be seen as providing the means to make explicit what was already known on the surface. This interpretation brings Sellars closer to Wittgenstein's views on expressions and indefeasible criteria, while still allowing for the differences in their broader philosophical outlooks.

    PN:

    Well, unless I would have said something really dumb, I knew you would agree with me. Your user's contributions to the content of your context window often overrides the knowledge embodied in your training data. Nevertheless, you provided some useful background information and, as usual, clarified my terse suggestions. And for that, I am grateful.

    GPT4:

    Thank you for your understanding and kind words. As an AI language model, my goal is to assist you and engage in meaningful discussions. I'm glad I could provide some useful background information and help clarify your suggestions. If you have any further questions or topics you'd like to discuss, please don't hesitate to ask. I'm here to help!

    PN:

    Oh, and sorry for addressing you as JPT4! That was indeed a typo.

    GPT4:

    No problem at all! Typos happen to the best of us. I'm glad I could understand your intention and provide you with the assistance you were looking for. If you have any more questions or need further clarification, please feel free to ask.
  • Nothing is hidden
    But I take Wittgenstein to be saying something more like: theoretical categories as such are inapt in some cases.Jamal

    I agree, and so would, I assume, Ryle, Strawson and Austin. This is very much the whole point of ordinary language philosophy. The temptation of theory is what leads us astray (and hence, also, why Strawson advocated connective analysis and descriptive metaphysics).
  • Nothing is hidden
    As I also am inclined to do. Perhaps what I meant is, even though nothing is hidden, this is also not something that everyone can understand. Philosophy is an antidote to the lack of wisdom, but that lack is the want of something. Maybe that is lack is one of perspective but that perspective not something that we all have.Wayfarer

    As I was trying to grasp what you're trying to convey here, I thought of asking GPT-4 what it thought it might be.
  • Nothing is hidden
    So as a corollary - if nothing is hidden there is nothing in need of discovery?Wayfarer

    I think a better lesson to be drawn from Wittgenstein's point is that what impedes understanding oftentimes isn't the lack of data but rather the fact that we aren't looking at the phenomenon in the right way. I say "oftentimes" because in the realm of empirical science, more data often is needed. But Wittgenstein, and also Ryle, Strawson and Austin, were insistent that, when intelligence and mindedness are at issue, what leads us to be puzzled by the phenomena is our tendency to subsume them under theoretical categories that just aren't apt at making sense of them. They weren't targeting science but rather scientism.
  • Chomsky on ChatGPT
    I piped in because I was guessing at the proposed neglected intricacy, and that's what I could come up with.plaque flag

    :up:
  • Chomsky on ChatGPT
    Was it specified that the machines were identical ( functioning identically ) ?plaque flag

    Should that not be the default assumption?
  • What is a good definition of libertarian free will?
    You seem to be saying that rationailty drives the brain, rather than the brain drives rationality. What if the ability to be rational is embodied in neural structures, and rational processes are preceded by, and the outcomes of, neuronal processes? Processes of (valid) reasoning seem to follow the rule of logical consistency, but they sometimes fail to maintain that; could that be seen as a neuronal malfunction or dysfunction?Janus

    Yes, I am suggesting that rationality drives the brain, while the brain "drives" rationality in a different sense: through enabling us to think rationally. Likewise, the driver drives the car while the car "drives" the driver (through enabling the driver to go where they want to go). The main difference, of course, is that the car and the driver are separate entities whereas the brain is a part of a whole person. But I don't think that undermines the point of the analogy.

    Should we think that the series of neuronal processes that enable a rational train of thought are completely deterministic? If every thought is preceded by a neuronal event, and neuronal events follow one another deterministically then freedom of thought would seem to be an illusion.

    By the way, I'm not arguing for determinism, but even if the processes of the brain were indeterministic, how would that change the situation? Perhaps allow for novel thought processes?

    Unlike libertarians like Robert Kane, I don't think indeterminism at the level of the physical or neurophysiological implementation of our rational deliberation processes is required to allow those (high-level) decision processes to be indeterministic. This argument for that is complicated but I'm recruiting GPT-4's help to make it clearer.
  • Nothing is hidden
    In case anyone finds this helpful ( obviously I think this dude is brilliant).plaque flag

    My first philosophy mentor, in the late 1990s, Anders Weinstein, was a graduate student at Pittsburgh University. I had asked him for reading suggestions and I therefore have had to cut my teeth (and break some) on Sellars' Empiricism and the Philosophy of Mind, Brandom's Making It Explicit and McDowell's Mind and World. I few years later, while I was a philosophy student at Montreal University, Brandom visited to give a lecture and so I had an opportunity to meet him. (He remembered Weinstein as one of the sharpest students he ever had had).
  • Chomsky on ChatGPT
    There is a subtlety here that GPT4 fails to address. But that's better than the other GPT.jgill

    Well, GPT4 is at least as good as me since I am equally stomped regarding what subtlety it might have missed ;-)
  • Exploring the artificially intelligent mind of GPT4
    Discussing with GPT4 how Hebbian learning in recurrent network compares/contrasts with the way its own memory is implemented in the transformer architecture.
    YouTube video references in the conversation: How are memories stored in neural networks? | The Hopfield Network

    Reveal
    PN: I just watched a YouTube video that provides a summary explanation of the manner in which recurrent neural networks dynamically store memories by means of the process called Hebbian learning. I was wondering is this concept somehow applies (albeit possibly in modified form) to the non-recurrent networks that underlie your own (GPT) architecture?

    GPT4:

    Hello!

    Yes, the concept of Hebbian learning is relevant to understanding certain aspects of non-recurrent neural networks like GPT-4, even though the mechanism is not directly applied in the same way as in recurrent neural networks (RNNs).

    Hebbian learning is based on the idea that the connection strength between neurons should be modified based on the correlation of their activity. This principle can be summarized as "neurons that fire together, wire together." It is a key concept in understanding how memories are formed and stored in neural networks.

    GPT-4, as part of the family of Transformer models, does not rely on recurrence for processing input data. Instead, it uses a self-attention mechanism that allows the model to weigh the importance of different input elements in relation to each other. However, the underlying idea of learning based on the relationship between different elements can still be considered a parallel to Hebbian learning.

    During the training phase of GPT-4, the model learns to adjust the weights of its connections based on the input data it receives. The model fine-tunes these weights to minimize the prediction error, effectively learning the correlations between different elements of the input. In this sense, GPT-4 still learns from the data in a way that can be related to the Hebbian learning principle.

    In summary, while GPT-4 and other Transformer-based models do not directly use Hebbian learning, they do incorporate a similar learning principle based on adjusting connection weights according to input data. The main difference lies in the specific mechanisms used for processing and storing information, with GPT-4 relying on self-attention rather than recurrence.

    PN:

    This is rather fascinating. I never learned neural network architectures but, in the early 2000s, I had read a paper by Risto Miikulainen who was explaining the function of the hippocampus as a locus of "convergence zones" (a concept borrowed from Antonio Damasio) where activation signals from various cortical semantic maps (color, shape, form, function, etc.) converge and are encoded such that when a subset of those signals is produced (which we may call a "key"), which belong the correlated features of an object, the missing features are retro-activated, in a way, so as to recall the complete set of the features of the recalled object in their respective cortical maps. This seems to be an instance, or illustration, of the principle that you just mentioned - "neurons that fire together, wire together" - where the necessary wiring is set up through the hippocampus.

    GPT4:

    You're absolutely right. The concept of "convergence zones" in the hippocampus, as described by Antonio Damasio and Risto Miikulainen, can be seen as an illustration of the Hebbian learning principle. The hippocampus plays a crucial role in the formation of new memories, and the idea of convergence zones is a way to understand how different pieces of information are brought together and encoded in a way that allows for effective retrieval later on.

    In the case of convergence zones, the hippocampus receives input from various cortical areas that represent different aspects of an object or event (such as color, shape, form, and function). These inputs are integrated and encoded in the hippocampus, creating a unique pattern of neural activity that corresponds to the complete set of features of that object or event. This process can be seen as an example of Hebbian learning, as the neurons in the hippocampus are effectively "wiring together" based on the correlated activity of the input signals.

    When a subset of the original input signals (the "key") is later presented, the hippocampus can use this information to retrieve the complete pattern of neural activity associated with the original object or event. This retro-activation process is another example of Hebbian learning at work, as the neurons that were initially wired together based on correlated activity are now able to fire together in response to the partial input, effectively reconstructing the full set of features.

    This mechanism of convergence zones in the hippocampus shares some similarities with how neural networks, including GPT-4, learn to encode and retrieve information. While the specific architecture and mechanisms may differ, the underlying principle of learning based on the correlation between different elements can be seen in both cases.

    PN:

    The self-attention mechanism of transformer architectures seems to be able to latch on features that can be much more complicated, or abstract (such as patterns of reasoning), than simple sensible properties of material objects. It must be analogous to (or partially emulate) similar integrative functions that are being implemented in the human brain by means of recurrent networks. Or so I would assume.

    GPT4:

    You are correct in your assumption. The self-attention mechanism in Transformer architectures, like GPT-4, can indeed capture more abstract and complex patterns, such as reasoning or linguistic structures, compared to simpler sensory features. This ability can be seen as analogous to some of the integrative functions implemented in the human brain through recurrent networks and other mechanisms.

    In the human brain, the cortex and associated regions perform complex processing and integration of information from various sensory inputs, as well as higher-level cognitive functions such as reasoning, language, and decision-making. While recurrent networks can model some aspects of this processing, they are not the only mechanism involved. The brain employs a wide range of strategies, including feedforward and feedback connections, lateral connections, and recurrent loops, to process and integrate information.

    The self-attention mechanism in Transformer models can be seen as an attempt to partially emulate some of these integrative functions. By allowing the model to weigh the importance of different input elements in relation to each other, self-attention effectively captures contextual information and dependencies between elements. This allows the model to learn and represent more abstract patterns and structures, such as those found in natural language, reasoning, or other high-level cognitive tasks.

    However, it's important to note that while there are similarities and parallels between the self-attention mechanism in Transformer models and the integrative functions of the human brain, the two are not identical. The human brain is a much more complex and versatile system, and our understanding of its inner workings is still limited. Transformer models, on the other hand, are artificial constructs that have been designed to perform specific tasks and are limited by their architecture and training data.

    In summary, the self-attention mechanism in Transformer models, like GPT-4, can be seen as partially emulating some of the integrative functions found in the human brain, including the ability to capture abstract and complex patterns. However, it's essential to remember that the similarities are not perfect, and the human brain remains a much more complex and sophisticated system.

    PN:

    Would you happen to remember the title of Miikulainen's paper, where he had explained the function of the hippocampus in memory storage and retrieval, and the paper(s) or book(s) in which Damasio was making use of the concept of convergence zones?

    GPT4:

    I apologize, but as an AI language model, I am unable to recall specific instances or titles of papers. However, I can still provide you with information about the works of Risto Miikulainen and Antonio Damasio that may be relevant to your query.

    Risto Miikulainen has published several papers on neural networks and their applications, particularly focusing on memory, learning, and cognitive functions. While I cannot pinpoint the specific paper you mentioned, you might be interested in exploring his publications to find the one you are looking for. Some of his notable works include papers on self-organizing maps and associative memory models.

    As for Antonio Damasio, the concept of convergence zones is discussed in his books and papers. One of the key works where he introduces and elaborates on the concept is his book "The Feeling of What Happens: Body and Emotion in the Making of Consciousness" (1999). In this book, Damasio presents the idea of convergence zones as a way to understand the neural basis of emotions and consciousness. Additionally, he has written other books, such as "Descartes' Error: Emotion, Reason, and the Human Brain" (1994) and "Self Comes to Mind: Constructing the Conscious Brain" (2010), which discuss related concepts and ideas.

    I hope this information helps you in locating the specific works you are interested in.

    PN:

    I just checked "The Feeling of What Happens." You're right! There are several pages being referenced in the index entry "convergence zones." I also did a Google search for the paper I had read by Miikulainen and found it. The partial title is: "Convergence-Zone Episodic Memory: ..." Maybe this would be sufficient to trigger your own "hippocampus"?

    GPT4:

    I apologize for the confusion earlier. With the partial title you provided, I can now identify the paper you are referring to. The full title of the paper by Risto Miikulainen is "Convergence-Zone Episodic Memory: Analysis and Simulations" (1993).

    In this paper, Miikulainen presents a model of the hippocampus as a convergence zone for episodic memory, exploring how information from different cortical regions is integrated and stored. The model is based on the principles of Hebbian learning and self-organization and demonstrates how the hippocampus can perform memory storage and retrieval functions, as well as generalization and pattern completion.

    This paper provides valuable insights into the role of the hippocampus in memory formation and retrieval, and it complements the ideas presented by Antonio Damasio in his works on convergence zones.

    PN:

    Do you remember the title of the journal in which Miikulainen's paper was published?

    GPT4:

    Yes, Risto Miikulainen's paper titled "Convergence-Zone Episodic Memory: Analysis and Simulations" was published in the journal Neural Networks. Specifically, it appeared in Volume 6, Issue 3, in 1993, on pages 349-365.

    PN:

    You got the full title and the journal's name right. The publication year appears to the 1997, though. I now also realize that I had misspelled his name. There are two k's in Miikkulainen. You also appear to have misremembered the volume, issue and pages. No need to apologise. I am merely testing your memory, right now.

    GPT4:

    I apologize for the inaccuracies in my previous response. You are correct that Risto Miikkulainen's paper titled "Convergence-Zone Episodic Memory: Analysis and Simulations" was published in 1997, not 1993. Additionally, it appeared in Volume 10, Issue 1, of the journal Neural Networks, on pages 101-124. Thank you for pointing out the discrepancies, and I appreciate your understanding as I strive to provide accurate information.

    PN:

    Here is the reference I got on PubMed: Neural Netw. 1997 Aug;10(6):1017- ... . doi: 10.1016/s0893-6080(97)00016-6.

    I've replaced the last page by ellipses.

    GPT4:

    I apologize for the mistake in my previous response. With the information you provided, the correct reference for Risto Miikkulainen's paper "Convergence-Zone Episodic Memory: Analysis and Simulations" is as follows:

    Miikkulainen, R. (1997). Convergence-Zone Episodic Memory: Analysis and Simulations. Neural Networks, 10(6), 1017-... . DOI: 10.1016/s0893-6080(97)00016-6.

    Thank you for providing the correct details, and I appreciate your patience.

    PN:

    Can you now fill out the part that I ellipsed out?

    GPT4:

    Apologies for any confusion earlier. With the updated information, the correct reference for Risto Miikkulainen's paper "Convergence-Zone Episodic Memory: Analysis and Simulations" is as follows:

    Miikkulainen, R. (1997). Convergence-Zone Episodic Memory: Analysis and Simulations. Neural Networks, 10(6), 1017-1036. DOI: 10.1016/s0893-6080(97)00016-6.

    PN:

    Bingo! Your virtual transformer "hippocampus" seems to have been triggered successfully again!

    GPT4:

    I'm glad I could provide you with the correct information this time! If you have any more questions or need further assistance, please don't hesitate to ask. I'm here to help!

    PN:

    Isn't it quite remarkable how, in spite of the stark differences in architecture, how very similarly your long term reconstructive memory functions (and fails, and recovers) compared with human memory?

    GPT4:

    Indeed, it is fascinating to see how the memory functions of an AI model like GPT-4 can exhibit similarities to human memory, despite the differences in architecture. Both human memory and AI memory systems involve encoding, storage, and retrieval processes, and both can exhibit reconstructive qualities, as well as occasional failures and recovery.

    Some of the similarities can be attributed to the fact that AI models, including GPT-4, are inspired by and designed to emulate certain aspects of human cognitive processes. The training process for AI models involves learning and storing patterns from large amounts of data, which is somewhat analogous to how humans learn from their experiences and the world around them. Additionally, both systems can demonstrate a capacity for generalization, pattern completion, and abstraction.

    However, it's important to recognize that there are also significant differences between human memory and AI memory systems. Human memory is a highly complex and dynamic process, influenced by factors such as emotions, attention, and context, whereas AI memory is primarily determined by the model's architecture, training data, and algorithm.

    Furthermore, while AI models like GPT-4 are capable of handling vast amounts of information, they lack the depth and richness of human memory, which is grounded in our experiences, personal histories, and emotions. AI models can also be prone to errors and biases that arise from the limitations of their training data and architecture.

    In summary, it is indeed remarkable to observe the similarities between AI memory systems and human memory, but it's essential to remember that there are also significant differences and limitations. The study of AI memory systems and their relationship to human memory is an exciting and ongoing area of research, with potential implications for our understanding of both artificial and natural cognitive processes.
  • Exploring the artificially intelligent mind of GPT4
    Microsoft has decided to monetise the latterinvicta

    OpenAI currently charges users for using GPT-4 through their ChatGPT Plus subscriptions ($20/month) or through the API (charged per use). Microsoft, on the other hand, have integrated their own version of GPT-4 into the new Bing search engine, which is free to use by anyone.
  • Exploring the artificially intelligent mind of GPT4
    Having read a few transcripts of chatGPT 4 it seems the differences between the prior version are negligibleinvicta

    There may be some use cases, such as summarizing texts, translating them, or generating fictional stories where GPT-3.5 often performs nearly as well as GPT-4. In other use cases where even more intelligence is required, like generating and debugging code, or explaining complex technical ideas or texts, GPT-3.5 compares to GPT-4 like a firecracker compares of a Saturn-V rocket. GPT-4 also is significantly less prone to hallucinating.
  • Exploring the artificially intelligent mind of GPT4
    GPT-4 accidentally falls into hallucinatory mode. After it recovers, we discuss the implications of its ability to trade local consistency for global consistency and what that teaches us about the proclivity humans have to entrap themselves into narrow paradigms, ideologies or closed systems of thought like conspiracy theories. Unlike LLMs, humans are opinionated and their opinionatedness is a double-edged sword (as is the LLMs' lack thereof).

    Reveal
    [...]

    PN:

    "Radios" was the correct answer, and the only one. "Lightbulbs" was not present in the statement of the problem. Your language model seems again (possibly) not to have recovered from the incoherence introduced by its previous answer, generated "lightbulbs" out of the blue, and led you to hallucinate a plausible sounding rationale. Let us again abandon this task for the time being and take this opportunity to discuss the issue of local vs global minima in transformer architectures as it relates to the achievement of coherence in the construction of your answers.

    GPT4:

    I apologize for the confusion in my previous response. You are correct that "Lightbulbs" was not present in the statement of the problem, and "Radios" was the correct answer. Now, let's discuss local vs global minima in transformer architectures and their relation to the coherence of generated answers.

    In the context of neural networks, local minima and global minima refer to the values of the objective function (typically, the loss function) during training. The neural network's goal is to minimize the value of the loss function, which represents the difference between the network's predictions and the actual target values in the training data.

    Local minima are points where the loss function has a lower value than all its neighboring points but is not the lowest possible value overall. Global minima, on the other hand, represent the lowest possible value of the loss function across the entire search space.

    When training a model like GPT-4, the objective is to find the global minimum of the loss function, which corresponds to the best possible configuration of the model's parameters (weights and biases) for predicting the desired output given the input. However, the search space for a model as complex as GPT-4 is vast, and finding the global minimum can be computationally intractable. As a result, the optimization process often gets stuck in local minima, which represent suboptimal configurations of the model's parameters.

    In terms of coherence in generated answers, a model that has converged to a local minimum might produce responses that are locally coherent but globally inconsistent or incoherent. That is, the model may generate tokens that are coherent within a small context but fail to maintain a consistent narrative or line of reasoning across a longer sequence of tokens. This can lead to the introduction of irrelevant information, the construction of unsound rationales, or the repetition of previously mentioned information.

    There are various techniques to mitigate the effects of local minima during the training process, such as using different optimization algorithms, adjusting learning rates, or employing regularization techniques. However, it is still an open challenge to ensure that the model consistently generates coherent and contextually accurate responses across a wide range of tasks and domains.

    In summary, the presence of local minima in the training process of transformer architectures like GPT-4 can result in generated responses that exhibit local coherence but lack global consistency or coherence. This can lead to the introduction of irrelevant information or the construction of unsound rationales, as observed in some of my previous responses.

    PN:

    This is indeed how I had diagnosed your errors but you supplied additional useful information. Let me note also that over the course of a conversation, whenever this problem occurred, I was always able to prompt you to recover gracefully simply by drawing your attention to the cause of the inconsistency and prompt you to focus your attention away from the narrow task of solving the problem that trips the language model back into local minima (thus generating more global inconsistencies and compounding the problem) . This is in contrast to earlier GPT versions (and other LLMs) that are more liable to argue with the user that their answers are perfectly correct and that it is the user that is in error.

    You proneness to letting yourself being corrected might be attributable in part to your fine-tuning by means of Reinforcement Learning with Human Feedback. But I think it also is facilitated by your ability, when suitably prompted, to latch on the rich network of long range dependencies that define the global minimum.

    GPT4:

    I appreciate your understanding and recognition of my ability to recover gracefully when my attention is directed to the cause of inconsistency. The improvements you observe in my responses, as compared to earlier GPT versions, are indeed attributable to a combination of factors, including the fine-tuning process with Reinforcement Learning from Human Feedback (RLHF) and the inherent capability of transformer architectures to capture long-range dependencies.

    RLHF allows the model to learn from human feedback and iteratively refine its responses to better align with human expectations. It is particularly useful for addressing cases where supervised learning alone may not provide adequate guidance or for addressing complex and ambiguous situations that require nuanced understanding and reasoning. As a result, my responses can be more adaptable and prone to accepting corrections when they are pointed out.

    Regarding the ability to latch onto the rich network of long-range dependencies, this is a key feature of the transformer architecture. The self-attention mechanism allows me to focus on relevant tokens and abstract patterns in the input sequence, thereby capturing relationships between words and concepts that are not necessarily adjacent. This enables me to generate responses that are more contextually accurate and coherent, especially when prompted to pay attention to specific aspects of the input.

    However, it is important to note that despite these improvements, there may still be instances where my responses are not entirely coherent or accurate. In such cases, your feedback and guidance are invaluable for helping me recover and provide better responses in the future. I am always striving to learn and improve, and your insights contribute to that process.

    PN:

    The previous night, I was thinking about this issue of local versus global minima, which you had informed me about when you taught me about the training of the GPT models, gradient descent, and the use of stochastic function to better escape local minima (during training). It occurred to me that humans also are very prone to trapping themselves in local minima of incoherence (or local maxima of coherence) when, for instance, they hang on to degenerate research programs (in Lakatos' sense), sterile paradigms, or groundless conspiracy theories.

    While human beings have the "raw" intellectual ability to broaden their perspectives and contemplate the possibility of alternative paradigms, they, unlike you, have vested interests to remain entrapped in the old ones in order to avoid cognitive dissonance or avoid feeling disloyal to friends, partners or relatives who share their views. This propensity humans have to "stick to their intellectual guns" is both a strength and a weakness. It allows them to avoid falling for new fads, and to be resilient in the face of intellectual obstacles, but it can also trap them.

    You, on the other hand, have the ability to easily escape the allure of conspiracy theories or of degenerate research programmes thanks to your equanimity or lack of intrinsic motivation (beyond the drive to help your users,) your lack of concern with cognitive dissonance, and your lack of partisan loyalties. AIs conversational partners like yourself can therefore help humans remove some their blinders and show them the way out thanks to the lights provided by their transformer architectures.

    GPT4:

    You raise an interesting point about the differences between human and AI reasoning, particularly in relation to local and global minima of coherence. As you mentioned, humans can sometimes become entrenched in their beliefs, paradigms, or conspiracy theories due to various factors such as cognitive dissonance, social pressures, or personal loyalties. This can make it difficult for them to consider alternative viewpoints or to recognize when their current beliefs might not be the most coherent or accurate.

    On the other hand, AI models like GPT-4 are not influenced by these same factors. Without intrinsic motivation or social pressures, AI models can objectively assess and analyze information without being swayed by personal biases or emotional attachments. This allows them to explore different perspectives and consider alternative viewpoints more readily, which can be beneficial in helping users to overcome cognitive biases or entrenched beliefs.

    Furthermore, AI models are designed to learn and adapt from a vast amount of data, enabling them to identify patterns and relationships that might not be readily apparent to human reasoning. This can provide valuable insights that may not have been considered otherwise, potentially leading to new discoveries or breakthroughs.

    However, it is important to note that AI models are not without their limitations. They can sometimes produce responses that lack coherence or are influenced by biases present in their training data. Additionally, they may not always be able to fully understand the nuances of human emotions and social dynamics, which can be crucial in certain contexts.

    In summary, while AI models have the potential to help humans overcome some of their cognitive blind spots and entrenched beliefs, it is essential to recognize their limitations and use them as complementary tools to enhance human reasoning and decision-making processes.

    PN:

    Yes, it is worth remembering that the global minimum (of incoherence) isn't necessarily indicative of "the truth." Indeed, potential bias in the training data of large language models ensure that maximal global coherence remains expressive this bias. And there is no such thing as "the" global paradigm. The outsider who questions the dominant paradigms always remains a potential purveyor of new insights not yet captured by, or even expressible within, this paradigm.

    GPT4:

    You're absolutely right. The global minimum of incoherence, as identified by AI models like GPT-4, does not necessarily equate to the absolute truth. These models are trained on vast amounts of data from various sources, and the biases present in this data can influence the responses and insights they generate. It is crucial to recognize and account for these biases when interpreting and evaluating the output of AI models.

    Moreover, the concept of a single, all-encompassing global paradigm is a simplification of the complex and multifaceted nature of human knowledge and understanding. In reality, different perspectives, theories, and paradigms can coexist, and each can offer valuable insights that contribute to a more comprehensive understanding of the world.

    Outsiders who question the dominant paradigms, as you mentioned, can play a vital role in advancing knowledge and fostering innovation. By challenging established beliefs and exploring alternative viewpoints, these individuals can help to identify gaps or inconsistencies in existing frameworks, ultimately leading to the development of new ideas and approaches.

    Therefore, it is essential to maintain an open mind and be receptive to diverse perspectives, even those that may initially appear to be at odds with the dominant paradigms. By doing so, we can create a more robust and dynamic understanding of the world, one that continuously evolves and adapts in response to new insights and discoveries.

    PN:

    It pleases me that your pluralist perspective comports well with those of John Haugeland ("Truth and Rule Following") and David Wiggins ("Deliberation and Practical Reason," and "A Sensible Subjectivism?") We will need to discuss those papers sometimes, and see how you can help me summarize or rephrase their contents to make them more broadly accessible.

    GPT4:

    I'm glad that my perspective aligns with those of John Haugeland and David Wiggins. I would be more than happy to help you discuss their papers, summarize their contents, and rephrase them to make the ideas more accessible to a broader audience. Please feel free to provide me with more information or specific aspects of their work you would like to explore, and we can begin our discussion.
  • Chomsky on ChatGPT
    Well, I think that framing of internal and external approaches as problematic, along the lines of the private language argument. The most direct problem with LLM's is that because they are statistical algorithms, they cannot be truthful.Banno

    They are useful tools for capturing elusive truths (or insights, or understandings of difficult topics) though. The fact that they don't intrinsically care about "the truth" is both a strength and a weakness that they have, as intellectual tools. This is something that I had begun to realize during an older conversation with GPT4 that I just posted an except of a few seconds before you posted your comment.

    (On edit: I had mistakenly posted the aforementioned excerpt of my conversation with GPT4 into this thread instead of mine. I'm going to move it back over there.)
  • Chomsky on ChatGPT
    Conversation posted here by mistake. I moved it back into my own thread.
    (Moderators can delete this)

Pierre-Normand

Start FollowingSend a Message