https://youtubetranscript.com/?v=SPnyxnvU4ko
Darwin and physics, high energy theoretical physics, agree that space-time is doomed. It’s not fundamental reality. And the search is on in the last 10 years among physicists to find structures entirely beyond space-time, not curled up inside space-time, beyond space-time. We’ve mistaken a headset for the truth because it’s easy if that’s all you’ve seen all your life is a headset and it’s hard to imagine something outside of it. Now we’re free, using mathematics, to ask what kind of structures could we posit beyond space-time. Hello everyone watching and listening. Today I’m speaking with author and cognitive neuroscientist, Dr. Donald Hoffman. We discussed Dr. Hoffman’s research on what we know as reality, why space-time itself is now considered by many a doomed framework of interpretation, and how consciousness might be best understood as a vast probability space within which we orient ourselves. Hello Dr. Hoffman, it’s very good to see you. I’ve been interested in your theory for a long time, partly because I’m quite attracted by the doctrine of pragmatism, which was really part of what I tried to discuss with Sam Harris many, many times. And it seems that your work bears, well, it’s a broad general interest, but it also bears on specific interests of mine because I’ve always been curious about the relationship between Darwinian concepts of truth and let’s say the concepts of truth put out by the more Newtonian, say, objective materialists. They don’t seem commensurate to me. And so would you start by explaining your theory, your broad theory of perception? I know that’ll take a while, but it’s a tricky, it’s a tricky theory. So do you want to lay it out for us to begin with? Most Darwinian scholars would agree that evolution shapes sensory systems to guide adaptive behavior, that is to keep organisms alive long enough to reproduce. But many also believe that in addition, evolution shapes us to see reality as it is, at least some aspects of reality that we need for survival. So that’s often among my colleagues in studying evolution with natural selection. They’ll say, yeah, seeing the truth will make you more fit in many cases. And so even though Darwin says it’s evolution shapes sensory systems just to keep you alive long enough to reproduce, many people think that seeing aspects of reality as it is will also make you more fit and make you more likely to reproduce. So I decided with my graduate students a few years ago to look into this. There are tools. Darwin’s theory is now a mathematical theory. We have the tools of evolutionary game theory that John Maynard Smith and others invented in the 1970s. And so it’s a wonderful theory. So Darwin’s ideas can now be tested with mathematical precision. And I thought that maybe what we would find is that evolution tries to do things on the cheap. If you have to spend more calories, then you have to go out and kill something to get those calories. And so there are selection pressures to do things cheaply and quickly, heuristics. So I went into it thinking that maybe that would make it so that many sensory systems didn’t see all of the truth. But I just wanted to check and see what would happen. To my surprise, when we actually started studying this, there came up principles that made me realize that the chance that we see reality as it is on Darwinian principles is essentially zero. And that was a stunning result. Why zero? Zero is a very low number. So why zero? That’s right. So it’s a bit technical, but in evolutionary theory, in the evolutionary game presentation of it, you think of evolution as like a game. And in a game, you’re competing with other players and you’re trying to get points. Now in the game of evolution, the way its model is there are these fitness payoff functions. Those are sort of the points that you can get for being in certain states and taking certain actions. And so these fitness payoffs are what guides the selection. They guide the evolution. And so we began to analyze those fitness payoffs. The fitness payoffs, to be very concrete about a fitness payoff, suppose that you’re a lion and you want to mate. Well, a steak won’t be very useful for you for that process. You’ll have very little fitness payoff for a steak if you’re a lion looking to mate. If you’re a lion that’s looking to eat and you’re hungry, then of course the steak will have high fitness payoffs for you. So fitness payoff depends on the organism, like a lion versus say a cow. A steak is of no fitness payoff for any cow for any purposes. Quite the contrary. Quite the contrary. That’s right. So the fitness payoff depends on the organism, its state, I mean hungry versus sated, for example, and the action, feeding, fighting, fleeing and mating, for example. So these fitness payoffs are functions of the world. They depend on the state of the world and its structure and the organism, its state and its action. So they’re complicated functions. And in some sense you could think that there’s just effectively one fitness payoff function. There’s this one big fitness payoff function which handles the world and all possible organisms and all possible states and actions. So there’s a big fitness payoff. The question is, but we can think about it as many fitness payoffs if we want to as well. The question is, suppose then, so this fitness payoff function takes as its starting point the state of the world, right? That’s the domain of the function. And the range of the function might be the fitness payoff value, say from zero to 100. Zero means you lose, 100 means you did as good as you could possibly do. So zero to 100, say. So it’s a function from the state of the world, cross organism, in the state and action, into this number, so zero to 100, zero to 1,000, whatever you want to use. So the question then is, does this function preserve information about the structure of the world? This is the function that’s guiding the evolution of our sensory systems. So does this function, if the function is what mathematicians call a homomorphism, a structure preserving map. So for example, the world might have an order relationship, like one is less than two is less than three, like a distance or a distance metric or something like that. To be a homomorphism would mean that if things were in a certain order in the world, the function would take them into that same order or some homomorphism of the order onto the states of the payoffs. So that’s the technical question. What is the probability that a generically chosen payoff function will be a homomorphism of a metric or total order or a partial order or a topology or measurable structure? Any structure that you can imagine the world might have, you can ask what is the probability that a generically chosen payoff function will preserve it. If it doesn’t preserve it, there’s no information in the payoff function to shape sensory systems to see that truth, to see that structure of the world. So what’s remarkable is that evolutionary theory is indifferent about the payoff functions. They don’t say they have to be a certain shape. In other words, every fitness payoff function that you could imagine is on equal footing on current evolutionary theory to every other one. There’s nothing in Darwin’s theory that says these are the fitness payoff functions and this is their structure. So what we had to do then is to say, okay, we have to just look at all possible fitness payoff functions and ask how many of them, what fraction of these payoff functions would preserve a total order or a metric or a measurable structure or whatever it might be. And here’s the remarkable and in retrospect obvious thing. For a payoff function to preserve a structure like a metric or a total order, it must satisfy certain equations. So you have to write down these equations that the homomorphism must satisfy, that the fitness payoff function must satisfy to be a homomorphism. Well, once you write down an equation, most payoff functions simply aren’t going to satisfy it. I mean, the equations are quite restrictive. And in fact, in the limit, as you look at a world that has an infinite number of states and payoff values that go from zero to infinity, the fraction of payoff functions that actually are homomorphic goes to zero precisely. All right. So this is going to be a somewhat meandering question because it’s a very complicated thing to get right. So people who think that the world is made out of self-evident facts underestimate the complexity of perception. And so here’s how I’ll make that case. And you can tell me what you think. You can imagine you could ask an engineer a simple question. Can you build a bridge? And you might think the fact of the bridge will be a fact. And the answer to the question, which would be yes or no, will be a fact. And that’s that. It’s all self-evident. It’s sort of like the behaviorists assuming that the stimulus was self-evident. It’s very much analogous to that. Okay. But here’s the problem. There’s a whole set of assumptions built into that question that people don’t even notice. And so let me walk through some of the assumptions. It’s like, well, I can’t build a bridge if you want it to last 50 million years. So I could build a bridge that would last a century or two centuries. I can’t build a bridge for no money with no labor, with materials that are just at hand. So the thing you define as a bridge is already subject to all sorts of constraints. Now you and I mutually understand those constraints without even having to speak about them. So I’m also going to assume that if I ask you, can you build a bridge, and you say yes, you’re also saying, I’m willing to work with you. I’m willing to work honestly. I’m willing to hire the right number of people. I’m not going to screw you during the construction. The bridge that we build, we both understand that human beings will be able to walk across it and as many as will fit on the bridge without the bridge falling down and also cars. And that means it’ll have to be about the same width as a car or a truck or four lanes of cars or trucks. And it’ll have to abide by all the building codes and so forth. There’s so many constraints in that question that it would take you an unlimited amount of time to list them all. And you don’t because you’re talking to an engineer and he’s a human being like you, uncultured like you. And so he understands the world like you do. And so there’s a hundred million things you don’t have to talk about. But they’re there. They’re constraining the set of facts that’s relevant to the issue and they’re constraining them seriously. Now those constraints, those are nested in an even higher order set of constraints, which are Darwinian. Right? Like, well, the axiomatic agreements that you and I come to as a consequence of our shared perceptions, our shared embodiment and our shared enculturation are a consequence of a broader process, which is essentially Darwinian. Now that Darwinian set of constraints is instantiated in motivational systems in part. So we might say, well, anything that you and I do together will have to be done while taking into account hunger and anger and fear and pain, the whole emotional potentiality of people, plus our fundamental motivational systems, the manner in which we lay out this particular task will have to satisfy all that. Now that’s also unspoken. Now when you talk about evolutionary game theory and pragmatic constraints, let’s say you talked about the lion who wants to mate and not eat, you’re referring to one motivational system or another, one governing sex per se and the other governing hunger. And then the manner in which the lion is going to perceive the world or the manner in which we’re going to perceive the world is going to be bounded by the operation of that motivational system. And the perception is going to be deemed sufficient if when we enact it, the motivational system is satiated. Fair enough? Okay. Okay. Now, but then there’s a more interesting issue that pertains to the big fitness payoff. So if you look at how the nervous system is structured, you have these underlying motivational systems, which are goal setting machines in which define the parameters within which a perception is valid. But all those systems have to interact together and they cause conflict. So if you’re hungry and tired, you don’t know whether you should get up and make a peanut butter sandwich or if you should just go to sleep and leave it till the morning. Like there’s inbuilt conflict. And part of the reason that the cortex evolved was to mediate subcortical conflicts. And then even at the cortical level, the manner in which you integrate your fundamental motivation and the manner in which I motivate integrate mine have to be integrated or will fight. And so I would say, and I don’t know if evolutionary theorists have dealt with this and it’s relevant to your theory that perception doesn’t map the real world. Is there a higher order set of integrated constraints that serves reproduction over the long run that all the lower order fitness payoffs are necessarily subordinate to? And I know this is a terribly complicated question. Is that the reality that perception serves? You made the case that perceptions will not map one to one on reality. And I suppose that’s partly because reality is infinitely complex. Right? You can fragment it infinitely and you can contextualize it infinitely. So it’s very hard to calibrate. All right. So we got to put that aside. But then I would say, well, maybe there’s another transcendent fundamental reality that’s Darwinian in nature that integrates everything with regards to optimized long term survival and perceptions are optimized to suit that. So I know that’s a terribly complicated question, but this is a terribly complicated subject. So I think we have to think a little out of the box on this question, because when we conclude that evolution shapes us not to see reality as it is, then the question is, well, what is it shaping our sensory systems to give us? As well as what is reality? Right? That question also comes up. Yeah. Absolutely. The way I like to think about it is that evolution shapes sensory systems to serve as a user interface. So like the desktop on your computer, for example. So when you’re actually working on a computer, in this metaphor, what you’re literally doing is toggling millions of voltages in a computer, in circuits. And you’re having to toggle them in very specific patterns, millions of them in exactly the right pattern. Well, if you had to do that by hand, if you had to deal with that reality and interface with that reality, one voltage and get it in exactly, well, it’d take you forever and you probably wouldn’t get it right. And you wouldn’t be able to write your email or edit your picture, whatever you’re doing on your computer. So we spend good money and people spend a lot of time building interfaces that allow you to be ignorant, completely ignorant. Most of us have no idea what’s under the hood in our laptops. We have no idea. We know that there’s circuits and software, but most of us have never studied it. And yet we’re able to very swiftly and expertly edit our images and send texts and emails and so forth without having any clue, literally no clue what’s under the hood. What’s the reality that we’re actually toggling? And so it seems that that’s what evolution has done for us, has given us an incredibly dumbed down interface. We call it space and time and physical objects. So we think of space and time as the fundamental reality and physical objects as truly existing in that objective reality. But it’s really just in this metaphor, a virtual reality headset. We’ve evolved a virtual reality headset that utterly hides the very nature of reality and on purpose, quote unquote, on purpose, so to speak, because it would be not flexible. We drown in the complexity. Right. OK, so some evidence for that, as far as I’m concerned, is the following. I mean, first of all, if you look at a desktop, it consists, let’s say, in part of folders. Now folders are actually something in the real world that you can pick up and we understand them. You can manipulate them. You can see how they operate by using your, as a consequence of your embodiment. And so that embodiment gives you a deep understanding of the function of a folder and then you can represent it abstractly and you can put it on a desktop and everyone understands what it means. And that understanding is something like able to map a certain set of functions for a certain set of purposes. That’s what understands and it’s a constrained set of purposes. This is what really struck me about reading the pragmatists. They said and and and Perce and and James studied Darwin deeply and they were the first philosophers to realize exactly what implications Darwinian theory had for both ontology and epistemology ontology, which is the study of reality for everyone listening. That was a real surprise. You could you could understand that Darwin’s theory might have epistemological implications, implications for the theory of knowledge. But the fact that it had implications for what reality is per se is something that very few scientists have yet grappled with. And the pragmatists always said, look, when you accept something as a fact, one of the things you don’t notice is that you set up conditions for that to be factual. And the fact is something like this definition will do during this time span for this very constrained set of operations. Fact. OK, but the problem with that is that’s not a dead objective fact just lying on the ground. That’s a fact by necessity nested inside a motivational system. So facts now all of a sudden become motivated facts. And that just wreaks havoc with the notion of objective, like of a distant objective materialism, because the facts are supposed to be separate from motivation. And the pragmatists, as far as I’m concerned, following Darwin, demonstrated incontrovertibly that that’s like like you pointed to. I think it’s an it’s an it’s an elegance. That’s actually impossible now because you have to constrain reality in order to perceive it, because it’s too complex. You drown in the details. Otherwise, you drown in the complexity. Now, you made the claim, and I want to interrogate this a bit, that there’s there’s really no direct relationship, let’s say, between the desktop icon that you think is an object when you look at the world and the actual world. But let me offer you an alternative and tell me what you think about this. So. There’s this idea. This is a weird way of approaching this, but I’m going to do it anyways. There is a very strange stream of primarily Catholic thought, I believe, that tried to wrestle with the idea of how God could become man. So because God, of course, is infinite and everywhere and man is finite and bounded. And so the question is, well, how do you establish a relationship between the infinite and the bounded? And that’s analogous to the same problem that we’re trying to solve. And they came up with this hypothesis of kenosis, which means emptying. And their notion was, well, Christ was God, but in some ways like a low resolution representation of God, an image of God. Right. So there was a correspondence, but not a totality, at least not in any one instance. Now, the reason I’m bringing that up is because it seems to me that when we perceive an object, that it isn’t completely without. You call it homomorphism with, I believe, with the underlying world, it’s just extremely low resolution, like it’s a it’s a low resolution functional tool. That’s what an object is. But and it’s now and I would say I would advance in support of that, for example. Obviously, the icons that we have on a computer screen, we can use and we treat them like they’re real and clearly they’re low resolution. But also when we watch an animated show, for example, like The Simpsons, we’re looking at cartoon like icons, right? They’re emptied even further than like if I saw a Simpson cartoon of you, it would be like a very low resolution representation of the UIC, which is a very low resolution representation of whatever the hell you are in actuality. Like it’s a seat. But I but I but I think there’s an element of that perception that’s an unbiased sampling of the underlying reality, although it’s bent to pragmatic ends, pragmatic motivational ends. Now, I don’t know what you think about that. I thought about it for a long time. I can’t find a hole in it, but I’m wondering what you think. Well, I think here’s an analogy that might help explain the way I see it. Suppose you’re playing a VR version of Grand Theft Auto. So you have a headset and bodysuit on and you’re playing a multiplayer Grand Theft Auto. You’re playing with someone in China and England and so forth. And I’m sitting there in my ride. I’ve got a steering wheel and gas pedal and dashboard. And I’m looking out and I see to my right, I can see a red Ferrari. And to my left, I see a green Mustang. Well, now, of course, what I’m really interacting with in this analogy is some supercomputer somewhere. Right. And if I looked inside that supercomputer and look for a red Ferrari, I would find no red Ferraris anywhere inside that supercomputer. I would find voltages. So there in that sense, the red Ferrari is a symbol in my headset in the in the game. And there’s nothing in the objective reality in this metaphor that that it’s a low resolution version of. There’s just literally a completely different kind of beast. OK, OK. There are no red Ferraris. OK, so let me ask you about that. So I get your point, especially Jermain with regards to the online game. But is it not the case that in that supercomputer architecture, there’s a pattern that is analogous to the red Ferrari pattern that’s the externalized representation of the pattern, let’s say, on your retina and then that propagates into your brain like there is a there is a there is a conservation of pattern. Now, that Ferrari pattern in the supercomputer would be a very tiny element of an infinite landscape of patterns in the computer. But it’s not and it’s definitely not a pattern of a car per se. Right. It’s a pattern of a representation of a car. And but it’s still got some correspondence with a pattern of voltages, let’s say, that does have some existence within the supercomputer architecture. Well, so in that case, I would say that there is a causal connection that what’s going on inside the supercomputer has a causal connection with the sequence of pixels that are being illuminated in my headset so that I see a red Ferrari. So there’s a causal connection. But if I asked, is there some sense in which there’s a homomorphism of structure between what’s going on inside the computer and what I’m seeing on the screen as a red Ferrari, I would say there’s probably no homomorphism at all. And in that sense, there’s we can’t think about it as like a low resolution version of something. So to be specific, the the electrons in the computer have no color. My Ferrari is red. The shape of the Ferrari and the shapes of the electrons or even the pattern of motion of the electrons is independent. And what’s going on in part is that the pattern of electrons in the supercomputer, they’re programmed to operate in a certain way to cause certain other things to happen in my headset to trigger voltages that trigger pixels to have certain colors. And so there’s a whole sequence, a whole cascade of events that are going on there. And so to say that there’s a homomorphism, I think is I think it’s just barking up the wrong tree. According to a recent report, planned parenthood continues to rake in billions, despite dwindling clients. The biggest takeaway here is that planned parenthood is generating vast profits, including millions in taxpayer funding. With the help of preborn, you and me, we are stealing their clientele, meaning the babies they are trying to kill. Preborn operates on a very slim budget as they rescue over 200 babies lives every day and they receive no government funding. Preborn’s network of clinics are situated in the darkest corners, competing head to head with the abortion giants. They need our help now more than ever. When you donate 1 per month trial period at Shopify.com slash jbp. Go to Shopify.com slash jbp now to grow your business no matter what stage you’re at. That’s Shopify.com slash jbp. Probability is that probability space equivalent to let’s say a realm of potential around a my students and I tried to model anxiety as a response to entropy. Okay, so imagine that what you have in front of you is a set of branching possibilities, some of which can be realized with comparatively less effort. So they’re more probable, let’s say given your current state, some of which are virtually impossibly distal, but in principle could be managed if you were smart enough and could gather the resources. But so you have a probability space in front of you, some of which is sort of at hand, like it’s pretty easy for me to pick up this pen. Right? So that’s a high probability pathway laid out in front of me. So I mean, the the the mythological motifs that I referred to insist that what people face is something akin to the pre cosmogonic chaos that God himself faced when the cosmos first sprang into being right and so that the way to construe the world isn’t as a place of clockwork automaton machines, self-evident objects, but as a realm of possibility that differs in probability. And then the issue becomes how do you best orient yourself so that you can contend properly with that probability landscape? Now, is that, am I walking on parallel ground here? We’re in broad agreement in that, in the sense that our theory of conscious agents, by writing down a probability space, it is a space of potentiality. For example, to be very, very concrete. Suppose my experiment is just to flip a coin twice, heads and tails. Well, what’s my probability space? Well, I could get heads, heads, heads, tails, tails, tails, or tails, heads. Right? So there’s four possible land on the edge. Yeah, right. Right. So yeah, yeah. Well, then I have to increase my probability space too, if I wanted to include that. But but now notice, I write down the probability space first, but I haven’t flipped my coin yet. So it’s, it’s, it’s the space of potential outcomes of things that I can do. And that’s what probability spaces are. And so yeah, okay. So when I write down a probability space for consciousness, it’s a probability space in which I’m thinking about, in the first instance, that it’s about what is the probability of this, I’ll experience green, or mint, or the sound of a trumpet, or so all these different conscious experiences. So the probability space is a space of all possible kinds of conscious experiences that this particular agent might have. And you can imagine that there’s, for some agents, maybe that’s, they’re simple, they only have the experience of red, period. That’s it. That’s all this agent has red. Another one, the other one can experience red and green. And then another one can have 10 trillion experiences, you could imagine agents with, and then they can be related, right? Well, maybe the red agent can be thought of as a subspace of the one that says red and 10 million other things. So we can now- Right, right. Depends on how articulated the organism is, right? So yeah, the simpler organisms, exactly, the probability space around them collapses. That’s right. And so, right, right. And so all the infinite number of potential probabilities that we see in front of us just collapse into maybe five choices, something like that. And sometimes, yeah. Okay, so you know, Carl Friston, so this is quite interesting. So I talked to Carl Friston about emotion, about hope, positive emotion, let’s say, incentive reward, positive emotion. So positive emotion in that sense is a reward that signals advancement towards a goal. Now I’d already been conceptualizing with my students as had Friston anxiety as a marker for the emergence of entropy. But Friston pointed out, now, and I want to make a connection between his thinking and yours here, Friston pointed out that you can map positive emotion with respect to entropy too, because if you’re looking for a desired outcome, so imagine you’re trying to get a grip on the world to bring about a certain reality. If you see yourself making a step towards that end such that the number of potential pathways to that end decreases somewhat, that produces a dopamine kick. And that’s a signal of reduced entropy in relationship. And it seems to me that entropy is always calculated in relationship to a goal, right? You’re saying, well, how entropic is the current space? And you can’t answer that. You have to say, how entropic is the current space in relationship to the ordered state that I’m trying to bring about as a consequence of my actions? And then now and then you’ll stumble across something that blows up in your face. Let’s say, like, I’ve always thought about this, like, imagine you’re driving your car to work. Okay, and you might say, well, what is your car? And the objective materialist would say, well, it’s a enclosed shell with four tires, it would give you a materialist description. But I would say, no, no, no, that’s not how your nervous system is responding at all. Your nervous system, for your nervous system, the car is a conveyance from point A to point B. So it’s a tool. And it’s a tool that signifies zero entropy, essentially, as long as it performs its function. And then let’s say your car breaks down, and now you’re on the side of the road. Now what happens to you is the probability space around you, I would say it becomes more distal, any of your desired goals become more expensive and harder to compute. Right? What’s wrong with my car? Was I an idiot for buying that car? Am I generally an idiot? Am I going to get in trouble with my boss? What’s going to happen to the rest of the day? No, what’s what’s going to happen when I go see the mechanic? Right? The landscape blows into a broader range of unconstrained potentialities. And that seems to be signaled by anxiety. And anxiety then prepares your body for a multitude of potential actions. And the problem with that is that it’s very physiologically costly. Right? So that stress and that’ll wear you to a frazzle. So, okay, so is any of that not in accord with the manner in which you are modeling your theory of conscious agents? Right. So in the theory of conscious agents, I should say that in addition to the probability space and the conscious experiences that it allows, there is the dynamics. It’s a Markov chain, Markovian dynamics, where you have these matrices that describe the probability if I’m experiencing red now, what’s the probability I’ll experience green the next time I haven’t experienced. So there’s a dynamical and when we do the analysis, it turns out that our Markovian dynamics need not have an entropic arrow of time. It can be a stationary dynamics in which the entropy does not increase. So entropy in this realm of- That’s kind of what you hope. Right. That’s one of the things that makes things constant, right? Is that you assume that the entropic transformation is negligible. That’s why you can ignore things, right? When you ignore things, when you ignore almost everything, you’re assuming that the entropic transformation is negligible. Well, what I’m saying is that it’s possible to model a reality in which entropy doesn’t increase, period. It’s not ignoring anything. That’s the nature of this deeper reality outside of space time. But then it turns out to be a theorem that if you take a projection of that non-entropic, there’s no arrow of time in the sense of increasing entropy of this Markovian dynamics. But if you take a projection of it by conditional probability, any projection of it, it’s a theorem that you will, as an artifact of projection, have the illusion of an arrow of time. You will get- Right. Well, is that because- Well, look, if you’re pursuing a pragmatic goal, things can fall apart and go wrong. And that is an increase in entropy within the universe defined by that goal. That may say nothing about entropy per se as a characteristic of broader reality. See, I’ve always had this issue with entropy because entropy always seemed to be, to be by necessity, subjectively defined. It has to be disorder in relationship to some positive state of order. And then you get back into the Darwinian problem at that point. Like if it’s, well, if it’s bounded by motivation, then it’s encapsulated within a Darwinian space. So, okay, so in terms of your conception of objects, let me try this out. So I’m looking at this teleprompter here, and you’re sitting in the middle of it. Now I’m treating that like a set of conditional probabilities, right? I’m presuming that what this machine is doing right now is very much predictive of what it’s going to do in a second. And I’m predicating my perception itself on that reality. Now, you know, it could burst into flames. Now I feel that the probability of that is very low, so I’m not going to perceive the machine that way. Now, you know, there are disorders, obsessive compulsive disorders, a good example, where people stop being able to reduce that probability landscape to predictable safety, and they start reacting to almost everything as if it’s unpredictably dangerous. And, you know, things are, so I had clients, for example, they would go into a building. And the first thing they would do is look for all the fire escapes. And what they asked me was, well, why don’t you do that? Because the building could burn down and people do get trapped in buildings, and that’s a horrible way to die. So the mystery isn’t why they did that. The mystery for them was why everyone didn’t do that all the time. And I actually do believe that the great mystery is why people aren’t scared out of their skulls all the time, not why they’re sometimes calm. But so can you imagine an object now, the object is surrounded by a probability distribution, I would say, and that probability distribution is all the things that object might turn into in some period of time, let’s say. And I would say to some degree, when you look at the object, you actually also perceive that probability space because, you know, although I see that this teleprompter is stable, it’s unstable enough and dynamic enough to provide me with a representation of you. And so I’m playing with the by by by seeing the object and interacting with it, I’m playing with the probability space around it. So the is it the case that you see the damn probability space when you look at the object? Well, I don’t know if we see it this the space itself, we certainly were estimating what we think are the probabilities for various good things and bad things to happen. But but I would say that this this whole business about entropy increasing and so forth. First, I should point out that Shannon entropy, which is what we’re talking about here, it turns out not to be the most general notion of entropy. There are mathematicians and physicists are looking at broader definitions of entropy. There’s something called the solace entropy and others. So there are technical reasons for why I mean, Shannon entropy is great and is very, very useful. And when I was talking about the entropy of our dynamical systems and not having, you know, increasing entropy, I was talking about Shannon entropy, but there are more more general notions of entropy that are that are important. So, so I would say that that the very whole the whole structure of needing to estimate probabilities and worrying about outcomes and rewards and so forth, from the point of view of our dynamics of conscious agents, all of that, in fact, all of Darwinian theory is an artifact of projection. So, so it so here’s a dynamic of conscious agents outside of space time, there need not be any competition, no limited resources, no, no arrow of time. And yet, when I take any projection of that dynamics to get a new Markovian dynamics that has lost just a little bit of information, I will have an arrow of time and it can look like separate organisms competing for resources and so forth. In other words, I mean, I love Darwin’s theory of evolution, but natural selection is very powerful. I think the entire theory is not a deep insight into reality. I think it’s an artifact of projection, the very arrow of time. You think about the arrow of time, it is the fundamental limited resource in evolutionary theory. Time is the fundamental limited resource. If I don’t get food and time, I die. If I don’t made in time, I don’t reproduce. And if I don’t breathe air and time, so time is the fundamental limited resource. And the arrow of time itself need not be fundamental. It could be entirely an artifact of projection. So that what that means is in this against again, Okay, well, then I’d like to know this is back to the most fundamental possible question we could be describing is, well, what’s the nature of reality itself? I mean, when I was debating with Sam Harris, we got hung up on this consistently because I wasn’t willing to use the same definition of truth that he was. He uses an objective materialist definition. And I think that, you know, truth flies like an arrow, let’s say it’s got a functional element to it. That’s that you cannot eradicate. There’s no way out of that with an objective materialism, as far as I can tell. Now, you said the Darwinian race and the arrow of time is just an artifact. But if I said, well, hold on a second. I don’t exactly know what you mean by artifact then because if I don’t act like there’s an arrow of time and restricted resources in that regard, then I’m going to die. And that’s real enough for me. You know, you might even say, well, my death has little to do with the fundamental structure of reality. But I would say, well, it has enough to do with it. So it happens to concern me. And, and, and so, you know, we start to get into a discussion about what constitutes reality itself. If it, if this is just a projection, what in principle would be real? Right. So on this theory, then consciousness is the fundamental reality and the conscious experiences that observers have as the fundamental reality. And the experience that we have of space and time is a projection of a much deeper reality. And that, and that projection, because it loses information is necessarily going to have artifacts in it. And among the artifacts are things like separate objects in space and time, space and time itself is an artifact. So, so one reason I’m not a materialist is because our best materialist theories, namely evolution of a natural selection, and also quantum field theory and Einstein’s theory of gravity, they tell us that space time has no operational meaning at 10 to the minus 33 centimeters or 10 to the minus 43 seconds. In other words, our theories, our scientific theories that are that are the foundation of our materialist ideas, tell us precisely the scope and the limits of materialism, materialism, that kind of materialism is fine down to the Planck scale, 10 to the minus 33 centimeters. And after that, it completely falls apart. It’s utterly irrelevant. That’s right. The space time physicalist matter kind of materialism falls apart. And it’s not because of religious ideas. I’m saying it is I’m just listening to the science. It’s science tells us space time has no meaning beyond the Planck scale. And that’s why the avant-garde high energy theoretical physicists are now looking for structures entirely outside of space time, not curled up inside space time, entirely beyond. So, so it’s in that sense that materialism, and by the way, this is, I should say this about all scientific theories. My view about all scientific theories is that each scientific theory starts with certain assumptions, the premises of the theory. And it says, if you grant me those assumptions, then I can explain all this wonderful stuff. Okay, so how did you come to that conclusion? Because that’s, see, see, this is, I’ve been trying to wrestle with this with regards to, say, the potential relationship between the integrity of the scientific process and an underlying transcendent ethic. So I think, for example, I talked to Richard Dawkins about this a little bit, although we didn’t get that far for a variety of reasons. But like, I think that to be a scientist, there’s certain things that you have to accept on faith. These would be equivalent to those axioms. And I’m not talking about necessarily a scientific theory here, as you were, but the practice of science itself. So for example, you have to act as if there is truth. You have to act as if the truth is discoverable. You have to act as if you can discover it, then you have to act as if you discovering the truth and communicating it is good. And none of that is provable scientifically. You have to start with those axioms before you can even make a move. And it could be wrong, you know? I mean, we think that delving into the structure of the world with integrity is redemptive. We think that knowledge is useful pragmatically. But, you know, we’ve invented all sorts of things that could easily wipe us out, like the hydrogen bomb perhaps being foremost among those. And so the evidence that that set of claims is true is sorely lacking. Or you could say it’s 50-50. That’s another way of thinking about it. But I’m very curious about how you came to the conclusion that scientific theories themselves have to be axiomatically predicated. How did you walk down that road? Because that’s not a road that very many people walk down. Well, if you just look at any scientific theory, say Einstein’s theory of special relativity, he says, let’s start with two assumptions that, you know, the speed of light is universal for all observers, and that the laws of physics are the same in all inertial frames. He says, if you grant me those two miracles, then then then the whole- Here we go. And you could do the same thing. And so does Riemann. Darwin starts off and says, grant me that there are organisms in space and time and resources, and these organisms are competing for resources. Now I’ll give you a theory. So, if you just look at any scientific theory, a good theory will make explicit the assumptions, but if it’s not, you can find what the assumptions are. So there’s no theory- Okay, so- There’s no theory of everything. Do you think that there’s- Is there any difference between- Technically, I’m thinking, philosophically, I don’t see any difference between the claim that a given theory has to have axioms that aren’t provable from within the frame of that theory. That’s Godel’s theorem, as far as I could tell, applied much more broadly. I don’t see any difference between that and the proposition that to get the game started, there has to be- It’s something akin to a miracle. I mean, because these axioms- Imagine that a miracle inside a system is defined as any occurrence that isn’t governed by the rules that apply within that system. That’s a good working definition. Now your proposition is, well, I don’t care what theory you’re coming up with, there’s going to be a set of axiomatic presuppositions that are a launching point. See, I also think those axiomatic presuppositions are where you put all the entropy. You say, grant me this. It’s like, well, that takes care of 95% of the mystery, so we’ll just shelve that invisibly, right? Because it’s hidden inside the axioms, and then you can go about manipulating the small remnant of trouble that you have left over. I also think this is why people don’t like to have their axioms challenged, eh? Because if you say, well, I’m not going to accept that, then you let loose all the demons that are encapsulated within those axioms, and they start roaming about again, and people don’t like that at all. A good scientist will want to have their assumptions made absolutely mathematically, precisely, and explicit. So they’re just laid out there, and they say, these are the assumptions of the theory, and given these assumptions, I can now prove this. And this is the glory of science, where we put down precisely what our assumptions are, and then we look at it mathematically, and we can get both the scope of those assumptions, how much can we do with those assumptions, and the limits. Like in the case of space-time, the limits are 10 to the minus 33 centimeters. Game over. By the way, it’s not that deep, in my view. It’s not 10 to the minus 33 trillion centimeters. It’s just 10 to the minus 33, and the game is over for space-time. So that’s a good antidote for dogmatism, because your own theory, a mathematically precise theory, will tell you the limits of your assumptions, and then say, okay, now you need to look for a broader framework with deeper assumptions. But they will be new assumptions, and so I view this as infinite job security for scientists, because we will never ever get a theory of everything. We’ll always have a theory of everything except our current assumptions, and I agree with you, that those assumptions will essentially be the whole bailiwick of what we’re doing. So there’s a reality, whatever it is. Now, this is for me something of an interesting mystery. Our theories in some sense don’t even scratch the surface of the truth, and yet, because this process will go on forever and will still essentially have measure zero of the truth, and yet Einstein’s theory and quantum theory gave us the technologies that are allowing you and me to talk across the country. Well, so you could say, well, you could say that partly what’s happening there is that the more sophisticated the theory, the broader range of probable states of any given object or system of objects can be predicted. It’s something like that. But Piaget pointed that out when he was talking about developmental improvement in children’s cognitive theories. And so, you know, if you look at someone like Thomas Kuhn, Kuhn presumed that we undertook multiple scientific revolutions, but there was no necessary progress. There were just different sets of axioms. And Piaget knew about Kuhn’s theory, by the way, but Piaget’s point was, no, you’ve got it slightly wrong, because there is a progression of theory in that a better theory allows you to predict everything the previous theory allowed you to predict, plus some additional things. Now, your point would be, well, we could just continue that movement upward forever, right? Because the landscape of potentiality is inexhaustible. And so again, you can have your cake and eat it too. We can learn more. Einstein got us farther than Newton, which doesn’t mean that Einstein’s axiomatic set is the final set. Okay, so let me put a twist in this. I’ve been thinking about this recently. I’ll be I’m writing a new book. And one of the things I’m doing in that book is doing an analysis of the story of Abraham. Abraham’s very interesting story. Okay, so Abraham is called out into the world, even though he, he sort of hung around his father’s tent till he’s like 70. So he, he had, he had utopia at hand, he didn’t have to do any work to get everything he needed. But that wasn’t good enough. So a voice comes to him. It’s the voice of conscience, I would say, and says, Look, you’ve got all this security, but that isn’t what you’re built for. Get the hell out there in the world. And so he does that. And then all hell breaks loose. It’s one bloody catastrophe after another starvation and tyranny and warfare and the necessity of sacrificing his son. It’s just like one bloody thing after another. Okay, but during that process, Abraham, Abraham continues to aim up and he makes the proper sacrifices. And the consequence of that is that God promises him that his descendants will be more numerous than the stars. So I was reading that from an evolutionary perspective, and I thought, okay, what’s happening here is that, that the narrative is trying to map out a pathway that maximizes reproductive fitness, all things considered. Now, the problem I have with theories like Dawkins, let’s say is Dawkins reduces, and you tell me if you think this is wrong, Dawkins implicitly reduces sex to lust, then he reduces reproduction to sex. And the problem with that is that reproduction is not exhausted by lust or sex, quite the contrary, especially in human beings, because not only do we have to chase women, let’s say, but then when we have children, we have to invest in them for like 18 years before they’re good for continual reproduction. And we have to interact with them in a manner that’s predicated on an ethos that improves the probability of their reproductive fitness. And so reproduction, see, this is something that the Darwinists, the casual Darwinists do very incautiously, as far as I’m concerned, because they identify the drive to reproduction with sex. And that’s a big mistake, because sex might ensure your reproduction proximally for one generation, but the pattern of behavior that you establish and instantiate in your offspring, which would be an ethos, might ensure your reproduction multi-generationally, you see, and that appears to be what’s being played out in this story of Abraham is that the unconscious mind, let’s say, trying to map the fitness landscape is attempting to determine what pattern of behavior is most appropriate if the goal is maximal reproductive fitness calculated across multiple generations or maybe across infinitely iterating generations. And so that points to something again, like you said earlier, you called it a general fitness what was it? I gotta get it here, big fitness payoff, right? And that could be the ethos to which all these subsidiary ethoses are integrated. See, okay, okay. So, well, so I’m wondering what you think about that is that, first of all, what you think about the proposition that evolutionary biologists, Dawkins is a good case in mind, have erred when they’ve too closely identified reproduction like with short-term sex. It’s like that isn’t a guarantee of reproduction. We wouldn’t invest in our children if that was the case. We would just leave them. The sex is done, we’ve reproduced. You need an ethos to guarantee reproductive fitness across time. Well, there’s several levels here. First, Dawkins, of course, understands that most reproduction is asexual, right? So sexual reproduction is a relatively recent thing. Most reproduction has been asexual. So Dawkins is very famous for talking about the selfish gene. And it’s really, when he talks about reproduction, it’s about genes reproducing themselves. It’s really not so much about sex. Sex is one way of having that happen, but bacteria do it without sex. And so the emphasis on sex was, I would say, Dawkins, of course, understands that sex isn’t fundamental. Now, when it comes to human motivations and mammal motivations, perhaps in that specific context, you might then be talking about it. But even there, when you start talking about sexual reproduction, there are many, many strategies that organisms use. So for example, some spiders will have just hundreds of babies and eat some of them. They’ll eat some of them and let the others do that. Having the babies is their only job. And after that, the babies are on their own. And so there are different strategies. So this is where Dawkins is quite famous, justifiably for his work on the selfish gene idea. That is, there are different strategies, but the only thing that matters in this framework is what is the probability that particular genes spread through the population in later generations. Sex came along, apparently, to deal with… Okay, as one of the pathways to that, right? That’s right. But there’s another framework in thinking about all this as well. So again, I love evolutionary theory, I think, in terms of models of evolution and so forth, of creatures and their behaviors. It’s an incredibly powerful theory. I’ve used it a lot. My book, Case Against Reality, talks about it in great detail. It’s a wonderful theory. But I think that from this deeper framework that science is now moving into beyond space-time, all of evolutionary theory, all of it is an artifact of projection. In other words, if you’re looking, like, from a spiritual point of view, for some deep principles, deep spiritual principles, evolution, I don’t think is deep enough. I think that all of it is an artifact of space-time projection. And if you’re going to be looking for deep principles about the spiritual traditions, talking about Abraham, and really thinking big, I think that thinking inside space-time is not big enough. You’ve got to step entirely outside of space-time. Space-time has all these artifacts. And we’re so used to being stuck in the headset. Well, there is an insistence upon that in the Judeo-Christian tradition, because God is conceptualized, what would you say, traditionally, as being entirely outside of time and space. And so whatever works for the human landscape and the divine landscape, they’re not the same. There’s a relationship between them, however, but they’re not the same. Okay, so now, okay, so let me ask you about that. Now, you have made the case, not least in this interview, that consciousness is primary. Now, consciousness uses these projections. So how do you reconcile the notion that consciousness is primary? And I want to make sure I’m not misreading what you’re saying, that consciousness is primary. But consciousness operates in the world with these projections. See, because this is the thing I grapple with, is that if survival itself is dependent on the utilization of a scheme of pragmatic projections, in what sense can we say that reality is something other than that? Because, see, part of this is something that Peirce and William James wrestled with too. It’s like, well, why make the claim that there is a reality outside of the human concern with survival and reproduction? If consciousness is the primary reality, and it’s using projections to orient itself so that it can survive and reproduce in the biological sense, how can you even begin to put forward a claim that there is a reality that transcends that? Like, on what grounds does it transcend it, in relationship to what? Right, so these are deep waters. And the idea that I’m playing with right now is that this consciousness is, there’s one ultimate infinite consciousness. And what is it up to? Knowing itself. But how do you know yourself? Well, there are certain theorems that say that no system can actually completely know itself. Right? So if this one infinite consciousness wants to know itself, all it can do is start looking at itself through different perspectives. So putting on different headsets. So space-time is one headset. And from that perspective, here’s a projection. So this is a projection of the one infinite consciousness. And in that perspective, it looks like evolution by natural selection. It looks like quantum field theory and so forth. And it looks like I need to play the game this way. But this is a trivial headset. This is actually, I think, one of the cheaper headsets. Okay, that’s very interesting. Okay, so one of the things, so while writing the book that I’m writing now, I’ve been walking through all these biblical narratives. And one of the things they do, every single narrative provides a different characterization of the infinite. There’s no real replication. It’s like, well, here’s a picture of the divine, and here’s another one, and here’s another one, and here’s another one. Now there’s an insistence that runs through the text. This unites the text that those are all manifestations of the same underlying reality. But it is definitely the case that what’s happening is that these are movies, so to speak, shot from the perspective of different directors. And it does seem to me akin to something coming to know itself. There’s this ancient Jewish idea. This is a great, it’s like a Zen cone. It’s a great little mystery. It says, so here’s the proposition. So God has traditionally imbued the following characteristics, omniscience, omnipresence, and omnipotence. What does that lack? And you know, you think, well, that’s a ridiculous question, because by definition, that lacks nothing. But the answer is limitation. That lacks limitation. And that’s actually the classical explanation for God’s creation of man, is that the unlimited needs the limited as a viewpoint. It has something to do with the development of, as you pointed out, I believe, it has something to do with the possibility of coming to, it’s something like conscious awareness. You see this in T.S. Eliot too. I don’t remember which poem where he talks about coming back to the point of origin, which is like the return to childhood, you know, that heavenly notion that to enter the kingdom of heaven, you have to become as a little child. It’s like, but there’s a transformation there so that that return to the point of origin is accompanied by an expansion of consciousness. It’s not a collapse back into childish unconsciousness. It’s the reattainment of a, what would you say? It’s the reattainment of the state of play. That’s a good way of thinking about it, that obtained when you were a child, but with conscious differentiated knowledge. So there is this tremendous narrative drive in the Western tradition towards differentiated, comprehensive understanding as a positive good. And that seems tied up with the continual drama between God and man. So, and I do think the scientific enterprise is an offshoot of that. That’s what it looks like to me historically. So, okay, so how in the world do you survive in psychology departments, given what you’re thinking about? Well, I’ve got the mathematics. So as long as, if I was just talking this stuff without any mathematical underpinnings to it, it would be dismissed, of course. But, but the, you know, we’ve, we’ve, in the case of the evolutionary stuff, we’ve published papers in the journal theoretical biology, for example, and elsewhere, where we actually put the mathematics out there. So it’s peer reviewed. And, and I think that it’s a bit surprising, but and I, you know, I, I’m a minority, a small minority, but you know, that’s, that’s the way science progresses. It, it, it proceeds one funeral at a time. And so it progresses by, by minorities of one. Exactly right. So, so, and, and, and scientists understand that, you know, you, you want to have independent ideas, think out of the box, make it mathematically precise. Most of our ideas will be nonsense, including mine, but you, you got to put them out there and push them and, and see, see what happens. I have, I’ll say in terms of I’ve, I’ve gotten some stiff pushback, for example, some philosophers have published papers recently where they give the following argument against my Darwinian theory. They’ll say, look, Hoffman uses evolutionary game theory to show that space and time and physical objects and organisms don’t exist. Well, he’s got himself what they say an unenviable dialectical situation. Either evolutionary game theory faithfully represents Darwin’s ideas or it doesn’t. They say, okay. So if it doesn’t, then he can’t use it to say that the organisms and resources are not fundamental in space time. And if it does faithfully represent Darwin’s ideas, well, Darwin’s ideas are that space time is fundamental in their organisms and resources. So, so it couldn’t possibly contradict that. So either way Hoffman is screwed. Right. There’s nothing he can do. So, so, and, and so, and that’s been published actually in high, high value philosophy journals. And my response is, is, is quite simple. It misunderstands science completely. Every scientific theory has, when you write it down mathematically, it has a scope and its limits and the mathematics tells you both the scope and the limits. So for example, just to be very concrete Einstein’s theory of gravity, right. And I think 1907 or so he had this, the big idea. If I was standing on a weighing machine in an elevator and all of a sudden the cord was cut and I was in freefall, the, I would all of a sudden be weightless. That was his big idea for his theory of gravity. It took him years, seven or eight years to actually make the mathematics, but he wrote, wrote down his field equations. So, so those field equations are Einstein’s mathematics to capture his idea that space time is fundamental and has certain properties. Well, a year after he published it, Schwarzschild, a German scientist discovered that they entail black holes. And we’ve eventually found out that this theory entails that space time itself has no operational meaning beyond 10 to the minus 33 centimeters. So we could use the same argument that’s been used against me against Einstein. Now look, okay, Einstein’s field equations, either they’re faithfully representing Einstein’s ideas or they’re not. So we can use the same argument against Einstein, they’ve been used against my theory. Now either Einstein’s field equations capture his ideas faithfully or they don’t. If they don’t, then we couldn’t use them to show that space time isn’t fundamental. And if they do, they couldn’t possibly show that space time isn’t fundamental. That last step is the wrong one. The equations are there to show you the limits of your concepts. They give you precise and that’s, so that’s what these philosophers have missed is that the equations that we write down tell us not just the scope, but the limits of our theories. And that’s, that’s why science is so valuable because it tells us your theory, your assumptions go this far and no further. So that’s all I’ve done with Darwin’s theory of evolution is to say this theory- That also sounds to me very much like a vindication of the fundamental claim of the pragmatists, which is that we accept something as true without noticing that what we mean is true in a timeframe with certain implications for instantiation and something like that. And so true is a lot more like, does the bridge stand up when a hundred cars go across it? It’s not some final comprehensive, all encompassing definition of the truth for all time. And you’ve already made the case that it can’t be because that truth is an ever receding goal. It’s always bounded. Okay. So when I came across that, I thought, okay, well bounded by what? And it’s, well, it’s bounded by our aim. And, and then that’s bounded by our motivation. And then that’s in nested inside a Darwinian world. Okay. Now let’s go after the game theory. Well, let me just say one thing about that. Sorry. Go ahead. Yeah. Yeah. I would just say that the very deep, deepest spiritual traditions really say that upfront. Like the Tao Te Ching starts off that says the Tao that can be spoken of is not the true Tao. Once you understand that, then go ahead and read the rest of it. That’s a good example, because that’s a great book. Yeah. That’s a great book. And I think that that’s also the way we should think about our science. The science that can be spoken of is not the final reality. But, but given that it’s a wonderful thing to do science, and we should do science, and we should do it very, very rigorously, but we should always understand that if we’re talking about a theory of everything, it should be with a wink and a nod, because there is no theory of everything that we can write down. Right. It’s the theory of everything that we’ve discovered so far maybe, but it will never be the final theory of everything. Right. And it might have a broader, broader range of potential applications as well, but that doesn’t mean that we’ve exhausted the landscape of comprehensive theories. Right. Okay. So now the philosophers that you described as objecting to your theory said that if evolutionary game theory is correct, and it models Darwin’s propositions appropriately, then well, so game theory is extremely interesting to me, although I wouldn’t say I’m an expert in its comprehension, but I understand it’s just, I believe, and it seems to me to be something like this, is that if you iterate interactions, an ethos of one form or another emerges. So for example, if you play tip for tap simulations, you find out that the best trading strategy is cooperate, but slap when necessary, and then forgive something like that. And so what it points to very interestingly is something like a concordance between objective reality insofar as objective reality is an emergent pattern coming out of iterative interactions and something like an ethos. So the first question I have is like, why are you interested in evolutionary game theory? And why do you think that it is a valid representative, a more differentiated representative, if I’ve got the language right, of Darwinian theory? Oh, well, I’m interested in it because that’s within the field of evolutionary theory itself. Evolutionary game theory is taken as the prize mathematical tool for really understanding things. So that’s just the framework of the science itself. Okay, so that’s accepted as far as you’re concerned. Yeah, I mean, of course, there’s always debate, but by the vast subs… But it’s the received opinion. So if I wanted to, as a scientist, if I wanted to analyze Darwin’s theory for this issue about truth, and I wanted to do it rigorously, the tool was evolutionary game theory. That was the tool to use. And that’s not because I think it’s the final word or the truth, it’s just our current state of play in the field. It’s the best we have. That’s the best we have. And I wanted to use the best tool we have. And that’s the way we’re always pulling ourselves up by the bootstraps in science, right? We always say, these are the best theories we have and the best tools we have so far. Of course, our goal is not to prove that we’re right. Our goal is to find the limits of our current theories and transcend them. So we’re looking for, are the best tools that will say, aha, Darwin goes this far and no further. Spacetime goes this far, you know, in high energy theoretical physics. Einstein’s wonderful theories, they’re incredible gift. They go to 10 to the minus 33 centimeters and they stop. That gift stops right there. And now we have to go entirely outside. And that will be the never ending pattern of science is that whatever the scientists are finding outside of spacetime, that will just be our next baby step and we’ll analyze that and then say, okay, what’s beyond that and beyond that. And science will continue to. So as long as you recognize that that’s the game, you’ll realize that there’s no theory of everything in science. And then the question is, who am I? Who are we that are able to do this game? And that’s a very interesting question. Well, you know, there’s lots of things I’d like to ask you about, but that’s a pretty good place to stop. And we’re damn near at an hour and 30. So I hope I have the privilege of furthering my discussion with you at some point in the not new, not too dear future. I would like to say, is there anything in closing that you would like to bring to the attention of the listening audience, the watching audience that you think that we needed to cover to make what we have covered comprehensible? Or is that also in your estimation, a good place to stop? I’ll just say one little thing, I guess, and that is, some people might think, well, he’s got this theory of consciousness out of outside of spacetime. So what, who cares? And the, and I would agree with that unless I did something more. So what we’re trying to do now is scientists to say, we have this mathematical model of consciousness outside of spacetime. We just published a proposal for how to actually test it. So we’re going to have a projection into spacetime. We’re working on that projection. We’d like to model the, the inner structure of the proton. We would like to have a dynamics of conscious agents that projects down and gives us what’s called the momentum distributions of quarks and gluons inside a proton at all the Bjork and X and Q squared, the different spatial and temporal resolutions that particle physicists have studied. So, and the reason we’re going there is not because I think that’s the most important application of a theory of consciousness. It’s the most accessible one. That’s the simplest part of our science right now. Ultimately, of course, the brain has the nice neural correlates of consciousness. We want to understand that, but that’s really complicated. So we’re going to go after, if we can model the proton and get it exactly right, get the momentum distributions to several decimal places, it doesn’t mean our theory is right, but it does mean it can’t be dismissed out of hand. And so that’s, that’s what our goal is to take a theory of consciousness, not just to airy fairy, wave our hands and the other, but to actually get in there and predict the structure of the, the inner structure of the proton with great detail. If we can do that, then I would say we then can start to move up, you know, to molecules and then ultimately to neural systems in the brain and try to understand the neural correlates of consciousness, but not the neural correlates. The brain does not cause consciousness on this model. The brain is merely a symbol inside the headset. Right. So, so, and in fact, I would say this neurons do not even exist when they’re not perceived. Neurons cause none of our behavior. And yet I’m, I’m a cognitive neuroscientist. And I think that we should study, we, we, neuroscience is wonderful and we need more funding for it because it’s more complicated than we thought. We thought we look inside the brain, we see neurons. That’s because that’s the reality. There are neurons. No, that’s, that’s the interface description of something that’s much, much more complicated. We have to reverse engineer neurons to this network of conscious agents outside of space time. So we need more funding for neuroscientists, much more complicated. So, so I would just a little brief, of course, as you can imagine, I’m talking about something that could take hours to go into detail, but, but just to, to put those out there and say, these are objections and people might have, so we’re headed. Okay. Well, I do, I do have one. Okay. I do have one other question that I guess I do have to throw it out. So you have a very radical conception of consciousness. What has that done for you? Existentially, do you think? I mean, you’re obviously thinking about the place of consciousness while you’re thinking about it existentially. You’re thinking about the place of consciousness in the cosmos and you regard it as a fundamental reality. So what has that done to the manner in which you contemplate your own, say, mortality or the purpose of your life? And what, what’s that done for you on that, on that side of things? Quite a bit. That it’s really hit me in the face because I’m intuitively as much a physicalist and a materialist as anybody else. I mean, I, right. I’m wired up to, to believe all that. And so it’s, it, it’s come as a terrible shock to me. My, my whole self image has had to change and it’s, and I, I just. In what direction, in what direction your self image changed? What changed? Well, I thought of myself as a little object in space time. Right, right, right. And the death of the body is, is ultimately the death of me. And now it’s, well, our best science says that this is, you know, my body is just an icon in a headset. So in some sense, it’s just an avatar. This body is just an avatar. And so death is more like taking off a headset. So, but, but my emotions don’t agree with that. So I’ve got this really interesting. Well, that’s probably just as well. Right. Yeah, exactly. So, so it’s, it’s, so I do spend a lot of time in meditation and my father was a Protestant minister, a fundamentalist Protestant minister. I was, so I was raised in the Christian church. And so I, I look at those points of view. I look at the Eastern mystical stuff. I meditate myself and my, my ultimate thinking about this is, as I said, we can never have a theory of everything. And that includes of who I am. So the question about who I am, my best guess right now is at the deepest level, I and you are in fact, the one consciousness just looking at itself through different avatars. So it’s really the one using a Jordan avatar to talk to the one, you know, Hoffman avatar. And that’s, that’s what’s, what’s, what’s going on here. And in that sense, so are you responsible for being the best possible avatar you can be, so to speak? Well, in some sense, within this projection, within this headset, morals of a certain kind are the rules of the road. But, but my guess is that when we take the headset off, we’ll just laugh. That was, that was, that was what we had to do in this headset. But that was, I am not this avatar. I am the consciousness that’s far that transcends space and time. Well, you know, with the next time we talk, maybe that’s a road we should wander down. Sure. We didn’t, we didn’t get into the metaphysics of ethics, let’s say during this conversation. And there’s plenty of that. That’s obviously a whole other area. Okay. Okay. Well, that would be good. All right. Well, so to everyone watching and listening, thank you very much for tuning into this podcast. I, as most of you know, I’m going to talk to Dr. Hoffman for another half an hour behind the daily wear plus platform. And I’m going to see if I can find out where in the world his interests stemmed from and how they initially manifested themselves and developed across time. We’ll do that as much as we can in half an hour. Thank you to the crew here up in Northern Ontario for journeying up here to do this podcast. Thank you, Dr. Hoffman, very much for your time today to the daily wire plus people for making this possible. That’s also much appreciated. And we’ll see all of you watching and listening hopefully on another podcast. Thank you very much, sir. Thank you, Jordan.