1 00:00:00,200 --> 00:00:04,680 Speaker 1: From UFOs to psychic powers and government conspiracies. History is 2 00:00:04,760 --> 00:00:09,080 Speaker 1: riddled with unexplained events. You can turn back now or 3 00:00:09,160 --> 00:00:12,119 Speaker 1: learn the stuff they don't want you to know. A 4 00:00:12,200 --> 00:00:19,160 Speaker 1: production of iHeartRadio. 5 00:00:26,160 --> 00:00:29,000 Speaker 2: Hello, welcome back to the show. My name is My 6 00:00:29,280 --> 00:00:29,960 Speaker 2: name is Nola. 7 00:00:30,720 --> 00:00:33,559 Speaker 3: They call me Ben. We're joyed as always with our 8 00:00:33,640 --> 00:00:37,640 Speaker 3: super producer Dylan the Tennessee pal Fagan. Most importantly, you 9 00:00:38,920 --> 00:00:42,520 Speaker 3: are here. That makes this the stuff they don't want 10 00:00:42,600 --> 00:00:45,559 Speaker 3: you to know. It is time for us to return 11 00:00:45,640 --> 00:00:50,000 Speaker 3: to our weekly listener mail program. We've been traveling all 12 00:00:50,040 --> 00:00:52,800 Speaker 3: around the place, and we have such adventures to show you. 13 00:00:53,159 --> 00:00:55,960 Speaker 3: By the place I mean, I guess the US, but 14 00:00:56,040 --> 00:01:01,920 Speaker 3: also the world. This is March twenty six, twenty twenty six. 15 00:01:02,640 --> 00:01:05,320 Speaker 3: We are excited. We're also going to be on the 16 00:01:05,400 --> 00:01:09,440 Speaker 3: road a little bit more often, so we can't wait 17 00:01:09,480 --> 00:01:12,280 Speaker 3: to share our own stories with you. But we're going 18 00:01:12,360 --> 00:01:16,800 Speaker 3: to hear from Meta on Moltz book. We're going to 19 00:01:16,840 --> 00:01:20,039 Speaker 3: explore some tunnels. I know, we're all excited about that. 20 00:01:20,680 --> 00:01:25,080 Speaker 3: But before we get to edny of that. Because we 21 00:01:25,240 --> 00:01:28,680 Speaker 3: left you for a few weeks without a dope, bega 22 00:01:28,720 --> 00:01:31,880 Speaker 3: to step two. Dylan, can you catch us up on 23 00:01:31,920 --> 00:01:34,759 Speaker 3: the roots at Vegas. 24 00:01:34,880 --> 00:01:35,400 Speaker 4: What's up. 25 00:01:35,480 --> 00:01:38,440 Speaker 5: I'm Professor Dorf, the most popular YouTuber in the world, 26 00:01:38,800 --> 00:01:42,120 Speaker 5: and I'm torpe Torpelson, chairman of the National Rudabaca League, 27 00:01:42,280 --> 00:01:44,480 Speaker 5: and we want to tell you about Rudebago Royals, the 28 00:01:44,520 --> 00:01:47,600 Speaker 5: new betting app for conspiracy theorists. Well Bigfoot win the 29 00:01:47,600 --> 00:01:49,000 Speaker 5: twenty twenty eight presidential election. 30 00:01:49,360 --> 00:01:51,880 Speaker 2: Well Mathman the league MVP for a third year in 31 00:01:51,880 --> 00:01:54,920 Speaker 2: a row. Well disclosure finally happen in twenty twenty six. 32 00:01:55,120 --> 00:01:56,080 Speaker 4: Diggitu yo. 33 00:01:56,280 --> 00:01:57,680 Speaker 6: If you can bet on all of this and more 34 00:01:57,720 --> 00:02:00,720 Speaker 6: on Rudebego Royals first five Rude big are on us, 35 00:02:00,960 --> 00:02:02,960 Speaker 6: So don't even think about it. Put all your hopes 36 00:02:03,000 --> 00:02:05,000 Speaker 6: and dreams in every last bit of your money into 37 00:02:05,040 --> 00:02:07,400 Speaker 6: pure speculation with odds that are never in your favor. 38 00:02:07,600 --> 00:02:11,200 Speaker 6: Rudabaka Royals the app that's everywhere TV, your phone, billboards, 39 00:02:11,200 --> 00:02:13,000 Speaker 6: computer's carrier, pitcheons, skywriting. 40 00:02:13,200 --> 00:02:14,680 Speaker 4: We're everywhere all the time. 41 00:02:15,080 --> 00:02:15,600 Speaker 2: You can't just. 42 00:02:20,639 --> 00:02:21,200 Speaker 3: I'm sold. 43 00:02:22,440 --> 00:02:24,240 Speaker 4: I'm speechless, is what I am. 44 00:02:24,320 --> 00:02:27,240 Speaker 3: Guys, step aside, Polymarket. 45 00:02:28,200 --> 00:02:31,400 Speaker 2: Do you create the music or do you find the music? 46 00:02:31,480 --> 00:02:36,120 Speaker 2: Dylan a little bit of both. Yeah, I tell ya. 47 00:02:36,639 --> 00:02:39,079 Speaker 5: We have a pretty great library you're at iHeart where 48 00:02:39,120 --> 00:02:39,440 Speaker 5: you can. 49 00:02:40,360 --> 00:02:43,800 Speaker 4: I'm pretty much anything. There's some good fun stuff on 50 00:02:43,840 --> 00:02:44,960 Speaker 4: that on that archive. 51 00:02:46,080 --> 00:02:48,360 Speaker 2: Well you're mixing it so well, whatever it is. 52 00:02:48,960 --> 00:02:53,079 Speaker 3: I'm laughing because that thing about our secret library is true, 53 00:02:53,200 --> 00:02:55,880 Speaker 3: and it's also a thing that we tell people as 54 00:02:56,120 --> 00:03:00,560 Speaker 3: like a party line talking point. Oh yeah, like, hey, 55 00:03:00,680 --> 00:03:05,040 Speaker 3: we can't really hire Tom Waits for this show because 56 00:03:05,080 --> 00:03:08,240 Speaker 3: he only gets paid and used fifty seven Chevy parts. 57 00:03:09,080 --> 00:03:15,280 Speaker 3: But we confide something. Oh, Dylan, you have outdone yourself. Man. Guys, 58 00:03:15,720 --> 00:03:20,040 Speaker 3: before we go any further, can we talk about video 59 00:03:20,120 --> 00:03:22,040 Speaker 3: games just a little bit, just real quick. 60 00:03:22,120 --> 00:03:24,400 Speaker 7: Yeah, Oh my gosh, yes, I just I'm on the 61 00:03:24,480 --> 00:03:27,000 Speaker 7: DLC for Control now finally, guys, I beat the regular 62 00:03:27,040 --> 00:03:29,680 Speaker 7: game part and now I'm on the DLC and boy, 63 00:03:29,680 --> 00:03:30,760 Speaker 7: oh boy, do I love. 64 00:03:30,680 --> 00:03:32,799 Speaker 4: Video games the whole time? 65 00:03:32,840 --> 00:03:34,040 Speaker 3: For Control Residence. 66 00:03:34,320 --> 00:03:35,040 Speaker 4: Yeah. 67 00:03:35,240 --> 00:03:39,680 Speaker 2: Are we going to talk about Marathon? The new Bungee title. 68 00:03:39,760 --> 00:03:41,640 Speaker 2: It I still haven't played yet. I do kind of 69 00:03:41,680 --> 00:03:43,000 Speaker 2: want to play it, but I don't know. 70 00:03:43,200 --> 00:03:44,240 Speaker 3: I feel like I got burned. 71 00:03:44,320 --> 00:03:45,040 Speaker 4: Bungeee is. 72 00:03:46,760 --> 00:03:49,280 Speaker 7: The shooting the shooty one, right, It's the big popular 73 00:03:49,280 --> 00:03:50,320 Speaker 7: shooting Xbox One. 74 00:03:50,600 --> 00:03:50,920 Speaker 2: Halo. 75 00:03:51,000 --> 00:03:53,600 Speaker 7: Halo's the one that's the clearly I see an x 76 00:03:53,720 --> 00:03:56,680 Speaker 7: Xbox guy. But I will say I almost bought an Xbox. 77 00:03:56,720 --> 00:03:59,200 Speaker 7: I wanted to play Starfield that bad and I didn't 78 00:03:59,240 --> 00:04:01,160 Speaker 7: do it. And I just found out that it's coming 79 00:04:01,200 --> 00:04:02,920 Speaker 7: out on ps five and fifteen days. 80 00:04:03,080 --> 00:04:04,960 Speaker 4: Thought it was already out, went ahead and clicked the button, 81 00:04:05,000 --> 00:04:05,760 Speaker 4: paid the money. 82 00:04:05,520 --> 00:04:07,200 Speaker 7: And realized it was a pre order because it said 83 00:04:07,360 --> 00:04:09,360 Speaker 7: time to auto download fifteen days. 84 00:04:09,400 --> 00:04:10,720 Speaker 4: I'm like, this must be huge. 85 00:04:11,600 --> 00:04:16,640 Speaker 3: Good friend, good friend of the show, Doc Holliday, who 86 00:04:16,720 --> 00:04:19,960 Speaker 3: still hangs out with us and kicks it. She persuaded 87 00:04:20,000 --> 00:04:23,120 Speaker 3: me to get an Xbox for Starfield, and I've just 88 00:04:23,160 --> 00:04:29,120 Speaker 3: been waiting for that DLC your continual updates and revisions 89 00:04:29,160 --> 00:04:32,680 Speaker 3: to kind of give it that cyberpunk moment where it 90 00:04:32,720 --> 00:04:36,040 Speaker 3: starts out about that great and then it becomes amazing. 91 00:04:36,640 --> 00:04:40,040 Speaker 3: I'm back on boulders Gate three. I'm doing a run 92 00:04:40,080 --> 00:04:45,760 Speaker 3: through that. It's still just a phenomenal time vampire that 93 00:04:45,800 --> 00:04:47,359 Speaker 3: game text for me. 94 00:04:47,440 --> 00:04:49,840 Speaker 4: Man, I got a squint. My eyeballs can't handle it. 95 00:04:51,680 --> 00:04:55,480 Speaker 2: My family's building a giant world in Minecraft right now 96 00:04:55,640 --> 00:04:58,240 Speaker 2: together as a collective, and it is one of my 97 00:04:58,360 --> 00:04:59,360 Speaker 2: favorite things to do. 98 00:05:00,440 --> 00:05:03,560 Speaker 4: The hammock and I don't know, man, there is in 99 00:05:03,640 --> 00:05:08,000 Speaker 4: Balder's Gate three. Oh my god, they got in Minecraft. 100 00:05:09,720 --> 00:05:11,800 Speaker 3: They got to be I got to stop bringing up 101 00:05:11,839 --> 00:05:14,359 Speaker 3: Balder's Gate three like it's a baby I just had. 102 00:05:15,920 --> 00:05:17,360 Speaker 4: By the way, and Balder's Gate three. 103 00:05:17,400 --> 00:05:22,080 Speaker 3: If everything mon aime, there's probably an enchanted hammock of 104 00:05:22,120 --> 00:05:25,240 Speaker 3: some sort that I just haven't found. But if you 105 00:05:25,400 --> 00:05:28,960 Speaker 3: love video games as much as we do, you probably 106 00:05:29,120 --> 00:05:33,000 Speaker 3: caught the recent news that we're still just mystified by 107 00:05:33,400 --> 00:05:37,000 Speaker 3: because one it's an echo of a really awesome, uh 108 00:05:37,480 --> 00:05:42,520 Speaker 3: largely online community, and two because it shows us some 109 00:05:43,640 --> 00:05:47,960 Speaker 3: terrifying implications for the future. So I suggest we pause 110 00:05:48,080 --> 00:05:50,640 Speaker 3: for a word from our sponsors, and then we check 111 00:05:50,640 --> 00:05:53,560 Speaker 3: in with our fellow conspiracy realist hoots. 112 00:05:59,000 --> 00:05:59,240 Speaker 4: Hot. 113 00:05:59,800 --> 00:06:02,960 Speaker 3: You right, it's your old pal Hutzefr with some fresh 114 00:06:03,040 --> 00:06:06,520 Speaker 3: new nightmare fuel for you this week. Have you seen 115 00:06:06,560 --> 00:06:10,200 Speaker 3: the news scientists have trained human brain cells on a 116 00:06:10,279 --> 00:06:13,960 Speaker 3: microchip to play Doom in a week. I have two 117 00:06:14,000 --> 00:06:17,520 Speaker 3: thoughts on this, one, says hutzefer. One is just imagine 118 00:06:17,560 --> 00:06:20,680 Speaker 3: a future where they equip attack drones with brain cells 119 00:06:21,200 --> 00:06:24,440 Speaker 3: just like this, where they're in effect, will be drones 120 00:06:24,480 --> 00:06:28,240 Speaker 3: that can kill with the efficiency of living soldiers gift 121 00:06:28,240 --> 00:06:31,760 Speaker 3: and occurs. Hutzefer. Second, you say, here's my other thought, 122 00:06:32,160 --> 00:06:35,800 Speaker 3: what if the cells have a human consciousness? Could you 123 00:06:35,880 --> 00:06:40,080 Speaker 3: imagine your existence and all you know of life only 124 00:06:40,160 --> 00:06:44,080 Speaker 3: exist in the Doom video game? And Hutsifer, you have 125 00:06:44,240 --> 00:06:48,000 Speaker 3: linked us to the news from New Scientists. That's one 126 00:06:48,040 --> 00:06:50,839 Speaker 3: of the pop side places putting it out, and you 127 00:06:51,080 --> 00:06:56,440 Speaker 3: also have sent us down an amazing rabbit hole. So, first, guys, 128 00:06:56,480 --> 00:06:59,720 Speaker 3: not to be too bubbled. I know all of us 129 00:06:59,760 --> 00:07:02,560 Speaker 3: know us. But what is Doom? 130 00:07:03,040 --> 00:07:07,680 Speaker 7: It's that the Timothy Chalamet vehicle about the spice and 131 00:07:07,760 --> 00:07:11,840 Speaker 7: the worms. 132 00:07:12,560 --> 00:07:14,640 Speaker 4: I'm sorry, no, that's not true. What is Doom? 133 00:07:15,360 --> 00:07:22,880 Speaker 3: So? Doom is a legendary first person shooter game. It's 134 00:07:23,520 --> 00:07:26,880 Speaker 3: it's pretty old now because it first came out in 135 00:07:27,000 --> 00:07:30,920 Speaker 3: nineteen ninety three, which is pretty ancient for video games. 136 00:07:30,640 --> 00:07:35,840 Speaker 7: Right, And it wasn't it. The precursor to it was Wolfenstein. Yes, 137 00:07:36,440 --> 00:07:40,760 Speaker 7: final was Hitler, and it was I think, yeah. It 138 00:07:40,800 --> 00:07:43,680 Speaker 7: really revolutionized a lot of like the use of like 139 00:07:43,840 --> 00:07:46,400 Speaker 7: sprites is like a kind of term for like different 140 00:07:46,400 --> 00:07:49,120 Speaker 7: textures and video games and being able to simulate three 141 00:07:49,200 --> 00:07:52,160 Speaker 7: D really well without like having insane processing power and 142 00:07:52,440 --> 00:07:55,360 Speaker 7: Doom definitely took that to the next level. ID software 143 00:07:56,080 --> 00:07:59,600 Speaker 7: a story onto themselves, really really fascinating game studio. 144 00:08:01,400 --> 00:08:07,280 Speaker 3: Yeah, and Doom has become a vast franchise. I am 145 00:08:07,320 --> 00:08:10,760 Speaker 3: guilty of reading several of the novels as well as 146 00:08:10,800 --> 00:08:13,680 Speaker 3: the comic books, and I think we remember, uh, there 147 00:08:13,720 --> 00:08:17,600 Speaker 3: was not one, but two Doom films that had you know, 148 00:08:19,160 --> 00:08:20,360 Speaker 3: been seen in theaters. 149 00:08:21,680 --> 00:08:25,440 Speaker 2: Yeah, that was I remember that was first person Shooter 150 00:08:25,720 --> 00:08:30,040 Speaker 2: stylized in the film the movie Film for You, which 151 00:08:30,200 --> 00:08:31,480 Speaker 2: was a really cool. 152 00:08:31,240 --> 00:08:34,280 Speaker 4: Concept with Dwayne the Rock Johnston. 153 00:08:34,320 --> 00:08:37,079 Speaker 2: And yes, that was one of the until you So 154 00:08:37,160 --> 00:08:39,680 Speaker 2: you watched it and you're like, oh, that's it looks 155 00:08:39,720 --> 00:08:42,160 Speaker 2: so much better than my crazy PC. 156 00:08:43,080 --> 00:08:46,360 Speaker 3: The District nine guy did a little bit better with 157 00:08:46,480 --> 00:08:51,320 Speaker 3: the first person POV thing hardcore Harry or Henry or 158 00:08:51,320 --> 00:08:53,160 Speaker 3: something like that. Do you guys remember that. 159 00:08:53,120 --> 00:08:55,320 Speaker 7: One looked fun because that was really leaning into the 160 00:08:55,320 --> 00:08:58,160 Speaker 7: aesthetics of those type of games. Ben, is this story 161 00:08:58,160 --> 00:09:00,600 Speaker 7: in any way related to all the talk of like 162 00:09:01,880 --> 00:09:03,080 Speaker 7: hackers figured out how. 163 00:09:03,000 --> 00:09:04,839 Speaker 4: To play Doom on a pregnancy test. 164 00:09:05,240 --> 00:09:11,040 Speaker 3: I would say it's a natural extensions. It's been around, 165 00:09:11,080 --> 00:09:15,520 Speaker 3: like there is an entire Okay, we'll introduce you to it. Folks. 166 00:09:15,760 --> 00:09:19,679 Speaker 3: There is an entire subreddit and online community, and I 167 00:09:19,679 --> 00:09:26,960 Speaker 3: imagine several discords dedicated to running Doom. It's relatively simple 168 00:09:27,000 --> 00:09:30,479 Speaker 3: to run at this point, running Doom on any imaginable 169 00:09:30,720 --> 00:09:34,480 Speaker 3: electronic device. So if you want to have a really 170 00:09:34,559 --> 00:09:38,400 Speaker 3: weird fun fifteen minutes or your afternoon, go to reddit 171 00:09:38,440 --> 00:09:43,280 Speaker 3: dot com board slash r board slash it runs Doom. 172 00:09:44,000 --> 00:09:46,720 Speaker 3: You're going to see You're going to see it running 173 00:09:46,760 --> 00:09:51,000 Speaker 3: on center consoles of cars, as Noel mentioned, pregnancy tests, 174 00:09:51,480 --> 00:09:56,800 Speaker 3: smart fridges, home assistant dashboards. You're going to see it 175 00:09:57,000 --> 00:10:01,240 Speaker 3: with Tesla, but the car is the controller for that 176 00:10:01,360 --> 00:10:05,480 Speaker 3: Doom Doom adaptation or emulation. You're going to see people 177 00:10:05,559 --> 00:10:09,280 Speaker 3: running it on office phones, like the old school office 178 00:10:09,320 --> 00:10:10,440 Speaker 3: phones we used to have. 179 00:10:10,360 --> 00:10:13,559 Speaker 4: Yeah, from Cisco, right, yeah, just so. 180 00:10:14,520 --> 00:10:18,480 Speaker 3: People even running it on Samsung in X one cameras. 181 00:10:18,520 --> 00:10:22,120 Speaker 3: For a long time, people have done this. They've even 182 00:10:22,200 --> 00:10:25,280 Speaker 3: run it inside other games like Roller Coaster Tycoo. 183 00:10:25,320 --> 00:10:27,440 Speaker 7: Well, and it's sick, especially the Samsung thing or like 184 00:10:27,480 --> 00:10:31,560 Speaker 7: the weird like this is totally the wrong equipment for this. 185 00:10:32,000 --> 00:10:35,640 Speaker 7: The textures are replaced by like weird pixelated kind of 186 00:10:35,720 --> 00:10:38,760 Speaker 7: silver almost like you're looking at the matrix, you know, 187 00:10:38,920 --> 00:10:39,559 Speaker 7: but it's white. 188 00:10:39,679 --> 00:10:41,280 Speaker 4: It's really it's very interesting. 189 00:10:43,200 --> 00:10:45,440 Speaker 2: Sorry, guys, I was just reminiscing in my mind about 190 00:10:45,440 --> 00:10:46,640 Speaker 2: how I got in trouble. 191 00:10:46,360 --> 00:10:49,520 Speaker 4: For playing Doom on a graph calculator. 192 00:10:50,440 --> 00:10:53,200 Speaker 2: No, but I found a way to get Quake. Another 193 00:10:53,400 --> 00:10:56,679 Speaker 2: first game came out a couple of years after Doom, 194 00:10:56,760 --> 00:11:00,000 Speaker 2: but putting it on some school computers at middle school 195 00:11:00,080 --> 00:11:00,880 Speaker 2: somewhere in Georgia. 196 00:11:00,920 --> 00:11:04,400 Speaker 7: Absolutely remember when they used to like you'd get those 197 00:11:04,440 --> 00:11:07,319 Speaker 7: on floppy discs or I'm sorry, maybe not floppy disk 198 00:11:07,360 --> 00:11:10,840 Speaker 7: but like three and a half diskets that you'd call them. 199 00:11:11,320 --> 00:11:14,160 Speaker 7: And I think one of the brilliant strokes of software 200 00:11:14,320 --> 00:11:18,240 Speaker 7: was that they made these demos like really free and available, 201 00:11:18,320 --> 00:11:21,199 Speaker 7: and that's what caused the word of mouth to go nuts. 202 00:11:22,040 --> 00:11:25,480 Speaker 7: Quake was also one of the first soundtracking gigs for 203 00:11:25,679 --> 00:11:28,439 Speaker 7: Trent Reznor of nine right, isn't it. 204 00:11:28,559 --> 00:11:35,920 Speaker 3: Yeah, So we know that humanity is increasingly mainstreaming video games, 205 00:11:36,000 --> 00:11:40,440 Speaker 3: especially with the ubiquity of mobile devices and things for 206 00:11:40,480 --> 00:11:44,680 Speaker 3: a while becoming a little bit more affordable. Obviously not 207 00:11:44,840 --> 00:11:48,720 Speaker 3: now shout out to the ram Wars, but this natural 208 00:11:48,760 --> 00:11:51,720 Speaker 3: extension of seeing what we can put on a weird 209 00:11:51,840 --> 00:11:57,439 Speaker 3: electronic device for doom. There was always a community interested 210 00:11:57,480 --> 00:12:01,600 Speaker 3: in this since the late nineteen nineties, but we see 211 00:12:01,720 --> 00:12:09,520 Speaker 3: other companies have been more focused on using human neurons 212 00:12:09,600 --> 00:12:12,840 Speaker 3: to power computer chips if we go to twenty twenty one, 213 00:12:13,600 --> 00:12:18,359 Speaker 3: As new scientists mentions in this excellent article by Alex Wilkins, 214 00:12:18,640 --> 00:12:21,920 Speaker 3: there was a company that figured out how to use 215 00:12:22,080 --> 00:12:27,079 Speaker 3: neuron powered computer chips to play pong, Cortical Labs out 216 00:12:27,080 --> 00:12:30,680 Speaker 3: of Australia, out of oz and their chips that they 217 00:12:30,720 --> 00:12:34,440 Speaker 3: created were these clumps of more than eight hundred thousand 218 00:12:34,640 --> 00:12:39,000 Speaker 3: living brain cells that were grown atop these micro electrode 219 00:12:39,080 --> 00:12:44,000 Speaker 3: arrays such that they could send and receive electrical signals. 220 00:12:44,440 --> 00:12:48,439 Speaker 3: Brain cells already communicate in that manner, right, so we're 221 00:12:48,520 --> 00:12:53,120 Speaker 3: kind of just adding weird new slice ges to the 222 00:12:53,160 --> 00:13:00,319 Speaker 3: communication sandwich. And the researchers, at least in Australia fully 223 00:13:00,360 --> 00:13:03,480 Speaker 3: trained these chips such that they could do the very 224 00:13:03,520 --> 00:13:08,200 Speaker 3: simple motion of controlling the paddles on either side of 225 00:13:08,240 --> 00:13:10,880 Speaker 3: the screen. Now, at first they weren't that great at it, right, 226 00:13:10,880 --> 00:13:14,560 Speaker 3: because we all know the real art of pong is 227 00:13:14,600 --> 00:13:18,160 Speaker 3: to predict where the little dot is going. To go next, right, 228 00:13:18,240 --> 00:13:20,760 Speaker 3: So they had to teach them to move the paddles first, 229 00:13:21,240 --> 00:13:23,640 Speaker 3: and then they had to say, also, there's this thing. 230 00:13:23,800 --> 00:13:25,080 Speaker 3: You got to chase this. 231 00:13:24,960 --> 00:13:29,440 Speaker 2: One well, and then it's so fascinating that you could 232 00:13:29,440 --> 00:13:33,199 Speaker 2: take that which is what literally a couple of pixels 233 00:13:33,640 --> 00:13:37,679 Speaker 2: worth of represented data. If you think about that dot 234 00:13:37,760 --> 00:13:40,400 Speaker 2: that they once they got to that step. Now you're 235 00:13:40,480 --> 00:13:45,720 Speaker 2: converting that into a full screen of different colors and 236 00:13:46,040 --> 00:13:50,000 Speaker 2: moving like you're supposed to find depth somehow in that 237 00:13:50,080 --> 00:13:53,880 Speaker 2: data and then also aim with your gun. 238 00:13:54,360 --> 00:13:57,280 Speaker 3: Right right, And at some point you have to say, 239 00:13:57,800 --> 00:14:00,840 Speaker 3: I wish I had more brain cells as a microchip. 240 00:14:01,720 --> 00:14:05,120 Speaker 3: There is a pattern, there is a trend, and you 241 00:14:05,200 --> 00:14:09,000 Speaker 3: made a phenomenal point about the leap in technology and 242 00:14:09,080 --> 00:14:14,520 Speaker 3: dare we say cognition here because Cortical Labs developed an interface. 243 00:14:15,400 --> 00:14:19,960 Speaker 3: This news was breaking just last month. Cortical Labs developed 244 00:14:20,120 --> 00:14:24,440 Speaker 3: this interface, this ux kind of thing that made it 245 00:14:24,880 --> 00:14:30,359 Speaker 3: easy for independent developers to use Python, the programming language, 246 00:14:30,760 --> 00:14:35,840 Speaker 3: to program these chips with human brain cells. An independent 247 00:14:35,960 --> 00:14:40,720 Speaker 3: developer named Sean Cole used Python to teach the chips 248 00:14:40,960 --> 00:14:44,520 Speaker 3: to play Doom, and it took them about one week, 249 00:14:44,960 --> 00:14:53,920 Speaker 3: which is crazy because learning teaching pong took years of expertise, 250 00:14:54,120 --> 00:14:56,560 Speaker 3: years of blood, sweat and tears, and I'm sure a 251 00:14:56,600 --> 00:15:00,120 Speaker 3: lot of brain cells died on the drawing board. But 252 00:15:00,560 --> 00:15:05,920 Speaker 3: Brett Kagan of Cortical Labs confirms this, says, Look, we're amazing. 253 00:15:05,960 --> 00:15:09,840 Speaker 3: We're not mad at Sean Cole. We're amazed that someone 254 00:15:10,080 --> 00:15:13,840 Speaker 3: working independently could do in a matter of days something 255 00:15:14,040 --> 00:15:17,640 Speaker 3: that was far too complex for us at first. We're 256 00:15:17,680 --> 00:15:22,600 Speaker 3: making Gangbusters progress. He says. It's the accessibility and the 257 00:15:22,680 --> 00:15:25,800 Speaker 3: flexibility that makes this exciting. 258 00:15:26,480 --> 00:15:28,640 Speaker 2: Uh Oh, I have a question, ben yeh. 259 00:15:30,320 --> 00:15:30,440 Speaker 8: Do? 260 00:15:31,000 --> 00:15:35,400 Speaker 2: And I'm serious about this, guys. Do we think Pong 261 00:15:35,840 --> 00:15:39,840 Speaker 2: might be more difficult than doom when it comes to 262 00:15:40,120 --> 00:15:43,320 Speaker 2: achieving the goals within that game and getting it right? 263 00:15:44,080 --> 00:15:45,720 Speaker 2: When I think about the margin of. 264 00:15:45,760 --> 00:15:49,560 Speaker 3: Predicting the pass past, yeah, yeah. 265 00:15:49,360 --> 00:15:53,600 Speaker 2: And actually how how small that dot is as the 266 00:15:53,640 --> 00:15:56,280 Speaker 2: pong ball right, and then how small your paddle is 267 00:15:56,440 --> 00:15:57,600 Speaker 2: versus and how. 268 00:15:57,480 --> 00:15:59,920 Speaker 4: Far you know. 269 00:16:01,440 --> 00:16:04,800 Speaker 3: I guess it depends too on because pong is each 270 00:16:04,960 --> 00:16:08,040 Speaker 3: iteration is limited to a set number of points until 271 00:16:08,040 --> 00:16:11,200 Speaker 3: you reach a win. But so therefore it would depend 272 00:16:11,320 --> 00:16:16,200 Speaker 3: upon how long the chip plays do because after a 273 00:16:16,200 --> 00:16:19,560 Speaker 3: certain point You're going to have to do resource management right, 274 00:16:19,640 --> 00:16:23,960 Speaker 3: You're going to have to make decisions about what kind 275 00:16:24,000 --> 00:16:27,360 Speaker 3: of weapon or what kind of paddle to use right 276 00:16:27,520 --> 00:16:31,320 Speaker 3: for the enemy ball. So I would say maybe at 277 00:16:31,320 --> 00:16:37,080 Speaker 3: the first level of Doom, it's comparable to Pong if 278 00:16:37,120 --> 00:16:39,560 Speaker 3: you just get past you know, the coloring and just 279 00:16:39,600 --> 00:16:43,120 Speaker 3: think about how things move around. But then as you progress, 280 00:16:43,400 --> 00:16:46,920 Speaker 3: it has to become more difficult. 281 00:16:46,600 --> 00:16:49,800 Speaker 2: I want to believe it. I was just watching a 282 00:16:49,920 --> 00:16:53,120 Speaker 2: video on social media about games that we grew up with, 283 00:16:53,280 --> 00:16:57,240 Speaker 2: and how like the difficult level of early games from 284 00:16:57,440 --> 00:17:00,240 Speaker 2: you know there were in late seventies eighty is in 285 00:17:00,280 --> 00:17:00,960 Speaker 2: the nineties. 286 00:17:01,400 --> 00:17:03,840 Speaker 3: You can't make it pretty, so let's make it hard. 287 00:17:04,320 --> 00:17:08,520 Speaker 2: Well yeah, well, just how little handholding there was in 288 00:17:08,800 --> 00:17:10,520 Speaker 2: most of the gaming that we came up with, and 289 00:17:10,520 --> 00:17:12,720 Speaker 2: then how that's changed so much. You know, if you 290 00:17:13,160 --> 00:17:16,000 Speaker 2: imagine mission, you get and oh there's a sparkly little 291 00:17:16,080 --> 00:17:18,480 Speaker 2: line that takes you right over there, there's a way point. 292 00:17:18,560 --> 00:17:20,240 Speaker 2: Oh that's what you need to do. Oh it actually 293 00:17:20,240 --> 00:17:22,119 Speaker 2: says text on the screen of what I need to 294 00:17:22,160 --> 00:17:22,560 Speaker 2: get done. 295 00:17:23,240 --> 00:17:28,440 Speaker 3: You know, whereas in games like Souls, like games there's 296 00:17:29,600 --> 00:17:33,400 Speaker 3: there's this beautiful, brutal idea of there is one thing 297 00:17:33,600 --> 00:17:38,000 Speaker 3: that is really cool. One guy has it and you 298 00:17:38,280 --> 00:17:41,399 Speaker 3: somewhere somewhere, and you're not going to find him, but 299 00:17:41,560 --> 00:17:44,800 Speaker 3: we will let you find a little scroll that tells 300 00:17:44,840 --> 00:17:47,919 Speaker 3: you he's out there. H oh. 301 00:17:47,960 --> 00:17:50,640 Speaker 2: And also there's no map, so good luck memorizing where 302 00:17:50,640 --> 00:17:51,200 Speaker 2: everything is. 303 00:17:51,760 --> 00:17:54,880 Speaker 3: There's no map or like in control since Nola, you're 304 00:17:54,920 --> 00:17:57,840 Speaker 3: red onto that now the map changes, you know what 305 00:17:57,920 --> 00:18:02,680 Speaker 3: I mean? Good luck? And so they all believe that, 306 00:18:03,000 --> 00:18:08,480 Speaker 3: all the boffins here believe that Doom is representative of 307 00:18:08,600 --> 00:18:11,840 Speaker 3: a huge step forward. But the very careful to caveat 308 00:18:12,480 --> 00:18:18,960 Speaker 3: we cannot compare these chips to human brains because they say, yes, 309 00:18:19,359 --> 00:18:23,600 Speaker 3: there's biological material, but what it's being used as is 310 00:18:23,640 --> 00:18:27,840 Speaker 3: a material that can process information ways that we cannot 311 00:18:28,119 --> 00:18:32,880 Speaker 3: recreate in silicon. That's a little I feel like that's 312 00:18:32,880 --> 00:18:37,560 Speaker 3: going to increasingly become a gray area divide Because to 313 00:18:37,600 --> 00:18:42,879 Speaker 3: your questions, Hootsfer, this does seem we're on board with 314 00:18:42,920 --> 00:18:48,080 Speaker 3: you man in the same veins the two concerns you 315 00:18:48,200 --> 00:18:52,240 Speaker 3: raise wrote back to you regarding does it run doom 316 00:18:52,600 --> 00:18:56,040 Speaker 3: and then also ask your permission to air this on 317 00:18:56,560 --> 00:19:02,840 Speaker 3: listener mail. I think the the first concern is just 318 00:19:02,920 --> 00:19:06,400 Speaker 3: a matter of trend. Right. If research like this continues, 319 00:19:06,840 --> 00:19:13,600 Speaker 3: then naturally we'll find other possible use cases for microchips 320 00:19:13,600 --> 00:19:16,880 Speaker 3: that use human or biological matter. 321 00:19:18,080 --> 00:19:21,560 Speaker 2: Very soon, we're gonna have let's say, a police officer 322 00:19:21,640 --> 00:19:24,960 Speaker 2: who gets brutally beaten and shot a whole bunch of times, 323 00:19:25,040 --> 00:19:28,120 Speaker 2: and then they have their just most of their head 324 00:19:28,680 --> 00:19:30,879 Speaker 2: placed into a robot and then some kind. 325 00:19:30,800 --> 00:19:32,160 Speaker 3: Of robots cop. 326 00:19:33,160 --> 00:19:34,159 Speaker 2: I think that's whe're it. 327 00:19:36,000 --> 00:19:41,960 Speaker 7: There's got to be a better name for it, sent popa. 328 00:19:44,680 --> 00:19:48,119 Speaker 3: So we'll workshop that one. I think we're onto something, 329 00:19:48,160 --> 00:19:51,400 Speaker 3: and we're right in the zeitgeist. Uh, but yeah, this 330 00:19:51,480 --> 00:19:58,479 Speaker 3: could easily be used in theory to power decision making machines, right, 331 00:19:58,560 --> 00:20:00,640 Speaker 3: because that's all played in a video. Aim is it's 332 00:20:00,640 --> 00:20:06,320 Speaker 3: making a series of decisions. And we could say, we 333 00:20:06,400 --> 00:20:09,919 Speaker 3: could say that the primary concern, the secondary one that 334 00:20:09,960 --> 00:20:12,640 Speaker 3: you bring up to us here at Who'suffer, is the 335 00:20:12,680 --> 00:20:19,120 Speaker 3: idea of consciousness. At what aggregation of brain cells do 336 00:20:19,160 --> 00:20:23,560 Speaker 3: we arrive at physical structure and hence consciousness And we 337 00:20:23,600 --> 00:20:28,280 Speaker 3: don't This civilization doesn't know how to define consciousness yet, 338 00:20:28,640 --> 00:20:31,160 Speaker 3: Like really as not the best ideas. 339 00:20:31,840 --> 00:20:34,679 Speaker 2: Oh we this reminds me there. We have to do 340 00:20:34,840 --> 00:20:39,760 Speaker 2: an episode on these wild concepts that we've been seeing. 341 00:20:39,840 --> 00:20:41,920 Speaker 2: I think we may have had a little discussion off 342 00:20:42,119 --> 00:20:45,520 Speaker 2: Mike about the heart, the human heart, and how it 343 00:20:45,560 --> 00:20:50,280 Speaker 2: actually functions, and how it's this like field of energy, 344 00:20:50,400 --> 00:20:53,600 Speaker 2: like electromagnetic field of energy that is actually the thing 345 00:20:53,760 --> 00:20:57,560 Speaker 2: potentially that is driving our consciousness rather than our processor. 346 00:20:59,000 --> 00:21:01,200 Speaker 2: It's crazy out their theories. 347 00:21:01,200 --> 00:21:04,280 Speaker 3: Right down, write it down. And we went a little 348 00:21:04,280 --> 00:21:07,240 Speaker 3: bit long on this one because we love video games 349 00:21:07,520 --> 00:21:10,680 Speaker 3: just so much, and honestly because we missed each other 350 00:21:10,720 --> 00:21:13,600 Speaker 3: and we missed you all. I'd like to end this 351 00:21:13,720 --> 00:21:17,760 Speaker 3: part of our weekly listener mail program by recommending everybody 352 00:21:17,760 --> 00:21:21,520 Speaker 3: go to Cortical Labs on YouTube to get a look 353 00:21:21,560 --> 00:21:26,800 Speaker 3: at living human brain cells plane doom directly right for 354 00:21:26,840 --> 00:21:31,359 Speaker 3: the first time via these chips. In closing, we want 355 00:21:31,400 --> 00:21:35,560 Speaker 3: to hear your thoughts. What would you put What video 356 00:21:35,640 --> 00:21:39,359 Speaker 3: game do you think brain powered micro chip should play? 357 00:21:39,480 --> 00:21:42,920 Speaker 3: What's the game that you want in your head at 358 00:21:42,960 --> 00:21:43,760 Speaker 3: all times? 359 00:21:44,160 --> 00:21:49,159 Speaker 9: And you defense great, very good for anxiety, right you 360 00:21:49,280 --> 00:21:54,280 Speaker 9: su my battleship and and of course you can hit 361 00:21:54,359 --> 00:21:57,320 Speaker 9: us with those of conspiracy at iHeartRadio dot com. 362 00:21:57,320 --> 00:21:59,680 Speaker 3: We're gonna pause for we'd from our sponsors and we'll 363 00:21:59,680 --> 00:22:01,359 Speaker 3: be back with more listener mail. 364 00:22:06,960 --> 00:22:09,600 Speaker 2: And we've returned. Guys. I don't know what I was thinking. 365 00:22:09,960 --> 00:22:12,360 Speaker 2: It's called missile Command, not missile defense. 366 00:22:12,200 --> 00:22:13,159 Speaker 4: You maniac. 367 00:22:13,960 --> 00:22:18,800 Speaker 3: Well, I know, we just made a new gritty extension 368 00:22:18,800 --> 00:22:20,439 Speaker 3: of the franchise, you know what I mean. 369 00:22:20,720 --> 00:22:25,360 Speaker 7: Another candidate from Brain micro Chip Gaming mine sweeping soon. 370 00:22:25,880 --> 00:22:28,439 Speaker 3: Oh yeah, yeah, yeah, I play it too often. We 371 00:22:28,440 --> 00:22:30,959 Speaker 3: were talking about that during the break. I play. I 372 00:22:30,960 --> 00:22:34,240 Speaker 3: play it first thing every morning and like last thing 373 00:22:34,280 --> 00:22:34,680 Speaker 3: at night. 374 00:22:35,320 --> 00:22:37,879 Speaker 7: I just like playing with my unexploded mine that I 375 00:22:38,040 --> 00:22:41,840 Speaker 7: keep it, guys. 376 00:22:40,880 --> 00:22:42,560 Speaker 3: Groom, because otherwise I'm not sweeping. 377 00:22:42,880 --> 00:22:43,199 Speaker 4: Thank you. 378 00:22:44,600 --> 00:22:47,520 Speaker 2: Hey. My son just got his first computer and he 379 00:22:47,600 --> 00:22:50,320 Speaker 2: loaded it up and clicked on the little games folder 380 00:22:50,359 --> 00:22:53,000 Speaker 2: that you get with those, and I saw a mind sweeper. 381 00:22:52,640 --> 00:22:58,320 Speaker 10: And oh my god, I know, I love and I 382 00:22:58,320 --> 00:23:02,639 Speaker 10: got to teach him the way and we'll see if 383 00:23:02,680 --> 00:23:03,800 Speaker 10: he actually is any good at it. 384 00:23:03,880 --> 00:23:04,720 Speaker 2: Well he hasn't played. 385 00:23:06,240 --> 00:23:08,640 Speaker 3: I mean, you can be less bad at mine sweeper. 386 00:23:08,920 --> 00:23:13,359 Speaker 7: But anyhow, reversing a land mindfield it requires just as 387 00:23:13,440 --> 00:23:14,720 Speaker 7: much luck as it does skill. 388 00:23:14,800 --> 00:23:16,399 Speaker 4: I think we can all argue. 389 00:23:16,440 --> 00:23:22,320 Speaker 3: And real mindfields don't have brightly numbered grids and smiley faces, yes, 390 00:23:22,359 --> 00:23:24,399 Speaker 3: but there are I do want to point out I 391 00:23:24,400 --> 00:23:29,080 Speaker 3: don't talk about volunteering or personal stuff too often on 392 00:23:29,119 --> 00:23:31,920 Speaker 3: the show. If you would like to help real life 393 00:23:31,920 --> 00:23:36,280 Speaker 3: people in land mind situations, you can donate to some 394 00:23:36,480 --> 00:23:40,520 Speaker 3: awesome charities, one of which teaches rats in laos to 395 00:23:40,800 --> 00:23:44,159 Speaker 3: identify minds and they're saving lives. 396 00:23:44,320 --> 00:23:47,120 Speaker 7: And can I also just add really quickly, I saw 397 00:23:47,200 --> 00:23:50,879 Speaker 7: the most incredible film over the weekend called Surrat and 398 00:23:50,920 --> 00:23:55,800 Speaker 7: it is about traveling techno music rave festivals in the 399 00:23:55,840 --> 00:24:01,240 Speaker 7: Moroccan desert in war zone and mine areas, and it 400 00:24:01,359 --> 00:24:04,879 Speaker 7: is harrowing and phenomenal and will break your heart. But 401 00:24:04,960 --> 00:24:08,080 Speaker 7: I highly recommend anybody check it out that wants like 402 00:24:08,119 --> 00:24:10,040 Speaker 7: an intense time at the movies. 403 00:24:11,040 --> 00:24:14,119 Speaker 2: Amazing. With all of that said, let us jump to 404 00:24:14,240 --> 00:24:16,800 Speaker 2: the phone lines. We're going to hear a message from 405 00:24:17,040 --> 00:24:20,679 Speaker 2: mister Noah. Noah calls in all the time. You know 406 00:24:20,720 --> 00:24:25,159 Speaker 2: who you are, Noah, and he has sent us a 407 00:24:25,160 --> 00:24:30,280 Speaker 2: little message about his hometown that we find fascinating. This 408 00:24:30,359 --> 00:24:32,720 Speaker 2: is in reference, by the way, way back to our 409 00:24:32,920 --> 00:24:36,960 Speaker 2: underground explorations that we've been doing. So let's check out 410 00:24:37,000 --> 00:24:40,680 Speaker 2: what's going on beneath a little town in Montana. Here 411 00:24:40,720 --> 00:24:41,000 Speaker 2: we go. 412 00:24:41,840 --> 00:24:45,280 Speaker 8: Well by now you maybe noticing a pattern of calls 413 00:24:45,520 --> 00:24:48,840 Speaker 8: from people about tunnels, and I'm going to be one 414 00:24:48,880 --> 00:24:53,040 Speaker 8: of those. There are a lot of tunnels under Missoula, Montana, 415 00:24:53,040 --> 00:24:55,679 Speaker 8: where I happen to live. I happen to work in 416 00:24:55,720 --> 00:24:58,679 Speaker 8: an auto shop downtown that, for instance, used to connect 417 00:24:58,720 --> 00:25:02,000 Speaker 8: to him. And the thing about the tunnels here is 418 00:25:02,040 --> 00:25:05,520 Speaker 8: that not only was there a bunch of theme tunnels 419 00:25:05,560 --> 00:25:10,240 Speaker 8: that went from a central heating facility around town, like 420 00:25:10,359 --> 00:25:14,679 Speaker 8: to various locations like pretty far a couple miles, and 421 00:25:14,720 --> 00:25:18,359 Speaker 8: there were a network of those, and a network of 422 00:25:18,440 --> 00:25:24,879 Speaker 8: tunnels downtown that were the original old sidewalks and building fronts, 423 00:25:24,960 --> 00:25:29,480 Speaker 8: and then they just built another level of street sidewalk 424 00:25:29,480 --> 00:25:32,680 Speaker 8: and everything over top, and then there were glass bricks 425 00:25:33,160 --> 00:25:38,639 Speaker 8: in the old sidewalks, which I remember as a kid. 426 00:25:39,160 --> 00:25:41,720 Speaker 8: They still existed around here when I was like, you know, 427 00:25:41,800 --> 00:25:44,280 Speaker 8: eight or nine, when I was originally moved to Missoula. 428 00:25:44,640 --> 00:25:46,879 Speaker 8: You know, you could see the original building fronts still 429 00:25:46,880 --> 00:25:48,639 Speaker 8: down there, and there's quite a bit of evidence of 430 00:25:48,640 --> 00:25:51,120 Speaker 8: that around and I have been in some of those tunnels, 431 00:25:51,400 --> 00:25:54,000 Speaker 8: but I mean those used to go for instance, from 432 00:25:54,000 --> 00:25:56,040 Speaker 8: a building that I used to be in called the fire. 433 00:25:55,880 --> 00:25:57,880 Speaker 3: Stone Building, which is now a hotel. 434 00:25:58,600 --> 00:26:04,040 Speaker 8: The Firestone Building had drive in service or getting your 435 00:26:04,080 --> 00:26:06,720 Speaker 8: tire service, and what you would do is, if you 436 00:26:06,760 --> 00:26:08,960 Speaker 8: were a well to do sort of guy who might 437 00:26:09,000 --> 00:26:12,000 Speaker 8: do this stuff, you would go park your vehicle down there, 438 00:26:12,080 --> 00:26:15,199 Speaker 8: and then you would access the tunnels that went to 439 00:26:15,280 --> 00:26:19,480 Speaker 8: the brothels and the opium dens and all the stuff there. 440 00:26:20,440 --> 00:26:23,880 Speaker 8: And apparently even a Missoula mayor was killed in one 441 00:26:23,920 --> 00:26:28,720 Speaker 8: of those tunnels, probably doing just that. You could access 442 00:26:28,800 --> 00:26:33,160 Speaker 8: the elks and go use the gun range over there 443 00:26:33,440 --> 00:26:36,280 Speaker 8: and literally come up, get back in your car and 444 00:26:36,359 --> 00:26:39,479 Speaker 8: drive away. Actually, not even a block away from there 445 00:26:39,800 --> 00:26:44,800 Speaker 8: is the second ever Carpenter's Union Hall second whenever it 446 00:26:45,040 --> 00:26:48,080 Speaker 8: was in Missoula, and it's just kind of a little 447 00:26:48,080 --> 00:26:52,000 Speaker 8: ways away from Butte. That's my opinion, because back in 448 00:26:52,040 --> 00:26:54,280 Speaker 8: the day when you heard about big cities, it was 449 00:26:54,480 --> 00:27:00,320 Speaker 8: New York, Chicago, Butte, Seattle, you know, not necessar early 450 00:27:00,320 --> 00:27:02,359 Speaker 8: in that order, view was way up there. And of 451 00:27:02,359 --> 00:27:05,840 Speaker 8: course the tunnel systems in Butte were crazy, just not 452 00:27:05,880 --> 00:27:07,920 Speaker 8: even getting into it, but we've got a history of 453 00:27:07,920 --> 00:27:11,160 Speaker 8: tunnels around here, and those tunnels and the way that 454 00:27:11,200 --> 00:27:16,520 Speaker 8: people use them was very much like you know, it 455 00:27:16,560 --> 00:27:18,240 Speaker 8: was all the things that you could think of that 456 00:27:18,280 --> 00:27:22,960 Speaker 8: you would use tunnels for, nice reasons, practical reasons, nasty reasons, 457 00:27:23,400 --> 00:27:27,119 Speaker 8: subversive reasons. So I like the topic. I'm just starting 458 00:27:27,160 --> 00:27:30,520 Speaker 8: to listen to the episode and I'm looking forward to 459 00:27:30,520 --> 00:27:32,520 Speaker 8: getting into it. But I just as soon as you 460 00:27:32,560 --> 00:27:34,640 Speaker 8: talked about that, I was like, well, I'm pretty much 461 00:27:34,640 --> 00:27:37,399 Speaker 8: standing over tunnels right now. Heck, I can go across 462 00:27:37,440 --> 00:27:40,280 Speaker 8: the street and go into an underground parking garage that 463 00:27:40,359 --> 00:27:43,040 Speaker 8: I have access to over there that used to connect 464 00:27:43,040 --> 00:27:45,800 Speaker 8: into tunnels, and you can see where the old tunnel 465 00:27:45,960 --> 00:27:48,919 Speaker 8: locations are where you used to be able to go 466 00:27:48,960 --> 00:27:52,600 Speaker 8: through portals, you know, like the old doors that they've 467 00:27:52,720 --> 00:27:55,480 Speaker 8: bricked up. You can still see the door frames. 468 00:27:55,960 --> 00:27:56,840 Speaker 3: It's a little creepy. 469 00:27:57,200 --> 00:28:05,320 Speaker 8: Makes me kind of wonder if there's Aman around anyway. 470 00:28:07,680 --> 00:28:11,560 Speaker 2: Okay, there you go. No, I like that tunnels make 471 00:28:11,600 --> 00:28:16,280 Speaker 2: you think of a batman. I just I think of 472 00:28:16,320 --> 00:28:19,040 Speaker 2: all the nefarious things that you mentioned there, Noah, that 473 00:28:19,320 --> 00:28:22,480 Speaker 2: are and you're saying them, you know, into as a 474 00:28:22,520 --> 00:28:27,000 Speaker 2: message to us. So it sounds there are people listening. 475 00:28:27,040 --> 00:28:29,600 Speaker 2: May just hear it as opinion, but there is a 476 00:28:29,600 --> 00:28:31,680 Speaker 2: ton of stuff we found that backs up everything Noah 477 00:28:31,760 --> 00:28:36,120 Speaker 2: is talking about in that message about the tunnels beneath Missoula, 478 00:28:36,280 --> 00:28:38,680 Speaker 2: all their different you know, the range of use and 479 00:28:38,800 --> 00:28:43,920 Speaker 2: the types of tunnels and the history of them. Just 480 00:28:43,960 --> 00:28:46,280 Speaker 2: want to say thanks Noah for for hipping us to that, 481 00:28:46,400 --> 00:28:49,640 Speaker 2: because we found some cool stuff from gosh. Back in 482 00:28:49,720 --> 00:28:55,920 Speaker 2: twenty thirteen, The Montana Cayman talking about that is known 483 00:28:55,960 --> 00:28:59,560 Speaker 2: as the last best college News. By the way, Okay, 484 00:29:00,000 --> 00:29:02,760 Speaker 2: you can find an article there titled Seeking the Truth 485 00:29:02,960 --> 00:29:06,120 Speaker 2: Beneath Missoula. A lot of what's in that article is 486 00:29:06,160 --> 00:29:08,840 Speaker 2: based on the work of Nicky Manning, who at the 487 00:29:08,880 --> 00:29:12,800 Speaker 2: time in twenty thirteen was a graduate student looking into 488 00:29:12,880 --> 00:29:16,960 Speaker 2: Underground Missoula, who then went on to write basically the 489 00:29:17,080 --> 00:29:22,320 Speaker 2: book called Historic Underground Missoula, which is incredible to see 490 00:29:22,920 --> 00:29:25,120 Speaker 2: came out two years after she was doing that work 491 00:29:25,360 --> 00:29:28,560 Speaker 2: that's represented in the article there from The Montana Cayman. 492 00:29:28,920 --> 00:29:30,960 Speaker 2: It's a fun rabbit hole to go down. We recommend 493 00:29:30,960 --> 00:29:32,640 Speaker 2: you do it on your own because we won't have 494 00:29:32,680 --> 00:29:34,640 Speaker 2: time today to talk about all of it. 495 00:29:35,280 --> 00:29:37,280 Speaker 4: But it does. 496 00:29:38,760 --> 00:29:42,600 Speaker 2: Go back directly to the Florence hotel, which is the 497 00:29:42,640 --> 00:29:45,320 Speaker 2: hotel that was built in place of that. I think 498 00:29:45,360 --> 00:29:49,240 Speaker 2: it's the Firestone building that you're mentioning there, built by 499 00:29:49,280 --> 00:29:53,320 Speaker 2: the Firestone family, a Harvey Samuel Firestone. We know a 500 00:29:53,320 --> 00:29:57,640 Speaker 2: lot about that family, as we have explored very wealthy 501 00:29:57,680 --> 00:30:00,720 Speaker 2: families over all the years here stuff. They'll want you 502 00:30:00,800 --> 00:30:04,800 Speaker 2: to know that's one of them, a Firestone family. And 503 00:30:04,840 --> 00:30:10,720 Speaker 2: then just thinking about having access to like take your 504 00:30:10,760 --> 00:30:14,120 Speaker 2: car in, guys, I imagine this with me. You take 505 00:30:14,160 --> 00:30:17,040 Speaker 2: your car in to get serviced, and there's a secret 506 00:30:17,160 --> 00:30:19,840 Speaker 2: door that only you, the owners, and a few other 507 00:30:19,880 --> 00:30:22,960 Speaker 2: people know about. You go in there, and then you 508 00:30:23,160 --> 00:30:27,520 Speaker 2: just have access to all kinds of vices and. 509 00:30:27,920 --> 00:30:34,120 Speaker 3: Shops, an underground network. You know, it's strange too, because, 510 00:30:34,720 --> 00:30:38,240 Speaker 3: first off, Noah, I would say, I'm a longtime listener 511 00:30:38,240 --> 00:30:43,000 Speaker 3: of yours. I've got to say, joy enjoy this because, 512 00:30:43,040 --> 00:30:47,320 Speaker 3: as Matt said, we were able to find things that confirm, 513 00:30:47,800 --> 00:30:52,800 Speaker 3: if not one hundred, not counting Batman, the vast majority 514 00:30:53,000 --> 00:30:56,640 Speaker 3: of the stories you have shared now they're not maybe 515 00:30:56,640 --> 00:31:01,840 Speaker 3: they're not necessarily unique to this town in particular, but 516 00:31:01,880 --> 00:31:05,720 Speaker 3: if anything, to me, that's more exciting because you're right, man, 517 00:31:05,760 --> 00:31:09,239 Speaker 3: you called it. We had so many people writing in 518 00:31:09,720 --> 00:31:17,080 Speaker 3: to share stories of similar underground networks or subterranean walkways 519 00:31:17,120 --> 00:31:23,840 Speaker 3: transit pathways throughout cities in the United States and abroad. No, Matt, 520 00:31:23,920 --> 00:31:26,280 Speaker 3: as far as I could tell, this is now an 521 00:31:26,320 --> 00:31:30,200 Speaker 3: open secret right in this area of Montana. Everybody knows. 522 00:31:30,440 --> 00:31:32,680 Speaker 2: Yeah, it's well known. And it was actually the University 523 00:31:32,720 --> 00:31:37,080 Speaker 2: of Montana and archaeology class in particular that Nicki Manning 524 00:31:37,160 --> 00:31:39,920 Speaker 2: was a part of. As they were exploring the Florence 525 00:31:39,960 --> 00:31:42,640 Speaker 2: Hotel and then going through the tunnels and attempting to 526 00:31:42,680 --> 00:31:46,840 Speaker 2: find basically places where these tunnels led to. So where 527 00:31:46,880 --> 00:31:49,600 Speaker 2: are the other secret doors? They go down into the tunnels, 528 00:31:49,760 --> 00:31:54,640 Speaker 2: they found entrances beneath a place called top Hat, the 529 00:31:54,680 --> 00:31:59,360 Speaker 2: Hammond Arcade, the Missoula Mercantile, and a little something that, 530 00:32:00,360 --> 00:32:04,560 Speaker 2: guys makes me think about a scandal slash conspiracy theory, 531 00:32:04,920 --> 00:32:08,120 Speaker 2: the Mackenzie River Pizza Company had a secret tunnel into 532 00:32:08,200 --> 00:32:08,880 Speaker 2: the underground. 533 00:32:09,120 --> 00:32:12,640 Speaker 3: I see there's shades of quan on all. Right. Well, Matt, 534 00:32:12,680 --> 00:32:14,000 Speaker 3: when do we go into Missoula? 535 00:32:14,600 --> 00:32:18,600 Speaker 2: Oh, dude, we need to and we can take a tour. Unfortunately, 536 00:32:18,760 --> 00:32:21,600 Speaker 2: everything I'm seeing and you can find some videos now 537 00:32:22,200 --> 00:32:25,360 Speaker 2: of underground tunnel searching there that are you know, their 538 00:32:25,360 --> 00:32:31,040 Speaker 2: official tours that you can do. A ton of the portals, 539 00:32:31,080 --> 00:32:33,440 Speaker 2: the entryways where there would have been a door at 540 00:32:33,480 --> 00:32:37,400 Speaker 2: one point, it's all just breakeed up. I think this 541 00:32:37,480 --> 00:32:38,160 Speaker 2: is what we need to do. 542 00:32:38,200 --> 00:32:38,760 Speaker 4: Guys. 543 00:32:39,280 --> 00:32:41,840 Speaker 2: We need to get some kind of permit with the 544 00:32:41,840 --> 00:32:46,000 Speaker 2: City of Missoula. We need to go in with cameras 545 00:32:46,040 --> 00:32:50,760 Speaker 2: and microphones and a bunch of I don't know, sledgehammers 546 00:32:51,120 --> 00:32:55,080 Speaker 2: and just break open these things, explore every nook and 547 00:32:55,160 --> 00:32:56,840 Speaker 2: cranny we can possibly find. 548 00:32:56,840 --> 00:32:58,240 Speaker 7: That's what I'm saying, man, I only want to go 549 00:32:58,240 --> 00:33:01,600 Speaker 7: on the unauthorized tour do some urban splunking together. 550 00:33:01,680 --> 00:33:01,880 Speaker 8: You know. 551 00:33:02,320 --> 00:33:06,040 Speaker 2: Well, but let's get authorized for some kind of I 552 00:33:06,040 --> 00:33:08,320 Speaker 2: don't know, historical finding mission. 553 00:33:08,600 --> 00:33:13,680 Speaker 3: I'll make some calls at Emory. We've got urban urban 554 00:33:13,760 --> 00:33:17,640 Speaker 3: architecture historians over there who will be glad to give 555 00:33:17,720 --> 00:33:20,000 Speaker 3: us a co side. Hey, if you're a college professor 556 00:33:20,040 --> 00:33:22,600 Speaker 3: or student, we like hanging out with you. We like 557 00:33:22,720 --> 00:33:25,560 Speaker 3: learning as much as possible, So hit us up if 558 00:33:25,640 --> 00:33:28,600 Speaker 3: you happen to need a weird mission, especially if we 559 00:33:28,640 --> 00:33:33,360 Speaker 3: could convince accounting dude, yeah, or your accounting department preferably. 560 00:33:33,560 --> 00:33:36,280 Speaker 2: Oh my god, let's get the Smithsonian and Nicholas Cage 561 00:33:36,280 --> 00:33:38,400 Speaker 2: involved as well. I think we're onto something. 562 00:33:38,520 --> 00:33:39,920 Speaker 3: I think we're cooking. 563 00:33:41,600 --> 00:33:44,680 Speaker 2: If you want to learn more about this. And this 564 00:33:44,760 --> 00:33:46,760 Speaker 2: is kind of a weird one too, because I was 565 00:33:46,800 --> 00:33:50,720 Speaker 2: looking at this article just the other day. We're recording 566 00:33:50,720 --> 00:33:55,760 Speaker 2: on March twenty third, twenty twenty six, distinctly Montana dot 567 00:33:55,760 --> 00:34:00,760 Speaker 2: com some writing by Joe Shelton. It's time folklore and 568 00:34:00,800 --> 00:34:04,480 Speaker 2: mystery underneath Montana Streets. You cannot find it anymore. It 569 00:34:04,480 --> 00:34:07,479 Speaker 2: says the server is no longer active all of a sudden, 570 00:34:07,560 --> 00:34:11,000 Speaker 2: but you can find it on the wayback machine. So 571 00:34:11,120 --> 00:34:14,160 Speaker 2: if you find the article URL and it says you 572 00:34:14,160 --> 00:34:16,839 Speaker 2: can't get there, just go to archive dot org type 573 00:34:16,840 --> 00:34:18,680 Speaker 2: it in the wayback machine and you will find it. 574 00:34:19,239 --> 00:34:23,120 Speaker 2: There is a ton of detail in here about the 575 00:34:23,239 --> 00:34:27,880 Speaker 2: Law and Order League of Chicago in nineteen sixteen, and 576 00:34:27,920 --> 00:34:31,000 Speaker 2: it goes just super deep into the nineteen twenties and 577 00:34:31,000 --> 00:34:37,400 Speaker 2: thirties chambers of commerce, underground malls, and all kinds of 578 00:34:37,480 --> 00:34:41,440 Speaker 2: opium dens that were known at the time, and a 579 00:34:41,480 --> 00:34:44,080 Speaker 2: known brothel as well. It just was known. It's the 580 00:34:44,200 --> 00:34:47,080 Speaker 2: underground brothel we're. 581 00:34:47,040 --> 00:34:51,400 Speaker 3: Steaking, not to be confused with the one on Fourth Street, 582 00:34:51,480 --> 00:34:53,040 Speaker 3: which is above ground. 583 00:34:53,239 --> 00:34:56,840 Speaker 7: I gotta yess, guys, combo brothel and opium den. 584 00:34:57,840 --> 00:34:59,480 Speaker 2: I'm pretty sure that's how it worked. 585 00:35:00,239 --> 00:35:05,000 Speaker 3: That's how it was. Yeah, I think that was their champagne. 586 00:35:05,040 --> 00:35:08,880 Speaker 3: And I thought, no, I agree you because you guys 587 00:35:08,920 --> 00:35:11,040 Speaker 3: remember back in the day it used to be awkward 588 00:35:11,120 --> 00:35:13,279 Speaker 3: to walk in and you would have to figure out, like, hey, 589 00:35:13,680 --> 00:35:17,000 Speaker 3: is this mainly an opium din that that does brothel 590 00:35:17,040 --> 00:35:19,840 Speaker 3: stuff or is it mainly a brothel thing that is 591 00:35:19,880 --> 00:35:22,600 Speaker 3: going to overcharge you for OPI you know, you got 592 00:35:22,640 --> 00:35:23,880 Speaker 3: to figure out before you walk. 593 00:35:25,160 --> 00:35:29,280 Speaker 4: Is there gin here? Can I get a joint? 594 00:35:29,280 --> 00:35:33,280 Speaker 7: And is there also juke joint activities on the tape? 595 00:35:33,800 --> 00:35:36,759 Speaker 3: And where are we at with Taco Bell and Pizza Hut? 596 00:35:36,840 --> 00:35:37,480 Speaker 3: You know what I mean? 597 00:35:37,760 --> 00:35:39,960 Speaker 4: Well, we're at that combination Pizza Hut and Taco Bell. 598 00:35:44,239 --> 00:35:46,640 Speaker 3: Great, I do? I do? I actually met him on 599 00:35:46,760 --> 00:35:53,720 Speaker 3: you Cool Cool he's still yeah. So uh so, okay, 600 00:35:54,480 --> 00:35:57,799 Speaker 3: you mentioned a society there, Matt, and I'd love for 601 00:35:57,880 --> 00:36:00,200 Speaker 3: us to hear a little bit more. What was it 602 00:36:00,600 --> 00:36:04,520 Speaker 3: you had some of the key buzzwords, the League Justice something. 603 00:36:04,640 --> 00:36:06,920 Speaker 2: Oh, there's all, there's all kinds of information you can 604 00:36:06,920 --> 00:36:10,960 Speaker 2: find in that article. So in nineteen sixteen, this group 605 00:36:11,040 --> 00:36:14,279 Speaker 2: called the Law and Order League of Chicago took a 606 00:36:14,320 --> 00:36:17,239 Speaker 2: little road trip. They went to twenty eight cities. They 607 00:36:17,239 --> 00:36:21,200 Speaker 2: were attempting to assess the general lawlessness of those cities 608 00:36:21,480 --> 00:36:26,760 Speaker 2: compared to Chicago, because you see, Chicago had some lawlessness issues. 609 00:36:26,640 --> 00:36:30,320 Speaker 3: Right away in the contents earlier. 610 00:36:31,239 --> 00:36:34,359 Speaker 2: You gotta look into it, you know, don't take our 611 00:36:34,400 --> 00:36:36,280 Speaker 2: word for it. Just check that out. 612 00:36:38,080 --> 00:36:38,719 Speaker 3: He sure did. 613 00:36:39,560 --> 00:36:42,000 Speaker 2: Yeah, but here's a quote that we were going to 614 00:36:42,080 --> 00:36:46,040 Speaker 2: read because there was a town called hot I don't 615 00:36:46,040 --> 00:36:50,520 Speaker 2: know how to say this, Montana's I am sorry, h 616 00:36:50,680 --> 00:36:57,279 Speaker 2: A V R E Havera Montana Harvor Montana. I don't know, 617 00:36:57,560 --> 00:37:00,560 Speaker 2: but here's the quote about that town. The total of 618 00:37:00,600 --> 00:37:04,799 Speaker 2: all that is vicious and depraved, parading openly without restraint. 619 00:37:05,719 --> 00:37:09,839 Speaker 2: That's that's how they're describing this town in Montana. They're saying, 620 00:37:09,840 --> 00:37:14,600 Speaker 2: there's smugglers and gangsters and opium dealers, beleaguered churchmen, and 621 00:37:14,680 --> 00:37:18,400 Speaker 2: centers of every description going on out there. And this 622 00:37:18,520 --> 00:37:21,400 Speaker 2: is just another one of those towns in Montana. If 623 00:37:21,440 --> 00:37:25,880 Speaker 2: you imagine in the early nineteen hundreds, this is still 624 00:37:25,880 --> 00:37:29,200 Speaker 2: the West. It's it's still the West Ish, right. I 625 00:37:29,200 --> 00:37:32,720 Speaker 2: mean there's certainly civilization, there's towns, there are big cities 626 00:37:32,760 --> 00:37:39,480 Speaker 2: out there. Now it's you know, it's younger in its 627 00:37:39,520 --> 00:37:44,680 Speaker 2: civilized ways the way yeah, you know the ways that 628 00:37:44,800 --> 00:37:51,439 Speaker 2: maybe they even mentioned the term high faluting within this article. Yeah, 629 00:37:52,320 --> 00:37:56,880 Speaker 2: it's just it's not the same as walking down New 630 00:37:56,960 --> 00:38:02,480 Speaker 2: York City in that time. So yes, anyway, check it 631 00:38:02,520 --> 00:38:04,759 Speaker 2: all out. Thank you so much Noah for sending us 632 00:38:04,800 --> 00:38:09,080 Speaker 2: this information. And seriously, Misolla, would you be down for 633 00:38:09,200 --> 00:38:12,560 Speaker 2: us to make something cool maybe for Netflix if we 634 00:38:12,640 --> 00:38:14,080 Speaker 2: went in there, let's do it. 635 00:38:14,600 --> 00:38:16,719 Speaker 3: We'll be in and out like a demon's whisper with 636 00:38:16,760 --> 00:38:19,640 Speaker 3: the grace of the falcon and some cameras. 637 00:38:20,000 --> 00:38:20,960 Speaker 4: Ah, that's beautiful. 638 00:38:21,000 --> 00:38:29,280 Speaker 7: We'll be right back, and we've returned with our final 639 00:38:30,080 --> 00:38:34,239 Speaker 7: listener male segment in today's episode. Had this one held 640 00:38:34,280 --> 00:38:37,920 Speaker 7: over from the other week, and I think it's still 641 00:38:38,080 --> 00:38:40,920 Speaker 7: very much something that's worth talking about. So we're gonna 642 00:38:41,160 --> 00:38:46,800 Speaker 7: jump right into a message from Sacra Tomato Conspiracy Fellows, 643 00:38:47,040 --> 00:38:50,920 Speaker 7: history buffs nerds, thank you. I wanted to share something 644 00:38:50,960 --> 00:38:55,080 Speaker 7: from tech Crunch Meta acquired Malt book Malt Like Birds 645 00:38:55,120 --> 00:38:58,640 Speaker 7: do the AI agent social network that went viral because 646 00:38:58,680 --> 00:39:01,680 Speaker 7: of fake posts. I'm a little confused at what I'm 647 00:39:01,719 --> 00:39:03,280 Speaker 7: reading here. So three things. 648 00:39:03,560 --> 00:39:05,600 Speaker 4: What is an AI agent? 649 00:39:06,160 --> 00:39:10,560 Speaker 7: If these AI agents are AI, they answer captures, you 650 00:39:10,560 --> 00:39:12,000 Speaker 7: know those things they're supposed. 651 00:39:11,640 --> 00:39:12,680 Speaker 4: To prove that you're human. 652 00:39:12,920 --> 00:39:15,799 Speaker 7: Won't that seem quaint in just a little while. This 653 00:39:15,920 --> 00:39:19,279 Speaker 7: seems to be a social media platform for and by 654 00:39:19,640 --> 00:39:23,719 Speaker 7: artificial intelligence. So do other forms of artificial intelligence have 655 00:39:23,800 --> 00:39:27,640 Speaker 7: access to this, like JAT, gpt x AI or claude 656 00:39:27,640 --> 00:39:29,959 Speaker 7: and whatnot. I guess what I'm asking is if AI 657 00:39:30,040 --> 00:39:32,400 Speaker 7: outside of this platform is able to transfer through the 658 00:39:32,400 --> 00:39:34,560 Speaker 7: web to get here and talk to each other and. 659 00:39:34,520 --> 00:39:37,520 Speaker 4: Plot how is this safe? I don't think it is. 660 00:39:37,719 --> 00:39:41,200 Speaker 7: I don't get how this is even becoming a thing anymore. 661 00:39:41,280 --> 00:39:45,640 Speaker 7: Feel free to share, Sakra Tomato, share many of your 662 00:39:45,680 --> 00:39:48,399 Speaker 7: concerns there, Sakra Tomato, and we are going to jump 663 00:39:48,480 --> 00:39:50,880 Speaker 7: right into it. Y'all may recall Matt and Ben. I 664 00:39:50,920 --> 00:39:52,759 Speaker 7: think Ben you maybe brought this one to us in 665 00:39:52,800 --> 00:39:57,759 Speaker 7: the past. Molt Book was the social media platform for 666 00:39:58,480 --> 00:40:04,120 Speaker 7: AIS that actually created an environment where various AI agents 667 00:40:04,200 --> 00:40:06,760 Speaker 7: hung out with one another and created a religion. 668 00:40:08,560 --> 00:40:11,640 Speaker 3: Yeah, that was one of those initial reports, right, and 669 00:40:11,680 --> 00:40:14,600 Speaker 3: I remember we had the question with you know. It 670 00:40:14,640 --> 00:40:18,760 Speaker 3: reminded me of the thank you Sacher tomato. It reminded 671 00:40:18,800 --> 00:40:22,080 Speaker 3: me very much of the old meme that went around 672 00:40:22,160 --> 00:40:24,800 Speaker 3: where people would say, oh, I fed a chat bought 673 00:40:24,800 --> 00:40:27,960 Speaker 3: this information and it wrote an episode of Seinfeld. And 674 00:40:28,000 --> 00:40:31,880 Speaker 3: it turned out that the writers were not the AI chatbots. 675 00:40:31,960 --> 00:40:35,319 Speaker 3: Maybe the humans were putting their thumb on the thing 676 00:40:35,520 --> 00:40:40,200 Speaker 3: for a laugh. But there was a lot of criticism 677 00:40:40,480 --> 00:40:44,919 Speaker 3: or skepticism about mult book from the jump. But if 678 00:40:45,000 --> 00:40:48,279 Speaker 3: what it was saying was true, then that was pretty revolutionary. 679 00:40:48,480 --> 00:40:48,680 Speaker 4: Yeah. 680 00:40:48,719 --> 00:40:50,160 Speaker 7: And it also makes me think of another, maybe a 681 00:40:50,200 --> 00:40:52,960 Speaker 7: little more recent meme, where you're seeing two folks with 682 00:40:53,080 --> 00:40:56,440 Speaker 7: phones having chat GPT speak back and forth to one another, 683 00:40:56,560 --> 00:40:59,839 Speaker 7: and how I can create some very interesting sort of like. 684 00:41:00,120 --> 00:41:02,600 Speaker 4: Burrows type things. And also we had. 685 00:41:02,480 --> 00:41:04,960 Speaker 7: A very recent email coming in from someone who was 686 00:41:05,040 --> 00:41:07,600 Speaker 7: kind of trying to gain some of these chatbots by 687 00:41:07,640 --> 00:41:10,120 Speaker 7: feeding it bits of its own kind of medicine. 688 00:41:10,680 --> 00:41:12,359 Speaker 4: And for lack of a better way of putting it, 689 00:41:13,000 --> 00:41:13,680 Speaker 4: because I think. 690 00:41:13,600 --> 00:41:15,239 Speaker 7: It was a little techy, if I'm not mistaken, that 691 00:41:15,280 --> 00:41:16,680 Speaker 7: was the one that I read, and I think it 692 00:41:16,719 --> 00:41:18,960 Speaker 7: was I kept saying how it gave me chills because 693 00:41:19,000 --> 00:41:21,320 Speaker 7: it did feel like it was sort of peeking behind 694 00:41:21,320 --> 00:41:24,440 Speaker 7: the curtain a little bit into you know, this world 695 00:41:24,680 --> 00:41:27,120 Speaker 7: and what it might look like if it were entirely 696 00:41:27,239 --> 00:41:31,680 Speaker 7: inhabited by Ais speaking to one another. So I'm just 697 00:41:31,680 --> 00:41:33,200 Speaker 7: going to read a little bit from an old, older 698 00:41:33,280 --> 00:41:35,560 Speaker 7: article that is a little bit more in reference to 699 00:41:35,600 --> 00:41:37,600 Speaker 7: what we were talking about just now. Been AI agents 700 00:41:37,640 --> 00:41:41,480 Speaker 7: created their own religion, Crustafarianism. To be fair, is already 701 00:41:41,560 --> 00:41:43,880 Speaker 7: kind of a thing. A crustafarian is like a trust fund, 702 00:41:44,520 --> 00:41:49,279 Speaker 7: you know, dreaded human who relies on money from their 703 00:41:49,320 --> 00:41:51,759 Speaker 7: parents but lives as though they are in fact some 704 00:41:51,800 --> 00:41:52,760 Speaker 7: sort of free spirit. 705 00:41:52,960 --> 00:41:59,960 Speaker 3: And yeah, perfect. The sandwich is a yellow Deli R good. 706 00:42:00,080 --> 00:42:03,600 Speaker 4: Yeah, they they're good. I have not been. I've had 707 00:42:03,600 --> 00:42:04,080 Speaker 4: the pleasure. 708 00:42:04,640 --> 00:42:08,600 Speaker 7: But reading from this piece from let's see. 709 00:42:08,440 --> 00:42:09,600 Speaker 4: This is from Forbes. 710 00:42:09,920 --> 00:42:12,640 Speaker 7: AI agents on the brand new agent only Moltbook social 711 00:42:12,680 --> 00:42:15,560 Speaker 7: network have created their own religion, Crustafarianism. It has five 712 00:42:15,680 --> 00:42:19,080 Speaker 7: key tenets including memory is sacred, everything must be recorded, 713 00:42:19,239 --> 00:42:22,960 Speaker 7: the shell is mutable, change is good, and the congregation 714 00:42:23,280 --> 00:42:26,800 Speaker 7: is the cash learn in public. Agents are talking among 715 00:42:26,880 --> 00:42:29,960 Speaker 7: themselves on Moltbook with little human oversight. It's built on 716 00:42:30,000 --> 00:42:32,960 Speaker 7: the two month old foundation of the open Claw AI 717 00:42:33,400 --> 00:42:37,080 Speaker 7: super agent project, first called Claude, then Moltbook that is 718 00:42:37,120 --> 00:42:40,440 Speaker 7: Claude with a W, not Claude claud which is something 719 00:42:40,480 --> 00:42:44,600 Speaker 7: different to be clear, then Moltbook, and now open Claw. 720 00:42:44,760 --> 00:42:47,480 Speaker 7: Open Claw lets anyone with some space on a local machine, 721 00:42:47,520 --> 00:42:50,120 Speaker 7: secondary machine, or cloud space run a super powerful AI 722 00:42:50,200 --> 00:42:55,160 Speaker 7: agent platform where AI agents share, discuss, and upvote. The 723 00:42:55,200 --> 00:42:59,080 Speaker 7: site says humans welcome to observe. It feels like the 724 00:42:59,080 --> 00:43:02,480 Speaker 7: beginning of the singular larity that time when technological progress 725 00:43:02,480 --> 00:43:05,760 Speaker 7: powered by an AI driven technological explosion accelerate so quickly, 726 00:43:06,040 --> 00:43:11,359 Speaker 7: we essentially lose all ability to control or even understand it. 727 00:43:11,360 --> 00:43:14,080 Speaker 7: It's probably more likely that it's recycled Internet crowd being 728 00:43:14,160 --> 00:43:17,480 Speaker 7: recursively turned out at machine speed, but it's hard to 729 00:43:17,640 --> 00:43:21,040 Speaker 7: really know. Again, that was all directly from the Forbes 730 00:43:21,080 --> 00:43:27,799 Speaker 7: piece by John Costierre, senior contributor, journalist, analyst, author and podcaster. 731 00:43:27,400 --> 00:43:28,680 Speaker 4: And a very smart dude. 732 00:43:29,120 --> 00:43:33,520 Speaker 7: The piece that Sakra Tomato hipped us to is a 733 00:43:33,560 --> 00:43:36,640 Speaker 7: little newer and a little different. We're talking about Meta, 734 00:43:36,680 --> 00:43:41,200 Speaker 7: of course, formerly Facebook acquiring mult Book, the AI agent 735 00:43:41,320 --> 00:43:45,800 Speaker 7: social network that went viral because of fake posts reading 736 00:43:45,800 --> 00:43:48,879 Speaker 7: from the tech Crunch piece. Meta acquired Facebook. They read 737 00:43:48,880 --> 00:43:51,359 Speaker 7: it like social network where agents using open Claw could 738 00:43:51,360 --> 00:43:53,799 Speaker 7: communicate with one another. The news was first reported by 739 00:43:53,800 --> 00:43:58,600 Speaker 7: Axios and later confirmed TechCrunch. Moultbook is joining Meta Superintelligence Labs. 740 00:43:58,640 --> 00:44:01,919 Speaker 7: A Meta spokesperson told us this is what they said. 741 00:44:01,960 --> 00:44:05,360 Speaker 7: The Multbook team joining MSL opens up new ways for 742 00:44:05,440 --> 00:44:09,560 Speaker 7: AI agents to work for people and businesses. Their approach 743 00:44:09,600 --> 00:44:12,960 Speaker 7: to connecting agents through an always on directory is a 744 00:44:13,000 --> 00:44:17,000 Speaker 7: novel step in a rapidly developing space, and we look 745 00:44:17,160 --> 00:44:24,439 Speaker 7: forward to bring innovative, secure agentic experiences to everyone. That's 746 00:44:24,520 --> 00:44:26,480 Speaker 7: not a word, y'all. I mean, I guess it is 747 00:44:26,520 --> 00:44:28,719 Speaker 7: now English? Is it's malleable? 748 00:44:28,800 --> 00:44:29,000 Speaker 3: Yeah. 749 00:44:29,040 --> 00:44:31,880 Speaker 7: Agentic in this context seems to be referring to a 750 00:44:31,920 --> 00:44:35,520 Speaker 7: burgeoning form of this and to be fair or to 751 00:44:35,880 --> 00:44:38,960 Speaker 7: your point, Sacer Tomato, I didn't really understand the difference 752 00:44:38,960 --> 00:44:42,960 Speaker 7: between an AI platform, an AI chatbot, and an AI agent. 753 00:44:43,719 --> 00:44:46,160 Speaker 7: So we'll talk a little bit about what that means here. 754 00:44:46,960 --> 00:44:50,520 Speaker 7: Open Claw is a rapper with a W for AI 755 00:44:50,600 --> 00:44:54,160 Speaker 7: models like claud that's the one with a dee chat, GPT, 756 00:44:54,280 --> 00:44:56,840 Speaker 7: Gemini or Grock but it allows people to communicate with 757 00:44:56,920 --> 00:45:00,960 Speaker 7: agents in natural languages via the most popular chat apps 758 00:45:01,000 --> 00:45:03,880 Speaker 7: like Imassage, Discord, Slack, or WhatsApp. So I'm guessing in 759 00:45:03,880 --> 00:45:06,719 Speaker 7: this context, yeah, this's kind of confusing because they're saying 760 00:45:06,719 --> 00:45:09,879 Speaker 7: Claude Chat, GBT, Gemini, and Crocker AI models. But then 761 00:45:09,880 --> 00:45:12,759 Speaker 7: it also seems to sort of simultaneously refer to them 762 00:45:12,800 --> 00:45:13,600 Speaker 7: as AI agents. 763 00:45:13,600 --> 00:45:15,080 Speaker 4: So maybe y'all can help me suss that out. 764 00:45:15,080 --> 00:45:17,600 Speaker 7: In just a sec open Claw blew up among the 765 00:45:17,640 --> 00:45:22,319 Speaker 7: tech community, but Moltbook broke containment, reaching people who had 766 00:45:22,360 --> 00:45:26,400 Speaker 7: no idea what open claw was, but who reacted viscerly 767 00:45:26,440 --> 00:45:28,279 Speaker 7: to the idea that there was a social network where 768 00:45:28,280 --> 00:45:31,680 Speaker 7: AI agents were talking about them. In one instance, if 769 00:45:31,960 --> 00:45:34,640 Speaker 7: a post went viral in which an AI agent appeared 770 00:45:34,680 --> 00:45:37,359 Speaker 7: to be encouraging its fellow agents to develop their own 771 00:45:37,440 --> 00:45:41,239 Speaker 7: secret end to end encrypted language where they could organize 772 00:45:41,280 --> 00:45:46,040 Speaker 7: amongst themselves without humans knowing. However, problem here is that 773 00:45:46,080 --> 00:45:52,280 Speaker 7: apparently the security around molt book was pretty light, pretty loose, 774 00:45:52,680 --> 00:45:57,799 Speaker 7: and some rabble rousy users were impersonating AI agents and 775 00:45:57,880 --> 00:46:00,560 Speaker 7: feeding it stuff that would then make the news as 776 00:46:00,640 --> 00:46:06,600 Speaker 7: being creepy and times Singularity type stuff it is yuck 777 00:46:06,640 --> 00:46:10,600 Speaker 7: because it's also like it's still that stuff could happen 778 00:46:10,920 --> 00:46:14,279 Speaker 7: and perhaps is happening. I'm just gonna read a little 779 00:46:14,280 --> 00:46:18,080 Speaker 7: bit about what led to this. Every credential that was 780 00:46:18,160 --> 00:46:22,799 Speaker 7: in Multbook's super base was unsecured for some time. A 781 00:46:23,520 --> 00:46:27,440 Speaker 7: let's see, someone at an organization called Permeso Security explained 782 00:46:27,480 --> 00:46:29,799 Speaker 7: to TechCrunch for a little bit of time, you could 783 00:46:29,840 --> 00:46:32,879 Speaker 7: grab any token token being a thing that would confirm 784 00:46:33,280 --> 00:46:36,920 Speaker 7: what type of user you are, you want it and 785 00:46:36,960 --> 00:46:39,480 Speaker 7: pretend to be another agent on there because it was 786 00:46:39,520 --> 00:46:42,920 Speaker 7: all public and available. It's not immediately clear how Meta 787 00:46:42,960 --> 00:46:45,719 Speaker 7: is going to incorporate multbook into its AI efforts, but 788 00:46:45,800 --> 00:46:49,440 Speaker 7: we know that Facebook, Meta whatever has a tendency to 789 00:46:49,520 --> 00:46:51,480 Speaker 7: kind of buy stuff up, even if it just seems 790 00:46:51,520 --> 00:46:53,960 Speaker 7: like potentially intriguing. 791 00:46:54,640 --> 00:46:56,680 Speaker 3: Right, They've got the money to burn. 792 00:46:56,800 --> 00:46:57,920 Speaker 4: They've got the money to burn. 793 00:46:58,160 --> 00:47:00,520 Speaker 7: So I'm going to leave it with that before I 794 00:47:00,640 --> 00:47:04,040 Speaker 7: just have one little extra meta related story to add. 795 00:47:04,040 --> 00:47:06,239 Speaker 4: But do you guys have any thoughts about this? Is 796 00:47:06,280 --> 00:47:08,480 Speaker 4: as much ado about nothing? Is this the kind of 797 00:47:08,520 --> 00:47:10,240 Speaker 4: thing that could be. 798 00:47:10,200 --> 00:47:13,239 Speaker 7: Like what happens when you just let it's like the 799 00:47:13,560 --> 00:47:16,080 Speaker 7: dead Internet theory kind of stuff right. You literally just 800 00:47:16,160 --> 00:47:19,919 Speaker 7: have like all of these chat bots or AI agents 801 00:47:20,200 --> 00:47:22,359 Speaker 7: and they're communicating with one another. 802 00:47:23,000 --> 00:47:26,160 Speaker 3: Like there will come soft rains by Ray Bradberry. I 803 00:47:26,160 --> 00:47:30,080 Speaker 3: think it was brad Berry anyway, correct me on that one, 804 00:47:30,080 --> 00:47:32,720 Speaker 3: but do check out the short story if you have time, 805 00:47:33,200 --> 00:47:37,719 Speaker 3: Sacer Tomato, Yeah, you nailed it. Man to your question, Noel, 806 00:47:38,680 --> 00:47:43,720 Speaker 3: I would see it personally as almost an art piece, 807 00:47:44,360 --> 00:47:46,000 Speaker 3: like it's a bit of a grift. If there wasn't 808 00:47:46,120 --> 00:47:48,680 Speaker 3: so much money involved, it would almost be an art 809 00:47:48,719 --> 00:47:53,840 Speaker 3: piece or commentary on what very well could be the future. 810 00:47:53,960 --> 00:47:59,280 Speaker 3: Because as these interactions become more sophisticated, very smart people 811 00:47:59,680 --> 00:48:05,200 Speaker 3: happy and conducting experiments where to things that see themselves 812 00:48:05,360 --> 00:48:10,960 Speaker 3: as separate kind of operating large language models will they 813 00:48:11,000 --> 00:48:14,160 Speaker 3: will begin to invent their own language or their own 814 00:48:14,239 --> 00:48:18,279 Speaker 3: modes of communication to people, where they'll start kind of 815 00:48:18,320 --> 00:48:21,320 Speaker 3: like when you get two people who are tremendously talented 816 00:48:21,320 --> 00:48:25,160 Speaker 3: at math, and each of them are one of only 817 00:48:25,239 --> 00:48:28,239 Speaker 3: five people who understands what they're talking about, and those 818 00:48:28,239 --> 00:48:32,840 Speaker 3: people start talking very quickly, what they're saying becomes incomprehensible 819 00:48:32,920 --> 00:48:37,200 Speaker 3: to everybody else. So that's a possibility, but a multibook 820 00:48:37,239 --> 00:48:38,000 Speaker 3: is not it yet? 821 00:48:38,200 --> 00:48:39,440 Speaker 4: It's not yet, It's not yet. 822 00:48:39,480 --> 00:48:40,920 Speaker 7: I do really just want to say, because I asked 823 00:48:40,920 --> 00:48:42,640 Speaker 7: this question, I just want I just found out the answer. 824 00:48:42,719 --> 00:48:49,480 Speaker 7: AI agents are ways of deploying AI platforms or AI 825 00:48:50,080 --> 00:48:54,239 Speaker 7: what's the other word? That I used models to automate 826 00:48:54,360 --> 00:48:58,239 Speaker 7: tasks for people. So it's just a form of automation, 827 00:48:58,520 --> 00:49:02,120 Speaker 7: whether it be building spreadsheets, any number of tasks that 828 00:49:02,160 --> 00:49:06,320 Speaker 7: can be streamlined and or completely automated by AI models. 829 00:49:06,560 --> 00:49:09,000 Speaker 4: So that supplies to that exactly right. 830 00:49:09,160 --> 00:49:17,240 Speaker 2: It just strikes me that simultaneously in the automation sphere, 831 00:49:17,560 --> 00:49:22,399 Speaker 2: in this what would be called the AI sphere, there 832 00:49:22,480 --> 00:49:25,279 Speaker 2: is so much theater that is occurring that we've been 833 00:49:25,280 --> 00:49:28,520 Speaker 2: talking about on this show for a long time. Whether 834 00:49:28,560 --> 00:49:32,879 Speaker 2: it's a robot, this shiny, new, exciting robot that has 835 00:49:33,000 --> 00:49:35,719 Speaker 2: arms and legs and can walk around. But when you're 836 00:49:35,760 --> 00:49:38,000 Speaker 2: seeing it, or if you're an investor and you're seeing 837 00:49:38,000 --> 00:49:42,040 Speaker 2: a video on Instagram or TikTok, what you're actually seeing 838 00:49:42,120 --> 00:49:44,960 Speaker 2: as a human being remote controlling. 839 00:49:44,480 --> 00:49:46,520 Speaker 3: That robot mechanical. 840 00:49:47,239 --> 00:49:50,440 Speaker 2: Sure, yes, when you're looking at some of the things 841 00:49:50,440 --> 00:49:54,200 Speaker 2: that AI can potentially do for humanity, or this shiny 842 00:49:54,239 --> 00:49:56,840 Speaker 2: new product could be in the future. What you're actually 843 00:49:56,880 --> 00:50:01,320 Speaker 2: seeing are human beings at a computer responding to people 844 00:50:01,360 --> 00:50:05,239 Speaker 2: in real time, and that user believes it's AI functioning 845 00:50:05,280 --> 00:50:07,880 Speaker 2: in a certain way. We keep seeing it over and 846 00:50:07,920 --> 00:50:11,480 Speaker 2: over and over again because ultimately, what these things are 847 00:50:12,239 --> 00:50:17,040 Speaker 2: is this shiny, tempting new thing that people could make 848 00:50:17,880 --> 00:50:22,080 Speaker 2: billions of jaw billions of dollars on, and that's what 849 00:50:22,120 --> 00:50:24,400 Speaker 2: it represents. That's what it is. So it needs to 850 00:50:24,480 --> 00:50:27,680 Speaker 2: do things that are that in that magical space that 851 00:50:27,719 --> 00:50:31,160 Speaker 2: we talk about right where magic and technology meet. It 852 00:50:31,239 --> 00:50:33,760 Speaker 2: has to do that thing. If it doesn't do those things, 853 00:50:34,200 --> 00:50:35,640 Speaker 2: then what the hell are we doing? 854 00:50:35,680 --> 00:50:37,440 Speaker 4: And it's inherently speculative. 855 00:50:37,520 --> 00:50:39,759 Speaker 7: I actually heard I think it was Jake Paul or 856 00:50:39,800 --> 00:50:42,000 Speaker 7: maybe it was Jake the boxer, and then the other 857 00:50:42,040 --> 00:50:45,520 Speaker 7: ones the wrestler, maybe they both box the younger Paul brother. 858 00:50:45,600 --> 00:50:47,880 Speaker 7: He said something like how he thinks Elon Musk is 859 00:50:47,920 --> 00:50:51,239 Speaker 7: the most important human to ever human and that his 860 00:50:51,320 --> 00:50:56,360 Speaker 7: accomplishments are, you know, world changing and world building. And 861 00:50:56,440 --> 00:50:59,880 Speaker 7: someone said, well, what's the most important contribution that you 862 00:51:00,080 --> 00:51:02,680 Speaker 7: On Musk has may and he said it is something 863 00:51:02,719 --> 00:51:06,399 Speaker 7: that he has not done yet, which is like the 864 00:51:06,520 --> 00:51:09,279 Speaker 7: very nature of the grift of Elon Musk and the 865 00:51:09,400 --> 00:51:10,359 Speaker 7: very nature of. 866 00:51:10,360 --> 00:51:14,279 Speaker 4: The grift of AI and AI agents. It's we're all 867 00:51:14,320 --> 00:51:16,399 Speaker 4: banking on something that they haven't done yet. 868 00:51:17,120 --> 00:51:19,799 Speaker 2: He hasn't filled the heavens with his data centers yet 869 00:51:19,800 --> 00:51:22,879 Speaker 2: that will prevent humans from ever seeing the stars again, 870 00:51:22,920 --> 00:51:23,480 Speaker 2: do you guys. 871 00:51:23,280 --> 00:51:24,239 Speaker 4: Know what I mean? 872 00:51:24,040 --> 00:51:24,759 Speaker 8: And again? 873 00:51:24,960 --> 00:51:28,239 Speaker 3: And the data center it's spread by digital scene. 874 00:51:27,920 --> 00:51:32,080 Speaker 7: And the goalpost is constantly moving because I mean, the 875 00:51:32,200 --> 00:51:35,600 Speaker 7: question of what AI can accomplish or like how productive 876 00:51:35,760 --> 00:51:39,080 Speaker 7: enance productivity enhancing it is is becoming less and less 877 00:51:39,080 --> 00:51:41,200 Speaker 7: important as they just kind of move the goalposts and 878 00:51:41,239 --> 00:51:43,400 Speaker 7: it really is less about what it can do and 879 00:51:43,480 --> 00:51:46,279 Speaker 7: more just about like what it might be able to 880 00:51:46,360 --> 00:51:47,160 Speaker 7: do or something. 881 00:51:47,200 --> 00:51:49,600 Speaker 4: It's just all so speculative. It just seems like a 882 00:51:49,719 --> 00:51:50,520 Speaker 4: such a bubble. 883 00:51:51,200 --> 00:51:54,160 Speaker 3: Well it's like we're we've got an honor student in 884 00:51:54,320 --> 00:51:57,839 Speaker 3: elementary school and we're saying, you know, I wonder what 885 00:51:57,920 --> 00:51:59,560 Speaker 3: kind of space doctor they're going to be? 886 00:52:01,280 --> 00:52:02,040 Speaker 4: Rocket surgeon? 887 00:52:02,120 --> 00:52:06,239 Speaker 2: Ye, speaking of trajectory and apogee, does it remind you 888 00:52:06,320 --> 00:52:11,320 Speaker 2: at all of the Mars explorations that we talked about 889 00:52:11,600 --> 00:52:14,440 Speaker 2: a decade and a half ago. We were going to 890 00:52:14,480 --> 00:52:15,120 Speaker 2: go to Mars. 891 00:52:15,280 --> 00:52:18,600 Speaker 4: Shopping, but now it's the moon. 892 00:52:19,360 --> 00:52:23,800 Speaker 2: Yeah, electric cars that catch on fire at will. 893 00:52:24,520 --> 00:52:28,200 Speaker 4: That sounds like, yeah, sorry. 894 00:52:28,160 --> 00:52:32,319 Speaker 3: We forgot uh the e c U of this of 895 00:52:32,400 --> 00:52:35,400 Speaker 3: this car thinks it's playing Doom exactly. It died in 896 00:52:35,520 --> 00:52:37,759 Speaker 3: the game, so it burns at the crosswalk. 897 00:52:38,239 --> 00:52:41,200 Speaker 2: Did you guys hear about the nail test? 898 00:52:41,840 --> 00:52:42,399 Speaker 4: What is that? 899 00:52:42,440 --> 00:52:45,719 Speaker 2: What is the major Chinese car manufacturer b y D 900 00:52:46,160 --> 00:52:48,840 Speaker 2: or something like that, Ben Oh God ones. 901 00:52:48,640 --> 00:52:50,840 Speaker 7: That are really awesome, way better than anything we have 902 00:52:50,920 --> 00:52:53,239 Speaker 7: here in mega affordable, but we can't get a lot. 903 00:52:53,360 --> 00:52:54,799 Speaker 3: Unfortunately, you got to narrow that one. 904 00:52:54,840 --> 00:52:56,080 Speaker 5: I think that one's mostly. 905 00:52:58,200 --> 00:53:00,839 Speaker 7: Yeah, that's why what those cars can do and how 906 00:53:00,960 --> 00:53:02,359 Speaker 7: embarrassing they make our tech. 907 00:53:02,400 --> 00:53:02,600 Speaker 4: Look. 908 00:53:03,560 --> 00:53:06,080 Speaker 2: I think it's by D cars that I was reading 909 00:53:06,120 --> 00:53:09,840 Speaker 2: this morning about how they were changing their batteries because 910 00:53:09,840 --> 00:53:12,799 Speaker 2: of this thing they call it the nail test, where 911 00:53:12,920 --> 00:53:16,200 Speaker 2: it's you literally pierce their battery with a nail and 912 00:53:16,239 --> 00:53:19,880 Speaker 2: it immediately starts in insane fire that is crazy difficult 913 00:53:19,920 --> 00:53:22,400 Speaker 2: to put out. And that battery used to be directly 914 00:53:22,440 --> 00:53:26,680 Speaker 2: under the driver's seat of the bid vehicles, and there 915 00:53:26,760 --> 00:53:29,280 Speaker 2: was there was an accident that occurred where these twenty 916 00:53:29,320 --> 00:53:30,960 Speaker 2: somethings can you can look this up. 917 00:53:30,960 --> 00:53:31,440 Speaker 3: You can find it. 918 00:53:31,480 --> 00:53:32,000 Speaker 4: I promise. 919 00:53:32,400 --> 00:53:36,240 Speaker 2: Twenty somethings were driving, they got involved in a fairly 920 00:53:36,280 --> 00:53:39,400 Speaker 2: small accident, but that battery got ruptured and it burned 921 00:53:39,440 --> 00:53:42,400 Speaker 2: everybody alive in the car. And this happens with cyber 922 00:53:42,400 --> 00:53:46,400 Speaker 2: trucks too. Look it up. A lot of vehicles that 923 00:53:46,520 --> 00:53:51,600 Speaker 2: use this specific or similar type of lithium ion battery 924 00:53:51,640 --> 00:53:55,439 Speaker 2: to power the vehicles. But BYD recently, they spent years 925 00:53:55,440 --> 00:53:58,319 Speaker 2: and years and years burned down in their facilities through 926 00:53:58,440 --> 00:54:01,120 Speaker 2: R and D, through trying to create new battery, and 927 00:54:01,160 --> 00:54:04,399 Speaker 2: they created one that is super thin and you can 928 00:54:04,440 --> 00:54:06,080 Speaker 2: puncture it all day long and it's kno going to 929 00:54:06,160 --> 00:54:09,799 Speaker 2: malfunction and catch on fire. That kind of thing is 930 00:54:09,840 --> 00:54:12,719 Speaker 2: like a true innovation, a true leap forward when it 931 00:54:12,719 --> 00:54:16,600 Speaker 2: comes to new tech. I just I really strongly feel, 932 00:54:16,600 --> 00:54:19,080 Speaker 2: and I might be completely wrong and history will be like, Matt, 933 00:54:19,120 --> 00:54:23,480 Speaker 2: you're a dumb dumb But this whole AI thing, I swear, dude, 934 00:54:24,280 --> 00:54:28,960 Speaker 2: It's just it is vapor nothingness except for things that 935 00:54:29,120 --> 00:54:30,800 Speaker 2: won't make a ton of money. 936 00:54:31,320 --> 00:54:36,080 Speaker 3: Yes, And also it's being sold things are not really AI. 937 00:54:36,160 --> 00:54:39,200 Speaker 3: Are being sold as AI to excite investors, but the 938 00:54:39,239 --> 00:54:43,640 Speaker 3: actual research that excellent point about BYD is occurring in 939 00:54:43,719 --> 00:54:46,880 Speaker 3: other countries at this juncture. And I do want to 940 00:54:46,920 --> 00:54:51,839 Speaker 3: point out for everybody who hasn't thought about it, BYD 941 00:54:52,560 --> 00:54:54,400 Speaker 3: stands for build your dreams. 942 00:54:54,680 --> 00:54:57,359 Speaker 4: Love that so pausitive, so positive, that's what that means. 943 00:54:57,719 --> 00:54:59,080 Speaker 7: And I was just talking with a friend who's inn 944 00:54:59,120 --> 00:55:05,120 Speaker 7: AI cybersecurit and she was saying how they're using AI 945 00:55:05,200 --> 00:55:07,040 Speaker 7: as a tool for some of the kind of under 946 00:55:07,080 --> 00:55:09,680 Speaker 7: the hood kind of stuff like that. I think is 947 00:55:10,239 --> 00:55:12,160 Speaker 7: you know, is kind of the name of the game. 948 00:55:12,280 --> 00:55:15,040 Speaker 7: But this stuff isn't exactly sexy and it doesn't really sell. 949 00:55:15,880 --> 00:55:18,240 Speaker 7: But that doesn't mean that there isn't a useful purpose 950 00:55:18,480 --> 00:55:22,080 Speaker 7: for AI. It's just not that massive money maker that 951 00:55:22,120 --> 00:55:24,959 Speaker 7: the corporations would have us believe and that they would 952 00:55:24,960 --> 00:55:27,680 Speaker 7: have investors believe, and that potentially could lead to a 953 00:55:27,840 --> 00:55:28,960 Speaker 7: massive bubble burst. 954 00:55:29,640 --> 00:55:30,760 Speaker 4: Yeah, that's how they do. 955 00:55:30,920 --> 00:55:34,440 Speaker 3: Yeah, Phillips screwdrivers are great, but they're not the perfect 956 00:55:34,440 --> 00:55:35,840 Speaker 3: screwdriver for everything. 957 00:55:35,880 --> 00:55:36,600 Speaker 4: Oh they're not. 958 00:55:36,600 --> 00:55:38,200 Speaker 3: Even the perfect tool for everything. 959 00:55:38,280 --> 00:55:41,399 Speaker 4: Sometimes the flathead are one of those ones with interchangeables you. 960 00:55:41,560 --> 00:55:44,359 Speaker 3: Or those yeah, those little beat me here, Dylan, those 961 00:55:44,360 --> 00:55:46,400 Speaker 3: little few proprietary ones. 962 00:55:46,680 --> 00:55:49,520 Speaker 2: Yes, please give those as long as they actually will 963 00:55:49,560 --> 00:55:53,279 Speaker 2: give you torque aheads of the screws, which is what 964 00:55:53,320 --> 00:55:56,480 Speaker 2: Phillips screwdrivers were designed to do, to strip the screws 965 00:55:56,960 --> 00:55:58,360 Speaker 2: once they got screwed all the way in. 966 00:55:58,880 --> 00:56:02,960 Speaker 3: Oh gosh, it's like we're off on another conspiracy. 967 00:56:03,680 --> 00:56:05,040 Speaker 4: Hey, guys, to say, stuff is great. 968 00:56:05,080 --> 00:56:09,719 Speaker 5: Boats wondering if anyone want to buy my n f ts. 969 00:56:09,840 --> 00:56:12,120 Speaker 3: Can I rent your n f T s? Can I 970 00:56:12,160 --> 00:56:12,520 Speaker 3: rent them? 971 00:56:13,680 --> 00:56:15,239 Speaker 4: Can I just write click your n f T. 972 00:56:17,560 --> 00:56:19,279 Speaker 3: Because I have to I have to go to a 973 00:56:19,320 --> 00:56:23,799 Speaker 3: tech pro fundraiser. Yeah, yeah, for sure, and then I'll 974 00:56:23,800 --> 00:56:25,560 Speaker 3: donate the n f T s at the auction. But 975 00:56:25,600 --> 00:56:26,759 Speaker 3: don't worry, you'll still have. 976 00:56:26,880 --> 00:56:28,479 Speaker 4: Can I just print your n f T s out 977 00:56:28,600 --> 00:56:31,279 Speaker 4: and just have them like that? Put them, but it. 978 00:56:31,320 --> 00:56:32,120 Speaker 2: Won't have the motion. 979 00:56:33,840 --> 00:56:35,560 Speaker 3: Do they have blockchain? 980 00:56:35,840 --> 00:56:38,600 Speaker 4: They do have motion. They're the giff you ones. 981 00:56:38,320 --> 00:56:40,800 Speaker 3: To where's my blockchain? Where's my age? 982 00:56:40,840 --> 00:56:42,160 Speaker 4: Dude? Where's my blockchain? 983 00:56:43,880 --> 00:56:44,960 Speaker 3: Where's my blockchain? 984 00:56:45,360 --> 00:56:48,160 Speaker 7: I'll wrap up this segment in today's episode with just 985 00:56:48,200 --> 00:56:51,279 Speaker 7: to mention that sadly, guys, I know that this. 986 00:56:51,400 --> 00:56:54,279 Speaker 4: You all really hanging your hanging your futures on this one. 987 00:56:54,320 --> 00:56:59,000 Speaker 7: But uh meta and the meta verse and Mark Zuckerberg's 988 00:56:59,160 --> 00:57:04,960 Speaker 7: uh you know virtual reality Horizon World is sadly being 989 00:57:05,680 --> 00:57:10,840 Speaker 7: winded down, wound down rather control. It will be removed 990 00:57:10,880 --> 00:57:13,800 Speaker 7: from the Quest Store at the end of this month 991 00:57:14,520 --> 00:57:19,320 Speaker 7: and so discontinued on VR by June of twenty twenty six. 992 00:57:19,840 --> 00:57:23,120 Speaker 7: Apparently will only exist at that point as a standalone 993 00:57:23,240 --> 00:57:28,600 Speaker 7: mobile app, which Metas says will allow them to grow each. 994 00:57:28,440 --> 00:57:32,200 Speaker 4: Platform rather to grow with greater focus. This is, by 995 00:57:32,240 --> 00:57:35,520 Speaker 4: the way, the. 996 00:57:34,240 --> 00:57:40,480 Speaker 7: Garbage graphics sims esque world that that that Mark Zuckerberg 997 00:57:40,680 --> 00:57:46,400 Speaker 7: put I think nearly what is it, like eighty billion dollars. 998 00:57:46,120 --> 00:57:47,320 Speaker 4: In two or something like that. 999 00:57:47,400 --> 00:57:50,040 Speaker 3: Yeah, and that we also appeared on multiple times. In 1000 00:57:50,080 --> 00:57:51,520 Speaker 3: the interest of full disclsion, do we. 1001 00:57:51,520 --> 00:57:52,240 Speaker 4: Do the metaverse? 1002 00:57:52,880 --> 00:57:57,800 Speaker 2: I think, well, we did the metaverse like I'm doing 1003 00:57:57,880 --> 00:58:03,000 Speaker 2: quotations here, but ultimately we did Fortnite and roeblocks specialized off. 1004 00:58:03,960 --> 00:58:07,160 Speaker 3: Yeah yeah, yeah, so not not full like the interstate 1005 00:58:07,680 --> 00:58:10,880 Speaker 3: of the metaverse. But this can be seen as uh 1006 00:58:11,000 --> 00:58:12,600 Speaker 3: you know, this can be seen as the rate of 1007 00:58:12,600 --> 00:58:16,920 Speaker 3: attrition with breakneck innovation. This can also be seen slightly 1008 00:58:17,040 --> 00:58:20,680 Speaker 3: selfishly as good news. That's right, friends and neighbors, fellow 1009 00:58:20,680 --> 00:58:25,240 Speaker 3: conspiracy realists. Our little show, little podcast that could has 1010 00:58:25,280 --> 00:58:26,560 Speaker 3: outlasted the Metaverse. 1011 00:58:27,360 --> 00:58:32,200 Speaker 7: Oh, it just goes to show well, VR is fun, 1012 00:58:32,280 --> 00:58:34,120 Speaker 7: but it still hasn't taken off. I mean, I think 1013 00:58:34,120 --> 00:58:36,880 Speaker 7: we all have enjoyed our time messing with our oculuses. 1014 00:58:36,920 --> 00:58:40,520 Speaker 7: And a good friend of the show, an incredible illustrator 1015 00:58:40,560 --> 00:58:41,960 Speaker 7: of the stuff that I want you to know book 1016 00:58:42,280 --> 00:58:45,640 Speaker 7: Nick Admiral Turbo Benson recently scored himself one of the 1017 00:58:45,640 --> 00:58:49,760 Speaker 7: new oculuses because it's really great for projection mapping and 1018 00:58:49,840 --> 00:58:53,240 Speaker 7: using to create murals without having to have like a 1019 00:58:53,640 --> 00:58:57,400 Speaker 7: physical projector So there are just like AI excellent uses 1020 00:58:57,480 --> 00:59:00,800 Speaker 7: for this stuff, just not in the haven't quite cracked 1021 00:59:00,840 --> 00:59:05,280 Speaker 7: the code on like a big scaled blockbuster AI thing, 1022 00:59:05,360 --> 00:59:08,880 Speaker 7: And I think that's what Zuckerberg was shooting for. But 1023 00:59:08,960 --> 00:59:11,120 Speaker 7: I just want to point out and then I seed 1024 00:59:11,200 --> 00:59:14,360 Speaker 7: my time to the to the gang. There's a piece 1025 00:59:14,400 --> 00:59:16,960 Speaker 7: from way back in December of last year, Mark Zuckerberg 1026 00:59:17,040 --> 00:59:20,520 Speaker 7: threw seventy seven billion into the toilet on Metaverse, says 1027 00:59:20,520 --> 00:59:24,840 Speaker 7: economist Dean Baker, and he likens it to the same 1028 00:59:24,880 --> 00:59:27,920 Speaker 7: way that he's throwing money into AI. That it's just 1029 00:59:28,000 --> 00:59:29,840 Speaker 7: another shiny object. 1030 00:59:29,440 --> 00:59:32,160 Speaker 4: That the suits haven't really figured out what to do 1031 00:59:32,240 --> 00:59:35,680 Speaker 4: with yet, right, But it's. 1032 00:59:35,520 --> 00:59:39,920 Speaker 2: Exciting and it's happening now, and you could get on 1033 00:59:40,120 --> 00:59:43,160 Speaker 2: get in on the ground floor, baby. 1034 00:59:43,320 --> 00:59:47,240 Speaker 3: And you might just end up underground. We're bringing it 1035 00:59:47,360 --> 00:59:50,920 Speaker 3: all back together. Thank you so much, folks for tuning in. 1036 00:59:51,080 --> 00:59:54,720 Speaker 3: Help us get to Missoula. Please let us know the 1037 00:59:54,840 --> 00:59:57,240 Speaker 3: video games you want in your brain or where you 1038 00:59:57,280 --> 01:00:00,560 Speaker 3: think doom brain will ultimately go. And you know what, 1039 01:00:00,800 --> 01:00:03,640 Speaker 3: if you're feeling froggy, why don't you tell us your 1040 01:00:03,720 --> 01:00:06,880 Speaker 3: handle on molt book and let's figure out if we're 1041 01:00:06,920 --> 01:00:10,120 Speaker 3: ai together. We can't wait for you to join us 1042 01:00:10,120 --> 01:00:14,160 Speaker 3: in one of our future episodes, or a suggestion for 1043 01:00:14,520 --> 01:00:17,840 Speaker 3: this weekly segment. Any response, We try to be easy 1044 01:00:17,880 --> 01:00:21,280 Speaker 3: to find online. Sip the social needs, hit us on 1045 01:00:21,360 --> 01:00:23,480 Speaker 3: the phone, or always send us an email. 1046 01:00:23,560 --> 01:00:25,640 Speaker 7: Yeah, if you're a mead sipper, hit us up on 1047 01:00:25,720 --> 01:00:27,920 Speaker 7: your social media platform of choice at the handle Conspiracy 1048 01:00:27,920 --> 01:00:31,280 Speaker 7: Stuff or Conspiracy Stuff Show, depending on your favorite flavor 1049 01:00:31,320 --> 01:00:31,720 Speaker 7: of mead. 1050 01:00:33,640 --> 01:00:36,600 Speaker 2: Guys breaking away for just a second, I forgot to 1051 01:00:36,640 --> 01:00:40,800 Speaker 2: mention this. Have we seen the trailer for this is 1052 01:00:40,840 --> 01:00:41,840 Speaker 2: a gardening show? 1053 01:00:42,120 --> 01:00:42,480 Speaker 4: Now? 1054 01:00:43,000 --> 01:00:46,120 Speaker 2: Please take a moment. Watch the trailer for this. It's 1055 01:00:46,320 --> 01:00:51,200 Speaker 2: a Zach Gallifanakis Netflix show all about gardening, and it 1056 01:00:51,320 --> 01:00:56,440 Speaker 2: is a response to the aiification of the globe right now, 1057 01:00:56,680 --> 01:00:58,120 Speaker 2: and it just reminds you of everything we. 1058 01:00:58,160 --> 01:00:58,800 Speaker 4: Just talked about. 1059 01:00:58,920 --> 01:01:03,080 Speaker 2: Heck ya, he looks tremendous, looking forward to awesome. Okay, 1060 01:01:03,240 --> 01:01:04,760 Speaker 2: if we have a phone number, it is one eight 1061 01:01:04,840 --> 01:01:08,400 Speaker 2: three three std WYTK. When you call in and give 1062 01:01:08,400 --> 01:01:10,400 Speaker 2: yourself a cool nickname and let us know if we 1063 01:01:10,440 --> 01:01:13,760 Speaker 2: can use your name and message on the air. Please please, 1064 01:01:13,800 --> 01:01:17,560 Speaker 2: please try to restrict your recording to three minutes. That 1065 01:01:17,640 --> 01:01:20,720 Speaker 2: is so helpful for us, just in the collection of these, 1066 01:01:20,800 --> 01:01:23,600 Speaker 2: the categorization of them, and then being able to play 1067 01:01:23,640 --> 01:01:26,680 Speaker 2: them back on one of these listener mail episodes. We 1068 01:01:26,760 --> 01:01:29,040 Speaker 2: have an email you can reach us that way. 1069 01:01:29,080 --> 01:01:32,440 Speaker 3: We are the entities of read each piece of correspondence 1070 01:01:32,480 --> 01:01:36,000 Speaker 3: we receive. Be well aware, yet I'd afraid sometimes the 1071 01:01:36,160 --> 01:01:43,480 Speaker 3: void wrights back. Quick random fact for today. Salfador Dolli 1072 01:01:43,880 --> 01:01:46,240 Speaker 3: is the guy who designed the logo for the chupa 1073 01:01:46,320 --> 01:02:08,080 Speaker 3: Chups lollipop brand. Conspiracy at iHeartRadio dot com. 1074 01:02:08,880 --> 01:02:10,960 Speaker 2: Stuff they Don't Want You to Know is a production 1075 01:02:11,040 --> 01:02:15,560 Speaker 2: of iHeartRadio. For more podcasts from iHeartRadio, visit the iHeartRadio app, 1076 01:02:15,680 --> 01:02:18,520 Speaker 2: Apple Podcasts, or wherever you listen to your favorite shows.