1 00:00:01,720 --> 00:00:07,160 Speaker 1: Cool media. Oh my god, you guys, it could happen here, 2 00:00:08,280 --> 00:00:13,040 Speaker 1: meaning our podcast, it could, it is, It's happened. 3 00:00:13,680 --> 00:00:17,320 Speaker 2: Robert, shouldn't you rename the podcast? It is happening here? 4 00:00:17,600 --> 00:00:19,720 Speaker 1: Yeah? Uh huh, that's that's a fun joke that I 5 00:00:19,760 --> 00:00:25,279 Speaker 1: only hear forty seven times a day. And the whole 6 00:00:25,280 --> 00:00:28,280 Speaker 1: point of the podcast was, well, initially I was a 7 00:00:28,320 --> 00:00:31,400 Speaker 1: crazy person saying a bunch of stuff would happen, and 8 00:00:31,440 --> 00:00:35,760 Speaker 1: now it's a bunch of that stuff happened, and even 9 00:00:35,800 --> 00:00:39,040 Speaker 1: more of it looks very likely, and so now I 10 00:00:39,159 --> 00:00:40,479 Speaker 1: just feel bad all the time. 11 00:00:41,000 --> 00:00:43,040 Speaker 3: It's going to be cooled. I fucking called it. I 12 00:00:43,080 --> 00:00:45,520 Speaker 3: fucking told you, bro, I said this was going to happen. 13 00:00:45,960 --> 00:00:49,040 Speaker 2: Why don't you rename the podcast? I just feel bad 14 00:00:49,040 --> 00:00:49,680 Speaker 2: all the time? 15 00:00:49,880 --> 00:00:52,199 Speaker 1: Yeah, why don't you rename the podcast? Robert should have 16 00:00:52,200 --> 00:00:56,120 Speaker 1: bought more stock and ammunition companies than he did and 17 00:00:56,280 --> 00:00:59,720 Speaker 1: DGI jeez, should I have bought stock and DGI? 18 00:01:00,160 --> 00:01:04,600 Speaker 3: Yeah, I'm gonna buy her a little DGI drone here, 19 00:01:04,800 --> 00:01:05,080 Speaker 3: you know. 20 00:01:05,120 --> 00:01:06,679 Speaker 1: Yeah, there we go a lot of people are going 21 00:01:06,760 --> 00:01:09,280 Speaker 1: to be buying little DGI drones here very soon. 22 00:01:09,480 --> 00:01:13,240 Speaker 3: James, I should point out that I'm buying one. It's 23 00:01:13,280 --> 00:01:14,679 Speaker 3: not capable of carrying a payload. 24 00:01:15,080 --> 00:01:18,199 Speaker 2: It's definitely a safer investment to pull out your furrow 25 00:01:18,240 --> 00:01:21,679 Speaker 2: one k Now, when the market's crashing, use that money, 26 00:01:21,720 --> 00:01:25,000 Speaker 2: buy drones. Those drones will be worth a lot more 27 00:01:25,000 --> 00:01:25,840 Speaker 2: in five years. 28 00:01:26,520 --> 00:01:30,119 Speaker 3: Or what is that? That is? That is the sound 29 00:01:30,120 --> 00:01:32,959 Speaker 3: of a sound investment, a box of bullets. 30 00:01:33,440 --> 00:01:35,200 Speaker 2: It's like how boomers used to like invest in like 31 00:01:35,319 --> 00:01:39,320 Speaker 2: silver or gold as like a stable current. No, we're 32 00:01:39,360 --> 00:01:43,800 Speaker 2: investing in DGI, like physical dghy drones. 33 00:01:44,640 --> 00:01:47,600 Speaker 1: We are investing in drones and boxes of gunpowder. 34 00:01:48,160 --> 00:01:50,400 Speaker 3: Yeah, you gotta get it in a bottle, rubbit in 35 00:01:50,400 --> 00:01:52,880 Speaker 3: a box, it can get light struck or get moist. 36 00:01:52,880 --> 00:01:55,520 Speaker 3: You want to get it in a special black black bottle. 37 00:01:55,480 --> 00:01:58,440 Speaker 1: James, I keep all of my gunpowder. And you know 38 00:01:58,480 --> 00:02:01,240 Speaker 1: how like people used to cocaine by wrapping it in 39 00:02:01,320 --> 00:02:02,680 Speaker 1: toilet paper and swallowing it. 40 00:02:03,000 --> 00:02:06,440 Speaker 3: No, sure, okay, well if you say sobody. 41 00:02:08,400 --> 00:02:13,320 Speaker 1: Speaking of toilet paper. Nate Silver has a newsletter and 42 00:02:13,360 --> 00:02:16,120 Speaker 1: it would be useful as toilet paper more so than 43 00:02:16,200 --> 00:02:17,520 Speaker 1: it is as a newsletter. 44 00:02:17,600 --> 00:02:20,480 Speaker 2: Sorry, I just got like PTSD flashbacks from twenty twenty four. 45 00:02:20,560 --> 00:02:21,919 Speaker 3: Would you said that it's okay. 46 00:02:22,120 --> 00:02:27,840 Speaker 1: Normally, my rule of thumb is every election, usually starting 47 00:02:27,840 --> 00:02:32,639 Speaker 1: in like December, the year before election year, I begrudgingly 48 00:02:33,000 --> 00:02:36,280 Speaker 1: fight down a series of panic attacks, vomit three or 49 00:02:36,280 --> 00:02:38,560 Speaker 1: four times in a bucket, and then head over to 50 00:02:38,639 --> 00:02:41,360 Speaker 1: Nate Silver's blog to see what he's saying about the polls. 51 00:02:41,760 --> 00:02:44,080 Speaker 1: And I do this. I hate that I keep having 52 00:02:44,360 --> 00:02:46,760 Speaker 1: I have regularly on election the years people were like 53 00:02:46,919 --> 00:02:49,400 Speaker 1: but he was always wrong. He's like, no, he's reasonably 54 00:02:49,400 --> 00:02:52,120 Speaker 1: good on polls. He's usually if you read what he's 55 00:02:52,120 --> 00:02:56,239 Speaker 1: saying about presidential polls, the reality bears out pretty close 56 00:02:56,280 --> 00:02:59,560 Speaker 1: to that. So I read him during elections and I 57 00:02:59,639 --> 00:03:02,720 Speaker 1: hate it because he's never been right about anything else. 58 00:03:02,760 --> 00:03:07,360 Speaker 1: But he's he's a gambler. He's a degenerate, filthy gambler. 59 00:03:07,720 --> 00:03:11,320 Speaker 1: And so when we're talking about degenerate, filthy gambler stuff, 60 00:03:11,560 --> 00:03:14,560 Speaker 1: and by god, election polls are the most degenerate type 61 00:03:14,600 --> 00:03:18,480 Speaker 1: of gambling that exists, he's worth reading. And then after 62 00:03:18,560 --> 00:03:21,560 Speaker 1: the election, no matter how well or badly it goes, 63 00:03:21,919 --> 00:03:24,280 Speaker 1: I ignore him again for four years. And I didn't get 64 00:03:24,320 --> 00:03:27,200 Speaker 1: to do that this year because on February twenty fifth, 65 00:03:27,240 --> 00:03:30,959 Speaker 1: twenty twenty five, Nate wrote a column called elon Musk 66 00:03:31,080 --> 00:03:33,079 Speaker 1: and Spiky Intelligence. 67 00:03:33,440 --> 00:03:35,720 Speaker 2: Spiky Intelligence. Am I hearing that right? 68 00:03:36,320 --> 00:03:37,480 Speaker 1: Spiky intelligence? 69 00:03:37,560 --> 00:03:37,760 Speaker 3: Yes? 70 00:03:38,040 --> 00:03:40,760 Speaker 1: And it very helpfully starts with a drawing that I'm 71 00:03:40,800 --> 00:03:44,600 Speaker 1: sure he used some AI, like he must have used 72 00:03:44,640 --> 00:03:47,440 Speaker 1: some AI like video software to do that, just like 73 00:03:47,520 --> 00:03:51,080 Speaker 1: shows you a kind of spiky star looking thing and 74 00:03:51,120 --> 00:03:54,640 Speaker 1: then like a blob with rounded edges. I can't begin 75 00:03:54,720 --> 00:03:58,200 Speaker 1: to imagine why Nate Silver thought that, like we needed 76 00:03:58,200 --> 00:03:59,120 Speaker 1: this illustrated. 77 00:03:59,240 --> 00:04:01,240 Speaker 3: I have to see this. Yeah, yeah, I would like 78 00:04:01,280 --> 00:04:02,520 Speaker 3: it to be shad look at this. 79 00:04:02,760 --> 00:04:06,240 Speaker 1: Why did you like, oh promise of AI we couldn't have? 80 00:04:08,280 --> 00:04:10,320 Speaker 3: Yes, wow, yeah, it just it looks like. 81 00:04:10,320 --> 00:04:13,080 Speaker 1: Maybe an amoeba if you, if you looks like an amoeba, 82 00:04:13,240 --> 00:04:15,080 Speaker 1: and then like a poorly drawn star. 83 00:04:15,280 --> 00:04:18,640 Speaker 3: Is it? This is? When is an actual thing? This 84 00:04:18,839 --> 00:04:20,400 Speaker 3: gifts you? Wait, this is a thing. 85 00:04:20,760 --> 00:04:22,960 Speaker 2: This is this is Boba and Kiki with a weird 86 00:04:22,960 --> 00:04:25,720 Speaker 2: like digital fuzz over the. 87 00:04:25,680 --> 00:04:27,000 Speaker 1: Fucker Boba and Kiki. 88 00:04:27,440 --> 00:04:29,039 Speaker 3: Yeah, okay, Garrison, Yeah. 89 00:04:28,920 --> 00:04:31,520 Speaker 2: It's a it's like a social experiment to like ask 90 00:04:31,640 --> 00:04:35,040 Speaker 2: people what like the emotional correspondence of each of these 91 00:04:35,040 --> 00:04:40,159 Speaker 2: shapes are like, which it's like a sure, like like 92 00:04:40,160 --> 00:04:42,880 Speaker 2: which one looks, which one looks nicer, which one looks meaner? 93 00:04:43,000 --> 00:04:45,839 Speaker 2: You know that sort of thing. I'm a Kiki type, 94 00:04:45,880 --> 00:04:48,760 Speaker 2: like like I I am a Kiky in terms in 95 00:04:48,839 --> 00:04:51,080 Speaker 2: terms of my behavior, I am Garrison. 96 00:04:51,200 --> 00:04:53,280 Speaker 1: Now that you bring up Rorshak, all I can think 97 00:04:53,279 --> 00:04:55,360 Speaker 1: of is how cool it would be if Rorshak from 98 00:04:55,360 --> 00:04:58,160 Speaker 1: The Watchman showed up in Nate Silver's house and did 99 00:04:58,200 --> 00:04:58,560 Speaker 1: his thing. 100 00:05:00,400 --> 00:05:05,600 Speaker 2: Unfortunately, I think Rorshak at night Silver friends. 101 00:05:06,720 --> 00:05:10,200 Speaker 1: Actually yeah, no, no, Nate would. But after them getting 102 00:05:10,240 --> 00:05:12,440 Speaker 1: along for like forty five minutes, Nate would take him 103 00:05:12,440 --> 00:05:15,440 Speaker 1: to an illegal card game, and Rorshak would murder everybody 104 00:05:15,480 --> 00:05:19,240 Speaker 1: in the room because they were gambling without a license. 105 00:05:21,520 --> 00:05:25,240 Speaker 2: So I'm assuming Nate's going to try to argue that 106 00:05:25,240 --> 00:05:29,000 Speaker 2: that Musk's intelligence is akin to the kiki drug here 107 00:05:29,240 --> 00:05:32,320 Speaker 2: as opposed to like the emphathetic right. 108 00:05:32,160 --> 00:05:34,240 Speaker 1: Now there, Actually, yes, there is a little bit of 109 00:05:34,240 --> 00:05:36,719 Speaker 1: that in there. He does not mention this Kiki and 110 00:05:36,800 --> 00:05:39,160 Speaker 1: Boba thing. I don't know if that's because I'm supposed 111 00:05:39,240 --> 00:05:42,400 Speaker 1: to just infer it from the image or if he's Okay, 112 00:05:42,560 --> 00:05:44,240 Speaker 1: we'll get your opinion on it. Is is Hee ripping these 113 00:05:44,279 --> 00:05:47,000 Speaker 1: people off because this doesn't count as enough for him 114 00:05:47,040 --> 00:05:49,159 Speaker 1: to be crediting them if this is the underpinning of 115 00:05:49,200 --> 00:05:52,000 Speaker 1: his stupid idea, which he credits to his stupid book 116 00:05:52,000 --> 00:05:53,320 Speaker 1: that he came up with later. But I'm just going 117 00:05:53,360 --> 00:05:54,960 Speaker 1: to start reading the stupid column. 118 00:05:55,240 --> 00:05:57,919 Speaker 3: Well, hit us with the second paragraph, because that fuck. 119 00:05:57,760 --> 00:05:59,240 Speaker 1: You haven't gotten paragraph one. 120 00:05:59,360 --> 00:06:01,400 Speaker 3: Jack radicalized me immediately. 121 00:06:01,440 --> 00:06:04,160 Speaker 1: There's been a debate raging on Twitter. Noah Smith can 122 00:06:04,240 --> 00:06:06,760 Speaker 1: run you through the parameters about the intelligence of the 123 00:06:06,760 --> 00:06:11,080 Speaker 1: platform's owner, Elon Musk. My contribution was to suggest and 124 00:06:11,120 --> 00:06:13,560 Speaker 1: then there's a little eye in parentheses because we need 125 00:06:13,560 --> 00:06:16,720 Speaker 1: that Elon is obviously pretty bright. And then there's two 126 00:06:16,800 --> 00:06:20,479 Speaker 1: eyes in parentheses. This shouldn't be conflated with moral judgment. 127 00:06:20,760 --> 00:06:24,800 Speaker 1: Highly intelligent people do lots of bad things. Okay, you'd 128 00:06:24,800 --> 00:06:27,760 Speaker 1: think this wouldn't be especially controversial, but since it involves 129 00:06:27,760 --> 00:06:30,960 Speaker 1: Elon and intelligence, well it was. Elon has run founded 130 00:06:31,040 --> 00:06:35,200 Speaker 1: or co foundeds Tesla, SpaceX, open Ai, neuralink Xai, PayPal, 131 00:06:35,320 --> 00:06:39,200 Speaker 1: and more recently Twitter. He's also managed to steer himself 132 00:06:39,240 --> 00:06:41,240 Speaker 1: into a position where he's now the de facto chief 133 00:06:41,240 --> 00:06:43,599 Speaker 1: of staff to the President of the United States. I 134 00:06:43,640 --> 00:06:46,120 Speaker 1: do not doubt that Elon has gotten lucky in various respects. 135 00:06:46,200 --> 00:06:49,360 Speaker 1: Some of these were long shot bets, and Walter Isaacson's 136 00:06:49,360 --> 00:06:52,080 Speaker 1: biography of Musk documents he thought he'd be ruined if 137 00:06:52,080 --> 00:06:54,880 Speaker 1: there had been one more failed SpaceX launch. The success 138 00:06:54,920 --> 00:06:57,839 Speaker 1: of some of these enterprises might also be debated. Twitter 139 00:06:57,920 --> 00:07:00,200 Speaker 1: was a canny play for cultural and political and ones. 140 00:07:00,200 --> 00:07:02,680 Speaker 1: But probably And he doesn't bring up in this whole 141 00:07:02,680 --> 00:07:04,719 Speaker 1: thing where he's talking about, like all a successful company, 142 00:07:04,880 --> 00:07:08,400 Speaker 1: not a word about the boring company, not a word 143 00:07:08,400 --> 00:07:09,920 Speaker 1: about hyper loop right. 144 00:07:09,800 --> 00:07:11,760 Speaker 3: Well, yeah, yeah, any of the failure ones. 145 00:07:11,760 --> 00:07:14,200 Speaker 1: His record does seem better if you ignore the two 146 00:07:14,360 --> 00:07:18,040 Speaker 1: massively publicized and invested absolute failures. 147 00:07:18,160 --> 00:07:20,280 Speaker 2: Yes, well, and last week, I know there was a 148 00:07:20,280 --> 00:07:22,760 Speaker 2: space X lunch. I'm sure it went well. I'm sure 149 00:07:22,760 --> 00:07:26,240 Speaker 2: it didn't fling debris all over lower. 150 00:07:26,560 --> 00:07:30,360 Speaker 1: I'm sure he didn't nearly destroy several commercial aircraft, also 151 00:07:30,720 --> 00:07:34,280 Speaker 1: crediting it like, yeah, I guess technically co founded open ai, 152 00:07:34,400 --> 00:07:36,200 Speaker 1: but not in a way that mattered. He just shot 153 00:07:36,240 --> 00:07:38,400 Speaker 1: down money in there and then kind of edged out. 154 00:07:39,000 --> 00:07:39,320 Speaker 3: Sure. 155 00:07:39,560 --> 00:07:42,800 Speaker 1: Yes, and is actively in a conflict with everybody who 156 00:07:42,840 --> 00:07:46,840 Speaker 1: did make open ai as prominent as it is. Again, 157 00:07:47,160 --> 00:07:49,240 Speaker 1: Nitla has to leave a lot out in order to 158 00:07:49,320 --> 00:07:50,720 Speaker 1: start making this case. 159 00:07:51,000 --> 00:07:53,080 Speaker 2: But so he's going to argue that, you know, we're 160 00:07:53,080 --> 00:07:55,920 Speaker 2: going to see how how well this co presidency goes. 161 00:07:55,960 --> 00:07:58,240 Speaker 2: But he's probably a pretty smart guy to get all 162 00:07:58,280 --> 00:07:59,040 Speaker 2: of this stuff done. 163 00:07:59,200 --> 00:08:01,240 Speaker 1: Yeah, and he's also saying, well, like maybe Twitter won't 164 00:08:01,240 --> 00:08:03,280 Speaker 1: be profitable, but we'll see how you know, he could 165 00:08:03,320 --> 00:08:05,960 Speaker 1: probably profit from being the de facto chief of staff. 166 00:08:06,120 --> 00:08:08,040 Speaker 1: Not a word from Nate about like, yeah, but he's 167 00:08:08,080 --> 00:08:10,800 Speaker 1: just like, that's just breaking the law. So why are 168 00:08:10,840 --> 00:08:13,679 Speaker 1: we Why aren't we including in our canny businessman guys 169 00:08:13,680 --> 00:08:17,160 Speaker 1: that get rich selling like shitloads of heroin for the cartels, 170 00:08:17,280 --> 00:08:21,000 Speaker 1: Because yes, if you are breaking the laws sometimes that 171 00:08:21,160 --> 00:08:22,760 Speaker 1: goes well for you financially. 172 00:08:23,440 --> 00:08:26,120 Speaker 2: Well, Walter White, they've done some bad things, Yeah, but. 173 00:08:26,680 --> 00:08:29,880 Speaker 1: You can't deny he was a brilliant method, you know. 174 00:08:32,080 --> 00:08:36,000 Speaker 1: But I don't care what Elon's SAT score is. Fourteen hundred. 175 00:08:36,040 --> 00:08:39,680 Speaker 1: According to Isaacson, he's clearly some sort of outlier in 176 00:08:39,679 --> 00:08:43,319 Speaker 1: many ways people would associate with intelligence, probably even a genius. 177 00:08:43,880 --> 00:08:47,120 Speaker 1: And yet when my first off, it becomes clear through 178 00:08:47,120 --> 00:08:49,680 Speaker 1: this that Nate does not consider a fourteen hundred to 179 00:08:49,720 --> 00:08:53,480 Speaker 1: be an impressive SAT score, and would normally be judgmental 180 00:08:53,480 --> 00:08:56,200 Speaker 1: of someone who had an SAT score of fourteen hundred 181 00:08:56,240 --> 00:08:59,280 Speaker 1: if it weren't for all of Elon's other genius accomplishments. 182 00:09:00,040 --> 00:09:01,760 Speaker 1: And yet, when my partner and I were heading to 183 00:09:01,800 --> 00:09:03,480 Speaker 1: dinner the other day and we saw some tweet that 184 00:09:03,520 --> 00:09:06,400 Speaker 1: Elon sent I forget which one because he tweets so much, 185 00:09:06,640 --> 00:09:09,760 Speaker 1: we were both like, man, he's such a dumb ass. Yes, 186 00:09:09,840 --> 00:09:12,320 Speaker 1: someone can be both a genius and a dumbass. Welcome 187 00:09:12,360 --> 00:09:14,120 Speaker 1: to what I call spikey intelligence. 188 00:09:14,200 --> 00:09:14,720 Speaker 2: Here we go. 189 00:09:15,240 --> 00:09:17,920 Speaker 1: This gets to like the core of what's annoying about 190 00:09:18,040 --> 00:09:20,640 Speaker 1: Nate is his need to He's one of these guys 191 00:09:20,679 --> 00:09:22,920 Speaker 1: you know what, you know what it is. He's an 192 00:09:23,000 --> 00:09:27,600 Speaker 1: intellectual enclosurist right where he's not confident to be like 193 00:09:27,679 --> 00:09:30,360 Speaker 1: everyone is very aware of the fact that no one 194 00:09:30,400 --> 00:09:32,720 Speaker 1: is good at everything, and that people have holes in 195 00:09:32,760 --> 00:09:36,439 Speaker 1: their competence, and that there are like brilliant surgeons who 196 00:09:36,440 --> 00:09:39,679 Speaker 1: are bad fathers or whatever, because there are different kinds 197 00:09:39,679 --> 00:09:43,560 Speaker 1: of intelligence. This is like a broadly common understanding. Nate 198 00:09:43,640 --> 00:09:45,439 Speaker 1: has to give it a name so that he can 199 00:09:45,480 --> 00:09:47,200 Speaker 1: sell his books. So he gives it the names. It's 200 00:09:47,240 --> 00:09:50,000 Speaker 1: like an intellectual. Now it's my idea. I'm the one 201 00:09:50,040 --> 00:09:52,200 Speaker 1: who came up with the concept that smart people can 202 00:09:52,240 --> 00:09:54,680 Speaker 1: be dumbasses. Stop it, Nate, it's annoying. 203 00:09:55,120 --> 00:09:59,320 Speaker 3: Capital ass capital I rights to trademark spiky intelligence. Yeah. 204 00:09:59,440 --> 00:10:04,320 Speaker 1: Yeah, Now he acknowledges that this isn't entirely original, and 205 00:10:04,360 --> 00:10:08,120 Speaker 1: then links to somebody without really like crediting them. Interestingly, 206 00:10:08,200 --> 00:10:10,280 Speaker 1: many of the instances online refer to people of the 207 00:10:10,520 --> 00:10:13,800 Speaker 1: on the autism spectrum. Musk has publicly stated that he 208 00:10:13,840 --> 00:10:17,079 Speaker 1: has Asperger syndrome. But the concept is simple. While intelligence 209 00:10:17,160 --> 00:10:20,480 Speaker 1: is a multi dimensional phenomenon, the scientific consensus is that 210 00:10:20,520 --> 00:10:23,679 Speaker 1: there's also something known as a G factor sometimes also 211 00:10:23,720 --> 00:10:27,400 Speaker 1: called general intelligence. As an empirical matter, most traits we'd 212 00:10:27,400 --> 00:10:31,080 Speaker 1: associate with intelligence are positively correlated. For instance, math and 213 00:10:31,160 --> 00:10:34,360 Speaker 1: verbal skills and the gire are correlated. The correlations are 214 00:10:34,400 --> 00:10:36,160 Speaker 1: loose enough that you'll wind up with all sorts of 215 00:10:36,160 --> 00:10:39,199 Speaker 1: different permutations on the spectrum of human behavior. And he's 216 00:10:39,240 --> 00:10:41,000 Speaker 1: just going into like he talks about like the absent 217 00:10:41,040 --> 00:10:44,480 Speaker 1: minded professor, Like it's all just these these very common 218 00:10:44,520 --> 00:10:48,320 Speaker 1: ideas that like, yeah, people are usually bad at more 219 00:10:48,360 --> 00:10:51,800 Speaker 1: things than they're good at, right, Like it's there's no 220 00:10:51,840 --> 00:10:57,160 Speaker 1: need to explain how Elon Musk has been successful at 221 00:10:57,200 --> 00:10:59,760 Speaker 1: certain things, but Nate does, and he has to keep 222 00:10:59,760 --> 00:11:01,840 Speaker 1: going back to. Like he makes a comment later in 223 00:11:01,880 --> 00:11:04,920 Speaker 1: here about how Musk is clearly a brilliant engineer. He 224 00:11:04,920 --> 00:11:07,440 Speaker 1: doesn't back this up with evidence. He just says that, Like, well, 225 00:11:07,440 --> 00:11:10,000 Speaker 1: if you read the book that Ashley Vance wrote, he 226 00:11:10,080 --> 00:11:13,000 Speaker 1: obviously signed off on a lot of great engineering moves, 227 00:11:13,160 --> 00:11:15,439 Speaker 1: which ignores the fact that, like he's not making any 228 00:11:15,520 --> 00:11:18,440 Speaker 1: of these decisions, Like he bought a company that already 229 00:11:18,480 --> 00:11:21,680 Speaker 1: had good automotive technology. He hired a bunch of rocket 230 00:11:21,679 --> 00:11:27,160 Speaker 1: engineers to design rockets. Elon is arguably good at hiring 231 00:11:27,200 --> 00:11:31,680 Speaker 1: in certain circumstances, and he is inarguably a great hype man, right, 232 00:11:31,840 --> 00:11:35,080 Speaker 1: Like that's the actual brilliance that Elon has is he 233 00:11:35,160 --> 00:11:37,760 Speaker 1: was very very good at hyping people up and getting 234 00:11:37,760 --> 00:11:40,440 Speaker 1: people to believe in him until he was too big 235 00:11:40,480 --> 00:11:43,320 Speaker 1: to fail. Like that's the one thing he actually did. 236 00:11:43,800 --> 00:11:46,280 Speaker 1: But Nate can't accept that because I think it kind of, 237 00:11:46,559 --> 00:11:49,160 Speaker 1: among other things, it kind of reveals what Nate is, 238 00:11:49,559 --> 00:11:52,319 Speaker 1: who is a guy who was really good at one 239 00:11:52,520 --> 00:11:56,520 Speaker 1: narrow thing and now has a career writing about everything 240 00:11:57,280 --> 00:12:01,040 Speaker 1: and he can't. That's like a dangerous thing for Nate 241 00:12:01,080 --> 00:12:02,679 Speaker 1: to think too hard about. 242 00:12:03,600 --> 00:12:08,440 Speaker 2: Let's learn more about Nate's spiky intelligence after these very 243 00:12:08,480 --> 00:12:10,120 Speaker 2: soft and soothing ads. 244 00:12:10,320 --> 00:12:21,920 Speaker 4: Yeah we're back. 245 00:12:22,040 --> 00:12:24,360 Speaker 1: I want to talk a little bit about the danger 246 00:12:24,520 --> 00:12:27,079 Speaker 1: of being a guy who gets famous for being really 247 00:12:27,120 --> 00:12:29,640 Speaker 1: good at one thing and then gets a job talking 248 00:12:29,679 --> 00:12:32,920 Speaker 1: about everything, because I've had a version of that experience. 249 00:12:33,080 --> 00:12:36,200 Speaker 1: And let me tell you, you're not ever going to be 250 00:12:36,600 --> 00:12:39,160 Speaker 1: competent to discuss all of the things that you can 251 00:12:39,200 --> 00:12:42,840 Speaker 1: make money talking about if you're a popular entertainer. No 252 00:12:42,920 --> 00:12:45,640 Speaker 1: one ever has been, and no one ever will be, 253 00:12:46,240 --> 00:12:48,240 Speaker 1: which is why what you ought to do is the 254 00:12:48,280 --> 00:12:51,640 Speaker 1: thing Nate initially tried to do, which is bring on 255 00:12:51,760 --> 00:12:55,319 Speaker 1: a bunch of people to like run a website with 256 00:12:55,360 --> 00:12:59,080 Speaker 1: you right where you cover more things than one. Unfortunately, 257 00:12:59,080 --> 00:13:01,600 Speaker 1: it turns out five thirty was a bad business venture. 258 00:13:01,679 --> 00:13:04,800 Speaker 1: It got massively overvalued, a company spent a shitload more 259 00:13:04,800 --> 00:13:07,040 Speaker 1: money on it that it was capable of making, and 260 00:13:07,120 --> 00:13:10,120 Speaker 1: now everyone's gotten laid off and Nate left years ago 261 00:13:10,240 --> 00:13:13,199 Speaker 1: to do his sub stack. You know, it's a tragic 262 00:13:13,280 --> 00:13:16,360 Speaker 1: case in the problem of like Hubris and the fact 263 00:13:16,400 --> 00:13:19,319 Speaker 1: that maybe a guy who's really good at gambling shouldn't 264 00:13:19,480 --> 00:13:23,520 Speaker 1: run an entire media enterprise. But Nate doesn't like thinking 265 00:13:23,559 --> 00:13:25,320 Speaker 1: about that. It isn't like thinking about the fact that 266 00:13:25,400 --> 00:13:28,240 Speaker 1: maybe the only thing Elon Musk was ever good at 267 00:13:28,559 --> 00:13:31,600 Speaker 1: was being the guy from the Music Man, because I 268 00:13:31,640 --> 00:13:34,680 Speaker 1: think Nate bought into Elon Musk for a significant period 269 00:13:34,679 --> 00:13:39,640 Speaker 1: of time, Right, many people clearly does. Yes, Yeah, there's 270 00:13:39,640 --> 00:13:41,680 Speaker 1: been this thing lately where a lot of folks on 271 00:13:41,720 --> 00:13:44,360 Speaker 1: the left have been like the Oh, you couldn't always 272 00:13:44,400 --> 00:13:46,600 Speaker 1: tell that he was a con man, You couldn't always 273 00:13:46,600 --> 00:13:48,720 Speaker 1: tell that he was this bad, like he was always 274 00:13:48,760 --> 00:13:51,600 Speaker 1: the worst. I was like, no, Like back in twenty 275 00:13:51,720 --> 00:13:54,679 Speaker 1: fourteen fifteen, when I was writing about the billionaires and 276 00:13:54,840 --> 00:13:57,440 Speaker 1: rich people that were evil, I was focusing on Jamie 277 00:13:57,520 --> 00:13:59,679 Speaker 1: Diamond because he had helped create the two thousand and 278 00:13:59,720 --> 00:14:02,120 Speaker 1: eight five financial collapse and he's seen it. He just 279 00:14:02,120 --> 00:14:04,560 Speaker 1: seemed obviously much worse than this guy who up to 280 00:14:04,600 --> 00:14:08,040 Speaker 1: that point was pretty much just making cars and rockets. 281 00:14:08,240 --> 00:14:11,240 Speaker 1: You know, you have two companies doing that. Musk was 282 00:14:11,280 --> 00:14:14,559 Speaker 1: not top of most people's radars for very good reason, 283 00:14:15,120 --> 00:14:17,840 Speaker 1: which gets to like, there's this thing that's been created 284 00:14:17,880 --> 00:14:20,080 Speaker 1: because of some of like the sinister beliefs that his 285 00:14:20,160 --> 00:14:23,800 Speaker 1: grandfather had and his like family background which has a 286 00:14:23,800 --> 00:14:26,880 Speaker 1: lot of white supremacy in it. To that this has 287 00:14:26,920 --> 00:14:29,920 Speaker 1: been Elon's sort of like grand plan from the beginning, 288 00:14:30,000 --> 00:14:32,680 Speaker 1: and that it's all come together for him, like as 289 00:14:32,720 --> 00:14:35,840 Speaker 1: if he's he's, you know, a Marvel or a James 290 00:14:35,840 --> 00:14:38,360 Speaker 1: bond Villa who's been executing this like thirty year plan 291 00:14:38,480 --> 00:14:39,320 Speaker 1: to get where he is. 292 00:14:39,680 --> 00:14:39,840 Speaker 5: Yeah. 293 00:14:39,920 --> 00:14:43,280 Speaker 1: Yeah, I think when you look at his cognition, like 294 00:14:43,480 --> 00:14:46,120 Speaker 1: he's not the same man he was ten years ago. 295 00:14:46,280 --> 00:14:48,200 Speaker 1: He's not the same guy he was when he started 296 00:14:48,320 --> 00:14:50,520 Speaker 1: dating Grimes. And I'm saying he was a good man 297 00:14:50,640 --> 00:14:54,640 Speaker 1: before then. I don't think he particularly ever was, but 298 00:14:54,760 --> 00:14:58,080 Speaker 1: he's clearly his brain has degraded, in part due to 299 00:14:58,160 --> 00:14:59,240 Speaker 1: contact through Twitter. 300 00:15:00,080 --> 00:15:02,760 Speaker 2: Yeah, and you can like measure this through his posting 301 00:15:03,000 --> 00:15:05,400 Speaker 2: as well, Like, yeah, like that the types of posts 302 00:15:05,400 --> 00:15:08,840 Speaker 2: you would make in twenty seventeen are like completely opposite 303 00:15:08,920 --> 00:15:12,080 Speaker 2: to the way that he would talk about certain social issues. Now, 304 00:15:12,120 --> 00:15:15,120 Speaker 2: Oh yeah, he's not like meming about like anarcho syndicalism. 305 00:15:15,400 --> 00:15:17,320 Speaker 1: Yeah, we get to a few of those things. But 306 00:15:17,320 --> 00:15:19,400 Speaker 1: I want to read another quote from NAT's article because 307 00:15:19,400 --> 00:15:21,760 Speaker 1: he's going to talk about his book on the Edge, 308 00:15:21,880 --> 00:15:25,160 Speaker 1: which quote describes a certain community of intelligent people that 309 00:15:25,240 --> 00:15:29,200 Speaker 1: I call the River. These people who occupy a range 310 00:15:29,200 --> 00:15:32,560 Speaker 1: of professions from AI research to poker to venture capital, 311 00:15:32,880 --> 00:15:37,080 Speaker 1: are bright, but in spiky ways. In Baron Cohen's dochotomy, 312 00:15:37,160 --> 00:15:40,240 Speaker 1: they lean heavily towards the systematic side of the equation. 313 00:15:40,520 --> 00:15:43,480 Speaker 1: They're good at abstract analytic reasoning, but they may lack 314 00:15:43,560 --> 00:15:47,120 Speaker 1: other forms of intelligence like empathy, judgment, and self awareness. 315 00:15:47,360 --> 00:15:51,160 Speaker 1: They also have some distinctive characteristics largely unrelated to intelligence. 316 00:15:51,200 --> 00:15:54,840 Speaker 1: For example, they tend to be extraordinarily competitive and somewhat contrariant. 317 00:15:54,880 --> 00:15:56,800 Speaker 1: And again, what you are talking about all of these 318 00:15:56,840 --> 00:15:59,600 Speaker 1: people number one, when he says AI research, he's not 319 00:16:00,000 --> 00:16:02,920 Speaker 1: talking about people who are doing like the gut level coating. 320 00:16:02,960 --> 00:16:08,320 Speaker 1: He's talking about Sam Altman, right, poker, venture capital. This 321 00:16:08,360 --> 00:16:12,440 Speaker 1: is all gambling. You're all talking about gamblers. The River 322 00:16:12,600 --> 00:16:16,880 Speaker 1: is just gamblers, Nate. It's people like you who who 323 00:16:16,920 --> 00:16:21,800 Speaker 1: put money on bets, and they are contrarian and competitive 324 00:16:22,040 --> 00:16:26,920 Speaker 1: because that's how gamblers are. That's the intelligence, that's the river. 325 00:16:27,360 --> 00:16:30,080 Speaker 1: Like he's thinking about it as like this specific chunk 326 00:16:30,360 --> 00:16:32,920 Speaker 1: of intellectuals who have You know, there's some dangers, but 327 00:16:32,960 --> 00:16:36,000 Speaker 1: they have great potential to make the world brilliant. You're like, no, no, no, no, 328 00:16:36,400 --> 00:16:39,120 Speaker 1: these are just people who like wind up shooting themselves 329 00:16:39,160 --> 00:16:43,880 Speaker 1: outside of a sports betting facility. Like that's the river, mate. 330 00:16:44,640 --> 00:16:47,360 Speaker 2: I have been turning into a monster during our friend 331 00:16:47,400 --> 00:16:51,000 Speaker 2: poker nights recently. It's tough garrison. 332 00:16:51,160 --> 00:16:52,440 Speaker 1: By the way, I've been meaning to talk to you 333 00:16:52,480 --> 00:16:55,040 Speaker 1: about wearing the full data makeup, because you know your 334 00:16:55,040 --> 00:16:57,280 Speaker 1: skin can't breathe. If you coat your whole body, You're 335 00:16:57,280 --> 00:16:58,720 Speaker 1: only supposed to put that on your face. 336 00:16:58,800 --> 00:17:01,520 Speaker 2: I don't do that every time I played Get. 337 00:17:01,360 --> 00:17:02,960 Speaker 1: A gold finger yourself care. 338 00:17:03,840 --> 00:17:06,119 Speaker 2: I don't put on the data makeup every time I 339 00:17:06,119 --> 00:17:10,320 Speaker 2: play poker, just that one time. Actually, no, I've done 340 00:17:10,359 --> 00:17:15,240 Speaker 2: that twice nowt never minds Okay, okay, he's becoming I 341 00:17:15,280 --> 00:17:17,920 Speaker 2: also have the little hats. I ordered a twelve pack 342 00:17:18,000 --> 00:17:20,439 Speaker 2: of like of like the little like poker visors to 343 00:17:20,480 --> 00:17:21,600 Speaker 2: complete the outfaced. 344 00:17:22,000 --> 00:17:22,800 Speaker 1: Of course she did. 345 00:17:22,920 --> 00:17:26,159 Speaker 3: Yeah, it would be rude, not too for better or worse. 346 00:17:27,160 --> 00:17:30,639 Speaker 1: This typology the river is associated with high achievement and 347 00:17:30,720 --> 00:17:34,440 Speaker 1: certain highly lucrative professions, especially tech and finance. It is 348 00:17:34,480 --> 00:17:38,560 Speaker 1: also associated with high variance bankman Free built FTX into 349 00:17:38,600 --> 00:17:41,200 Speaker 1: a company that investors valued at thirty two billion before 350 00:17:41,240 --> 00:17:44,040 Speaker 1: the House of cards collapsed again because he was a dambler. 351 00:17:44,320 --> 00:17:45,080 Speaker 3: He's a con man. 352 00:17:45,240 --> 00:17:48,240 Speaker 1: Yeah, and again Nate can't just accept oh, he was 353 00:17:48,280 --> 00:17:50,679 Speaker 1: never actually very smart. He just got really lucky for 354 00:17:50,720 --> 00:17:53,439 Speaker 1: a while and then gave it and then gambled it 355 00:17:53,440 --> 00:17:56,119 Speaker 1: all away because he wasn't actually as smart as anyone thought. 356 00:17:56,280 --> 00:17:59,280 Speaker 1: Nate says, I interviewed SBF several times for the book, 357 00:17:59,320 --> 00:18:01,040 Speaker 1: and I can tell you that he very much falls 358 00:18:01,040 --> 00:18:03,200 Speaker 1: into the genius but dumbass category. 359 00:18:03,800 --> 00:18:08,440 Speaker 3: How about just dumbassy, lucky dumbass. It's not hard. 360 00:18:08,640 --> 00:18:12,000 Speaker 1: What's the genius? Where did he prove that? 361 00:18:12,160 --> 00:18:14,760 Speaker 2: I mean he proved that by fooling Nate Silver, a 362 00:18:14,760 --> 00:18:18,720 Speaker 2: man who probably values his own intelligence like a great deal. 363 00:18:19,000 --> 00:18:20,840 Speaker 3: I mean, that's the whole thing, right, Nate Silver Khan. 364 00:18:21,480 --> 00:18:23,679 Speaker 3: It would be ego death to admit that there were 365 00:18:23,760 --> 00:18:25,119 Speaker 3: just some lucky, dumb white dudes. 366 00:18:25,480 --> 00:18:26,080 Speaker 4: Yeah. 367 00:18:26,119 --> 00:18:28,320 Speaker 1: If a guy had won like one of the lotteries, 368 00:18:28,560 --> 00:18:30,600 Speaker 1: was like a billion and a half dollars, right, got 369 00:18:30,640 --> 00:18:33,479 Speaker 1: crazy rich and then lost it all in two weeks 370 00:18:33,480 --> 00:18:36,359 Speaker 1: because he just kept putting half a million dollars at 371 00:18:36,400 --> 00:18:39,280 Speaker 1: a time on twenty one black at a roulette table 372 00:18:39,280 --> 00:18:42,560 Speaker 1: in Vegas, and I would be like, well, obviously he's 373 00:18:42,600 --> 00:18:45,440 Speaker 1: a genius, but he's also kind of a dumbass. How 374 00:18:45,480 --> 00:18:47,520 Speaker 1: else could he have made the money in the first place? 375 00:18:47,560 --> 00:18:49,479 Speaker 1: And I was like, no, he got lucky, and then 376 00:18:49,520 --> 00:18:52,359 Speaker 1: he gambled it all away because he's he doesn't have 377 00:18:52,400 --> 00:18:56,879 Speaker 1: good judgment. Yeah, so it's important to avoid two pitfalls 378 00:18:56,880 --> 00:19:00,359 Speaker 1: when encountering people with spiky intelligence, namely, either there are 379 00:19:00,359 --> 00:19:02,840 Speaker 1: worst traits nor their best ones tell the whole story. 380 00:19:03,160 --> 00:19:06,040 Speaker 1: And I don't disagree with that. However, it's a meaningless 381 00:19:06,040 --> 00:19:09,879 Speaker 1: statement because that's true of every human being ever born. Yeah, 382 00:19:10,000 --> 00:19:13,200 Speaker 1: but clearly Nate doesn't feel that way because only I think. 383 00:19:13,440 --> 00:19:16,199 Speaker 1: The undercurrent here is that only people like this in 384 00:19:16,280 --> 00:19:19,640 Speaker 1: Nate's mind are worth talking about, because only gamblers bring 385 00:19:19,680 --> 00:19:21,120 Speaker 1: the world forward, right. 386 00:19:21,080 --> 00:19:23,040 Speaker 3: Yeah, No one else deserves empathy. 387 00:19:23,680 --> 00:19:28,359 Speaker 1: Yeah, yes, like you're just addicted to putting money on 388 00:19:28,440 --> 00:19:32,760 Speaker 1: sports games and elections, Nate Silver. Anyway, So here's the 389 00:19:32,760 --> 00:19:34,960 Speaker 1: two things he wants to warn us up or wants 390 00:19:34,960 --> 00:19:38,320 Speaker 1: people to avoid. Elon is highly intelligent in several ways, 391 00:19:38,359 --> 00:19:40,800 Speaker 1: but that does not mean that everything he does is brilliant. 392 00:19:40,920 --> 00:19:43,480 Speaker 1: Some things he does are exceptionally dumb or dangerous, and 393 00:19:43,520 --> 00:19:46,960 Speaker 1: we shouldn't make excuses for them. But likewise, it's absurd 394 00:19:46,960 --> 00:19:49,439 Speaker 1: to suggest that Elon isn't brilliant in many respects just 395 00:19:49,480 --> 00:19:51,840 Speaker 1: because he isn't in others. And if he has merely 396 00:19:52,000 --> 00:19:55,000 Speaker 1: very good SAT scores, I don't care. Nobody does. It's 397 00:19:55,040 --> 00:19:58,200 Speaker 1: not high school. Nobody cares about his SAT. 398 00:19:57,920 --> 00:20:01,600 Speaker 2: Skills, like like like fifty, like fifty five or something 399 00:20:01,680 --> 00:20:02,360 Speaker 2: like what are we doing? 400 00:20:02,840 --> 00:20:06,120 Speaker 1: Yeah, yeah, yeah, yeah, you are a middle aged man. 401 00:20:06,240 --> 00:20:08,399 Speaker 1: I don't even know what my SAT score was. 402 00:20:09,119 --> 00:20:11,240 Speaker 3: I was gonna say, like, I never took an SAT, 403 00:20:11,359 --> 00:20:13,479 Speaker 3: but I spent more than a decade in full time education, 404 00:20:13,600 --> 00:20:16,440 Speaker 3: and anyone who ever told me that SAT schools I 405 00:20:16,440 --> 00:20:19,560 Speaker 3: immediately hated and never took them. Seriously. I've spent almost. 406 00:20:19,240 --> 00:20:21,840 Speaker 1: Twenty years asking people questions for a living, and I've 407 00:20:21,880 --> 00:20:24,280 Speaker 1: never asked anyone there SAT. Sorry, Garrison. 408 00:20:24,520 --> 00:20:26,840 Speaker 2: Although SAT might not be like a stable metric for 409 00:20:26,960 --> 00:20:30,840 Speaker 2: evaluating intelligence, surely Nate has an alternative method. 410 00:20:32,040 --> 00:20:35,199 Speaker 1: Absolutely not, Garrison, just how much well he doesn't he 411 00:20:35,280 --> 00:20:37,719 Speaker 1: doesn't have an alternative method seeing what you might call 412 00:20:37,800 --> 00:20:41,480 Speaker 1: an infographic, because the next section of the article is 413 00:20:41,520 --> 00:20:46,920 Speaker 1: a quick inventory of Elon's intelligence. So first he admits 414 00:20:47,119 --> 00:20:49,640 Speaker 1: he tried to track Elon down for his stupid book, 415 00:20:49,720 --> 00:20:52,119 Speaker 1: but he couldn't get him to talk to him. Because Elon, 416 00:20:52,359 --> 00:20:55,359 Speaker 1: even I have to say, Elon does have better shit 417 00:20:55,480 --> 00:20:59,679 Speaker 1: to do than talk to Nate Silver, because Elon is 418 00:21:00,160 --> 00:21:02,639 Speaker 1: using ketamine to a near fatal degree, and that is 419 00:21:02,720 --> 00:21:05,200 Speaker 1: a better use of his time than talking to Nate Silver. 420 00:21:06,160 --> 00:21:09,920 Speaker 1: So since he can't actually talk to two Musk, he's 421 00:21:09,960 --> 00:21:13,520 Speaker 1: going to model and extrapolate from quote many other Silicon 422 00:21:13,600 --> 00:21:16,359 Speaker 1: Valley big wigs I have met, okay helping him. And 423 00:21:16,359 --> 00:21:18,960 Speaker 1: this is the fact that quote Musk maintains an extremely 424 00:21:19,000 --> 00:21:22,199 Speaker 1: public profile. He's turned X into a running diary of 425 00:21:22,200 --> 00:21:25,600 Speaker 1: his innermost thoughts and in addition to that, the biographies 426 00:21:25,640 --> 00:21:27,800 Speaker 1: of the guy one more caveat. Here I will try 427 00:21:27,800 --> 00:21:31,239 Speaker 1: to evaluate the overall trajectory of Elon's career, not just 428 00:21:31,280 --> 00:21:34,480 Speaker 1: his recent antics. So we got down here, and the 429 00:21:34,520 --> 00:21:38,119 Speaker 1: next segment is dimensions were Musk has exceptionally high or 430 00:21:38,160 --> 00:21:41,440 Speaker 1: genius level intelligence. So finally Nate's going to prove it, 431 00:21:41,640 --> 00:21:44,480 Speaker 1: and I'm gonna I'm going to show you guys, how 432 00:21:44,520 --> 00:21:47,280 Speaker 1: he how he chooses to do that, what the evidence 433 00:21:47,280 --> 00:21:48,480 Speaker 1: he gives us here is. 434 00:21:48,520 --> 00:21:50,560 Speaker 2: And I think this is something that we should reveal 435 00:21:50,560 --> 00:21:52,919 Speaker 2: to the audience after these ads. 436 00:21:53,080 --> 00:22:05,720 Speaker 1: Good point care, all right, we're back. So let's look 437 00:22:05,720 --> 00:22:08,480 Speaker 1: at what Nate shows as is the chief dimension where 438 00:22:08,560 --> 00:22:13,560 Speaker 1: Musk has shown high or genius level intelligence. Reading that 439 00:22:13,600 --> 00:22:17,320 Speaker 1: first line, man, So the first words under this are 440 00:22:17,359 --> 00:22:21,320 Speaker 1: cognitive load capacity in overall horsepower slash ram. He's always on, 441 00:22:21,440 --> 00:22:24,679 Speaker 1: I mean literally look at how often he's tweeting, and 442 00:22:24,720 --> 00:22:28,680 Speaker 1: then a huge graph that shows the density of tweets 443 00:22:28,720 --> 00:22:31,600 Speaker 1: posted and win which has been used by other people 444 00:22:31,640 --> 00:22:34,480 Speaker 1: to prove that since sometime in late twenty twenty two, 445 00:22:34,520 --> 00:22:37,080 Speaker 1: he's almost never gone more than about three hours without 446 00:22:37,119 --> 00:22:41,720 Speaker 1: posting a tweet. Like it's just a solid red after 447 00:22:41,760 --> 00:22:44,080 Speaker 1: he buys the site. This like graph of like when 448 00:22:44,080 --> 00:22:47,879 Speaker 1: he makes his posts, he's never offline. Now he's not sleeping. 449 00:22:48,280 --> 00:22:50,399 Speaker 2: So this is a graph of Elon Musk's tweets from 450 00:22:50,400 --> 00:22:54,399 Speaker 2: twenty fourteen to twenty twenty four showing the time of 451 00:22:54,560 --> 00:22:57,639 Speaker 2: day and when a post is posted, represented by small 452 00:22:57,680 --> 00:23:02,440 Speaker 2: red dots and Yes, around twenty twenty two, the thickness 453 00:23:02,440 --> 00:23:04,879 Speaker 2: of the red increases dramatically. 454 00:23:05,160 --> 00:23:07,280 Speaker 1: It's almost just a straight REDPA like the. 455 00:23:07,240 --> 00:23:12,240 Speaker 2: Period of where he must be sleeping in this yeah. 456 00:23:11,280 --> 00:23:15,480 Speaker 1: Is very concerning. No, he sometimes sleeps from about six 457 00:23:15,600 --> 00:23:18,760 Speaker 1: to nine am, as far as we can tell, but 458 00:23:18,920 --> 00:23:22,040 Speaker 1: not regularly or often. It's like a streak at twenty 459 00:23:22,040 --> 00:23:24,880 Speaker 1: twenty three where he just isn't sleepy, he's not sleeping. 460 00:23:24,920 --> 00:23:30,399 Speaker 1: And again, he's on drugs. People, I think they're probably prescripted. 461 00:23:30,440 --> 00:23:33,320 Speaker 1: I think I'm certain he's on ketamine that has been prescribed. 462 00:23:33,359 --> 00:23:35,560 Speaker 1: When you're this rich, you just get whatever drugs you 463 00:23:35,600 --> 00:23:37,520 Speaker 1: want to do, recreationally prescribed. 464 00:23:37,720 --> 00:23:37,920 Speaker 3: Right. 465 00:23:38,480 --> 00:23:42,480 Speaker 1: But this is drug user behavior. I don't say that 466 00:23:42,720 --> 00:23:45,920 Speaker 1: to judge drug users. I say that as someone who 467 00:23:45,920 --> 00:23:50,159 Speaker 1: had a drug problem, like this is drug user behavior. 468 00:23:50,320 --> 00:23:54,679 Speaker 1: And specifically Silver he's using sobriety as possible. Sorry, And 469 00:23:54,720 --> 00:23:59,000 Speaker 1: specifically Silver is using this as an evidence of MUCKs intelligence. Yeah, 470 00:23:59,440 --> 00:23:59,760 Speaker 1: it's not. 471 00:24:00,000 --> 00:24:03,480 Speaker 2: He's scaling his Twitter activity as a sign that he 472 00:24:03,560 --> 00:24:05,439 Speaker 2: must be like a special type of person. 473 00:24:06,320 --> 00:24:11,600 Speaker 1: He's railing adderall and eating ketamine lozenges all day every day. 474 00:24:11,680 --> 00:24:13,480 Speaker 1: That's what this is a sign of. And no one 475 00:24:13,560 --> 00:24:16,960 Speaker 1: is allowed to take his phone away anyway. Here's how 476 00:24:17,080 --> 00:24:20,240 Speaker 1: Nate explains why this is smart. In NBA terms, we 477 00:24:20,400 --> 00:24:23,400 Speaker 1: say this as a player with an exceptionally high motor, 478 00:24:23,800 --> 00:24:26,400 Speaker 1: and this is undoubtedly a valuable trait as the world 479 00:24:26,480 --> 00:24:29,879 Speaker 1: becomes more complex. Lack's fall. I was simultaneously doing an 480 00:24:29,880 --> 00:24:32,840 Speaker 1: extensive book media tour, running the election model, trying to 481 00:24:32,840 --> 00:24:36,080 Speaker 1: build up Silver bulletin, plus some intensive consulting work. Even 482 00:24:36,080 --> 00:24:38,080 Speaker 1: if I mostly kept my wits about me, it was 483 00:24:38,119 --> 00:24:40,520 Speaker 1: an incredible amount of mental and physical strain that would 484 00:24:40,520 --> 00:24:43,359 Speaker 1: only have been sustainable for a short burst. But Elon 485 00:24:43,480 --> 00:24:45,960 Speaker 1: is taking on I don't know, approximately a thousand times 486 00:24:46,000 --> 00:24:48,119 Speaker 1: more stress than that, and has done so for years. No, 487 00:24:48,240 --> 00:24:52,920 Speaker 1: he's not, He just tweets he has a massive number one. 488 00:24:53,240 --> 00:24:56,000 Speaker 1: All of the businesses are being run by people who 489 00:24:56,080 --> 00:24:59,040 Speaker 1: are specialists in those businesses. He gets called on to 490 00:24:59,119 --> 00:25:01,359 Speaker 1: sit in meetings in and say yes or notice stuff, 491 00:25:01,359 --> 00:25:05,760 Speaker 1: and occasionally tells him to do something crazy that causes issues. Right, 492 00:25:06,160 --> 00:25:10,200 Speaker 1: and they're not running smoothly. Tesla's lost more value now 493 00:25:10,240 --> 00:25:13,919 Speaker 1: than it gained after the election, and SpaceX just had 494 00:25:13,960 --> 00:25:18,199 Speaker 1: a giant rocket explode again. The boring company has not 495 00:25:18,880 --> 00:25:22,080 Speaker 1: done anything other than make a useless hole underneath Vegas 496 00:25:22,080 --> 00:25:25,119 Speaker 1: and the hyper loop is nothing right like this. This 497 00:25:26,040 --> 00:25:31,199 Speaker 1: is just full of shit, Nate, Like what you have 498 00:25:31,359 --> 00:25:34,879 Speaker 1: just described, Running an election model that's functional, going on 499 00:25:34,920 --> 00:25:38,439 Speaker 1: a book tour and consulting and writing a newsletter is 500 00:25:38,520 --> 00:25:41,240 Speaker 1: more work than I credit Elon Musk with actually doing. 501 00:25:41,320 --> 00:25:43,160 Speaker 1: Oh yeah, more actual effort work. 502 00:25:43,480 --> 00:25:46,360 Speaker 2: Risk is mostly like sitting in an occasional meeting doing 503 00:25:46,440 --> 00:25:49,399 Speaker 2: drugs and injecting random women with his sperm yes, and 504 00:25:49,520 --> 00:25:50,320 Speaker 2: sending tweets. 505 00:25:50,359 --> 00:25:53,600 Speaker 3: He doesn't do the injecting. I think, oh, god, Garrison, that. 506 00:25:53,640 --> 00:25:58,440 Speaker 1: Comes up to no, no oh, and it's crazy how 507 00:25:58,440 --> 00:26:01,040 Speaker 1: it does. Right before he posts the graph of how 508 00:26:01,119 --> 00:26:02,200 Speaker 1: much Elon tweets. 509 00:26:02,400 --> 00:26:05,720 Speaker 3: Oh, okay, God, there it is, okay, okay. 510 00:26:05,880 --> 00:26:08,600 Speaker 1: Politics and social media poison a lot of people's brains. 511 00:26:08,680 --> 00:26:11,480 Speaker 1: Having that much wealth and power has to be intoxicating, 512 00:26:11,600 --> 00:26:15,160 Speaker 1: especially if Muska ostracizes people who might keep him grounded 513 00:26:15,359 --> 00:26:18,720 Speaker 1: more sympathetically. He's taking on an incredible array of responsibilities, 514 00:26:18,880 --> 00:26:21,480 Speaker 1: doing several really hard jobs at once, each of which 515 00:26:21,480 --> 00:26:23,560 Speaker 1: would be stressful on their own, while still managing to 516 00:26:23,640 --> 00:26:27,600 Speaker 1: father thirteen children. I'm tweeting hundreds of times per week. Again, 517 00:26:27,960 --> 00:26:32,640 Speaker 1: equivalent efforts tweeting hundreds of times a week and fathering 518 00:26:32,720 --> 00:26:36,240 Speaker 1: thirteen children. He's not a father to them. No, he 519 00:26:36,400 --> 00:26:39,960 Speaker 1: just he contributed by it. He didn't even have sex. 520 00:26:40,080 --> 00:26:43,560 Speaker 3: Yeah, yeah, he's usually the lowest possible effort way to 521 00:26:43,640 --> 00:26:44,280 Speaker 3: have a child. 522 00:26:45,240 --> 00:26:48,359 Speaker 1: Wait, Like, I'm gonna guess most of the people with 523 00:26:48,440 --> 00:26:52,159 Speaker 1: penises listening to this come like, that's not a big effort. 524 00:26:52,280 --> 00:26:54,080 Speaker 1: You wouldn't include that. It's like, what did I get 525 00:26:54,080 --> 00:26:57,160 Speaker 1: done this week? Well, in addition to working forty hours, 526 00:26:57,200 --> 00:26:57,920 Speaker 1: I jacked off. 527 00:26:57,960 --> 00:27:04,399 Speaker 2: That's a little transphobic. This is this is a joke. Anyway, 528 00:27:04,480 --> 00:27:05,400 Speaker 2: continue it said. 529 00:27:06,640 --> 00:27:10,639 Speaker 1: I'm just saying it doesn't count as work. No, not 530 00:27:10,720 --> 00:27:13,720 Speaker 1: from us, and less you're a sex worker than it does. Okay, 531 00:27:13,920 --> 00:27:16,080 Speaker 1: Like especially I know a lot of male porn stars. 532 00:27:17,520 --> 00:27:19,520 Speaker 1: That's that is a difficult part of the job. That's 533 00:27:19,520 --> 00:27:22,560 Speaker 1: why they inject their penises directly with erection drugs that 534 00:27:22,640 --> 00:27:23,359 Speaker 1: kill their hearts. 535 00:27:24,000 --> 00:27:26,639 Speaker 2: I would like to get into more of Silver's like 536 00:27:27,160 --> 00:27:32,119 Speaker 2: justification for why why he associates this this high tweet 537 00:27:32,119 --> 00:27:33,600 Speaker 2: load with like intelligence. 538 00:27:34,440 --> 00:27:39,240 Speaker 1: Well, because it shows rapid cognition in thin slicing ability okay, 539 00:27:39,760 --> 00:27:44,280 Speaker 1: m hmm, right, yeah, sure sure. Indeed, in a capitalist 540 00:27:44,359 --> 00:27:46,960 Speaker 1: system with a significant premium, I'm being first to market. 541 00:27:47,040 --> 00:27:49,720 Speaker 1: Making decent judgments fast is often more important than making 542 00:27:49,760 --> 00:27:54,600 Speaker 1: better judgments slowly. Canonically, vcs imagine themselves rapidly filtering through 543 00:27:54,600 --> 00:27:57,880 Speaker 1: potential founders, as though on Shark Tank, relying on well 544 00:27:57,920 --> 00:28:00,760 Speaker 1: known gut instinct. But this also gets people in trouble, 545 00:28:00,760 --> 00:28:06,120 Speaker 1: as it has for Elon what is shark Tank's success rate? Yeah, 546 00:28:07,000 --> 00:28:08,639 Speaker 1: I bet there's a quick answer to that. 547 00:28:09,119 --> 00:28:12,480 Speaker 3: Yeah, and that's consuering. The air has built in free 548 00:28:12,520 --> 00:28:14,920 Speaker 3: television advertising for any product. 549 00:28:14,560 --> 00:28:21,000 Speaker 1: Means than fifty percent of deals they're successfully closed. God, yeah, 550 00:28:21,080 --> 00:28:25,119 Speaker 1: so I I it out. All This tweeting also shows 551 00:28:25,160 --> 00:28:28,639 Speaker 1: abstract problem solving capability. This is related to the idea 552 00:28:28,680 --> 00:28:32,879 Speaker 1: of creativity, though in Musk's case it seemingly doesn't manifest 553 00:28:32,920 --> 00:28:37,160 Speaker 1: itself an artistic prowessly ceeingly, you know what, I'll give 554 00:28:37,200 --> 00:28:37,560 Speaker 1: it to date. 555 00:28:37,640 --> 00:28:38,640 Speaker 2: There, I'll give it to day. 556 00:28:39,680 --> 00:28:42,160 Speaker 1: I don't disagree with you there. And then, of course, 557 00:28:42,320 --> 00:28:46,320 Speaker 1: instrumental rationality philosophy nerds like to distinguish between two types 558 00:28:46,320 --> 00:28:50,600 Speaker 1: of rationality. Instrumental rationality is aligning means with ens basically 559 00:28:50,600 --> 00:28:52,960 Speaker 1: figuring out the most efficient ways to get what you want. 560 00:28:53,280 --> 00:28:55,000 Speaker 1: For this category, I think you have to point towards 561 00:28:55,040 --> 00:28:58,280 Speaker 1: the scoreboard, must have some unparalleled accomplishments, and isn't about 562 00:28:58,280 --> 00:29:00,800 Speaker 1: to let anybody stand in his way. It's also a 563 00:29:00,840 --> 00:29:04,280 Speaker 1: category often associated with manipulativeness or even being an asshole, 564 00:29:04,360 --> 00:29:07,840 Speaker 1: not one for nice guys. Now, and again, if Musk's 565 00:29:07,840 --> 00:29:11,400 Speaker 1: actual goal is his stated goal getting to Mars, then 566 00:29:12,000 --> 00:29:15,440 Speaker 1: backing the political party that is actively doing as much 567 00:29:15,520 --> 00:29:18,560 Speaker 1: damage to the biosphere as possible, ensuring that it will 568 00:29:18,600 --> 00:29:21,560 Speaker 1: not have the carrying capacity necessary to make any kind 569 00:29:21,920 --> 00:29:25,720 Speaker 1: of off world civilization likely, I would argue is a 570 00:29:25,760 --> 00:29:28,479 Speaker 1: stupid decision. But he doesn't actually want us to get 571 00:29:28,560 --> 00:29:31,120 Speaker 1: to Mars, right, He just wants to be in charge 572 00:29:31,160 --> 00:29:31,760 Speaker 1: of everything. 573 00:29:32,040 --> 00:29:34,720 Speaker 2: No, he wants to run his businesses with no government interference. 574 00:29:34,760 --> 00:29:35,480 Speaker 2: That's really it. 575 00:29:35,560 --> 00:29:38,640 Speaker 1: Yes it is, yes, yes, yes, and he has been 576 00:29:38,760 --> 00:29:43,320 Speaker 1: very successful at that. But again it's the successive brute force. 577 00:29:43,400 --> 00:29:45,520 Speaker 1: It's the same way as like, if you hire a 578 00:29:45,560 --> 00:29:47,920 Speaker 1: thousand people who are willing to like break the kneecaps 579 00:29:47,960 --> 00:29:50,240 Speaker 1: of a guy who annoys you, like you could say, 580 00:29:50,320 --> 00:29:52,600 Speaker 1: like I'm very smart when it comes to hurting people 581 00:29:52,600 --> 00:29:54,440 Speaker 1: who annoy me. But really, you just have a lot 582 00:29:54,440 --> 00:29:56,680 Speaker 1: of dudes who can beat people up for you. Like, 583 00:29:57,240 --> 00:29:59,480 Speaker 1: is that intelligence or did you just have enough money 584 00:29:59,480 --> 00:29:59,920 Speaker 1: to hire? 585 00:30:00,520 --> 00:30:02,160 Speaker 2: Or are you just a mob boss? Right? 586 00:30:02,200 --> 00:30:05,560 Speaker 1: Are you just a mob boss? And a mob boss? 587 00:30:05,680 --> 00:30:07,760 Speaker 1: No one is allowed to attack because it's going to 588 00:30:07,760 --> 00:30:10,880 Speaker 1: be domestic terror to fuck up a Tesla store soon, 589 00:30:11,040 --> 00:30:14,040 Speaker 1: you know. Anyway, we need ghost dog. 590 00:30:14,320 --> 00:30:16,320 Speaker 2: It's pretty it's pretty upsetting because you know, a few 591 00:30:16,320 --> 00:30:17,880 Speaker 2: weeks ago, I was having a little bit of a 592 00:30:18,120 --> 00:30:21,400 Speaker 2: resist live moment and I actually ashed my clothes cigarette 593 00:30:21,880 --> 00:30:25,360 Speaker 2: on a parked Tesla. Felt pretty cool about it. But 594 00:30:25,480 --> 00:30:28,400 Speaker 2: now I guess I can't even do that. It's too dangerous. 595 00:30:28,440 --> 00:30:29,560 Speaker 1: Now, No, you can't. 596 00:30:29,560 --> 00:30:31,280 Speaker 2: I could face substantial charges. 597 00:30:31,760 --> 00:30:34,320 Speaker 3: You might want to text resist to a certain five 598 00:30:34,360 --> 00:30:36,600 Speaker 3: digit number or something. That's probably the best way to 599 00:30:36,640 --> 00:30:37,560 Speaker 3: solve this garrison. 600 00:30:37,880 --> 00:30:40,280 Speaker 1: I just text resist to every single person in my 601 00:30:40,320 --> 00:30:43,560 Speaker 1: phone book every day. I mean, it takes about seven hours. 602 00:30:43,640 --> 00:30:45,400 Speaker 1: I have fallen behind on work. 603 00:30:46,320 --> 00:30:48,320 Speaker 3: You know, it's the only thing we can do to 604 00:30:48,360 --> 00:30:48,960 Speaker 3: my pastus. 605 00:30:49,680 --> 00:30:53,120 Speaker 2: The quickest path to intelligence is having a horrible sleep 606 00:30:53,160 --> 00:30:57,040 Speaker 2: deprivation and drug problem. Apparently, or at least that is 607 00:30:57,080 --> 00:30:59,120 Speaker 2: how you show for it. It's funny because I saw 608 00:30:59,480 --> 00:31:03,880 Speaker 2: Allan Johnson, the billionaire who who's eating his son's blood 609 00:31:03,960 --> 00:31:06,280 Speaker 2: or now plasma. Oh we had a dead guy posted 610 00:31:06,280 --> 00:31:08,960 Speaker 2: his only self study on like the damaging effects of 611 00:31:08,960 --> 00:31:12,080 Speaker 2: sleep deprivation, and I'm pretty sure Musk like retweeted it 612 00:31:12,120 --> 00:31:15,720 Speaker 2: with like with like an emoji or something like yeah, dude, dude, 613 00:31:16,160 --> 00:31:18,800 Speaker 2: your brain is completely sure. 614 00:31:19,080 --> 00:31:21,560 Speaker 1: Now you you are, you are fried. You are the 615 00:31:21,600 --> 00:31:23,200 Speaker 1: most cooked a man has ever been. 616 00:31:23,360 --> 00:31:26,720 Speaker 3: It's an interesting study, Like there there is legitimately interesting 617 00:31:26,760 --> 00:31:28,480 Speaker 3: things to look at Ela Musk's brain. 618 00:31:28,840 --> 00:31:31,880 Speaker 1: Well, yes, and there's a lot of actual scientific data 619 00:31:31,960 --> 00:31:36,600 Speaker 1: put together like exhaustively by researchers studying how not just 620 00:31:36,640 --> 00:31:40,480 Speaker 1: sleep deprivation but like wealth and power impact the brain. 621 00:31:41,040 --> 00:31:43,400 Speaker 1: And like all of it makes a strong case that 622 00:31:43,480 --> 00:31:46,400 Speaker 1: Elon Musk at this point has done more damage to 623 00:31:46,440 --> 00:31:48,960 Speaker 1: his brain than like a career one of those career 624 00:31:49,080 --> 00:31:51,920 Speaker 1: WWE wrestlers who like kills their whole family and then 625 00:31:51,920 --> 00:31:54,360 Speaker 1: shoots themselves in the chest so someone can study their 626 00:31:54,360 --> 00:31:54,960 Speaker 1: brain later. 627 00:31:55,520 --> 00:32:00,880 Speaker 2: Yeah, I mean, well before before close, I do want 628 00:32:00,920 --> 00:32:05,720 Speaker 2: to say, before any psychologists or sociologists or like linguists 629 00:32:05,760 --> 00:32:08,000 Speaker 2: get mad at me. Yes, I know Boba and kiki 630 00:32:08,640 --> 00:32:13,280 Speaker 2: is is that is this is a shape language like 631 00:32:13,720 --> 00:32:18,280 Speaker 2: correlation test I myself as well as Nate here, I 632 00:32:18,320 --> 00:32:20,560 Speaker 2: have kind of expanded it's it's it's usage to like 633 00:32:20,800 --> 00:32:24,640 Speaker 2: projecting even more like human or like like emotional qualities 634 00:32:24,680 --> 00:32:28,400 Speaker 2: onto these shapes or onto these specific words. So please, sociologists, 635 00:32:28,480 --> 00:32:30,920 Speaker 2: leave me alone, do not do not do that message 636 00:32:30,920 --> 00:32:41,280 Speaker 2: me your favorite French. I'm I'm afraid it's already too late. 637 00:32:41,320 --> 00:32:44,240 Speaker 2: I think I already hear like twelve different reddators typing. 638 00:32:45,080 --> 00:32:47,760 Speaker 2: But yes, I think Nate's just using that image there 639 00:32:47,760 --> 00:32:50,960 Speaker 2: as like a metaphor to like show how, you know, 640 00:32:51,120 --> 00:32:57,440 Speaker 2: aggressive or manipulative Musk's own intelligence is as symbolized by 641 00:32:57,760 --> 00:33:00,440 Speaker 2: by a kiki as opposed to you know, maybe maybe 642 00:33:00,480 --> 00:33:02,240 Speaker 2: like a Bill Gates, which might be more like a 643 00:33:02,280 --> 00:33:06,280 Speaker 2: boba intelligence type. Okay, a little softer, a little bit 644 00:33:06,280 --> 00:33:07,160 Speaker 2: more philanthropy. 645 00:33:07,320 --> 00:33:11,240 Speaker 1: You know, I just got finished reading nothing but rationalist 646 00:33:11,320 --> 00:33:15,520 Speaker 1: and Zizian literature for two straight weeks, about a quarter 647 00:33:15,520 --> 00:33:18,200 Speaker 1: of a million words by my last count, Garrison, I 648 00:33:18,200 --> 00:33:20,440 Speaker 1: don't have an enemy to do this again. I'm going 649 00:33:20,480 --> 00:33:22,920 Speaker 1: to get back to my Hitler books. You know where 650 00:33:22,920 --> 00:33:26,440 Speaker 1: things make sense, where the world's safe. 651 00:33:28,440 --> 00:33:31,840 Speaker 3: Yeah, I'm returning to writing about the Syrian Civil War, 652 00:33:31,920 --> 00:33:34,600 Speaker 3: which is my comparative happy place. 653 00:33:34,880 --> 00:33:36,520 Speaker 1: Ah, the Syrians Civil War. 654 00:33:40,040 --> 00:33:41,960 Speaker 3: It's a really great world. I do wonder if he's 655 00:33:42,000 --> 00:33:45,040 Speaker 3: trying to avoid some kind of intellectual property thing by 656 00:33:45,160 --> 00:33:48,120 Speaker 3: using that little filter that he used over to Boo BERANKI. 657 00:33:47,880 --> 00:33:50,400 Speaker 1: No, because it would be if it's actually not fair 658 00:33:50,520 --> 00:33:53,080 Speaker 1: use now as opposed to if you just mentioned that 659 00:33:53,160 --> 00:33:55,680 Speaker 1: thing he doesn't yah because he doesn't talk about them, 660 00:33:55,760 --> 00:33:57,880 Speaker 1: then it is fair use, right, and he could use 661 00:33:57,920 --> 00:34:00,400 Speaker 1: like a little clip of it is and to illustrate. 662 00:34:00,160 --> 00:34:02,360 Speaker 3: Point, yeah, like I did with Manu Chow. 663 00:34:02,520 --> 00:34:05,200 Speaker 1: Anyway, this is all I want to say again about 664 00:34:05,280 --> 00:34:07,760 Speaker 1: Nate silver until twenty twenty eight. And if you know 665 00:34:07,800 --> 00:34:10,839 Speaker 1: what the upside if democracy really does die is, we'll 666 00:34:10,880 --> 00:34:12,680 Speaker 1: never have to talk about him again. 667 00:34:14,680 --> 00:34:17,719 Speaker 2: If Trump and Musk really take over fully and do 668 00:34:17,800 --> 00:34:20,240 Speaker 2: a full coup, we never have to talk about Nate Silvan. 669 00:34:20,400 --> 00:34:24,520 Speaker 1: Nine minutes from now, I'm wearing a Curtis Yarvin T shirt. 670 00:34:25,760 --> 00:34:28,320 Speaker 3: They'll be doing a SOD numbers and he will still 671 00:34:28,360 --> 00:34:32,759 Speaker 3: be analyzing that data. They straight regime capture of Nate Silver. 672 00:34:33,000 --> 00:34:35,320 Speaker 1: Well, it doesn't seem possible that Trump could have gotten 673 00:34:35,320 --> 00:34:37,200 Speaker 1: one hundred and four percent of the vote. 674 00:34:37,040 --> 00:34:38,680 Speaker 3: But there's a spikey percentages. 675 00:34:39,280 --> 00:34:40,839 Speaker 1: There's a spikey percentages. 676 00:34:42,000 --> 00:34:44,760 Speaker 2: Why can't Nate Silver just like run like Trump's casino 677 00:34:44,880 --> 00:34:47,839 Speaker 2: or something? Right, this is just like just like put 678 00:34:47,920 --> 00:34:48,360 Speaker 2: him away. 679 00:34:48,920 --> 00:34:52,440 Speaker 1: I understand if Nate because Nate's rich, he doesn't need 680 00:34:52,520 --> 00:34:54,560 Speaker 1: to do the other stuff. And if he was like 681 00:34:54,880 --> 00:34:57,680 Speaker 1: just doing sports betting analysis forever, I'd be like, well, hey, 682 00:34:57,760 --> 00:35:00,399 Speaker 1: that's what he loves. Right. If I had Nate's so money, 683 00:35:00,400 --> 00:35:02,799 Speaker 1: I'd probably just write novels for the rest of my 684 00:35:02,840 --> 00:35:05,160 Speaker 1: life because that's what I like to do. I don't 685 00:35:05,280 --> 00:35:08,359 Speaker 1: understand why he keeps writing about politics. He's not good 686 00:35:08,400 --> 00:35:10,799 Speaker 1: at it, and he can't like it. He needs to 687 00:35:10,840 --> 00:35:11,520 Speaker 1: feel special. 688 00:35:11,560 --> 00:35:14,600 Speaker 3: He wants to feel like a special boy who knows 689 00:35:14,640 --> 00:35:16,160 Speaker 3: the answers that no one else does. 690 00:35:16,800 --> 00:35:19,719 Speaker 1: All right, Well, anyway, this is us making fun of 691 00:35:19,800 --> 00:35:21,960 Speaker 1: Nate Silver, so you don't well, you can still make 692 00:35:21,960 --> 00:35:23,520 Speaker 1: fun of it. You don't have to read him. We 693 00:35:23,600 --> 00:35:24,399 Speaker 1: did that for you. 694 00:35:24,480 --> 00:35:29,160 Speaker 5: Good Night, It could happen Here is a production of 695 00:35:29,200 --> 00:35:32,160 Speaker 5: cool Zone Media. For more podcasts from cool Zone Media. 696 00:35:32,440 --> 00:35:35,680 Speaker 5: Visit our website Coolzonemedia dot com, or check us out 697 00:35:35,760 --> 00:35:39,200 Speaker 5: on the iHeartRadio app, Apple Podcasts, or wherever you listen 698 00:35:39,239 --> 00:35:41,879 Speaker 5: to podcasts. You can now find sources for it could 699 00:35:41,880 --> 00:35:45,280 Speaker 5: happen here, listed directly in episode descriptions. Thanks for listening.