1 00:00:00,280 --> 00:00:02,840 Speaker 1: Brought to you by the reinvented two thousand twelve camera. 2 00:00:03,160 --> 00:00:07,600 Speaker 1: It's ready. Are you welcome to Stuff you Should Know 3 00:00:08,200 --> 00:00:11,960 Speaker 1: from house Stuff Works dot Com. Hey, and welcome to 4 00:00:12,000 --> 00:00:15,080 Speaker 1: the podcast. I'm Josh Clark and guess who's with me? 5 00:00:15,480 --> 00:00:19,160 Speaker 1: That would be Mr Charles W. Chuck Bryant, who, based 6 00:00:19,200 --> 00:00:23,680 Speaker 1: on his headwear Today has Headwear Choice Today, apparently has 7 00:00:23,760 --> 00:00:26,920 Speaker 1: joined the Cuban Revolution, Right, is that correct? US Friday 8 00:00:26,960 --> 00:00:29,680 Speaker 1: Hat Day? So I'm doing my best Fidel Castro. Yeah, 9 00:00:30,640 --> 00:00:32,680 Speaker 1: va shay. It's called a combat cat. Feel like that. 10 00:00:32,920 --> 00:00:35,560 Speaker 1: It is. It's very cool. It's very cool, Chuck. You 11 00:00:35,640 --> 00:00:38,920 Speaker 1: want to know something cooler? Yes, Okay, So actually I 12 00:00:38,920 --> 00:00:41,199 Speaker 1: don't know if cools your right word. Maybe horrific is 13 00:00:41,240 --> 00:00:44,959 Speaker 1: a better word. Okay. Uh. There's a study conducted here 14 00:00:44,960 --> 00:00:47,280 Speaker 1: in the States. And of course you know that the 15 00:00:47,360 --> 00:00:50,000 Speaker 1: United States, like most other countries, have a long history 16 00:00:50,040 --> 00:00:54,840 Speaker 1: of you know, um, well meaning but really misplaced medical 17 00:00:54,920 --> 00:00:59,720 Speaker 1: experiments or psychological experiments by giving LSD to unsuspecting a 18 00:00:59,800 --> 00:01:02,160 Speaker 1: mayor ends, which we've talked about right exactly. This one 19 00:01:02,280 --> 00:01:07,040 Speaker 1: was a little different, um. This one involved separating twins 20 00:01:07,800 --> 00:01:10,559 Speaker 1: who were up for adoption at birth in the state 21 00:01:10,600 --> 00:01:13,800 Speaker 1: of New York. And uh, there were I think thirteen 22 00:01:13,840 --> 00:01:17,560 Speaker 1: sets of twins in one set of triplets, and they 23 00:01:17,560 --> 00:01:20,920 Speaker 1: were all separated through this one adoption agency as part 24 00:01:21,000 --> 00:01:25,319 Speaker 1: of a study of nature versus nurture. So like, the 25 00:01:25,360 --> 00:01:28,120 Speaker 1: only thing the adopting parents knew was that their kid 26 00:01:28,200 --> 00:01:33,800 Speaker 1: was part of an ongoing um child psychology study. And 27 00:01:33,840 --> 00:01:37,040 Speaker 1: so these researchers were allowed access to these kids over 28 00:01:37,080 --> 00:01:40,800 Speaker 1: their their lifetimes. And then it went from the sixties 29 00:01:40,840 --> 00:01:44,040 Speaker 1: to I think night and um, the guy who was 30 00:01:44,120 --> 00:01:47,000 Speaker 1: running the show, his name was Peter Newbauer, right, he 31 00:01:47,040 --> 00:01:53,800 Speaker 1: was a child psychologist. He um apparently realized that if 32 00:01:53,840 --> 00:01:57,840 Speaker 1: he were to publish this study. Uh, basically he'd beat 33 00:01:57,920 --> 00:02:01,920 Speaker 1: lynched right right that by the time rolled around, people 34 00:02:01,960 --> 00:02:04,680 Speaker 1: didn't think too highly of separating twins, Like the the 35 00:02:04,720 --> 00:02:09,240 Speaker 1: ethics of experimentation had changed enough, not based on the results, 36 00:02:09,240 --> 00:02:10,560 Speaker 1: just space on the fact that he did this to 37 00:02:10,600 --> 00:02:13,640 Speaker 1: because he under yeah right, so Um, basically what he 38 00:02:13,720 --> 00:02:15,920 Speaker 1: did was take all of the research. He had the 39 00:02:15,919 --> 00:02:18,720 Speaker 1: study it was ready to be published, and he sealed 40 00:02:18,760 --> 00:02:21,920 Speaker 1: it and um it cannot be open until twenty six 41 00:02:22,800 --> 00:02:25,680 Speaker 1: and it's sitting in the archives at Yale university. I 42 00:02:25,680 --> 00:02:28,840 Speaker 1: imagine twenty six. He imagined he'd be long dead by then, Right, 43 00:02:28,960 --> 00:02:31,160 Speaker 1: so in six we're gonna find out a lot about 44 00:02:31,240 --> 00:02:33,880 Speaker 1: nature versus nurture. I will be long dead, but you might. 45 00:02:34,200 --> 00:02:36,920 Speaker 1: I'm supposed to make it, as you know. That's what 46 00:02:37,000 --> 00:02:39,639 Speaker 1: your death clock. Death clock says. So I don't think 47 00:02:39,680 --> 00:02:43,680 Speaker 1: so I'll be. I've got my vagas. Odds are against that. Yeah, 48 00:02:43,760 --> 00:02:46,680 Speaker 1: so um okay, So Chuck, that's a that's an example 49 00:02:46,720 --> 00:02:51,359 Speaker 1: of a really terrible experiment. Yeah. Right, have you heard 50 00:02:51,360 --> 00:02:54,720 Speaker 1: of savants, autistic savants? I have, Indeed you have. Okay, 51 00:02:55,040 --> 00:03:00,000 Speaker 1: they actually provide a much less horrible, uh natural experiment, 52 00:03:00,240 --> 00:03:04,720 Speaker 1: perfect natural experiment to study the brain. Right, Okay, you 53 00:03:04,720 --> 00:03:06,600 Speaker 1: want to talk about savants for a second, because I'm 54 00:03:06,600 --> 00:03:10,880 Speaker 1: going to explain later how they make this perfect experiment. Sure, Josh, 55 00:03:10,919 --> 00:03:15,120 Speaker 1: you know, autistic savants are people who are mentally deficient 56 00:03:15,120 --> 00:03:18,240 Speaker 1: in some areas but excel and others, like a lot 57 00:03:18,320 --> 00:03:20,480 Speaker 1: of times and others. That kid that plays the piano, 58 00:03:20,680 --> 00:03:22,639 Speaker 1: you've seen him in the jazz trio, I have not. 59 00:03:23,000 --> 00:03:26,480 Speaker 1: He's fifteen or so now, and when he first started playing, 60 00:03:26,480 --> 00:03:29,960 Speaker 1: he was really young and very advanced musically. So and 61 00:03:30,000 --> 00:03:32,880 Speaker 1: he's autistic savant. So that's one good example. Yeah, music 62 00:03:32,919 --> 00:03:35,960 Speaker 1: comes out a lot in savants. Um. Um. There's a 63 00:03:35,960 --> 00:03:39,160 Speaker 1: guy named Blind Tom. He was this African American guy 64 00:03:39,160 --> 00:03:42,360 Speaker 1: at the turn of the last century, not hippie rob 65 00:03:42,520 --> 00:03:48,000 Speaker 1: or not Blind Tom, and he was severely um autistic, 66 00:03:48,520 --> 00:03:51,080 Speaker 1: and he he could play pretty much any piece of 67 00:03:51,160 --> 00:03:55,560 Speaker 1: music that he heard once on the piano. Um. Interesting, Well, 68 00:03:55,560 --> 00:03:59,440 Speaker 1: it's autistic. Savant is different than autism though, aren't those two? Sure? 69 00:03:59,600 --> 00:04:02,760 Speaker 1: Not every buddy who is a savant is autistic and 70 00:04:02,800 --> 00:04:06,240 Speaker 1: not everybody who's autistic is a savant, so there is like, yes, 71 00:04:06,360 --> 00:04:09,200 Speaker 1: that's a good point there. There is a very I 72 00:04:09,240 --> 00:04:14,520 Speaker 1: guess a subgroup called autistic savants, and perhaps the most 73 00:04:14,520 --> 00:04:21,720 Speaker 1: famous savant um is a guy named Kim Peak Rainman. Yeah, yeah, 74 00:04:21,720 --> 00:04:24,560 Speaker 1: he's the real Rainman, is what they call him. I 75 00:04:24,600 --> 00:04:26,520 Speaker 1: thought they just made that up. No, now you were 76 00:04:26,520 --> 00:04:31,040 Speaker 1: dead on, Chuck. You have an amazing intuition savant. Yeah 77 00:04:31,120 --> 00:04:34,479 Speaker 1: maybe I don't think so, but maybe I'm terrible at math. 78 00:04:34,520 --> 00:04:37,560 Speaker 1: So yeah, that comes into play too as well. Um. 79 00:04:37,560 --> 00:04:39,960 Speaker 1: But Kim Peak is this guy who the guy who 80 00:04:39,960 --> 00:04:44,800 Speaker 1: wrote Rainman, Barrymorrow met uh in four and in Night 81 00:04:45,120 --> 00:04:47,080 Speaker 1: the movie came out. So he was very much based 82 00:04:47,120 --> 00:04:50,960 Speaker 1: on Kim Peak. Yeah, the guy can if you tell 83 00:04:51,040 --> 00:04:54,040 Speaker 1: him your birthday, your birthdate, I'll tell you what day 84 00:04:54,040 --> 00:04:56,760 Speaker 1: of the week you were born on. Um. He apparently 85 00:04:57,080 --> 00:05:00,840 Speaker 1: Um has read twelve thousand books around that, around that 86 00:05:00,880 --> 00:05:06,520 Speaker 1: he started reading and memorizing things at fourteen months. Wow. Um. 87 00:05:06,600 --> 00:05:10,880 Speaker 1: But he has severe brain damage, developmental brain damage. So 88 00:05:10,960 --> 00:05:13,640 Speaker 1: he can't like, button his own shirt. He can't care 89 00:05:13,680 --> 00:05:15,960 Speaker 1: for himself. Luckily, he's got a really good dad who 90 00:05:16,000 --> 00:05:18,360 Speaker 1: cares for him. Sure. But the cool thing about this 91 00:05:18,400 --> 00:05:22,800 Speaker 1: story is after Barrymorrow won an oscar, he gave it 92 00:05:22,839 --> 00:05:26,320 Speaker 1: to Kim Peake, did he write the screenplay? And so 93 00:05:26,600 --> 00:05:29,599 Speaker 1: Kim Peak carries it around everywhere he goes. Yeah, and 94 00:05:29,600 --> 00:05:31,880 Speaker 1: that cool. Yeah, that's awesome. Alright. So the reason why 95 00:05:32,160 --> 00:05:35,560 Speaker 1: savants and there have been some really spectacular ones throughout 96 00:05:35,560 --> 00:05:39,640 Speaker 1: the ages, UH provides such a great natural experiment for 97 00:05:39,720 --> 00:05:43,880 Speaker 1: us to investigate the brain is because they, most of them, 98 00:05:44,120 --> 00:05:47,760 Speaker 1: they almost exclusively have left UH damage to the left 99 00:05:47,800 --> 00:05:50,920 Speaker 1: hemisphere of the brain. And you know, just the very 100 00:05:50,960 --> 00:05:53,240 Speaker 1: fact that they can excel in math but can't button 101 00:05:53,279 --> 00:05:57,839 Speaker 1: their own shirts. It provides this kind of um, certain 102 00:05:57,960 --> 00:06:00,960 Speaker 1: framework to compare the rest of our brains to right. 103 00:06:01,080 --> 00:06:03,840 Speaker 1: You know, it's it's a it's an excellent comparison, right right, 104 00:06:03,880 --> 00:06:06,000 Speaker 1: And the left side is more about detail, correct, and 105 00:06:06,040 --> 00:06:08,040 Speaker 1: the right side is more about the big picture. Do 106 00:06:08,040 --> 00:06:11,359 Speaker 1: you love the lateralization of brain functions? Do? Well? I 107 00:06:11,440 --> 00:06:14,479 Speaker 1: like the brain period because it's still so mysterious. You know, 108 00:06:15,040 --> 00:06:17,920 Speaker 1: it's amazing how little we know still about the brain. Yeah, 109 00:06:18,000 --> 00:06:20,599 Speaker 1: it's amazing and disconcerting. Yes, at the same time, I 110 00:06:20,640 --> 00:06:23,400 Speaker 1: predict the next fifty years are going to see tremendous 111 00:06:23,400 --> 00:06:26,640 Speaker 1: advances in our understanding of the brain, in part because 112 00:06:26,680 --> 00:06:28,800 Speaker 1: of the study of savants, right right. So yeah, you 113 00:06:28,839 --> 00:06:31,880 Speaker 1: were talking about the lateralization of brain function. Yeah, you're right. 114 00:06:32,120 --> 00:06:35,120 Speaker 1: Left is the detail oriented side and right sees the 115 00:06:35,120 --> 00:06:39,560 Speaker 1: big picture. Right. Um. And so there's some people who 116 00:06:39,600 --> 00:06:43,160 Speaker 1: are studying savants, um. And you know, like I said, 117 00:06:43,160 --> 00:06:45,039 Speaker 1: one of the reasons why they are interesting is because 118 00:06:45,040 --> 00:06:47,560 Speaker 1: almost all of them have damage in one form or 119 00:06:47,600 --> 00:06:50,520 Speaker 1: another to the left side of the brain. Uh. And 120 00:06:50,600 --> 00:06:53,640 Speaker 1: even more suspicious is is you can you can, um, 121 00:06:53,880 --> 00:06:55,680 Speaker 1: maybe get in a car wreck or have a stroke, 122 00:06:55,760 --> 00:06:58,279 Speaker 1: and if if you the left side is impaired people 123 00:06:58,320 --> 00:07:00,920 Speaker 1: have been known to basically come out of a savant 124 00:07:01,560 --> 00:07:05,680 Speaker 1: and sometimes autistic savant. Interesting, right, Um. So one of 125 00:07:05,680 --> 00:07:07,760 Speaker 1: the people that I'd like to talk about today who's 126 00:07:07,800 --> 00:07:11,600 Speaker 1: studying savants is Dr Allen Snyder, yes Schneider, as I 127 00:07:11,680 --> 00:07:14,280 Speaker 1: like to call him. He is an ex pad American 128 00:07:14,800 --> 00:07:17,080 Speaker 1: who runs the Center for the Mind that's the British 129 00:07:17,080 --> 00:07:21,520 Speaker 1: spelling of Center in Sydney, Australia. And he is he's 130 00:07:21,560 --> 00:07:24,800 Speaker 1: a very eccentric person. It sounds like it, he really is. 131 00:07:25,240 --> 00:07:29,280 Speaker 1: But he's been studying savants for years and um he 132 00:07:29,320 --> 00:07:32,600 Speaker 1: has come up with a theory about mindsets and it's 133 00:07:32,640 --> 00:07:37,120 Speaker 1: based on the lateralization of brain function. I love that. Um, yeah, 134 00:07:37,120 --> 00:07:41,320 Speaker 1: the mindset. Basically, his theory is that mindsets are created 135 00:07:41,360 --> 00:07:46,720 Speaker 1: their personal um basically definitions on your experience. Right. So 136 00:07:46,880 --> 00:07:49,760 Speaker 1: if you see, uh, you know, a bear in the woods, 137 00:07:49,840 --> 00:07:51,720 Speaker 1: well let's a little let's come. Let's say a dog 138 00:07:52,560 --> 00:07:55,240 Speaker 1: in your driveway. You'll note things about the dog, that 139 00:07:55,280 --> 00:07:56,840 Speaker 1: he's furry, that he has a tail, he walks on 140 00:07:56,920 --> 00:07:59,680 Speaker 1: four legs, that kind of thing, and your brain kind 141 00:07:59,680 --> 00:08:01,440 Speaker 1: of store us that away. So next time you don't 142 00:08:01,440 --> 00:08:04,160 Speaker 1: see a dog, you think Oh my gosh, what's that. 143 00:08:05,680 --> 00:08:08,400 Speaker 1: I've never seen one of those. Yeah, he calls the mindsets, right, 144 00:08:08,440 --> 00:08:12,640 Speaker 1: so chuck. Uh, when we when we're basically assaulted with 145 00:08:12,680 --> 00:08:16,440 Speaker 1: stimuli all times, all the time, raw data basically UM 146 00:08:16,440 --> 00:08:19,640 Speaker 1: from like the humming of a fluorescent light, UM to 147 00:08:19,960 --> 00:08:24,240 Speaker 1: you know, UM, conversations that we overhear in restaurants, that 148 00:08:24,320 --> 00:08:29,240 Speaker 1: kind of thing, colors, actions, taste, smells. Yeah, We're we're 149 00:08:29,360 --> 00:08:32,440 Speaker 1: constantly assaulted with the reinput. Right. We have this thing 150 00:08:32,520 --> 00:08:35,760 Speaker 1: called latent inhibition, which is a brain process. They're still 151 00:08:35,800 --> 00:08:38,280 Speaker 1: again trying to get a handle on, UM, but latent 152 00:08:38,320 --> 00:08:41,800 Speaker 1: inhibition is basically the process by which we filter out 153 00:08:41,840 --> 00:08:44,320 Speaker 1: stuff we already know, so if we can identify, So 154 00:08:44,360 --> 00:08:47,480 Speaker 1: we're not constantly focused on the buzzing of a fluorescent 155 00:08:47,559 --> 00:08:49,959 Speaker 1: lights exactly, we're hearing all the voices in a restaurant. 156 00:08:50,000 --> 00:08:53,160 Speaker 1: Obviously that'd be maddening, right And actually, as a side note, UM, 157 00:08:53,240 --> 00:08:59,000 Speaker 1: schizophrenics uh have very low latent inhibition indeed, so they're 158 00:08:59,120 --> 00:09:01,520 Speaker 1: constantly assault to with all of this stuff. But they 159 00:09:01,559 --> 00:09:06,320 Speaker 1: also have the added, UM horrible side effect of UM 160 00:09:06,400 --> 00:09:10,520 Speaker 1: attaching meaning to to these snippets of conversation. Right, So 161 00:09:10,679 --> 00:09:14,679 Speaker 1: saying specifically, you're hearing voices, um, and you're not able 162 00:09:14,720 --> 00:09:18,000 Speaker 1: to externalize or internalize, meaning you can't tell the voices 163 00:09:18,040 --> 00:09:21,679 Speaker 1: are coming from your head and you're attaching meaning to them. 164 00:09:21,720 --> 00:09:27,280 Speaker 1: That's schizophrenia, right, That's horrible. So it's Snyder's belief and 165 00:09:27,480 --> 00:09:31,440 Speaker 1: I'm pretty sure the medical establishment at large is that, Um, 166 00:09:31,679 --> 00:09:34,680 Speaker 1: we're we're getting all of this raw data. It's being 167 00:09:34,720 --> 00:09:38,920 Speaker 1: accepted into our right hemisphere, right, which sees the big picture, 168 00:09:39,480 --> 00:09:42,520 Speaker 1: and uh, it sends it over to the left hemisphere, 169 00:09:43,200 --> 00:09:46,360 Speaker 1: which processes it into details which we hang on to. 170 00:09:46,880 --> 00:09:49,520 Speaker 1: This this interplay between the left hemisphere and the right 171 00:09:49,559 --> 00:09:52,680 Speaker 1: hemisphere creates those mindsets you were talking about, right, which, 172 00:09:52,760 --> 00:09:55,319 Speaker 1: like you said, is how we can see a dog 173 00:09:56,120 --> 00:09:58,400 Speaker 1: and come to understand what a dog is, and then 174 00:09:58,720 --> 00:10:00,880 Speaker 1: later on when we see another agus just say, oh, 175 00:10:00,960 --> 00:10:06,200 Speaker 1: that's a dog. Like we kind of categorize things in packets, right. Sure. So, UM, 176 00:10:06,320 --> 00:10:10,920 Speaker 1: we say all that to say this, if supposedly we 177 00:10:11,240 --> 00:10:13,160 Speaker 1: have damage to the left side of the brain, the 178 00:10:13,160 --> 00:10:15,840 Speaker 1: detail oriented brain, all we're doing is getting raw data 179 00:10:16,080 --> 00:10:19,120 Speaker 1: and we're not able to create these mindsets. There's this 180 00:10:19,200 --> 00:10:23,079 Speaker 1: wonderful article by a guy named Lawrence Osburne UM, And 181 00:10:23,120 --> 00:10:24,559 Speaker 1: it was in the New York Times in two thousand 182 00:10:24,600 --> 00:10:27,000 Speaker 1: three called Savant for a Day and he spent the 183 00:10:27,080 --> 00:10:29,720 Speaker 1: day with Alan Snyder, and the whole article is very long, 184 00:10:29,760 --> 00:10:33,520 Speaker 1: but it's definitely worth reading. Um. He chronicles, you know, 185 00:10:33,960 --> 00:10:35,880 Speaker 1: his day with Alan Snyder. And one of the things 186 00:10:35,880 --> 00:10:38,559 Speaker 1: that Snyder mentions is that some of the savants that 187 00:10:38,640 --> 00:10:40,360 Speaker 1: he studies when they come to see him at the 188 00:10:40,360 --> 00:10:43,440 Speaker 1: Center for the Mind, they may have been there dozens 189 00:10:43,480 --> 00:10:46,840 Speaker 1: of times, but they can get lost every single time 190 00:10:46,960 --> 00:10:50,000 Speaker 1: just because of the change of shadows. It looks different. 191 00:10:50,040 --> 00:10:54,199 Speaker 1: They're getting different input, so they'll get lost because it 192 00:10:54,320 --> 00:10:56,600 Speaker 1: doesn't look the same way it did that last time, 193 00:10:56,640 --> 00:10:59,360 Speaker 1: and they can't form mindsets saying this is the direction 194 00:10:59,360 --> 00:11:02,720 Speaker 1: I'm going right. Right, So since people with left left 195 00:11:02,920 --> 00:11:07,080 Speaker 1: hemisphere brain damage uh tend to be savants, right or 196 00:11:07,120 --> 00:11:11,280 Speaker 1: people who are savants have that condition. Um. Snyder has 197 00:11:11,520 --> 00:11:15,560 Speaker 1: actually come up with a a theory that all of 198 00:11:15,679 --> 00:11:18,840 Speaker 1: us are savants. If you get struck on the head 199 00:11:18,880 --> 00:11:22,480 Speaker 1: and your left hemisphere is damaged, you you could become 200 00:11:22,520 --> 00:11:25,440 Speaker 1: a savant. So we're all potential savants, right, And basically 201 00:11:25,480 --> 00:11:28,599 Speaker 1: the left side that that helps create these mindsets that 202 00:11:28,640 --> 00:11:30,960 Speaker 1: pays attention to these details and hangs onto him are 203 00:11:31,360 --> 00:11:34,800 Speaker 1: keeping us from being savants. Right, So how do you 204 00:11:34,840 --> 00:11:39,079 Speaker 1: investigate something like this? Well, he uses a process called 205 00:11:39,120 --> 00:11:44,040 Speaker 1: transcranial magnetic stimulation. We're gonna call it TMS. Yes, much 206 00:11:44,080 --> 00:11:48,240 Speaker 1: easier for our purposes. Uh. TMS was originally designed Josh 207 00:11:48,280 --> 00:11:53,160 Speaker 1: to examine brain functions during cranial surgery. And what it 208 00:11:53,200 --> 00:11:57,439 Speaker 1: does is it focuses magnetic pulses to either suppress or 209 00:11:57,520 --> 00:12:00,280 Speaker 1: enhance the electrical functions of the brain. Yeah, it's depends 210 00:12:00,320 --> 00:12:03,520 Speaker 1: on the frequency of the pulses, right, absolutely. And we 211 00:12:03,520 --> 00:12:06,000 Speaker 1: were talking, you know, privately, and I thought it sounded 212 00:12:06,080 --> 00:12:10,880 Speaker 1: very relaxing as your brain was being massaged and disappointed. 213 00:12:11,240 --> 00:12:13,560 Speaker 1: I was disappointed because you said that you don't feel anything. 214 00:12:13,559 --> 00:12:16,200 Speaker 1: You're not supposed to feel anything. It sounds very nice 215 00:12:16,200 --> 00:12:18,200 Speaker 1: to me. It does kind of, but I think that 216 00:12:18,320 --> 00:12:20,680 Speaker 1: you could probably get something like what you're describing it, 217 00:12:20,760 --> 00:12:23,640 Speaker 1: like Brookstone maybe, So don't don't fear. Yeah, that's where 218 00:12:23,640 --> 00:12:29,000 Speaker 1: you're just sure sharper image. Perhaps they're under are they anyway? 219 00:12:29,600 --> 00:12:32,720 Speaker 1: Alan Snyder started using t MS because he found this 220 00:12:32,920 --> 00:12:36,079 Speaker 1: curious little side effect of people that were getting tested 221 00:12:36,120 --> 00:12:41,920 Speaker 1: with tms UM had some cognitive malthfunctions, so like speech 222 00:12:41,920 --> 00:12:45,960 Speaker 1: impediments exactly while while this thing was trained on their brain, right, right, 223 00:12:46,040 --> 00:12:48,720 Speaker 1: But it also had some uh if you did, if 224 00:12:48,720 --> 00:12:50,440 Speaker 1: you put this on an average person at some pretty 225 00:12:50,480 --> 00:12:52,760 Speaker 1: cool results. Yeah, this is what Snyder has been doing. 226 00:12:52,880 --> 00:12:55,840 Speaker 1: This is his new experiment, right, and it's very cool. Um. 227 00:12:57,200 --> 00:13:00,240 Speaker 1: Of the people, the normal folks, let's call him, that 228 00:13:00,360 --> 00:13:03,720 Speaker 1: he exposed to t M s UM, they displayed artistic 229 00:13:03,760 --> 00:13:06,200 Speaker 1: and quantitative abilities that they didn't seem to have before. 230 00:13:06,960 --> 00:13:09,559 Speaker 1: So right on the money, it seems like it's actually 231 00:13:09,600 --> 00:13:12,120 Speaker 1: tapping into a part of our brain that we have 232 00:13:12,320 --> 00:13:14,880 Speaker 1: and we don't use, which sort of backs up his theory. Right, 233 00:13:15,040 --> 00:13:17,160 Speaker 1: and some of the some of the things he puts 234 00:13:17,200 --> 00:13:20,320 Speaker 1: people through, well, uh, he he uses TMS on them, 235 00:13:20,760 --> 00:13:23,800 Speaker 1: which apparently we're kind of It looks a lot like 236 00:13:23,920 --> 00:13:27,679 Speaker 1: a shower cap has a bundle of magnetic wires in it, 237 00:13:28,000 --> 00:13:30,880 Speaker 1: a thinking cap, if you will, Yeah, which is kind 238 00:13:30,880 --> 00:13:34,640 Speaker 1: of an inaccurate moniker, but an unfortunate one. The press 239 00:13:34,679 --> 00:13:36,640 Speaker 1: is kind of put on it, right. They had to 240 00:13:36,720 --> 00:13:39,680 Speaker 1: label it, right, you got to you gotta get people 241 00:13:39,679 --> 00:13:41,800 Speaker 1: to read, right, which is why we used it in 242 00:13:41,800 --> 00:13:46,560 Speaker 1: the title of the arti um. So depending on where 243 00:13:46,600 --> 00:13:49,840 Speaker 1: you put it on the skull, um, it's going to 244 00:13:49,920 --> 00:13:52,960 Speaker 1: affect that that very localized region of the brain. So 245 00:13:53,000 --> 00:13:55,440 Speaker 1: of course Snyder is interested in training this on the 246 00:13:55,520 --> 00:13:58,720 Speaker 1: left hemisphere of the brain. Sure um. And he's actually 247 00:13:59,120 --> 00:14:01,920 Speaker 1: using a low freak and see, so he's depressing the 248 00:14:02,400 --> 00:14:06,400 Speaker 1: left brains function and reportedly, like you said, people are 249 00:14:06,400 --> 00:14:08,480 Speaker 1: showing results. One of the things he likes to get 250 00:14:08,520 --> 00:14:13,000 Speaker 1: people to do is draw animals. And apparently with those 251 00:14:13,679 --> 00:14:17,280 Speaker 1: who show a reaction to t MS, they they their 252 00:14:17,360 --> 00:14:20,520 Speaker 1: drawings tend to get better or more realistic, more lifelike. 253 00:14:21,160 --> 00:14:26,280 Speaker 1: And Snyder's theory is that this, this drawing from memory 254 00:14:26,360 --> 00:14:30,840 Speaker 1: is not based on the preconceived notions that you already 255 00:14:30,840 --> 00:14:34,480 Speaker 1: have that would come from the left hemisphere of the brain. Curiously, 256 00:14:34,560 --> 00:14:37,040 Speaker 1: he also has found that people can, um, you know, 257 00:14:37,200 --> 00:14:40,480 Speaker 1: ordinary people were talking about can identify prime numbers from sight. 258 00:14:40,600 --> 00:14:45,560 Speaker 1: I love that field and words, I believe proof reading, 259 00:14:45,680 --> 00:14:48,240 Speaker 1: grammatical errors all of a sudden out of nowhere. So 260 00:14:48,320 --> 00:14:51,840 Speaker 1: over the course of this TMS therapy or whatever, um, 261 00:14:51,880 --> 00:14:55,520 Speaker 1: they're getting progressively better at these tasks. Right, but only 262 00:14:55,600 --> 00:14:58,440 Speaker 1: last about an hour though, correct, Yes, and it may 263 00:14:58,480 --> 00:15:00,480 Speaker 1: not happen at all. There's an argue went out there 264 00:15:00,520 --> 00:15:02,640 Speaker 1: that like, if you draw fourteen cats in a row, 265 00:15:02,680 --> 00:15:05,760 Speaker 1: they're going to get better. That that may or may 266 00:15:05,760 --> 00:15:08,720 Speaker 1: not be true, but it is pretty interesting data that 267 00:15:08,760 --> 00:15:11,520 Speaker 1: he's coming up with. And I don't think arguments like 268 00:15:11,560 --> 00:15:13,480 Speaker 1: that are are really putting the kai bosh on his 269 00:15:13,560 --> 00:15:16,920 Speaker 1: investigations using t MS, right, I don't think so. Now, which, 270 00:15:17,000 --> 00:15:19,960 Speaker 1: by the way, also I understand you said has has 271 00:15:20,040 --> 00:15:22,800 Speaker 1: just been approved by the FDA for use in treating depression, 272 00:15:22,920 --> 00:15:25,720 Speaker 1: so g Yeah, Josh. They studied uh, three hundred people 273 00:15:25,760 --> 00:15:30,400 Speaker 1: that had clinical depression in Philadelphia, and they found out 274 00:15:30,440 --> 00:15:32,920 Speaker 1: that people that underwent the t MS therapy were twice 275 00:15:32,960 --> 00:15:36,160 Speaker 1: as likely to go into remissions. And they're also now 276 00:15:36,800 --> 00:15:40,680 Speaker 1: this is just as of last week, I think, are 277 00:15:40,720 --> 00:15:44,680 Speaker 1: studying asking for stroke victims to volunteer for studies with teams. Yeah. 278 00:15:44,920 --> 00:15:48,280 Speaker 1: Apparently with with depression. UM, if you train it on 279 00:15:48,320 --> 00:15:52,480 Speaker 1: the frontal lobe, I believe, uh, they and you put 280 00:15:52,480 --> 00:15:54,920 Speaker 1: it on a high frequency, they've actually shown that it 281 00:15:55,920 --> 00:16:01,320 Speaker 1: restructures the brain like your neurons are are structured. UM. 282 00:16:01,360 --> 00:16:03,400 Speaker 1: And of course in the frontal lobe. That's where your 283 00:16:03,880 --> 00:16:08,040 Speaker 1: your ability to regulate mood is, right, So that's just weird, 284 00:16:08,120 --> 00:16:10,680 Speaker 1: but very hopeful it is. It makes you wonder if 285 00:16:10,720 --> 00:16:13,560 Speaker 1: this thing could be the key to making people smarter, 286 00:16:13,760 --> 00:16:17,520 Speaker 1: curing brain disease. Yeah. Interesting, so I guess. I guess 287 00:16:17,520 --> 00:16:20,040 Speaker 1: the takeaway from this one is the next time you 288 00:16:20,080 --> 00:16:23,320 Speaker 1: meet a savant of any kind and uh, here she 289 00:16:23,440 --> 00:16:26,400 Speaker 1: tries to impress you with their you know, mathematical or 290 00:16:26,840 --> 00:16:29,880 Speaker 1: musical skills, you can think to yourself, I could do 291 00:16:29,960 --> 00:16:32,880 Speaker 1: that too if I had, let's break damage. Sure, you're 292 00:16:32,920 --> 00:16:37,160 Speaker 1: not so special exactly, so Chuck, that would be uh, 293 00:16:37,240 --> 00:16:39,280 Speaker 1: what's the thinking cap? And couldn't make me a genius? 294 00:16:39,320 --> 00:16:42,160 Speaker 1: The answer is no, not really, But that's what you 295 00:16:42,160 --> 00:16:43,600 Speaker 1: would type in if you wanted to go to how 296 00:16:43,680 --> 00:16:47,160 Speaker 1: stuff works dot com? Righted um, and I think you 297 00:16:47,240 --> 00:16:50,360 Speaker 1: had something you wanted to say to everybody? Well, yeah, Josh, 298 00:16:50,480 --> 00:16:52,640 Speaker 1: is pretty exciting. Before we get to a listener mail, 299 00:16:53,240 --> 00:16:56,880 Speaker 1: we are launching a blog. Not just you and I, 300 00:16:57,000 --> 00:16:59,480 Speaker 1: but I believe six or seven blogs on the website, 301 00:16:59,640 --> 00:17:01,680 Speaker 1: a whole mess of them, whole mess of them, and 302 00:17:01,720 --> 00:17:04,320 Speaker 1: they gave you and I, as you know, our own 303 00:17:04,320 --> 00:17:07,120 Speaker 1: little blog called stuff you Should Know, although the whole 304 00:17:07,240 --> 00:17:09,320 Speaker 1: entire blog section is called stuff you Should Know. Don't 305 00:17:09,320 --> 00:17:13,919 Speaker 1: get confused. Yes it is. I hadn't noticed that they 306 00:17:14,000 --> 00:17:17,679 Speaker 1: named it after. So we would like our listeners to 307 00:17:18,119 --> 00:17:20,600 Speaker 1: get active. This is a call out to our listeners 308 00:17:20,600 --> 00:17:22,639 Speaker 1: to get on the blogs. We're going to be discussing 309 00:17:22,640 --> 00:17:26,679 Speaker 1: all kinds of cool stuff that isn't long enough to 310 00:17:26,720 --> 00:17:30,359 Speaker 1: make into a full episode, so like shorter topics on there. 311 00:17:30,440 --> 00:17:32,560 Speaker 1: And we'll also be talking about the shows that we do, 312 00:17:33,160 --> 00:17:36,359 Speaker 1: uh every Tuesday and Thursday or release. So yeah, and 313 00:17:36,440 --> 00:17:39,760 Speaker 1: actually we've picked up on a couple of listeners listener 314 00:17:39,840 --> 00:17:43,440 Speaker 1: male suggestions. Sure we've written on so keep those ideas 315 00:17:43,480 --> 00:17:46,160 Speaker 1: coming too, because you know, Chuck and I can only 316 00:17:46,200 --> 00:17:48,240 Speaker 1: do so much, right, So go to go to the 317 00:17:48,480 --> 00:17:50,920 Speaker 1: website and look for blogs. It should be pretty easy 318 00:17:50,960 --> 00:17:52,119 Speaker 1: to find. We'll have a you are all for you 319 00:17:52,240 --> 00:17:56,320 Speaker 1: very shortly and uh enjoy talk to each other, connect 320 00:17:56,480 --> 00:17:58,879 Speaker 1: what nicely? Don't show thanks? Okay, So you know what 321 00:17:58,920 --> 00:18:04,920 Speaker 1: this is, right it listener mail time? Yes it is, yes, 322 00:18:04,960 --> 00:18:07,359 Speaker 1: it is uh so Josh, this week we heard from 323 00:18:07,440 --> 00:18:10,520 Speaker 1: a man named Jason divin Ere. You pronounced it? You know, 324 00:18:10,920 --> 00:18:13,440 Speaker 1: I don't. I've never met him or email palace. He 325 00:18:13,520 --> 00:18:16,080 Speaker 1: works for how Stuff Works up in Chicago. Okay, so 326 00:18:16,119 --> 00:18:18,920 Speaker 1: this is an insider deal, but that's fine because Jason 327 00:18:18,960 --> 00:18:21,760 Speaker 1: did right. This is about the moon landing episode and 328 00:18:21,800 --> 00:18:25,040 Speaker 1: whether or not it was faked. Jason is a three 329 00:18:25,040 --> 00:18:27,239 Speaker 1: time space camper, which is kind of cool. I hope 330 00:18:27,280 --> 00:18:30,560 Speaker 1: you ribbed them for that. And full time nerd self 331 00:18:30,560 --> 00:18:33,320 Speaker 1: professed and he said he was excited to see a 332 00:18:33,359 --> 00:18:36,359 Speaker 1: podcast about the moon landing. When you were talking about 333 00:18:36,400 --> 00:18:38,480 Speaker 1: dust on the Moon, he said in the photos video, 334 00:18:38,560 --> 00:18:41,200 Speaker 1: dust appears to be clouding or kicked up more than 335 00:18:41,280 --> 00:18:43,720 Speaker 1: dust would be on Earth. This would occur because the 336 00:18:43,760 --> 00:18:46,960 Speaker 1: particles are airborne longer due to lack of gravity. One 337 00:18:47,040 --> 00:18:49,840 Speaker 1: six to gravity. By the way, yeah, it's not no growing. 338 00:18:50,880 --> 00:18:54,120 Speaker 1: What didn't fit was when you insinuated that to recreate 339 00:18:54,119 --> 00:18:58,600 Speaker 1: this effect on Earth that would require a vacuumized sound stage. Josh, 340 00:18:58,640 --> 00:19:01,160 Speaker 1: Apparently the air has nothing to do with it. On Earth, 341 00:19:01,160 --> 00:19:03,000 Speaker 1: the dust particles will rise and fall at the same 342 00:19:03,080 --> 00:19:05,800 Speaker 1: rate regardless of the presence of air in the room. 343 00:19:05,840 --> 00:19:07,879 Speaker 1: The only effect air would have on a falling object 344 00:19:07,960 --> 00:19:11,359 Speaker 1: to provide resistance when you're dealing with something as small 345 00:19:11,359 --> 00:19:14,400 Speaker 1: as tiny rocks that make up this dust. Air resistance 346 00:19:14,400 --> 00:19:16,560 Speaker 1: would be such a small factor it would not be 347 00:19:16,960 --> 00:19:21,160 Speaker 1: perceptible to the naked eye. So Jason fully geeked out 348 00:19:21,560 --> 00:19:24,560 Speaker 1: setis straight. That's awesome. And on that note, with the 349 00:19:24,560 --> 00:19:26,359 Speaker 1: moon landing, we had a bunch of a bunch of 350 00:19:26,359 --> 00:19:30,240 Speaker 1: people right in about the MythBusters episode where they tested 351 00:19:30,280 --> 00:19:33,040 Speaker 1: out some of these theories and they actually shot uh 352 00:19:33,440 --> 00:19:35,840 Speaker 1: beam of light a laser which guess is a beam 353 00:19:35,880 --> 00:19:38,720 Speaker 1: of light, uh, and it there are these reflectors that 354 00:19:38,760 --> 00:19:41,800 Speaker 1: they left on the moon and it bounced back and 355 00:19:41,840 --> 00:19:45,320 Speaker 1: they saw this, so they pretty much proved absolutely that 356 00:19:45,400 --> 00:19:47,119 Speaker 1: we did land on the Moon. And I don't have 357 00:19:47,160 --> 00:19:50,119 Speaker 1: a list of everyone that wrote in telling us about 358 00:19:50,119 --> 00:19:52,200 Speaker 1: that show, but it was a lot of folks. Yes, 359 00:19:52,240 --> 00:19:57,320 Speaker 1: so you haven't marked lots of listeners, yes, thank you. Well, 360 00:19:57,320 --> 00:20:01,560 Speaker 1: if you want to become Chucks or my email buddy, uh, 361 00:20:01,600 --> 00:20:04,080 Speaker 1: you can send us an email about anything you like 362 00:20:04,560 --> 00:20:12,719 Speaker 1: uh at stuff podcast at how stuff works dot com 363 00:20:12,720 --> 00:20:15,159 Speaker 1: for more on this and thousands of other topics. Is 364 00:20:15,200 --> 00:20:19,919 Speaker 1: it how stuff works dot com. MHM brought to you 365 00:20:19,960 --> 00:20:23,320 Speaker 1: by the reinvented two thousand twelve camera. It's ready are 366 00:20:23,400 --> 00:20:23,520 Speaker 1: you