1 00:00:00,080 --> 00:00:03,320 Speaker 1: Welcome to today's edition of the Clay Travis and Buck 2 00:00:03,360 --> 00:00:08,119 Speaker 1: Sexton Show podcast O. Welcome in Thursday edition Clay Travis 3 00:00:08,200 --> 00:00:11,600 Speaker 1: buck Sexton Show. I appreciate all of you hanging out 4 00:00:11,600 --> 00:00:14,520 Speaker 1: with us. I'm headed up to the Sweet sixteen. Buck 5 00:00:14,560 --> 00:00:17,119 Speaker 1: doesn't even know what the Sweet sixteen is. As the 6 00:00:17,200 --> 00:00:20,279 Speaker 1: NCAA tournament rolls on, I will be in New York 7 00:00:20,320 --> 00:00:23,599 Speaker 1: City before for too long for games tonight, Michigan State 8 00:00:23,680 --> 00:00:28,480 Speaker 1: going up against Who's Michigan State playing Kansas State, Tennessee 9 00:00:28,480 --> 00:00:31,560 Speaker 1: against FAU. I cannot wait for Madison Square Garden. That 10 00:00:31,680 --> 00:00:34,080 Speaker 1: is where I am headed. And it's interesting I'll be 11 00:00:34,120 --> 00:00:37,879 Speaker 1: headed to New York City where Buckett doesn't appear that 12 00:00:37,920 --> 00:00:41,000 Speaker 1: there is going to be an indictment of Donald Trump, 13 00:00:41,159 --> 00:00:45,960 Speaker 1: at least this week, which makes you wonder what might 14 00:00:46,080 --> 00:00:50,480 Speaker 1: be going on with Alvin Bragg's investigation and whether there 15 00:00:50,560 --> 00:00:54,320 Speaker 1: is now pushback emerging. Given, as we have discussed on 16 00:00:54,360 --> 00:00:59,240 Speaker 1: this program, the extremely novel and I think flight frankly 17 00:01:00,120 --> 00:01:03,560 Speaker 1: likely to be overturned theory upon which he is trying 18 00:01:03,600 --> 00:01:08,440 Speaker 1: to bring cases against Donald Trump. The data again, the 19 00:01:08,480 --> 00:01:11,120 Speaker 1: evidence this week reflects that it does not appear that 20 00:01:11,160 --> 00:01:13,240 Speaker 1: there's going to be an indictment this week, I would say, 21 00:01:13,280 --> 00:01:14,920 Speaker 1: and I'm curious what you think about this right off 22 00:01:14,959 --> 00:01:19,040 Speaker 1: the top, Buck, The longer we go without an official 23 00:01:19,040 --> 00:01:22,280 Speaker 1: indictment coming down, the more I think it's fair to 24 00:01:22,319 --> 00:01:26,440 Speaker 1: say there are behind the scenes shenanigans going on that 25 00:01:26,480 --> 00:01:29,959 Speaker 1: are putting into question whether Alvin Bragg is going to 26 00:01:30,000 --> 00:01:32,640 Speaker 1: be bringing charges in this case or not. Would you 27 00:01:32,680 --> 00:01:36,360 Speaker 1: analyze it that way? Oh yeah, I mean delay does 28 00:01:36,440 --> 00:01:40,240 Speaker 1: not favor the decision here because, first of all, for 29 00:01:40,280 --> 00:01:44,480 Speaker 1: a prosecutor, it just creates a further sense that you 30 00:01:44,560 --> 00:01:47,800 Speaker 1: are uncertain because the facts aren't really in dispute here. 31 00:01:47,880 --> 00:01:51,760 Speaker 1: That's what's so, that's what's interesting about this. It's not like, 32 00:01:51,800 --> 00:01:54,280 Speaker 1: I mean Trump whether he had a Stormy daniels a 33 00:01:54,320 --> 00:01:56,600 Speaker 1: fair enough people dispute that, but the payment is not 34 00:01:56,680 --> 00:02:02,040 Speaker 1: in dispute, right, and really very minimal bookkeeping aspects of 35 00:02:02,040 --> 00:02:04,560 Speaker 1: this or what people are discussing. So yeah, the longer 36 00:02:04,600 --> 00:02:07,120 Speaker 1: you wait on this one, if you're brag, I think, 37 00:02:07,320 --> 00:02:09,440 Speaker 1: the less likely it is to come down. And this 38 00:02:09,520 --> 00:02:12,600 Speaker 1: might be one of the biggest non story stories in 39 00:02:12,680 --> 00:02:15,480 Speaker 1: recent news cycle history. I do wonder what it's going 40 00:02:15,520 --> 00:02:18,720 Speaker 1: to do for Trump camp member he's official, he's a 41 00:02:18,720 --> 00:02:21,960 Speaker 1: candidate Trump fundraising. I've seen some of the text messages 42 00:02:21,960 --> 00:02:24,480 Speaker 1: and emails going out on this to raise funds. Sure, 43 00:02:24,600 --> 00:02:27,239 Speaker 1: you've seen some two Clay They probably, I would guess 44 00:02:27,280 --> 00:02:29,800 Speaker 1: they will end up raising more money in the last 45 00:02:29,840 --> 00:02:32,680 Speaker 1: two weeks of the first quarter of this year than 46 00:02:32,720 --> 00:02:35,519 Speaker 1: they did in the entire rest of the quarter put together. 47 00:02:35,560 --> 00:02:37,240 Speaker 1: That's just a guess. But they're gonna have to make 48 00:02:37,280 --> 00:02:41,240 Speaker 1: that public. We'll see. So that is there. We're gonna 49 00:02:41,240 --> 00:02:44,560 Speaker 1: talk with Julie Kelly, who has been phenomenal on so 50 00:02:44,639 --> 00:02:48,400 Speaker 1: many of these legal related issues at one thirty Eastern. 51 00:02:48,520 --> 00:02:51,280 Speaker 1: She will join us, will get the absolute latest there. 52 00:02:51,440 --> 00:02:55,079 Speaker 1: Other things that are going on, we have got continued 53 00:02:55,200 --> 00:03:01,240 Speaker 1: fallout over the absurdity of transgender men transgender women that 54 00:03:01,280 --> 00:03:04,520 Speaker 1: as men who are identifying as women continuing to dominate 55 00:03:04,560 --> 00:03:06,760 Speaker 1: women's athletics. We're gonna have some fun with some of 56 00:03:06,760 --> 00:03:10,720 Speaker 1: the clips that are absurd out there that as this 57 00:03:10,800 --> 00:03:13,359 Speaker 1: continues to become more and more of an issue all 58 00:03:13,400 --> 00:03:17,560 Speaker 1: over the country, the absurdity of it all continues to grow. 59 00:03:18,240 --> 00:03:22,320 Speaker 1: But right now, as we are speaking, I believe there 60 00:03:22,400 --> 00:03:26,480 Speaker 1: is a hearing on Capitol Hill surrounding TikTok, which has 61 00:03:26,520 --> 00:03:32,320 Speaker 1: turned into a major embrolio coming back from Italy. I'll 62 00:03:32,400 --> 00:03:37,720 Speaker 1: break it out there over what should we do about TikTok? Now, 63 00:03:37,760 --> 00:03:40,280 Speaker 1: I want to give buck for the audience out there 64 00:03:40,800 --> 00:03:44,280 Speaker 1: a quick sort of synopsis of what TikTok is, because 65 00:03:44,360 --> 00:03:48,200 Speaker 1: I am quite confident that this audience is not very 66 00:03:48,240 --> 00:03:51,280 Speaker 1: active on TikTok. I'm just gonna I'm just gonna make 67 00:03:51,720 --> 00:03:55,320 Speaker 1: a wild guess there. But chances are your kids or 68 00:03:55,480 --> 00:04:00,880 Speaker 1: grandkids are on TikTok, and they probably are fairly active 69 00:04:00,960 --> 00:04:06,280 Speaker 1: there because it is wildly popular with young kids. If 70 00:04:06,320 --> 00:04:10,160 Speaker 1: your kids or grandkids are ages let's say twelve to seventeen, 71 00:04:10,840 --> 00:04:13,840 Speaker 1: there's almost one hundred percent chance that they are active 72 00:04:13,960 --> 00:04:18,080 Speaker 1: on TikTok. I think that's probably the foundational power base 73 00:04:18,240 --> 00:04:22,320 Speaker 1: of TikTok in general. Buck, TikTok right now is where 74 00:04:22,360 --> 00:04:28,480 Speaker 1: Facebook was in this country over a decade ago, when people, 75 00:04:28,880 --> 00:04:31,240 Speaker 1: when a huge number of people were on it, on 76 00:04:31,360 --> 00:04:33,279 Speaker 1: it a lot, and on it for a long period 77 00:04:33,279 --> 00:04:35,440 Speaker 1: of time. There are one hundred and fifty million Americans 78 00:04:35,480 --> 00:04:38,880 Speaker 1: who at least sometimes use TikTok. But if you look 79 00:04:38,920 --> 00:04:42,839 Speaker 1: at the data and it's It's particularly popular among young people. 80 00:04:42,880 --> 00:04:46,000 Speaker 1: It's kind of taken Snapchat's play. Snapchat was I'm too 81 00:04:46,040 --> 00:04:48,920 Speaker 1: young and cool for Facebook, but now TikTok is I'm 82 00:04:48,960 --> 00:04:52,560 Speaker 1: too young and cool for Snapchat. But it has taken over. 83 00:04:52,640 --> 00:04:56,599 Speaker 1: It is the you know, the zeitgeist of social media 84 00:04:56,680 --> 00:04:58,720 Speaker 1: right now is TikTok with one hundred and fifty million 85 00:04:58,800 --> 00:05:02,120 Speaker 1: US users, and people will sit on at clay for hours, 86 00:05:02,520 --> 00:05:05,400 Speaker 1: correct hours. It is attentive. And I will tell you 87 00:05:05,440 --> 00:05:10,480 Speaker 1: that your kids or grandkids use TikTok like you might 88 00:05:10,760 --> 00:05:14,520 Speaker 1: use Google. In other words, they just go into TikTok 89 00:05:14,520 --> 00:05:17,440 Speaker 1: and type in a query, a question, and then they'll 90 00:05:17,480 --> 00:05:22,120 Speaker 1: just run with whatever. The feedback they get there is buck. 91 00:05:22,160 --> 00:05:24,479 Speaker 1: I don't know how active you are on TikTok. We 92 00:05:24,560 --> 00:05:29,520 Speaker 1: have got OutKick does a TikTok account that I'm amazed. 93 00:05:29,560 --> 00:05:32,640 Speaker 1: Sometimes we have videos that we put up, some of 94 00:05:32,640 --> 00:05:35,880 Speaker 1: them do eight or ten million views. I mean, it 95 00:05:36,040 --> 00:05:39,000 Speaker 1: is unbelievable when one of these things catches fire, how 96 00:05:39,040 --> 00:05:43,120 Speaker 1: massive the influence can be. The difference. I would say 97 00:05:43,440 --> 00:05:44,960 Speaker 1: for those of you out there who might say I'm 98 00:05:44,960 --> 00:05:46,920 Speaker 1: never gonna be on TikTok, some of you are probably 99 00:05:46,920 --> 00:05:48,919 Speaker 1: like I'm fed up with Facebook. I don't trust Twitter. 100 00:05:48,960 --> 00:05:51,000 Speaker 1: I'm not gonna get on Instagram. I understand all that 101 00:05:51,560 --> 00:05:58,440 Speaker 1: the difference is Facebook, Twitter, Instagram, Snapchat all American owned companies. 102 00:05:58,920 --> 00:06:02,359 Speaker 1: TikTok is a Chinese company, and the Chinese Communist Party 103 00:06:02,760 --> 00:06:07,400 Speaker 1: has in many ways control over every Chinese company to 104 00:06:07,640 --> 00:06:11,160 Speaker 1: some degree, certainly to a degree that doesn't exist in 105 00:06:11,200 --> 00:06:14,599 Speaker 1: the United States at all. So the question that is 106 00:06:14,600 --> 00:06:17,840 Speaker 1: out there and being debated today is I think an 107 00:06:17,839 --> 00:06:21,599 Speaker 1: important one. What should happen with TikTok. You'll remember Trump 108 00:06:21,680 --> 00:06:25,880 Speaker 1: tried to force TikTok's American assets to be sold to 109 00:06:26,120 --> 00:06:29,320 Speaker 1: I believe it was Oracle, and was it Facebook or 110 00:06:29,360 --> 00:06:31,320 Speaker 1: Oracle and somebody else that was going to buy the 111 00:06:31,320 --> 00:06:34,680 Speaker 1: TikTok American assets, and then that thing fell apart over 112 00:06:34,760 --> 00:06:39,400 Speaker 1: court rulings. Now the Biden administration is actually taking essentially 113 00:06:39,440 --> 00:06:42,920 Speaker 1: the same position as Trump and demanding that this not 114 00:06:43,040 --> 00:06:46,560 Speaker 1: be a Chinese owned company because there are fears that 115 00:06:46,760 --> 00:06:51,200 Speaker 1: China is using TikTok to spy on American users. So 116 00:06:51,560 --> 00:06:53,440 Speaker 1: that is the backdrop that is what is going on 117 00:06:53,560 --> 00:06:57,080 Speaker 1: here right now. Buck, What is your perspective on what 118 00:06:57,200 --> 00:07:00,040 Speaker 1: should happen with TikTok based on your knowledge of the 119 00:07:00,080 --> 00:07:03,160 Speaker 1: app and where it lies in the larger social media universe. 120 00:07:03,640 --> 00:07:06,480 Speaker 1: I'm I'm a little bit of a dissident on this issue. 121 00:07:06,680 --> 00:07:09,840 Speaker 1: I think that there there's one congressman out of New 122 00:07:09,920 --> 00:07:15,200 Speaker 1: York I'm forget, thank you, Jamal Bowman who thinks that 123 00:07:15,200 --> 00:07:18,600 Speaker 1: there's a bit of anti China hysteria around this. I 124 00:07:18,640 --> 00:07:21,080 Speaker 1: actually think he's right, and I and this is something 125 00:07:21,080 --> 00:07:22,880 Speaker 1: and I understand I'm out there on my own on 126 00:07:22,920 --> 00:07:25,760 Speaker 1: this one. But first of all, let me let me 127 00:07:25,800 --> 00:07:28,920 Speaker 1: just make my case before people start yelling at Sexton, 128 00:07:29,200 --> 00:07:33,560 Speaker 1: left wing New York. Yeah. So, so here's what I'll 129 00:07:33,600 --> 00:07:36,840 Speaker 1: say about it. Um. First of all, if you're worried 130 00:07:36,880 --> 00:07:43,000 Speaker 1: about brainwashing of your children and and also about politicization 131 00:07:43,520 --> 00:07:47,760 Speaker 1: and weaponization, we haven't even touched Okay, we've seen what's 132 00:07:47,760 --> 00:07:50,400 Speaker 1: going on with Twitter, thank you Elon. We haven't even 133 00:07:50,440 --> 00:07:54,680 Speaker 1: touched Facebook. We haven't even touched Instagram, we haven't touched YouTube, 134 00:07:54,760 --> 00:07:58,800 Speaker 1: we haven't touched Google. They are far more powerful, far 135 00:07:58,920 --> 00:08:01,680 Speaker 1: more influential, and as you saw during COVID and the 136 00:08:01,680 --> 00:08:04,880 Speaker 1: twenty twenty election and other areas, you look into a 137 00:08:05,040 --> 00:08:08,360 Speaker 1: much bigger threat to our freedom of speech, our freedom 138 00:08:08,400 --> 00:08:11,360 Speaker 1: of you know, democracy, whatever. I mean, it's those are 139 00:08:11,600 --> 00:08:16,080 Speaker 1: for me such a bigger challenge than the idea that 140 00:08:16,200 --> 00:08:19,480 Speaker 1: China is going to be programming your kids over the 141 00:08:19,480 --> 00:08:22,840 Speaker 1: next few decades to be anti American and weak. I 142 00:08:22,840 --> 00:08:24,800 Speaker 1: want to say, what do you think YouTube is doing? 143 00:08:24,880 --> 00:08:27,640 Speaker 1: By the way, I mean, so, that's my first you know, 144 00:08:27,680 --> 00:08:31,080 Speaker 1: that's my first line of So your first line, and 145 00:08:31,600 --> 00:08:34,400 Speaker 1: I think I agree as a parent, is that social 146 00:08:34,440 --> 00:08:40,400 Speaker 1: media is troubling to you. TikTok is not distinguishingly troubling 147 00:08:40,440 --> 00:08:43,800 Speaker 1: to you relative to the overall social media environment. Yeah, 148 00:08:43,960 --> 00:08:47,480 Speaker 1: I think that TikTok is not. And I look, I'm 149 00:08:47,520 --> 00:08:50,360 Speaker 1: not I'm not somebody who's actively posting on TikTok Um. 150 00:08:50,520 --> 00:08:52,320 Speaker 1: There is a TikTok account that picks up some of 151 00:08:52,360 --> 00:08:54,640 Speaker 1: my stuff, but I'm not, you know, using it that way. 152 00:08:54,920 --> 00:08:56,839 Speaker 1: But I do go on there. They make a lot 153 00:08:56,840 --> 00:08:58,600 Speaker 1: of great little videos about how to see her the 154 00:08:58,640 --> 00:09:01,679 Speaker 1: perfect Ribi and other things. It is very addictive. I 155 00:09:01,720 --> 00:09:03,160 Speaker 1: will tell you, be prepared to lose a lot of 156 00:09:03,160 --> 00:09:06,000 Speaker 1: time using it. But okay, I a point one other 157 00:09:06,040 --> 00:09:08,760 Speaker 1: social media platforms and anybody who knows anything about Google 158 00:09:08,800 --> 00:09:12,200 Speaker 1: and YouTube would agree with me. They're infinitely more powerful 159 00:09:12,280 --> 00:09:15,400 Speaker 1: right now and more influential in online conversation I think 160 00:09:15,400 --> 00:09:17,400 Speaker 1: than TikTok is TikTok is powerful, don't get me right. 161 00:09:17,440 --> 00:09:22,040 Speaker 1: I shouldn't say infinitely, but substantially, substantially more powerful. The 162 00:09:22,120 --> 00:09:24,720 Speaker 1: second point is okay, So now we have the government 163 00:09:24,760 --> 00:09:27,559 Speaker 1: saying we don't like to content on a platform, so 164 00:09:27,600 --> 00:09:29,960 Speaker 1: we're going to ban it. And I understand it's really 165 00:09:29,960 --> 00:09:32,480 Speaker 1: easy when you think of it as the Chinese Communist Party. 166 00:09:32,840 --> 00:09:36,200 Speaker 1: They are our opponents on the world stage, to be sure, 167 00:09:36,520 --> 00:09:38,040 Speaker 1: and that's real and I feel that, and we talk 168 00:09:38,080 --> 00:09:40,439 Speaker 1: about that here all the time. But does anyone think 169 00:09:40,440 --> 00:09:43,760 Speaker 1: that democrats that their mindset is only on this one? 170 00:09:43,800 --> 00:09:46,040 Speaker 1: Should we ban the platform that we don't like. Look 171 00:09:46,080 --> 00:09:50,120 Speaker 1: what they did to Parlor with Amazon Web Services, knocking 172 00:09:50,160 --> 00:09:53,040 Speaker 1: them down. Democrats will always try to find a way 173 00:09:53,080 --> 00:09:55,400 Speaker 1: to destroy platforms they don't like. And when we go 174 00:09:55,440 --> 00:09:58,600 Speaker 1: along with them, which a lot of Republicans again I'm 175 00:09:58,640 --> 00:10:00,400 Speaker 1: on my own on this, when we go along with them, 176 00:10:00,480 --> 00:10:03,319 Speaker 1: you're encouraging that. You know. It's interesting we banned Russia 177 00:10:03,360 --> 00:10:06,600 Speaker 1: today in this country as the Russian cable news propaganda 178 00:10:06,679 --> 00:10:08,080 Speaker 1: channel and I kind of felt the same way. I 179 00:10:08,120 --> 00:10:11,240 Speaker 1: was like, CNN is far more destructive to America than 180 00:10:11,320 --> 00:10:15,000 Speaker 1: Russia today. Yeah, I understand foreign ownership. Third point and 181 00:10:15,040 --> 00:10:18,280 Speaker 1: maybe the best point here in my defense on this issue. 182 00:10:18,920 --> 00:10:21,360 Speaker 1: The plan that they have now I have not privy 183 00:10:21,400 --> 00:10:23,640 Speaker 1: to negotiations in the day to day of it is 184 00:10:23,679 --> 00:10:28,680 Speaker 1: called Project Texas. And they're going to have all data 185 00:10:28,720 --> 00:10:32,520 Speaker 1: generated by Americans on servers run and operated and owned 186 00:10:32,520 --> 00:10:36,240 Speaker 1: by Oracle, an American company, And this would be an 187 00:10:36,240 --> 00:10:41,239 Speaker 1: Austin Texas and only Oracle and the US based employees 188 00:10:41,240 --> 00:10:43,920 Speaker 1: of TikTok would have access to any of those servers, 189 00:10:44,040 --> 00:10:47,240 Speaker 1: any of those data, and would be subject to you know, 190 00:10:47,280 --> 00:10:49,600 Speaker 1: audits and you know, look at where the backup recording 191 00:10:49,640 --> 00:10:52,040 Speaker 1: is going and everything else. Now you could say those 192 00:10:52,080 --> 00:10:56,080 Speaker 1: TikTok employees, well, you know, oh, they're just gonna try 193 00:10:56,080 --> 00:10:58,839 Speaker 1: to find ways to back to where TikTok. I mean, 194 00:10:59,120 --> 00:11:02,080 Speaker 1: other social media companies have employees that also could very 195 00:11:02,120 --> 00:11:05,360 Speaker 1: easily backdoor information to China in that circumstance, Right, if 196 00:11:05,360 --> 00:11:08,719 Speaker 1: you're going to say that US based infrastructure and architecture 197 00:11:08,760 --> 00:11:12,120 Speaker 1: is subject to Chinese espionage, I'm going to say absolutely. 198 00:11:12,760 --> 00:11:15,800 Speaker 1: But there are plenty of Chinese nationals that are high 199 00:11:15,800 --> 00:11:18,840 Speaker 1: tech universities and everything else. So I don't see why. 200 00:11:18,880 --> 00:11:20,920 Speaker 1: I guess I don't see why TikTok is so much 201 00:11:20,960 --> 00:11:23,600 Speaker 1: worse in that context than some of these other social 202 00:11:23,600 --> 00:11:26,360 Speaker 1: media companies. And I will say this, I am. I mean, 203 00:11:26,400 --> 00:11:29,760 Speaker 1: it's like me. I think Glenn Greenwald. There's like five 204 00:11:29,800 --> 00:11:31,760 Speaker 1: of us who take this position right now, but I'm 205 00:11:31,760 --> 00:11:34,320 Speaker 1: one of them. Yeah, So I actually am on the 206 00:11:34,360 --> 00:11:38,720 Speaker 1: flip side here, and I appreciate and understand all of 207 00:11:38,760 --> 00:11:41,840 Speaker 1: those arguments. I think they are. I think they're solid. 208 00:11:42,360 --> 00:11:46,480 Speaker 1: Here's what I would say. China doesn't allow Twitter, they 209 00:11:46,520 --> 00:11:50,520 Speaker 1: don't allow Facebook, they don't allow Google. I don't think 210 00:11:50,559 --> 00:11:55,600 Speaker 1: they allow Instagram. They don't allow American companies access to 211 00:11:55,760 --> 00:12:00,319 Speaker 1: the Chinese marketplace to compete in their country. And my 212 00:12:00,400 --> 00:12:04,200 Speaker 1: position is just not even focused on TikTok in general. 213 00:12:04,559 --> 00:12:09,400 Speaker 1: If China won't open its market for fair competition in 214 00:12:09,520 --> 00:12:15,000 Speaker 1: social media again, YouTube not allowed, Twitter not allowed. I 215 00:12:15,000 --> 00:12:18,520 Speaker 1: don't think Google hardly is allowed, Facebook certainly is, and 216 00:12:18,640 --> 00:12:21,160 Speaker 1: all these companies are not allowed that are big social 217 00:12:21,200 --> 00:12:25,040 Speaker 1: media behemoths in the United States. My thing is, if 218 00:12:25,080 --> 00:12:29,320 Speaker 1: they won't let us have access to their citizens, and 219 00:12:29,400 --> 00:12:32,679 Speaker 1: they're demanding that they continue to have access to our citizens. 220 00:12:32,920 --> 00:12:36,840 Speaker 1: We're playing on an unequal playing field from a competitive basis, 221 00:12:37,160 --> 00:12:40,679 Speaker 1: and so I would demand that TikTok either pull out 222 00:12:40,720 --> 00:12:44,160 Speaker 1: of the United States or probably more likely sell its 223 00:12:44,280 --> 00:12:47,800 Speaker 1: United States assets in some way in the United States. 224 00:12:48,160 --> 00:12:49,760 Speaker 1: That would be That would be my position if I 225 00:12:49,840 --> 00:12:52,880 Speaker 1: were right now on Capitol Again, I think that's a 226 00:12:53,000 --> 00:12:55,880 Speaker 1: very also, I think that's a strong argument, But we 227 00:12:56,240 --> 00:12:59,120 Speaker 1: all end up in this same place. They're talking about 228 00:12:59,120 --> 00:13:01,440 Speaker 1: a band to be cleared everybody. Yeah, I'm saying I 229 00:13:01,440 --> 00:13:03,920 Speaker 1: think a ban is a bad idea, and I don't 230 00:13:03,920 --> 00:13:05,800 Speaker 1: thinking me say this too. I don't think a ban 231 00:13:05,880 --> 00:13:09,720 Speaker 1: would work. Buck kids are smarter than all of us 232 00:13:09,760 --> 00:13:12,000 Speaker 1: are when it comes to getting access to tech. I 233 00:13:12,040 --> 00:13:14,679 Speaker 1: just I don't understand how we would suddenly stop people 234 00:13:14,679 --> 00:13:18,560 Speaker 1: from using TikTok. You know, I think a reasonable national 235 00:13:18,600 --> 00:13:23,840 Speaker 1: security and US interest compromise is if all the assets 236 00:13:23,840 --> 00:13:28,680 Speaker 1: are based here, it's US personnel, American citizens, Oracle the 237 00:13:28,720 --> 00:13:31,720 Speaker 1: company that's backing up all the servers and data not 238 00:13:31,760 --> 00:13:33,880 Speaker 1: just for you and for your companies, but you know, 239 00:13:33,920 --> 00:13:36,200 Speaker 1: for the US government, a lot of capacity, right, I 240 00:13:36,200 --> 00:13:41,000 Speaker 1: mean Oracle could the Amazon Web Services doesn't matter. Larry 241 00:13:41,000 --> 00:13:43,400 Speaker 1: Ellison's a pretty big Republican for people out there who 242 00:13:43,840 --> 00:13:47,079 Speaker 1: runs Oracle effectively. Right, So I'm not disagreeing with the 243 00:13:47,080 --> 00:13:49,719 Speaker 1: people that say that TikTok something has to change about it. 244 00:13:49,760 --> 00:13:53,160 Speaker 1: I just think this notion of a band sets a 245 00:13:53,280 --> 00:13:56,560 Speaker 1: precedent that I don't like, is heavy handed in a 246 00:13:56,600 --> 00:14:01,920 Speaker 1: way that's that's unhelpful. And I think that also, you know, 247 00:14:01,960 --> 00:14:04,880 Speaker 1: there are some areas where we're gonna want to say, yeah, fine, 248 00:14:05,040 --> 00:14:09,720 Speaker 1: the trade practices issue certainly something we should address. But also, 249 00:14:10,559 --> 00:14:14,040 Speaker 1: you know, if we're going to be shutting out external 250 00:14:14,320 --> 00:14:17,199 Speaker 1: media from the rest of the world, then we're definitely 251 00:14:17,240 --> 00:14:18,920 Speaker 1: going to be as a country in a position where 252 00:14:18,920 --> 00:14:21,560 Speaker 1: we start telling other places, hey, you guys are so 253 00:14:21,640 --> 00:14:23,920 Speaker 1: closed off and you're not allowing you know, the global 254 00:14:23,960 --> 00:14:26,520 Speaker 1: Internet to be truly free. And they're gonna say, you mean, 255 00:14:26,520 --> 00:14:28,640 Speaker 1: you guys are the ones who banned TikTok, Right. So 256 00:14:28,920 --> 00:14:31,920 Speaker 1: I don't like that precedent either over the long term, 257 00:14:31,920 --> 00:14:34,160 Speaker 1: but I look, I understand why there are concerns about 258 00:14:34,160 --> 00:14:36,680 Speaker 1: it now. I also just think, I mean, I just 259 00:14:36,720 --> 00:14:39,880 Speaker 1: think Google and YouTube are there are the big this 260 00:14:39,960 --> 00:14:42,480 Speaker 1: is the big dragon that nobody has stepped up to understand. 261 00:14:42,640 --> 00:14:45,880 Speaker 1: They are programming your kids right now. They are programming 262 00:14:45,920 --> 00:14:49,200 Speaker 1: them with anti americanism, and they are American, and they're 263 00:14:49,240 --> 00:14:52,400 Speaker 1: here on us soil. YouTube and this is just me 264 00:14:52,520 --> 00:14:56,720 Speaker 1: speaking as a dad. YouTube and TikTok are the two 265 00:14:57,520 --> 00:15:00,920 Speaker 1: social media apps that my boys in by far the 266 00:15:01,000 --> 00:15:05,120 Speaker 1: most time on. They're not on Facebook, they're not on Instagram, 267 00:15:05,320 --> 00:15:07,480 Speaker 1: they're not really on Twitter, and we try to keep 268 00:15:07,520 --> 00:15:10,680 Speaker 1: them off social media in general. But they're on YouTube 269 00:15:11,000 --> 00:15:13,680 Speaker 1: and they are on TikTok. I think it's hard to 270 00:15:13,720 --> 00:15:15,600 Speaker 1: keep them off both. Just by way of giving everyone 271 00:15:15,680 --> 00:15:18,280 Speaker 1: a sense of the context here, YouTube has two hundred 272 00:15:18,360 --> 00:15:22,080 Speaker 1: million based US used. Yeah, right, so not TikTok is 273 00:15:22,680 --> 00:15:24,680 Speaker 1: up there, but it's not quite in the same at 274 00:15:24,720 --> 00:15:28,040 Speaker 1: the same category yet, so we'll see. We'll see. Also, 275 00:15:28,120 --> 00:15:30,160 Speaker 1: just don't trust the Biden administration. So when they call 276 00:15:30,240 --> 00:15:32,360 Speaker 1: for a band, I'm like, I don't know, what are 277 00:15:32,360 --> 00:15:34,800 Speaker 1: you abandon? Well, you know what's funny is the Biden 278 00:15:34,840 --> 00:15:37,200 Speaker 1: administration politically is kind of in an interesting spot. We 279 00:15:37,280 --> 00:15:41,080 Speaker 1: come back buck. We can mention this because they're overwhelmingly 280 00:15:41,080 --> 00:15:44,520 Speaker 1: supported by gen Z, and gen Z would be furious 281 00:15:44,560 --> 00:15:46,720 Speaker 1: at the Biden administration if they take away and then 282 00:15:46,760 --> 00:15:48,560 Speaker 1: what do we say, Gen Alpha or whatever the next 283 00:15:48,560 --> 00:15:52,360 Speaker 1: generation is, If they take away these kids opportunity and 284 00:15:52,400 --> 00:15:54,560 Speaker 1: ability to access TikTok, I think it's kind of fascinating. 285 00:15:54,600 --> 00:15:56,480 Speaker 1: But if this is where you get a lot of 286 00:15:56,600 --> 00:15:59,720 Speaker 1: Republicans and Democrats, and this is you listen to Republican 287 00:15:59,760 --> 00:16:02,520 Speaker 1: members the Congress and they're all hammer on this issue. 288 00:16:02,520 --> 00:16:04,280 Speaker 1: By the way, I think they're let's come back to 289 00:16:04,280 --> 00:16:06,040 Speaker 1: this as well, Clay, because they're much bigger issues with 290 00:16:06,120 --> 00:16:09,040 Speaker 1: China that no one is talking about and no one 291 00:16:09,080 --> 00:16:11,600 Speaker 1: wants to tackle. I think this also becomes the shiny 292 00:16:11,600 --> 00:16:14,520 Speaker 1: object that they can distract people with it. Anyway, don't 293 00:16:14,520 --> 00:16:17,080 Speaker 1: fall for the free phone deals from Verizon, AT and 294 00:16:17,120 --> 00:16:19,520 Speaker 1: T your Team Mobile. It's just another trick by companies 295 00:16:19,800 --> 00:16:22,320 Speaker 1: trying to hold onto your business. Instead, get a brand 296 00:16:22,360 --> 00:16:25,000 Speaker 1: new iPhone twelve from Pure Talk for just twelve bucks 297 00:16:25,000 --> 00:16:27,600 Speaker 1: a month at zero percent interest and no contract. You 298 00:16:27,640 --> 00:16:30,760 Speaker 1: can cancel or leave Pure Talk at any time. Get 299 00:16:30,760 --> 00:16:33,880 Speaker 1: a new iPhone Ultra Fast five G service and cut 300 00:16:33,880 --> 00:16:36,640 Speaker 1: your cell phone bill in half. Switching is easy, and 301 00:16:36,680 --> 00:16:39,000 Speaker 1: you can keep your same cell phone number just dial 302 00:16:39,080 --> 00:16:42,480 Speaker 1: pound two five zero and say Clay and Buck. Pure Talks, 303 00:16:42,680 --> 00:16:44,720 Speaker 1: US based customer service team will help you make the 304 00:16:44,760 --> 00:16:47,000 Speaker 1: switch in as little as ten minutes. Choose some a 305 00:16:47,040 --> 00:16:49,560 Speaker 1: variety of unlimited talk and text plans starting at thirty 306 00:16:49,560 --> 00:16:52,080 Speaker 1: dollars a month with plenty of high speed data, all 307 00:16:52,120 --> 00:16:55,520 Speaker 1: backed by one hundred percent money back guarantee. Dial pound 308 00:16:55,600 --> 00:16:58,680 Speaker 1: two five zero, say Clay and Buck for this special offer. 309 00:16:58,960 --> 00:17:02,640 Speaker 1: That's pound two five zero say Clay and Buck. Restricting 310 00:17:02,640 --> 00:17:07,280 Speaker 1: supply sees site for details Making sense in an Insane world. 311 00:17:07,480 --> 00:17:10,800 Speaker 1: Clay Travis and Buck Sexton. Welcome at to Clay and Buck. 312 00:17:10,800 --> 00:17:13,439 Speaker 1: We're talking about TikTok and the Possible band. We've also 313 00:17:13,520 --> 00:17:17,200 Speaker 1: got our eyes looking forward. Looking ahead, I should say 314 00:17:17,359 --> 00:17:19,159 Speaker 1: to whether or not there's going to be anything of 315 00:17:19,200 --> 00:17:21,840 Speaker 1: this looming Trump indictment. Also want to tell you that 316 00:17:21,880 --> 00:17:25,800 Speaker 1: in the second hour of the program you may have 317 00:17:25,840 --> 00:17:28,680 Speaker 1: seen this. Five hundred thousand kids weren't able to go 318 00:17:28,720 --> 00:17:33,879 Speaker 1: to school for three days because the Los Angeles teachers 319 00:17:33,960 --> 00:17:36,600 Speaker 1: unions went on strike demanding more money. I want to 320 00:17:36,640 --> 00:17:39,000 Speaker 1: tell you a little bit about some of the some 321 00:17:39,080 --> 00:17:43,280 Speaker 1: of the individuals involved in calling for that strike and 322 00:17:43,359 --> 00:17:45,960 Speaker 1: their politics and what's going on with teachers unions in 323 00:17:46,040 --> 00:17:49,520 Speaker 1: major cities across the country, which I think at this point, Clay, 324 00:17:49,520 --> 00:17:52,440 Speaker 1: everybody should realize, after what they did during COVID, they 325 00:17:52,440 --> 00:17:55,560 Speaker 1: are villains in this drama of our lives in America. 326 00:17:55,640 --> 00:17:57,840 Speaker 1: Teachers unions are not doing not doing the right thing 327 00:17:57,880 --> 00:18:01,199 Speaker 1: by the kids, no doubt. Times prescious. If you're a 328 00:18:01,240 --> 00:18:04,240 Speaker 1: business owner, especially a small business owner, Get Refunds dot 329 00:18:04,240 --> 00:18:07,280 Speaker 1: Com understands that. That's why they've developed a fast way 330 00:18:07,600 --> 00:18:09,760 Speaker 1: for you to determine if your business qualifies for the 331 00:18:09,760 --> 00:18:14,320 Speaker 1: IRS tax refund program, something called the Employee Retention Credit 332 00:18:14,480 --> 00:18:16,919 Speaker 1: or the ARC. You can go to Get Refunds dot 333 00:18:17,000 --> 00:18:19,320 Speaker 1: Com to get started less than ten minutes. See if 334 00:18:19,320 --> 00:18:23,080 Speaker 1: your business qualifies for ARC assistance, your business may be 335 00:18:23,160 --> 00:18:26,080 Speaker 1: eligible for a payroll tax refund of up to twenty 336 00:18:26,119 --> 00:18:30,080 Speaker 1: six thousand dollars per employee kept on payroll during COVID. 337 00:18:30,160 --> 00:18:33,920 Speaker 1: Get Refunds dot Com already helped clients claim over three 338 00:18:34,160 --> 00:18:37,000 Speaker 1: billion dollars in refunds. They may be able to help 339 00:18:37,080 --> 00:18:40,320 Speaker 1: your business to no upfront charges. They don't get paid 340 00:18:40,480 --> 00:18:44,480 Speaker 1: until your business gets its refund. Many businesses believe they 341 00:18:44,520 --> 00:18:48,800 Speaker 1: won't qualify based off incomplete or outdated information. Don't let 342 00:18:48,800 --> 00:18:52,080 Speaker 1: this opportunity pass you buy only available for a limited 343 00:18:52,080 --> 00:18:55,240 Speaker 1: amount of time. Go to get refunds dot com. That's 344 00:18:55,280 --> 00:19:00,600 Speaker 1: get refunds dot com. Sleeve Travis and Buck Sexon on 345 00:19:00,720 --> 00:19:04,280 Speaker 1: the front lines of truth. Welcome back in Clay Travis 346 00:19:04,320 --> 00:19:08,080 Speaker 1: buck Sexton Show, Thursday edition of the program. I put 347 00:19:08,119 --> 00:19:11,840 Speaker 1: up a poll question. I'm curious. I genuinely do not 348 00:19:12,000 --> 00:19:14,560 Speaker 1: know how this poll question is going to go, but 349 00:19:14,640 --> 00:19:19,120 Speaker 1: it's pretty simple, pretty straightforward. Do you believe TikTok should 350 00:19:19,160 --> 00:19:21,520 Speaker 1: be banned in the United States so long as a 351 00:19:21,640 --> 00:19:27,840 Speaker 1: Chinese company maintains ownership? I do. At Buck disagrees. There 352 00:19:27,920 --> 00:19:30,880 Speaker 1: is a great deal debate going on right here right now. 353 00:19:30,920 --> 00:19:35,000 Speaker 1: To me, it's less about China having ownership than it 354 00:19:35,160 --> 00:19:39,159 Speaker 1: is the Chinese communist government not allowing our social media 355 00:19:39,240 --> 00:19:42,720 Speaker 1: companies to have access to their market, So why should 356 00:19:42,720 --> 00:19:45,800 Speaker 1: we allow them to have access to our market? That's 357 00:19:45,880 --> 00:19:49,600 Speaker 1: my angle. I do have concerns about spying and all 358 00:19:49,600 --> 00:19:53,520 Speaker 1: the algorithmic ridiculousness that that's going on, But I agree 359 00:19:53,520 --> 00:19:57,800 Speaker 1: with Bucking that there's a lot of stuff that's being 360 00:19:57,800 --> 00:20:01,120 Speaker 1: fed to your kids and grandkids every day to make 361 00:20:01,160 --> 00:20:04,840 Speaker 1: them think that America is awful from American based corporations. 362 00:20:04,920 --> 00:20:07,600 Speaker 1: So the idea that China is going to do something 363 00:20:07,720 --> 00:20:13,439 Speaker 1: different than what YouTube or Twitter before Elon Musk bought it, 364 00:20:13,480 --> 00:20:16,000 Speaker 1: for instance, are doing. I question in general. I mean, 365 00:20:16,040 --> 00:20:18,840 Speaker 1: if you went back and looked at the the whatever 366 00:20:18,880 --> 00:20:22,280 Speaker 1: it was one hundred thousand dollars spent by Russian bought 367 00:20:22,320 --> 00:20:25,640 Speaker 1: farm accounts in the twenty sixteen election, and you looked 368 00:20:25,640 --> 00:20:27,600 Speaker 1: at most of it, I mean you could tell this 369 00:20:27,680 --> 00:20:31,640 Speaker 1: was written by some guy in abasement in Belarus on 370 00:20:31,680 --> 00:20:35,119 Speaker 1: behalf of you know, the Russian equivalent of the FSB 371 00:20:35,320 --> 00:20:37,679 Speaker 1: or something. I mean that it was ridiculous. It was 372 00:20:37,720 --> 00:20:40,600 Speaker 1: like no vote for Hill or a twenty sixteen vote 373 00:20:40,640 --> 00:20:44,000 Speaker 1: for Throw. I mean it was really clearly like translated 374 00:20:44,080 --> 00:20:48,439 Speaker 1: from the cyrillic into English by some translator program. And 375 00:20:48,520 --> 00:20:51,159 Speaker 1: they also spent almost no money, which anyone who has 376 00:20:51,160 --> 00:20:53,840 Speaker 1: ever bought ads on social media when you see they 377 00:20:53,840 --> 00:20:56,800 Speaker 1: spent like one hundred and twenty thousand dollars, it's almost 378 00:20:56,800 --> 00:20:59,240 Speaker 1: impossible to have any impact on Facebook at one hundred 379 00:20:59,240 --> 00:21:00,960 Speaker 1: and twenty k. But by the way, I would say 380 00:21:00,960 --> 00:21:03,640 Speaker 1: on this issue of TikTok, I'm certainly persuadable, and I'm 381 00:21:03,720 --> 00:21:08,960 Speaker 1: very aware of the espionage that's not only the CIA. 382 00:21:09,119 --> 00:21:10,959 Speaker 1: For God's sake, I mean, I think you're probably more 383 00:21:10,960 --> 00:21:14,320 Speaker 1: aware than almost than anyone. If if people understood how 384 00:21:14,400 --> 00:21:18,760 Speaker 1: much has already been stolen from American companies in terms 385 00:21:18,760 --> 00:21:26,760 Speaker 1: of proprietary technology, sensitive personal information, massive hacks of terabytes 386 00:21:27,040 --> 00:21:31,040 Speaker 1: upon terabytes of the most sensitive information. You can imagine, 387 00:21:31,160 --> 00:21:33,560 Speaker 1: that's been going on for twenty years. So I just 388 00:21:33,800 --> 00:21:36,800 Speaker 1: you know, oh, it's this is so bad because TikTok 389 00:21:36,880 --> 00:21:38,679 Speaker 1: is going to be this this portal to all of 390 00:21:38,680 --> 00:21:41,639 Speaker 1: our data and information. The Internet is the portal for 391 00:21:41,760 --> 00:21:44,880 Speaker 1: your data and information because the Chinese are hacking every 392 00:21:44,880 --> 00:21:48,560 Speaker 1: company they can of value and stealing everything of value 393 00:21:48,600 --> 00:21:51,399 Speaker 1: they can. And they've been doing it since Clay and 394 00:21:51,440 --> 00:21:53,439 Speaker 1: I were in college, or Clay was in college and 395 00:21:53,480 --> 00:21:56,760 Speaker 1: I was in nursery school, but close enough. I will 396 00:21:56,800 --> 00:22:00,119 Speaker 1: tell you this, Buck, I had a conversation recently. I 397 00:22:00,200 --> 00:22:01,639 Speaker 1: meant to share this with you because I thought you 398 00:22:01,680 --> 00:22:05,959 Speaker 1: would find it intriguing. These algorithms are constantly feeding kids 399 00:22:06,160 --> 00:22:09,680 Speaker 1: all sorts of different videos, all sorts of different content. 400 00:22:10,720 --> 00:22:13,720 Speaker 1: Do you know I talk to my kids friends recently 401 00:22:14,359 --> 00:22:17,960 Speaker 1: these are kids, you know, like ten to sixteen years old, 402 00:22:18,040 --> 00:22:19,719 Speaker 1: kind of in the window. I've got an eight year 403 00:22:19,760 --> 00:22:23,080 Speaker 1: old to a fifteen year old. Do you know who 404 00:22:23,160 --> 00:22:27,479 Speaker 1: the number one most influential person at least according to 405 00:22:27,520 --> 00:22:30,480 Speaker 1: the focus group that is my kids and their friends 406 00:22:31,160 --> 00:22:35,199 Speaker 1: is right now on any of these social media platforms. 407 00:22:35,400 --> 00:22:40,760 Speaker 1: Mister beast, he's up there. It's actually Andrew Tate. They 408 00:22:40,920 --> 00:22:44,280 Speaker 1: all see Andrew Tate videos all the time, and I 409 00:22:44,320 --> 00:22:46,960 Speaker 1: think this is this is look. I just tell you, 410 00:22:47,040 --> 00:22:49,760 Speaker 1: like a yeah, you know that the Andrew Tate Clay 411 00:22:49,960 --> 00:22:54,080 Speaker 1: they just extended another thirty days in prison without charge 412 00:22:54,440 --> 00:22:57,800 Speaker 1: so they can find out whether he should be charged. Yeah, 413 00:22:58,000 --> 00:23:00,880 Speaker 1: Chris is going into month four of this guy being 414 00:23:00,880 --> 00:23:04,680 Speaker 1: held in a hellish remaining in prison. If they if 415 00:23:04,680 --> 00:23:07,320 Speaker 1: he was so guilty and there was such bad stuff, 416 00:23:07,840 --> 00:23:09,920 Speaker 1: they need four months to come up with whether or 417 00:23:09,960 --> 00:23:12,720 Speaker 1: not they're even going to charge. I mean this, look, 418 00:23:12,760 --> 00:23:14,800 Speaker 1: I was gonna say it. America is not great with 419 00:23:14,800 --> 00:23:17,399 Speaker 1: the justice system. You don't want to get arrested anywhere 420 00:23:17,400 --> 00:23:19,240 Speaker 1: else though, I can tell you. I want to tell 421 00:23:19,280 --> 00:23:21,280 Speaker 1: you this, this is like a ten or fifteen or 422 00:23:21,320 --> 00:23:24,000 Speaker 1: maybe even twenty year prediction because these kids and these 423 00:23:24,000 --> 00:23:26,080 Speaker 1: are boys. Right, I've got three boys, so I'm talking 424 00:23:26,119 --> 00:23:28,240 Speaker 1: to boys. I'm not saying what girls are seeing. I 425 00:23:28,280 --> 00:23:30,680 Speaker 1: have no idea. I don't have teenage girls or young 426 00:23:30,720 --> 00:23:32,920 Speaker 1: girls or anything like that, so I'm sure they get 427 00:23:32,920 --> 00:23:36,760 Speaker 1: fed different things on the algorithm. I think that these 428 00:23:36,840 --> 00:23:40,159 Speaker 1: young boys today, and I'm talking about this as a 429 00:23:40,280 --> 00:23:44,120 Speaker 1: dad of three young boys, I think that they are 430 00:23:44,359 --> 00:23:48,679 Speaker 1: told that they are awful as men, right, and I 431 00:23:48,720 --> 00:23:50,720 Speaker 1: mean awful as men when they're like seven and eight 432 00:23:50,800 --> 00:23:54,960 Speaker 1: years old. That a lot of them are rejecting what 433 00:23:55,000 --> 00:23:59,080 Speaker 1: they are being taught in school and desperately seeking out 434 00:23:59,280 --> 00:24:04,200 Speaker 1: masculine figures that do not apologize for being men. And 435 00:24:04,240 --> 00:24:07,040 Speaker 1: I think there's a desperate yearning, to say nothing, of 436 00:24:07,080 --> 00:24:10,359 Speaker 1: young boys who are growing out, growing up without masculine 437 00:24:10,359 --> 00:24:13,880 Speaker 1: influence in their household. Right. Maybe they're getting raised by Grandma, 438 00:24:14,080 --> 00:24:16,480 Speaker 1: maybe they're getting raised by mom. They don't really know 439 00:24:16,600 --> 00:24:19,760 Speaker 1: what men are like. That's a huge percentage, unfortunately, of 440 00:24:19,880 --> 00:24:23,640 Speaker 1: young boys in this country. And interestingly, buck what they 441 00:24:23,680 --> 00:24:28,639 Speaker 1: show is young boys suffer far more from the absence 442 00:24:28,680 --> 00:24:31,280 Speaker 1: of a father in the household than young girls do. 443 00:24:31,520 --> 00:24:34,400 Speaker 1: When they look at what outcomes kids have. Girls tend 444 00:24:34,440 --> 00:24:37,240 Speaker 1: to be far more resilient. Maybe it's because there's a 445 00:24:37,240 --> 00:24:39,560 Speaker 1: mother figure and they can see someone exactly who they 446 00:24:39,600 --> 00:24:43,400 Speaker 1: want to live up to. Boys are desperately yearning right 447 00:24:43,440 --> 00:24:47,480 Speaker 1: now for male authority figures to teach them that it's 448 00:24:47,560 --> 00:24:50,960 Speaker 1: okay to be a boy. And I hear this, not 449 00:24:50,960 --> 00:24:53,920 Speaker 1: only yeah, it's not only okay to be a boy, 450 00:24:54,520 --> 00:24:56,560 Speaker 1: it should be proud of it. You should be proud 451 00:24:56,600 --> 00:25:00,359 Speaker 1: to become a man. Corren. Men are looking for people 452 00:25:00,640 --> 00:25:06,040 Speaker 1: to pattern themselves after. That involve strength, that involve courage, 453 00:25:06,080 --> 00:25:09,880 Speaker 1: that involve the kind of person that anybody would want 454 00:25:09,880 --> 00:25:13,520 Speaker 1: to have around, anybody would want to work with, being near, 455 00:25:14,280 --> 00:25:17,720 Speaker 1: having their community, having their society, and we are being 456 00:25:17,840 --> 00:25:21,520 Speaker 1: there is an active program to de masculine, demasculinize us, 457 00:25:21,520 --> 00:25:23,359 Speaker 1: if that is a word I can use, and we 458 00:25:23,400 --> 00:25:25,880 Speaker 1: can all see it. Just on the TikTok issue again, 459 00:25:25,920 --> 00:25:28,080 Speaker 1: to bring it full circle, I think one of the 460 00:25:28,119 --> 00:25:32,640 Speaker 1: things here that isn't being grappled with seriously enough play 461 00:25:32,760 --> 00:25:36,560 Speaker 1: is that the the algorithm is meant to be entirely 462 00:25:36,640 --> 00:25:40,240 Speaker 1: addictive based on what you watch. Yeah, how long are 463 00:25:40,240 --> 00:25:43,200 Speaker 1: you want it? How that? I mean, it's pretty clear 464 00:25:43,240 --> 00:25:46,080 Speaker 1: what where Andrew Tate is taking off on TikTok tu 465 00:25:46,200 --> 00:25:48,199 Speaker 1: buck Is. That's where a lot of these kids are 466 00:25:48,240 --> 00:25:51,600 Speaker 1: being exposed TikTok and YouTube. He is I think the 467 00:25:51,680 --> 00:25:56,040 Speaker 1: most popular, the most popular figure on these apps. So 468 00:25:56,080 --> 00:25:58,040 Speaker 1: he was all of them as you know, it's it's 469 00:25:58,080 --> 00:26:01,240 Speaker 1: pirate accounts. Yeah put it people who take things that 470 00:26:01,280 --> 00:26:03,719 Speaker 1: he said and sharing them and then they go viral. 471 00:26:04,080 --> 00:26:08,200 Speaker 1: So's he's still very Tate is still very much in 472 00:26:08,320 --> 00:26:12,120 Speaker 1: the minds of particularly young men who are by the way, 473 00:26:12,160 --> 00:26:14,719 Speaker 1: that's the apparatus wants to control young men more than 474 00:26:14,760 --> 00:26:18,440 Speaker 1: anything else, because who's really a problem for you in society. 475 00:26:18,119 --> 00:26:20,240 Speaker 1: When the young men all get upset with how a 476 00:26:20,320 --> 00:26:23,359 Speaker 1: society is going, things can happen. They want to be controlled. 477 00:26:23,359 --> 00:26:25,320 Speaker 1: They want them docile, they want them to shut up 478 00:26:25,359 --> 00:26:27,720 Speaker 1: and do as they're told. Right, that's the people that 479 00:26:27,760 --> 00:26:30,160 Speaker 1: are telling you to eat the bugs, have low tea, 480 00:26:30,760 --> 00:26:32,760 Speaker 1: cry all the time about all the bad things your 481 00:26:32,800 --> 00:26:36,560 Speaker 1: ancestors did or didn't do. Whatever. Those people want you 482 00:26:36,600 --> 00:26:39,160 Speaker 1: to be ashamed of your masculinity. But I just think 483 00:26:39,320 --> 00:26:42,120 Speaker 1: on the yeah, the Tate, there's some of that that's 484 00:26:42,119 --> 00:26:44,720 Speaker 1: happening on TikTok. But Clay they keep saying, Oh, they're 485 00:26:44,720 --> 00:26:48,560 Speaker 1: gonna brainwash our kids. The brainwashing is already occurring. And 486 00:26:48,600 --> 00:26:51,000 Speaker 1: what TikTok is actually showing people, if you get on 487 00:26:51,040 --> 00:26:53,239 Speaker 1: it and you start to pay attention to it, is 488 00:26:53,280 --> 00:26:57,760 Speaker 1: what American youth, through their schools, through their entertainment, through 489 00:26:58,000 --> 00:27:00,760 Speaker 1: what they want to see more of. Yeah, people seem 490 00:27:00,840 --> 00:27:02,560 Speaker 1: to think that this is just like there's not some 491 00:27:02,680 --> 00:27:05,159 Speaker 1: Chinese Communist Party figure who's showing up who was like 492 00:27:05,160 --> 00:27:08,120 Speaker 1: America is very evil. I'm like, you should hate America. 493 00:27:08,240 --> 00:27:10,600 Speaker 1: They're showing people things. Where does all the lives of 494 00:27:10,640 --> 00:27:13,879 Speaker 1: TikTok stuff come from? Obviously in the name. They're showing 495 00:27:13,880 --> 00:27:16,240 Speaker 1: what's already happening in this country, and kids are being 496 00:27:16,280 --> 00:27:19,080 Speaker 1: trained to believe by the society around them. This is 497 00:27:19,080 --> 00:27:20,719 Speaker 1: what I want to see more of. This is what 498 00:27:20,760 --> 00:27:23,600 Speaker 1: I believe in. This is so it's a mirror in 499 00:27:23,640 --> 00:27:26,879 Speaker 1: many ways to our current problems. As it is also 500 00:27:26,960 --> 00:27:29,639 Speaker 1: I think of an issue for long term programming in 501 00:27:29,680 --> 00:27:32,960 Speaker 1: CCP intervention. I'm not saying that's not a risk, but 502 00:27:33,119 --> 00:27:36,400 Speaker 1: right now it's showing us how actually destabilize a lot 503 00:27:36,400 --> 00:27:39,840 Speaker 1: of America's foundation has become. I saw a quote there's 504 00:27:39,840 --> 00:27:42,000 Speaker 1: a good New York magazine article. I need to send 505 00:27:42,000 --> 00:27:44,280 Speaker 1: you the link buck that I read while I was 506 00:27:44,320 --> 00:27:46,560 Speaker 1: over in Italy. You know, late at night, everybody's in 507 00:27:46,600 --> 00:27:50,040 Speaker 1: bed asleep, you're getting ready, and I'll kind of scroll 508 00:27:50,160 --> 00:27:53,119 Speaker 1: through and be reading all sorts of different stuff, and 509 00:27:53,320 --> 00:27:57,480 Speaker 1: there was an unbelievable quote from a left wing mom 510 00:27:57,800 --> 00:28:00,920 Speaker 1: in Brooklyn, and just to read this to you because 511 00:28:00,920 --> 00:28:04,200 Speaker 1: I think it's so great. As Tate was blowing up, 512 00:28:04,280 --> 00:28:06,760 Speaker 1: this is her kid. Well, i'll tell this story when 513 00:28:06,760 --> 00:28:08,560 Speaker 1: we come back, and we'll also take some of your calls, 514 00:28:08,600 --> 00:28:11,080 Speaker 1: because I think you guys are going to really enjoy 515 00:28:11,160 --> 00:28:12,880 Speaker 1: this story. Some of the quotes from this New York 516 00:28:12,920 --> 00:28:15,480 Speaker 1: magazine article are pretty incredible about Andrew Tate. But what 517 00:28:15,520 --> 00:28:17,680 Speaker 1: I would say as we go to break is there 518 00:28:17,800 --> 00:28:21,639 Speaker 1: is a desperate desire for masculinity in this country and 519 00:28:21,880 --> 00:28:26,560 Speaker 1: kids are craving it, young boys to a degree that 520 00:28:26,720 --> 00:28:30,359 Speaker 1: may never have existed in this country before. What figures 521 00:28:30,400 --> 00:28:34,000 Speaker 1: are going to fill the void in their lives that 522 00:28:34,040 --> 00:28:37,440 Speaker 1: they want to emulate. It's a big question. I think 523 00:28:37,480 --> 00:28:40,120 Speaker 1: for a lot of them, it's become Andrew Tate, which 524 00:28:40,160 --> 00:28:41,920 Speaker 1: is fascinating in and of itself, But I'll read these 525 00:28:42,000 --> 00:28:43,680 Speaker 1: quotes and we'll discuss. We'll take some of your calls 526 00:28:43,680 --> 00:28:45,400 Speaker 1: when we come back. In the meantime, it's at all 527 00:28:45,440 --> 00:28:47,920 Speaker 1: The Towers Foundation delivers on its promise to do good 528 00:28:47,920 --> 00:28:50,920 Speaker 1: and never forget the sacrifices America's heroes have made for US. 529 00:28:51,200 --> 00:28:54,480 Speaker 1: Heroes like US Air Force Senior Master Sergeant Israel del 530 00:28:54,520 --> 00:28:57,479 Speaker 1: Toro Junior. He promised his father he'd take care of 531 00:28:57,480 --> 00:29:00,320 Speaker 1: his family and his son, that he wouldn't grow up 532 00:29:00,360 --> 00:29:03,560 Speaker 1: without a father. When his humby exploded in Afghanistan, he 533 00:29:03,600 --> 00:29:06,400 Speaker 1: suffered severe burns over more than eighty percent of his 534 00:29:06,480 --> 00:29:10,040 Speaker 1: body with only a fifteen percent chance of survival, Del 535 00:29:10,080 --> 00:29:13,320 Speaker 1: Toro focused on the promises he made. Later, he became 536 00:29:13,360 --> 00:29:17,240 Speaker 1: the first fully disabled airmen permitted to re enlist and 537 00:29:17,600 --> 00:29:20,800 Speaker 1: a gold medal winner in the Invictus Games. Now he 538 00:29:20,880 --> 00:29:23,600 Speaker 1: shares his story to help others. Thanks to the Tunnel 539 00:29:23,640 --> 00:29:27,440 Speaker 1: to Towers Foundation and your generosity, del Toro received a 540 00:29:27,480 --> 00:29:31,480 Speaker 1: mortgage free home. It's a mortgage free smart home too 541 00:29:31,920 --> 00:29:35,120 Speaker 1: so important. Home has a variety of features to help 542 00:29:35,200 --> 00:29:38,800 Speaker 1: him live more independently. You can help heroes like del 543 00:29:38,840 --> 00:29:42,640 Speaker 1: Toro when they need it most. Both Buck and myself 544 00:29:42,760 --> 00:29:46,160 Speaker 1: donate eleven dollars a month to Tunnel to Towers. Please 545 00:29:46,160 --> 00:29:48,800 Speaker 1: do the same and join us. Go online to t 546 00:29:49,080 --> 00:29:53,800 Speaker 1: twot dot org. That's t the number two t dot org. 547 00:29:54,400 --> 00:29:58,520 Speaker 1: He gets to know the guys outside the issues with 548 00:29:58,760 --> 00:30:01,960 Speaker 1: Clay and Buck podcasts. Fight it on the iHeart app 549 00:30:02,080 --> 00:30:05,520 Speaker 1: or wherever you get your podcasts. Welcome back to Play 550 00:30:05,520 --> 00:30:07,840 Speaker 1: and Buck. We got a bunch of lines lit up, 551 00:30:08,000 --> 00:30:10,240 Speaker 1: people fired up about this one. I think on a 552 00:30:10,320 --> 00:30:12,880 Speaker 1: number of fronts. One because when it comes to TikTok, 553 00:30:12,920 --> 00:30:16,680 Speaker 1: we all know that social media has massive you can 554 00:30:16,680 --> 00:30:18,640 Speaker 1: opt out, and when it comes to TikTok, I would 555 00:30:18,720 --> 00:30:22,680 Speaker 1: I would wage your ninety nine percent of our esteemed 556 00:30:22,680 --> 00:30:26,320 Speaker 1: and beloved radio audience is not actively on TikTok. Maybe 557 00:30:26,920 --> 00:30:30,960 Speaker 1: ninety five, but ninety five to ninety nine percent, But 558 00:30:31,040 --> 00:30:36,080 Speaker 1: one hundred and fifty million Americans are. They're overwhelmingly very young. 559 00:30:36,320 --> 00:30:39,960 Speaker 1: It is left wing dominated. You would, And I did 560 00:30:40,000 --> 00:30:42,360 Speaker 1: try to post a little bit during COVID Clay I 561 00:30:42,440 --> 00:30:45,280 Speaker 1: got shut down. Couldn't talk about masks, couldn't you know, 562 00:30:45,320 --> 00:30:49,280 Speaker 1: I got strikes. So it was just as sensorious on 563 00:30:49,360 --> 00:30:53,280 Speaker 1: COVID stuff as as the other platforms but I think 564 00:30:53,280 --> 00:30:55,480 Speaker 1: people it's really affecting because they realize something's going on 565 00:30:55,560 --> 00:30:58,120 Speaker 1: with kids. Kids will sit I mean I see this 566 00:30:58,200 --> 00:31:01,120 Speaker 1: in airports and restaurants. They will sit and they will 567 00:31:01,120 --> 00:31:04,760 Speaker 1: go through TikTok for an hour, two hours, They'll they'll 568 00:31:04,800 --> 00:31:06,640 Speaker 1: just sit there, just going through, going through, going through. 569 00:31:06,640 --> 00:31:09,400 Speaker 1: I mean, and you know, look, people said the same 570 00:31:09,440 --> 00:31:11,680 Speaker 1: thing about Facebook when it started out, and Facebook was 571 00:31:11,840 --> 00:31:14,960 Speaker 1: very sticky, so to speak, as a platform. Originally, people 572 00:31:15,000 --> 00:31:17,040 Speaker 1: spent you know, tons of time there. The average was 573 00:31:17,080 --> 00:31:19,520 Speaker 1: like two hours a day on faith. I haven't. I mean, 574 00:31:19,760 --> 00:31:23,360 Speaker 1: Facebook has fallen off tremendously from what it used to be. Anyway, 575 00:31:23,400 --> 00:31:24,560 Speaker 1: we got do you want to get some of these 576 00:31:24,560 --> 00:31:26,800 Speaker 1: calls because people let me read this. We read this 577 00:31:26,880 --> 00:31:29,600 Speaker 1: quote from Andrew Tate. From this, we're also gonna pull 578 00:31:29,640 --> 00:31:31,720 Speaker 1: some some Tate audios. People can hear some of the stuff. 579 00:31:31,760 --> 00:31:34,800 Speaker 1: Go ahead if you if you aren't familiar with Andrew Tate, Um, 580 00:31:34,800 --> 00:31:37,440 Speaker 1: we will kind of familiarize you with it. But he 581 00:31:37,640 --> 00:31:40,400 Speaker 1: is as popular if you are. If you have a 582 00:31:40,480 --> 00:31:42,760 Speaker 1: kid twelve to seventeen years old, there's almost one hundred 583 00:31:42,760 --> 00:31:46,160 Speaker 1: percent chance they will know it. Um. And here as 584 00:31:46,240 --> 00:31:50,440 Speaker 1: Tate was blowing up. This is a left wing mom 585 00:31:50,600 --> 00:31:54,600 Speaker 1: who becomes concerned that her son is becoming a huge 586 00:31:54,640 --> 00:31:57,920 Speaker 1: fan of Andrew Tate. As Tate was blowing up, Ruby 587 00:31:58,000 --> 00:32:01,400 Speaker 1: made the connection and grew concerned even more so when 588 00:32:01,520 --> 00:32:05,320 Speaker 1: Charlie told her Tate was being interviewed on Tucker Carlson 589 00:32:05,600 --> 00:32:09,560 Speaker 1: and suggested they watched together. Funny, this is the quote 590 00:32:09,560 --> 00:32:11,760 Speaker 1: that I thought you would love, buck, She and her 591 00:32:11,840 --> 00:32:15,840 Speaker 1: husband said, No, here's her quote. I'm like a Brooklyn 592 00:32:16,000 --> 00:32:20,800 Speaker 1: far left person. If Tucker Carlson interviewed Barack Obama, who's 593 00:32:20,880 --> 00:32:24,400 Speaker 1: probably my favorite person on Earth, I wouldn't watch that. 594 00:32:24,920 --> 00:32:27,880 Speaker 1: Charlie her Son tried to explain, which I loved that quote. 595 00:32:28,040 --> 00:32:31,360 Speaker 1: Charlie Brooklyn Mom, Barack Obama's her favorite person on Earth. 596 00:32:31,600 --> 00:32:35,600 Speaker 1: Charlie tried to explain, saying that liking Tate had nothing 597 00:32:35,600 --> 00:32:38,920 Speaker 1: to do with politics, but he would watch these videos 598 00:32:38,920 --> 00:32:40,720 Speaker 1: all the times. Mom said one of the things that 599 00:32:40,760 --> 00:32:43,720 Speaker 1: she noticed was he became far more self reliant, and 600 00:32:43,760 --> 00:32:46,080 Speaker 1: he started doing pushups and sit ups all the time. 601 00:32:46,320 --> 00:32:48,720 Speaker 1: Because that's one of the things that Andrew Tate argues 602 00:32:48,800 --> 00:32:52,560 Speaker 1: is you have to have your mind and body physically strong. 603 00:32:53,200 --> 00:32:55,680 Speaker 1: And what is interesting to me, buck Is is you 604 00:32:55,720 --> 00:32:57,880 Speaker 1: talk about what's being banned and what kids are seeing. 605 00:32:58,360 --> 00:33:03,640 Speaker 1: Boys young boys are being influenced by people like Andrew 606 00:33:03,760 --> 00:33:06,320 Speaker 1: Tate that might be different than who you think they 607 00:33:06,360 --> 00:33:09,160 Speaker 1: would be influenced by. And some of the lessons of 608 00:33:09,240 --> 00:33:13,800 Speaker 1: Andrew Tate agree with Yes, we'll also I think it's 609 00:33:13,840 --> 00:33:17,920 Speaker 1: interesting that they hold everyone to this, that this standard 610 00:33:18,520 --> 00:33:20,720 Speaker 1: that you know, they'll they'll say, well do you agree 611 00:33:20,720 --> 00:33:23,120 Speaker 1: with this, and they'll pull something where Tate used to 612 00:33:23,160 --> 00:33:27,040 Speaker 1: be as he was becoming a more famous individual. Um, 613 00:33:27,080 --> 00:33:29,240 Speaker 1: they'll pull something where he said, you know, he said 614 00:33:29,280 --> 00:33:32,480 Speaker 1: stuff that was for me would be would does cross 615 00:33:32,520 --> 00:33:34,840 Speaker 1: the line. It's something I wouldn't No one's saying that 616 00:33:34,880 --> 00:33:36,520 Speaker 1: you agree with that. You know there's other people who 617 00:33:36,600 --> 00:33:40,200 Speaker 1: were per with anybody right right. It's like, you know, 618 00:33:40,280 --> 00:33:45,600 Speaker 1: Charles Barkley has some incredible free Mike Tyson. Okay, Mike Tyson, 619 00:33:45,640 --> 00:33:47,680 Speaker 1: everybody has a plan until they're punched the face A 620 00:33:47,840 --> 00:33:51,880 Speaker 1: brilliant and timeless phrase. YEA, that doesn't mean that I 621 00:33:51,960 --> 00:33:55,240 Speaker 1: agree with everything Mike Tyson says. Everything Mike Tyson does. 622 00:33:55,560 --> 00:33:58,320 Speaker 1: You know Mike Tyson's past, but the man does have 623 00:33:58,920 --> 00:34:01,120 Speaker 1: some wisdom that has been passed onto a whole lot 624 00:34:01,160 --> 00:34:02,680 Speaker 1: of people. So I think it's interesting that they do 625 00:34:02,720 --> 00:34:05,040 Speaker 1: this thing where when you say, well, I like the 626 00:34:05,080 --> 00:34:09,000 Speaker 1: following tenets of Tate if you will, they say, well, 627 00:34:09,000 --> 00:34:10,920 Speaker 1: he did all this other stuff, or they say he 628 00:34:11,000 --> 00:34:14,160 Speaker 1: was involved in in webcam businesses for example, which you 629 00:34:14,160 --> 00:34:16,560 Speaker 1: know if you read any stories. Yeah, I think that's gross. 630 00:34:16,680 --> 00:34:19,800 Speaker 1: I think that's exploitative. I opposed that one hundred percent. 631 00:34:20,440 --> 00:34:21,839 Speaker 1: But does that mean that I have to say he's 632 00:34:21,880 --> 00:34:24,120 Speaker 1: wrong when he tells somebody to do a lot of 633 00:34:24,120 --> 00:34:25,799 Speaker 1: sit ups and push ups and be a man and 634 00:34:25,840 --> 00:34:29,440 Speaker 1: defend women you don't like. I think it's interesting because 635 00:34:30,680 --> 00:34:34,319 Speaker 1: is saying this stuff, and what strikes my kids and 636 00:34:34,360 --> 00:34:36,520 Speaker 1: what they respond to so well and their friends buck, 637 00:34:36,680 --> 00:34:40,680 Speaker 1: is they live in such a fear laden society that 638 00:34:40,880 --> 00:34:43,520 Speaker 1: Andrew Tate is one of the people out there saying 639 00:34:43,560 --> 00:34:45,800 Speaker 1: you need to toughen up, you need to be able 640 00:34:45,840 --> 00:34:49,359 Speaker 1: to handle difficulties in life, because the real world ain't all, 641 00:34:49,600 --> 00:34:53,440 Speaker 1: you know, rainbows and unicorns. Yeah. I think that's especially 642 00:34:53,440 --> 00:34:56,080 Speaker 1: for young men these days. I want to say, we've 643 00:34:56,080 --> 00:34:59,120 Speaker 1: got the Tate clip here. Is just the kind of 644 00:34:59,160 --> 00:35:01,239 Speaker 1: stuff that clan our talking about when we're picking and 645 00:35:01,320 --> 00:35:03,759 Speaker 1: choosing from what he says that has some wisdom to it. 646 00:35:03,760 --> 00:35:07,000 Speaker 1: Plate twenty nine. The baseline of masculinity as a whole 647 00:35:07,520 --> 00:35:09,399 Speaker 1: is the thing that makes a good man a man 648 00:35:09,880 --> 00:35:12,040 Speaker 1: is that he does what he doesn't want to do. 649 00:35:12,200 --> 00:35:14,279 Speaker 1: He doesn't want to work, and he works anyway. He 650 00:35:14,320 --> 00:35:16,360 Speaker 1: doesn't want to go to war, and he fights anyway. 651 00:35:16,480 --> 00:35:18,320 Speaker 1: He doesn't want to get up, he gets up anyway. 652 00:35:18,320 --> 00:35:19,680 Speaker 1: That's the whole point of it. We didn't want to 653 00:35:19,719 --> 00:35:22,600 Speaker 1: die in the Titanic. Guess what happened. We died the Titanic. 654 00:35:22,840 --> 00:35:24,080 Speaker 1: You can't sit there as a man and say you 655 00:35:24,120 --> 00:35:25,839 Speaker 1: don't feel like it. You're not allowed to not feel 656 00:35:25,840 --> 00:35:29,560 Speaker 1: like it. You're supposed to do it anyway regardless. So 657 00:35:31,040 --> 00:35:33,640 Speaker 1: I sit here, I say that people need to hear that. Yeah, 658 00:35:33,640 --> 00:35:38,000 Speaker 1: that's that's important. It's a great message for young men. 659 00:35:38,480 --> 00:35:41,319 Speaker 1: And you sit here and you say, okay. Also, I 660 00:35:41,320 --> 00:35:43,279 Speaker 1: want to you know, if it wasn't Biden, if it 661 00:35:43,320 --> 00:35:45,080 Speaker 1: was Trump in the White House, I know it's Romania 662 00:35:45,120 --> 00:35:47,800 Speaker 1: that's holding him. The timing does seem suspicious to a 663 00:35:47,880 --> 00:35:51,840 Speaker 1: lot of people. Would there be a black US citizen, 664 00:35:51,920 --> 00:35:53,919 Speaker 1: which is Tate is a US citizen. By the way, 665 00:35:54,080 --> 00:35:57,319 Speaker 1: would it be a black American held in like the 666 00:35:57,440 --> 00:36:01,160 Speaker 1: Romanian gulag for months on end with no actual presentation 667 00:36:01,200 --> 00:36:04,520 Speaker 1: of charges, without a word of protest from the State Department. Really, 668 00:36:05,160 --> 00:36:07,680 Speaker 1: just thinking about the way Britney Griner was defended in 669 00:36:07,719 --> 00:36:10,200 Speaker 1: Russia compared to the way that Andrew Tate is being 670 00:36:10,200 --> 00:36:15,200 Speaker 1: defended in Romania. I mean, Britney Griner admitted that she 671 00:36:15,280 --> 00:36:19,359 Speaker 1: took drugs into Russia because she's a black lesbian who 672 00:36:19,440 --> 00:36:23,400 Speaker 1: voted for Joe Biden. The entire Democrat apparatus demanded that 673 00:36:23,440 --> 00:36:28,000 Speaker 1: she get out of prison. Andrew Tate American citizen mixed race, 674 00:36:28,080 --> 00:36:30,960 Speaker 1: Like you SAIDs dad's black, mom's white. It's almost like 675 00:36:31,080 --> 00:36:33,960 Speaker 1: nobody cares that he's being held with no charges in Romania. 676 00:36:34,000 --> 00:36:38,040 Speaker 1: Seems kind of counterproductive. And also it's like they're not 677 00:36:38,080 --> 00:36:40,520 Speaker 1: going to trade the Lord of War for Tate? Are 678 00:36:40,600 --> 00:36:43,680 Speaker 1: they remember that? That's what they did Victor Bout they 679 00:36:43,800 --> 00:36:47,000 Speaker 1: decided to make a trade. There's no trade going on here. 680 00:36:47,040 --> 00:36:49,239 Speaker 1: But also there are no charges because he's being held 681 00:36:49,280 --> 00:36:52,600 Speaker 1: indefinitely in a Romanian prison and he maintains he's one 682 00:36:52,640 --> 00:36:55,280 Speaker 1: hundred percent innocent, and a lot of people are saying 683 00:36:55,600 --> 00:36:58,560 Speaker 1: other politics behind this decision that we're not even seeing. 684 00:36:59,080 --> 00:36:59,920 Speaker 1: We'll look into more of the