1 00:00:04,440 --> 00:00:12,280 Speaker 1: Welcome to tech Stuff, a production from iHeartRadio. Hey there, 2 00:00:12,280 --> 00:00:15,520 Speaker 1: and welcome to tech Stuff. I'm your host, Jonathan Strickland. 3 00:00:15,560 --> 00:00:18,639 Speaker 1: I'm an executive producer with iHeart Podcasts and how the 4 00:00:18,760 --> 00:00:21,400 Speaker 1: tech are you. It's time for some tech news that 5 00:00:21,440 --> 00:00:24,880 Speaker 1: caught my eye this week. We are on February twenty third, 6 00:00:25,120 --> 00:00:29,400 Speaker 1: twenty twenty four. Let's get to it now. Yesterday was 7 00:00:29,560 --> 00:00:32,360 Speaker 1: a bit of a doozy for cellular carriers here in 8 00:00:32,360 --> 00:00:36,520 Speaker 1: the United States. Disruptions and service plagued AT and T customers, 9 00:00:36,560 --> 00:00:40,839 Speaker 1: and reportedly some folks who patronize other carriers had similar problems. 10 00:00:41,080 --> 00:00:44,479 Speaker 1: It was a massive outage and it lasted for several hours. 11 00:00:44,880 --> 00:00:48,160 Speaker 1: Those with Wi Fi capable phones could still make calls 12 00:00:48,159 --> 00:00:51,560 Speaker 1: over Wi Fi networks, but once out of the electromagnetic 13 00:00:51,560 --> 00:00:55,280 Speaker 1: embrace of the router, they found themselves without servants. Speculation 14 00:00:55,480 --> 00:00:58,520 Speaker 1: range from a technical glitch to a possible cyber attack 15 00:00:58,560 --> 00:01:01,680 Speaker 1: as being the cause of the issue. That last possibility 16 00:01:01,720 --> 00:01:05,160 Speaker 1: got the authorities involved, as more than seventy thousand dollars 17 00:01:05,240 --> 00:01:09,720 Speaker 1: reports flooded in from cities like Houston and Dallas and Texas, 18 00:01:09,920 --> 00:01:13,480 Speaker 1: Los Angeles. Out in California, Chicago and Illinois and my 19 00:01:13,640 --> 00:01:17,040 Speaker 1: hometown of Atlanta, Georgia. As for those other carriers, T 20 00:01:17,240 --> 00:01:20,240 Speaker 1: Mobile and Cricket Wireless reported some of their customers had 21 00:01:20,280 --> 00:01:23,959 Speaker 1: also mentioned a loss of service, and curiously, it seemed 22 00:01:23,959 --> 00:01:26,480 Speaker 1: to happen after those customers had tried to call AT 23 00:01:26,560 --> 00:01:29,160 Speaker 1: and T customers. I should also add that Cricket Wireless 24 00:01:29,160 --> 00:01:32,479 Speaker 1: actually piggybacks off of AT and t's network, so maybe 25 00:01:32,480 --> 00:01:34,959 Speaker 1: it's not such a big surprise that they had some 26 00:01:35,160 --> 00:01:39,480 Speaker 1: issues with their customers. The Federal Communications Commission, or FCC, 27 00:01:40,000 --> 00:01:43,640 Speaker 1: has launched an investigation into the incident. The US Cybersecurity 28 00:01:43,640 --> 00:01:46,560 Speaker 1: and Infrastructure Security Agency also looked into it, plus the 29 00:01:46,760 --> 00:01:50,960 Speaker 1: Department of Homeland Security started to investigate it. But this morning, 30 00:01:51,040 --> 00:01:53,600 Speaker 1: AT and T says the problem that came down to, quote, 31 00:01:53,680 --> 00:01:57,400 Speaker 1: the application and execution of an incorrect process used as 32 00:01:57,440 --> 00:02:00,880 Speaker 1: we were expanding our network end quote. That's according to 33 00:02:00,920 --> 00:02:03,560 Speaker 1: an actual AT and T release. This lines up with 34 00:02:03,760 --> 00:02:07,680 Speaker 1: an earlier analysis from the Cybersecurity Agency. They said that 35 00:02:07,760 --> 00:02:10,600 Speaker 1: the most likely cause of the issue was a fault 36 00:02:10,639 --> 00:02:14,240 Speaker 1: in a cloud computing configuration error. While it took hours 37 00:02:14,280 --> 00:02:16,600 Speaker 1: to restore service to all customers, AT and T says 38 00:02:16,680 --> 00:02:19,880 Speaker 1: that process is now complete so hopefully if you have 39 00:02:19,919 --> 00:02:22,079 Speaker 1: AT and T you don't have any issues right now. 40 00:02:22,600 --> 00:02:25,920 Speaker 1: And now a quick update on Cruise. Not Tom Cruise, 41 00:02:26,160 --> 00:02:29,280 Speaker 1: but rather the autonomous taxi company that famously got in 42 00:02:29,360 --> 00:02:32,320 Speaker 1: trouble when one of its vehicles struck and then dragged 43 00:02:32,360 --> 00:02:35,320 Speaker 1: a pedestrian for twenty feet or so while the car 44 00:02:35,360 --> 00:02:40,040 Speaker 1: was pulling over, and then allegedly Cruz was attempting to 45 00:02:40,160 --> 00:02:43,400 Speaker 1: cover up the incident by not sharing their video of 46 00:02:43,440 --> 00:02:47,080 Speaker 1: the actual accident. Anyway, the state of California has revoked 47 00:02:47,120 --> 00:02:50,920 Speaker 1: Cruise's license from operating within that state, but the company 48 00:02:51,000 --> 00:02:54,160 Speaker 1: is looking at other cities and is reportedly preparing to 49 00:02:54,200 --> 00:02:58,200 Speaker 1: return to service with actual drivers in the seats as well. 50 00:02:58,400 --> 00:03:01,880 Speaker 1: Spokesman Pat Morrissey said, our goal is to relaunch in 51 00:03:01,960 --> 00:03:07,120 Speaker 1: one city with manually driven vehicles and supervised testing as 52 00:03:07,200 --> 00:03:10,040 Speaker 1: soon as possible, once we have taken steps to rebuild 53 00:03:10,080 --> 00:03:13,040 Speaker 1: trust with regulators and the public. End quote. Now. I 54 00:03:13,040 --> 00:03:15,800 Speaker 1: would like to think that our trust would require a 55 00:03:15,840 --> 00:03:18,800 Speaker 1: lot of work to rebuild, but I just don't know, 56 00:03:18,919 --> 00:03:22,280 Speaker 1: because some folks just love them sell some technology. But 57 00:03:22,320 --> 00:03:24,880 Speaker 1: then again, Cruz has to turn things around if it 58 00:03:24,960 --> 00:03:27,720 Speaker 1: is to remain a viable business. Since that accident, the 59 00:03:27,760 --> 00:03:31,440 Speaker 1: company has had several top leaders jump ship, and it 60 00:03:31,480 --> 00:03:35,280 Speaker 1: has also downsized more than twenty percent. So they definitely 61 00:03:35,280 --> 00:03:37,560 Speaker 1: need to make things start to happen, or else there's 62 00:03:37,560 --> 00:03:40,560 Speaker 1: not going to be a company anymore. Google is back 63 00:03:40,560 --> 00:03:44,280 Speaker 1: in the hot seat because it's newly renamed AI, which 64 00:03:44,320 --> 00:03:47,000 Speaker 1: is now known as Gemini, has a tendency to go 65 00:03:47,080 --> 00:03:49,920 Speaker 1: a little overboard for the sake of diversity. Now, I 66 00:03:49,920 --> 00:03:53,760 Speaker 1: think most of us understand that in most contexts, diversity 67 00:03:54,120 --> 00:03:57,560 Speaker 1: is important. It brings fresh perspectives and methods to bear. 68 00:03:58,040 --> 00:04:02,200 Speaker 1: Generally speaking, we all benefit when we concentrate on making 69 00:04:02,240 --> 00:04:07,080 Speaker 1: sure we are championing diversity. But if you wanted to 70 00:04:07,120 --> 00:04:09,680 Speaker 1: create a historically accurate portrayal of a group of people, 71 00:04:09,760 --> 00:04:13,640 Speaker 1: such as, oh, I don't know Nazi soldiers, you might 72 00:04:13,680 --> 00:04:16,840 Speaker 1: be shocked if the AI generated image included people of 73 00:04:16,920 --> 00:04:22,680 Speaker 1: color dressed in Nazi uniforms. The Nazis famously were not inclusive. 74 00:04:23,040 --> 00:04:25,280 Speaker 1: This sounds a bit like Godwin's law, right, I mean, 75 00:04:25,279 --> 00:04:27,719 Speaker 1: that's the law that says any conversation on the Internet 76 00:04:27,800 --> 00:04:31,920 Speaker 1: will inevitably gravitate toward Hitler as folks disagree with one another. 77 00:04:32,200 --> 00:04:35,360 Speaker 1: But the genesis of the complaints gets a little bit 78 00:04:35,360 --> 00:04:39,120 Speaker 1: more murky because they start in the anti diversity crowd, 79 00:04:39,480 --> 00:04:43,360 Speaker 1: or folks who are kind of like white rights people, 80 00:04:43,520 --> 00:04:46,760 Speaker 1: or they just are I guess they would position themselves 81 00:04:46,800 --> 00:04:49,679 Speaker 1: as anti woke. But these were people who are complaining 82 00:04:49,680 --> 00:04:52,279 Speaker 1: that if you asked Gemini to create images of like, say, 83 00:04:52,360 --> 00:04:55,359 Speaker 1: an American woman, that it would create some images that 84 00:04:55,400 --> 00:04:57,640 Speaker 1: included people of color, and maybe you'd end up with 85 00:04:57,680 --> 00:05:00,000 Speaker 1: more people of color than white people. Maybe you'd even 86 00:05:00,120 --> 00:05:02,080 Speaker 1: end up with a group of people and there's no 87 00:05:02,640 --> 00:05:06,440 Speaker 1: white person represented in that group. This has led to 88 00:05:06,839 --> 00:05:09,760 Speaker 1: some of these people arguing that Google is pushing anti 89 00:05:09,839 --> 00:05:13,560 Speaker 1: white propaganda on the public, which, you know, let's be serious. 90 00:05:13,600 --> 00:05:15,919 Speaker 1: I mean, the examples I've looked at at least have 91 00:05:16,440 --> 00:05:19,760 Speaker 1: had white folks in those examples. It's not like they were, 92 00:05:20,279 --> 00:05:23,440 Speaker 1: you know, completely left out. And goodness knows, America has 93 00:05:23,480 --> 00:05:26,880 Speaker 1: a very long history of producing media that studiously ignored 94 00:05:26,920 --> 00:05:29,400 Speaker 1: the fact that people who are not white also live 95 00:05:29,440 --> 00:05:33,120 Speaker 1: in America. See also the sitcom Friends. You would think 96 00:05:33,160 --> 00:05:37,600 Speaker 1: that New York was almost entirely white based on that show. Anyway, 97 00:05:38,120 --> 00:05:41,240 Speaker 1: the Nazi thing indicates that perhaps Google has been a 98 00:05:41,240 --> 00:05:46,000 Speaker 1: little heavy handed when trying to force diversity into image generation. Right, Like, 99 00:05:46,040 --> 00:05:49,560 Speaker 1: there are different ways to have diversity work its way 100 00:05:49,600 --> 00:05:53,760 Speaker 1: into your product, and brute force is not necessarily the 101 00:05:53,800 --> 00:05:57,200 Speaker 1: best way. You can almost imagine that the algorithm has 102 00:05:57,240 --> 00:05:59,640 Speaker 1: a little like switch at the end of the sequence, 103 00:05:59,760 --> 00:06:02,800 Speaker 1: as is the group of people that you've created diverse 104 00:06:02,920 --> 00:06:06,400 Speaker 1: if not insert people of color. Now, obviously that is 105 00:06:06,440 --> 00:06:09,960 Speaker 1: not appropriate in every context, and honestly it points out 106 00:06:10,080 --> 00:06:13,400 Speaker 1: that true diversity and fairness is really hard to achieve, 107 00:06:13,720 --> 00:06:17,160 Speaker 1: even if we're just talking virtually. That doesn't mean that 108 00:06:17,200 --> 00:06:19,800 Speaker 1: we shouldn't bust our butts to try and achieve it. 109 00:06:20,200 --> 00:06:23,240 Speaker 1: And I would recommend Adam Robertson's article on the Verge 110 00:06:23,279 --> 00:06:28,200 Speaker 1: titled Google apologizes for missing the mark after Gemini generated 111 00:06:28,360 --> 00:06:31,640 Speaker 1: racially diverse Nazis. If you'd like to read more, it's 112 00:06:31,680 --> 00:06:35,280 Speaker 1: a very good article. In other Google Gemini news, Google 113 00:06:35,320 --> 00:06:39,440 Speaker 1: incorporated an AI writing generator into Google Chrome this week. 114 00:06:39,760 --> 00:06:42,840 Speaker 1: Now to use this, you actually first must enable it. 115 00:06:43,080 --> 00:06:44,919 Speaker 1: You have to go into settings to do this, and 116 00:06:44,960 --> 00:06:47,400 Speaker 1: I did that this morning, but I haven't really had 117 00:06:47,400 --> 00:06:49,520 Speaker 1: a chance to take this out for a spin to 118 00:06:49,560 --> 00:06:51,800 Speaker 1: see what it's like. I'll be sure to report back 119 00:06:51,839 --> 00:06:54,240 Speaker 1: once I've had some time with it. I did try 120 00:06:54,600 --> 00:06:58,000 Speaker 1: the tab organizer option. There's an option that's supposed to 121 00:06:58,320 --> 00:07:01,719 Speaker 1: reorganize your tab so that they make sense, but I 122 00:07:01,760 --> 00:07:05,080 Speaker 1: had already kind of grouped my tabs into relevant research. 123 00:07:05,520 --> 00:07:08,479 Speaker 1: So I guess there just wasn't anything left for Gemini 124 00:07:08,520 --> 00:07:11,920 Speaker 1: to organize because nothing happened. But you know, it's either 125 00:07:11,920 --> 00:07:14,679 Speaker 1: that or it just didn't work anyway. You can also 126 00:07:14,720 --> 00:07:17,880 Speaker 1: do things like you can click within a text field online. 127 00:07:18,120 --> 00:07:21,200 Speaker 1: Let's say you found an online form, like maybe you're 128 00:07:21,280 --> 00:07:24,520 Speaker 1: leaving a review for a product, and you can choose 129 00:07:24,840 --> 00:07:28,960 Speaker 1: help me write by right clicking, you know, write write. 130 00:07:29,520 --> 00:07:33,239 Speaker 1: That will get Gemini to either write or rewrite your stuff. 131 00:07:33,560 --> 00:07:36,520 Speaker 1: So maybe you want to sound particularly educated while you're 132 00:07:36,560 --> 00:07:39,000 Speaker 1: writing your next review about a pair of sneakers or whatever, 133 00:07:39,240 --> 00:07:42,160 Speaker 1: you can use this option. I have not tried this yet, 134 00:07:42,200 --> 00:07:45,800 Speaker 1: so I don't know what it's like, but it's good 135 00:07:45,800 --> 00:07:47,760 Speaker 1: to note that if you do use this feature, Google 136 00:07:47,800 --> 00:07:51,480 Speaker 1: will collect some data about the whole process, like what 137 00:07:51,640 --> 00:07:55,160 Speaker 1: site you're on and what you wrote or had Gemini 138 00:07:55,280 --> 00:07:59,400 Speaker 1: write for you, and further this then could be reviewed 139 00:07:59,400 --> 00:08:02,800 Speaker 1: by human beings further along the chain, as a way 140 00:08:02,840 --> 00:08:04,680 Speaker 1: to make sure that Gemini is on track and is 141 00:08:04,720 --> 00:08:07,880 Speaker 1: improving over time, So there could be human review added 142 00:08:07,920 --> 00:08:11,040 Speaker 1: into this, So for the privacy minded, you might want 143 00:08:11,080 --> 00:08:14,600 Speaker 1: to think twice before you activate this feature, you know, 144 00:08:14,920 --> 00:08:18,800 Speaker 1: and otherwise the things you write are possibly going to 145 00:08:18,800 --> 00:08:21,559 Speaker 1: be reviewed by human beings, and if you're writing stuff 146 00:08:21,560 --> 00:08:24,360 Speaker 1: that you'd rather human beings not actually see, which I 147 00:08:24,400 --> 00:08:26,080 Speaker 1: don't know why you'd be writing on the internet if 148 00:08:26,080 --> 00:08:29,040 Speaker 1: that's the case, but maybe don't use this. The website 149 00:08:29,080 --> 00:08:31,120 Speaker 1: Reddit is about to go public, so it's about to 150 00:08:31,160 --> 00:08:34,240 Speaker 1: hold its initial public offering, as in, you know, become 151 00:08:34,400 --> 00:08:37,400 Speaker 1: a publicly traded company. This is a really big deal 152 00:08:37,440 --> 00:08:40,600 Speaker 1: for several reasons. Reddit the site that lets users create 153 00:08:40,640 --> 00:08:43,480 Speaker 1: and manage pages called subreddit. It's about pretty much any 154 00:08:43,520 --> 00:08:46,400 Speaker 1: topic you can imagine, and probably a few you can't. 155 00:08:47,000 --> 00:08:50,160 Speaker 1: It has served as a sort of gathering ground for 156 00:08:50,280 --> 00:08:52,920 Speaker 1: interesting links to all sorts of stuff online, and it's 157 00:08:52,920 --> 00:08:55,880 Speaker 1: been around for years, but it has never turned a profit. 158 00:08:56,360 --> 00:09:00,200 Speaker 1: In fact, it lost more than ninety million dollars last year. However, 159 00:09:00,240 --> 00:09:02,720 Speaker 1: it did post a twenty one percent growth in revenue 160 00:09:02,840 --> 00:09:05,800 Speaker 1: year over year, so it made more money, which meant 161 00:09:05,840 --> 00:09:11,120 Speaker 1: it lost less money overall. It also recently went through 162 00:09:11,120 --> 00:09:14,240 Speaker 1: a really rocky situation where the company changed its Application 163 00:09:14,320 --> 00:09:18,080 Speaker 1: Programmer Interface policy or API policy, and in the process 164 00:09:18,120 --> 00:09:21,719 Speaker 1: it alienated a whole bunch of its user base. You know, 165 00:09:21,800 --> 00:09:25,920 Speaker 1: Reddit saw users deliberately sabotaged several popular subreddits as a 166 00:09:25,920 --> 00:09:28,720 Speaker 1: result of that, and the experience left many Reddit users 167 00:09:28,840 --> 00:09:31,439 Speaker 1: sour on the site. Now that being said, Reddit has 168 00:09:31,520 --> 00:09:34,200 Speaker 1: offered something of an olive branch, at least to the 169 00:09:34,240 --> 00:09:37,120 Speaker 1: most prolific Reddit users. They will have the chance to 170 00:09:37,120 --> 00:09:39,680 Speaker 1: purchase shares and Reddit before it goes active on the 171 00:09:39,679 --> 00:09:43,400 Speaker 1: stock market. That means you can potentially buy shares at 172 00:09:43,440 --> 00:09:45,560 Speaker 1: a lower price than what it will be when it 173 00:09:45,760 --> 00:09:49,280 Speaker 1: is trading on the stock market, though that's not a guarantee. 174 00:09:49,480 --> 00:09:53,800 Speaker 1: They can be pretty exclusive, as you know insiders in 175 00:09:53,880 --> 00:09:57,280 Speaker 1: this case. So there might be some prolific redditors out 176 00:09:57,280 --> 00:09:59,480 Speaker 1: there who still hold a grudge against Reddit for those 177 00:09:59,559 --> 00:10:02,680 Speaker 1: changes the API. But I have seen people really fall 178 00:10:02,760 --> 00:10:04,480 Speaker 1: over themselves for the chance to be part of an 179 00:10:04,480 --> 00:10:07,440 Speaker 1: exclusive group, even if that group isn't really impressive. So 180 00:10:07,720 --> 00:10:09,840 Speaker 1: I'm guessing a lot of the folks who have the 181 00:10:09,920 --> 00:10:13,280 Speaker 1: opportunity to make those purchases will in the interest of 182 00:10:13,280 --> 00:10:16,840 Speaker 1: full disclosure, I'm definitively not one of those people. I 183 00:10:16,920 --> 00:10:19,480 Speaker 1: use Reddit for research, but I do so pretty much 184 00:10:19,720 --> 00:10:23,040 Speaker 1: only as a lurker. I think I may have posted 185 00:10:23,040 --> 00:10:26,520 Speaker 1: on Reddit perhaps half a dozen times total across my 186 00:10:26,760 --> 00:10:29,640 Speaker 1: entire history with the site, and it's almost always as 187 00:10:29,640 --> 00:10:32,800 Speaker 1: a comment to an existing thread. But in other interesting news, 188 00:10:32,840 --> 00:10:37,160 Speaker 1: Open Ai co founder and former CEO and then reinstated CEO, 189 00:10:37,559 --> 00:10:41,000 Speaker 1: Sam Altman is one of the top shareholders on Reddit. 190 00:10:41,240 --> 00:10:43,600 Speaker 1: He boasts an eight point seven percent share in the 191 00:10:43,600 --> 00:10:46,800 Speaker 1: company that's more than twice as much as CEO Steve 192 00:10:46,880 --> 00:10:50,840 Speaker 1: Huffman possesses of his own company. This means that Altman 193 00:10:50,920 --> 00:10:55,560 Speaker 1: can actually outvote the CEO on shareholder matters. Altman's shares 194 00:10:55,640 --> 00:11:00,400 Speaker 1: means he's the third largest shareholder overall. The company's Advance Publications, 195 00:11:00,440 --> 00:11:03,439 Speaker 1: which owns Conde Nast among other things, and the Chinese 196 00:11:03,440 --> 00:11:07,040 Speaker 1: company ten Cent, which owns pretty much everything, are ahead 197 00:11:07,040 --> 00:11:10,480 Speaker 1: of him. Also probably not related, but Reddit has started 198 00:11:10,520 --> 00:11:13,439 Speaker 1: to negotiate with AI companies to allow them to use 199 00:11:13,480 --> 00:11:16,680 Speaker 1: Reddit to help train AI in return for paying a 200 00:11:16,679 --> 00:11:19,520 Speaker 1: hefty amount of money. So Google recently forked over sixty 201 00:11:19,559 --> 00:11:22,840 Speaker 1: million dollars and have agreed to do that per year 202 00:11:23,240 --> 00:11:25,840 Speaker 1: for the privilege of using Reddit to train AI. Which 203 00:11:25,880 --> 00:11:29,040 Speaker 1: is interesting that Sam Altman has such a prominent place 204 00:11:29,080 --> 00:11:32,120 Speaker 1: in the shareholders and Reddit is taking money from Google, 205 00:11:32,320 --> 00:11:35,360 Speaker 1: a competitor to open Ai, for this training privilege. I 206 00:11:35,440 --> 00:11:37,440 Speaker 1: really wonder if open ai will have to cough up 207 00:11:37,480 --> 00:11:40,640 Speaker 1: money to do the same thing. I guess we'll find out. Okay, 208 00:11:40,679 --> 00:11:42,800 Speaker 1: we're going to take a quick break to thank our sponsors. 209 00:11:42,800 --> 00:11:55,320 Speaker 1: We'll be back with more news after this. Okay, we're back. 210 00:11:55,640 --> 00:11:58,240 Speaker 1: So Yahoo is shaking things up over at in gadget. 211 00:11:58,440 --> 00:11:59,920 Speaker 1: If you're not familiar with the inn gadget, that's a 212 00:11:59,920 --> 00:12:03,000 Speaker 1: site that reports on tech topics and offers up stuff 213 00:12:03,080 --> 00:12:06,640 Speaker 1: like product reviews. The company laid off ten people in 214 00:12:06,679 --> 00:12:10,920 Speaker 1: the editorial department, including top leadership positions like the editor 215 00:12:10,960 --> 00:12:14,120 Speaker 1: in chief, and it also split the editorial department into 216 00:12:14,200 --> 00:12:17,600 Speaker 1: two groups, so one group will focus exclusively on news 217 00:12:17,640 --> 00:12:21,040 Speaker 1: and features and the other one on reviews. Reportedly, this 218 00:12:21,120 --> 00:12:23,559 Speaker 1: is to realign the editorial staff in an effort to 219 00:12:23,679 --> 00:12:26,560 Speaker 1: drive more traffic to the site, because more traffic means 220 00:12:26,679 --> 00:12:29,920 Speaker 1: more advertising revenue. And Gadget's general manager issued a memo 221 00:12:29,960 --> 00:12:33,720 Speaker 1: that said the changes will quote increase our velocity in 222 00:12:33,760 --> 00:12:37,600 Speaker 1: the quote that raises huge red flags for me as 223 00:12:37,600 --> 00:12:40,760 Speaker 1: someone who used to write articles for a website. Increase 224 00:12:40,800 --> 00:12:44,400 Speaker 1: our velocity sounds a lot like making deadlines much tighter, 225 00:12:44,880 --> 00:12:49,439 Speaker 1: requiring writers to write far more pieces per given amount 226 00:12:49,480 --> 00:12:52,040 Speaker 1: of time, and giving writers less time to research and 227 00:12:52,120 --> 00:12:54,680 Speaker 1: understand a topic, which makes it harder to write something 228 00:12:54,720 --> 00:12:57,280 Speaker 1: that isn't just a regurgitated press release. And that doesn't 229 00:12:57,320 --> 00:13:01,400 Speaker 1: really help anybody. Now. I don't have any attachment to Engadget. 230 00:13:01,400 --> 00:13:03,320 Speaker 1: I don't know any people who work there or used 231 00:13:03,320 --> 00:13:06,079 Speaker 1: to work there, but I really hope everyone find it, finds, 232 00:13:06,120 --> 00:13:08,480 Speaker 1: you know, gainful employment. They end up in a good place, 233 00:13:08,520 --> 00:13:10,439 Speaker 1: because it's a rough time to be out of work 234 00:13:10,600 --> 00:13:13,199 Speaker 1: as a web based writer. Now I probably should have 235 00:13:13,240 --> 00:13:15,960 Speaker 1: put this next story after the cruise story, but here 236 00:13:15,960 --> 00:13:18,559 Speaker 1: we are. So. The State of California has put Alphabet's 237 00:13:18,559 --> 00:13:22,360 Speaker 1: application to expand Robotaxi services in the San Francisco and 238 00:13:22,400 --> 00:13:27,680 Speaker 1: Los Angeles areas until June nineteenth. The California Public Utilities 239 00:13:27,679 --> 00:13:32,280 Speaker 1: Commission states that the application is suspended for further staff review. Now, 240 00:13:32,320 --> 00:13:35,120 Speaker 1: this is a standard procedure if the agency does not 241 00:13:35,280 --> 00:13:38,439 Speaker 1: approve an application within thirty days of it having been submitted. 242 00:13:38,800 --> 00:13:41,840 Speaker 1: So it's not like this is totally unusual. It's not 243 00:13:41,920 --> 00:13:45,560 Speaker 1: like it. It's definitely an indication that California is pushing 244 00:13:45,600 --> 00:13:48,760 Speaker 1: back against alphabet and WEIMO. My guess is that the 245 00:13:48,800 --> 00:13:53,240 Speaker 1: state of California is reevaluating their stance on autonomous vehicle 246 00:13:53,280 --> 00:13:56,120 Speaker 1: operation in the wake of the cruise incident as well 247 00:13:56,120 --> 00:13:59,400 Speaker 1: as other accidents involving autonomous vehicles, and that it might 248 00:13:59,440 --> 00:14:02,000 Speaker 1: be a wild for it issues any approvals to expand 249 00:14:02,040 --> 00:14:06,319 Speaker 1: service until the state legislation decides upon suitable rules and regulations. 250 00:14:06,800 --> 00:14:08,400 Speaker 1: Go back just a couple of years, and a lot 251 00:14:08,400 --> 00:14:11,240 Speaker 1: of car manufacturers boasted of plans to switch over to 252 00:14:11,360 --> 00:14:14,520 Speaker 1: only making electric vehicles in the near future. Mercedes Benz, 253 00:14:14,559 --> 00:14:16,800 Speaker 1: for example, said the plan was to go all EV 254 00:14:17,000 --> 00:14:19,440 Speaker 1: by the year twenty thirty, but now the company is 255 00:14:19,520 --> 00:14:22,040 Speaker 1: backing out of that promise. So the problem you see 256 00:14:22,240 --> 00:14:24,760 Speaker 1: is that folks are not buying as many evs as 257 00:14:24,800 --> 00:14:27,840 Speaker 1: the companies thought they would, and so the companies are 258 00:14:27,880 --> 00:14:32,000 Speaker 1: scared that going all EV will just devastate their industry, 259 00:14:32,320 --> 00:14:34,520 Speaker 1: and so companies are starting to change their tune, saying 260 00:14:34,560 --> 00:14:37,440 Speaker 1: that they won't be quite as aggressive in pushing evs 261 00:14:37,440 --> 00:14:40,040 Speaker 1: to the market, and that they will continue to produce 262 00:14:40,160 --> 00:14:43,960 Speaker 1: gas and hybrid vehicles instead. That's a blow to environmentalists 263 00:14:44,000 --> 00:14:46,520 Speaker 1: who are really hoping to see consumer transportation make a 264 00:14:46,560 --> 00:14:49,800 Speaker 1: really big change. But the market dictates where these companies 265 00:14:49,800 --> 00:14:52,080 Speaker 1: are going to go, and it's hard to argue that 266 00:14:52,120 --> 00:14:55,320 Speaker 1: a company should produce itself out of business. That doesn't 267 00:14:55,360 --> 00:14:58,280 Speaker 1: do really anyone any good. Now, one could argue that 268 00:14:58,360 --> 00:15:01,880 Speaker 1: issues like vehicle costs and charging station availability are really 269 00:15:01,880 --> 00:15:04,880 Speaker 1: having an impact, but in fact it's a really complex 270 00:15:05,000 --> 00:15:07,480 Speaker 1: challenge and maybe it's just too complex to follow through 271 00:15:07,480 --> 00:15:09,800 Speaker 1: on promises that were made just a few years ago. 272 00:15:09,960 --> 00:15:13,400 Speaker 1: Blue Sky, one of the federated social platform alternatives to 273 00:15:13,640 --> 00:15:16,360 Speaker 1: x slash Twitter, made a really big move this week. 274 00:15:16,720 --> 00:15:19,400 Speaker 1: The company will now allow you to choose who hosts 275 00:15:19,440 --> 00:15:23,479 Speaker 1: your data. So by default, blue Sky hosts all your information, 276 00:15:23,600 --> 00:15:25,680 Speaker 1: but you can designate another party to host all of 277 00:15:25,760 --> 00:15:28,040 Speaker 1: your data, and that includes stuff like all the stuff 278 00:15:28,080 --> 00:15:30,560 Speaker 1: you post, all of your likes, all of your follows, etc. 279 00:15:31,160 --> 00:15:34,320 Speaker 1: And that even includes yourself. If you want to store 280 00:15:34,400 --> 00:15:37,040 Speaker 1: your data on your own server, you can do that 281 00:15:37,480 --> 00:15:39,800 Speaker 1: and it doesn't have to be forever. Blue Sky structure 282 00:15:39,800 --> 00:15:42,640 Speaker 1: allows you to port data from host to host, so 283 00:15:42,760 --> 00:15:44,960 Speaker 1: if you change your mind, you can move your stuff. 284 00:15:45,120 --> 00:15:47,840 Speaker 1: Now this is huge since most social platforms require you 285 00:15:47,920 --> 00:15:50,520 Speaker 1: to host all of your information on their own servers 286 00:15:50,640 --> 00:15:54,000 Speaker 1: because your data is valuable and the platform wants to 287 00:15:54,080 --> 00:15:56,720 Speaker 1: leverage it. Being able to choose where it all goes 288 00:15:56,800 --> 00:15:59,240 Speaker 1: gives more power to the user, and if the user 289 00:15:59,280 --> 00:16:02,080 Speaker 1: determines that they're the current host is doing something shady 290 00:16:02,360 --> 00:16:05,360 Speaker 1: or it's a poor steward of their information, they can 291 00:16:05,440 --> 00:16:07,560 Speaker 1: change where it's stored. If you want to read more, 292 00:16:07,640 --> 00:16:10,560 Speaker 1: there's a post on blue Sky. It is titled blue 293 00:16:10,560 --> 00:16:13,920 Speaker 1: Sky an Open Social Web and it has all the details. 294 00:16:14,240 --> 00:16:16,480 Speaker 1: Our next story is cool because it suggests we can 295 00:16:16,520 --> 00:16:20,480 Speaker 1: achieve incredible boosts and processing power without having to fight 296 00:16:20,600 --> 00:16:23,640 Speaker 1: against the laws of quantum mechanics to do it. A 297 00:16:23,720 --> 00:16:28,040 Speaker 1: UC Riverside Associate professor named Hungwai Seng has published a 298 00:16:28,120 --> 00:16:33,480 Speaker 1: paper titled Simultaneous and Heterogeneous multi Threading and it explains 299 00:16:33,520 --> 00:16:37,600 Speaker 1: this concept. So basically saying points out that modern computers 300 00:16:37,600 --> 00:16:41,280 Speaker 1: typically have multiple processors. You know, you might have a 301 00:16:41,320 --> 00:16:45,640 Speaker 1: CPU and a GPU, you might have hardware accelerators for 302 00:16:45,680 --> 00:16:48,560 Speaker 1: stuff like machine learning or AI, and all of these 303 00:16:48,600 --> 00:16:51,400 Speaker 1: handle information separately, kind of like as if they are 304 00:16:51,440 --> 00:16:54,560 Speaker 1: in a silo, and then information can travel from processor 305 00:16:54,600 --> 00:16:58,720 Speaker 1: to processor. But this means the overall process is inefficient. 306 00:16:59,120 --> 00:17:02,080 Speaker 1: So Saying In his team demonstrated that by multi threading 307 00:17:02,120 --> 00:17:05,520 Speaker 1: processes across all of these different types of units, he 308 00:17:05,560 --> 00:17:10,359 Speaker 1: could speed things up considerably. He nearly doubled processing speed 309 00:17:10,600 --> 00:17:13,000 Speaker 1: while at the same time almost cutting the amount of 310 00:17:13,080 --> 00:17:17,080 Speaker 1: energy needed to complete those processes by half. So not 311 00:17:17,160 --> 00:17:19,919 Speaker 1: only does it boost computational speeds, does so in a 312 00:17:20,040 --> 00:17:23,280 Speaker 1: really energy efficient way. It's pretty insane that he could 313 00:17:23,280 --> 00:17:26,240 Speaker 1: get nearly double the output just by using multi threading 314 00:17:26,240 --> 00:17:29,720 Speaker 1: across processors without having to do a single hardware upgrade. 315 00:17:29,920 --> 00:17:33,399 Speaker 1: And this has the potential to create enormous benefits. So 316 00:17:33,480 --> 00:17:37,000 Speaker 1: imagine that you were able to double the output and 317 00:17:37,200 --> 00:17:40,880 Speaker 1: half the energy requirements of a data server farm for example. 318 00:17:41,200 --> 00:17:43,200 Speaker 1: Like that would mean that you would need less water 319 00:17:43,320 --> 00:17:45,880 Speaker 1: to cool the data farm, you would be using less 320 00:17:46,000 --> 00:17:48,400 Speaker 1: energy to run it, so it'd be more cost efficient. 321 00:17:48,600 --> 00:17:51,359 Speaker 1: You'd be able to double the output. That's incredible. Of course, 322 00:17:51,760 --> 00:17:54,560 Speaker 1: we could probably just screw that up by doubling the 323 00:17:54,640 --> 00:17:57,520 Speaker 1: number of servers we're storing in that server farm, and 324 00:17:57,560 --> 00:17:59,879 Speaker 1: then we're right back to where we started. But you know, 325 00:18:00,080 --> 00:18:02,879 Speaker 1: you get the point. Satan himself is actually cautioned that 326 00:18:03,040 --> 00:18:04,639 Speaker 1: a lot more work has to be done to figure 327 00:18:04,640 --> 00:18:09,040 Speaker 1: out when this methodology will be practical and applicable, because 328 00:18:09,080 --> 00:18:12,119 Speaker 1: it might be that for some applications it's ideal, but 329 00:18:12,200 --> 00:18:15,199 Speaker 1: for others we might see far less impressive gains and 330 00:18:15,240 --> 00:18:18,199 Speaker 1: that it wouldn't make sense to implement it. So it 331 00:18:18,280 --> 00:18:20,200 Speaker 1: is good to keep that in mind. But still pretty 332 00:18:20,280 --> 00:18:22,760 Speaker 1: darn cool. Now, before I sign off, I do have 333 00:18:22,800 --> 00:18:24,720 Speaker 1: an article i'd like to recommend to y'all, and this 334 00:18:24,800 --> 00:18:27,520 Speaker 1: one is from Wired. It's by Lauren Good and it's 335 00:18:27,560 --> 00:18:31,119 Speaker 1: titled tech Job interviews are out of Control. So the 336 00:18:31,119 --> 00:18:33,600 Speaker 1: piece explores how the culture in the tech sector has 337 00:18:33,680 --> 00:18:36,280 Speaker 1: changed dramatically over the last couple of years. You know, 338 00:18:36,320 --> 00:18:39,960 Speaker 1: once upon a time, developers and engineers found themselves courted 339 00:18:40,000 --> 00:18:43,520 Speaker 1: by headhunters and they were constantly being lured to greener 340 00:18:43,560 --> 00:18:47,639 Speaker 1: pastures on a regular basis with bigger compensation packages and benefits. 341 00:18:47,880 --> 00:18:50,879 Speaker 1: But now an interview process might require proving your worth 342 00:18:50,960 --> 00:18:53,880 Speaker 1: by working on a project without you know, actually being 343 00:18:53,880 --> 00:18:57,600 Speaker 1: an employee. So there you are doing real work for 344 00:18:57,680 --> 00:19:00,359 Speaker 1: free in order to prove that you're worth high hiring, 345 00:19:00,480 --> 00:19:02,720 Speaker 1: and meanwhile you're taking up time and energy you could 346 00:19:02,800 --> 00:19:06,000 Speaker 1: be using to actually land a gig, which is pretty 347 00:19:06,040 --> 00:19:08,120 Speaker 1: crazy stuff. But the full article goes into way more 348 00:19:08,160 --> 00:19:11,199 Speaker 1: detailed it's well worth reading. As always, I have no 349 00:19:11,200 --> 00:19:13,879 Speaker 1: connection to Wired or to Lauren Good. It's just a 350 00:19:13,920 --> 00:19:15,920 Speaker 1: good article and I think you should check it out. 351 00:19:16,359 --> 00:19:20,040 Speaker 1: That's it for the tech News for February twenty third, 352 00:19:20,160 --> 00:19:22,760 Speaker 1: twenty twenty four. I hope you are all well. I'll 353 00:19:22,760 --> 00:19:32,080 Speaker 1: talk to you again really soon. Tech Stuff is an 354 00:19:32,080 --> 00:19:37,600 Speaker 1: iHeartRadio production. For more podcasts from iHeartRadio, visit the iHeartRadio app, 355 00:19:37,760 --> 00:19:44,800 Speaker 1: Apple Podcasts, or wherever you listen to your favorite shows.