1 00:00:01,400 --> 00:00:04,720 Speaker 1: Entrepreneur in businessman Bill Gates and his business partner Paul 2 00:00:04,760 --> 00:00:09,240 Speaker 1: Allen founded and built the world's largest software business, Microsoft, 3 00:00:09,560 --> 00:00:14,440 Speaker 1: through technological innovation, keen business strategy, and aggressive business tactics. 4 00:00:14,920 --> 00:00:17,400 Speaker 1: In the process, Gates became one of the richest men 5 00:00:17,400 --> 00:00:21,239 Speaker 1: in the world. In February, Gates announced that he was 6 00:00:21,239 --> 00:00:25,079 Speaker 1: stepping down as Microsoft's chairman to focus on charitable work 7 00:00:25,120 --> 00:00:28,440 Speaker 1: at his foundation, the Bill of Melinda Gates Foundation, and 8 00:00:28,480 --> 00:00:31,080 Speaker 1: Gates sat down with David Rubinstein, co founder of the 9 00:00:31,120 --> 00:00:34,040 Speaker 1: Carlisle Group and hosts of the Bloomberg television show Peer 10 00:00:34,080 --> 00:00:39,160 Speaker 1: to Peer Conversations to discuss global warming, carbon emissions, regulating 11 00:00:39,200 --> 00:00:41,879 Speaker 1: big tech, and why he thinks the impossible foods and 12 00:00:41,920 --> 00:00:45,360 Speaker 1: beyond meat can help the environment. For about twenty years 13 00:00:45,440 --> 00:00:48,479 Speaker 1: or so, you've been the wealthiest man in the world. 14 00:00:49,640 --> 00:00:53,000 Speaker 1: But because you've given away so much money. Recently, Jeff 15 00:00:53,040 --> 00:00:56,680 Speaker 1: Bezos became wealthier. Do you think if you had stayed 16 00:00:56,680 --> 00:01:01,240 Speaker 1: in college and gotten your college degree. I mean, you 17 00:01:01,240 --> 00:01:04,120 Speaker 1: don't feel inadequate now because being only the second wealthiest 18 00:01:04,120 --> 00:01:06,200 Speaker 1: man in the world was at right. No, I mean 19 00:01:06,800 --> 00:01:10,119 Speaker 1: it's a sign that I haven't given them money way 20 00:01:10,160 --> 00:01:13,280 Speaker 1: fast enough to drop out of the top ten, you know, 21 00:01:13,360 --> 00:01:18,679 Speaker 1: and the market has been strong. Actually, the market has 22 00:01:18,680 --> 00:01:23,480 Speaker 1: been strong. Microsoft is up thirty this year. So to 23 00:01:23,600 --> 00:01:27,360 Speaker 1: what do you attribute that? Uh, the company, you know, 24 00:01:27,400 --> 00:01:31,080 Speaker 1: it's doing super well. Such in Ndala is a great CEO. 25 00:01:32,360 --> 00:01:34,920 Speaker 1: You know, the whole dream of the importance of software 26 00:01:35,319 --> 00:01:38,880 Speaker 1: has really come true. The five most valuable companies in 27 00:01:38,880 --> 00:01:43,240 Speaker 1: the world are these technology companies. Microsoft, you know, has 28 00:01:43,280 --> 00:01:46,560 Speaker 1: a good share of that. I get to spend about 29 00:01:46,600 --> 00:01:49,560 Speaker 1: a sixth of my time now is over at Microsoft. 30 00:01:50,080 --> 00:01:53,240 Speaker 1: So recently you said that the biggest mistake you've made 31 00:01:53,280 --> 00:01:59,200 Speaker 1: professionally was that, um, Microsoft should have had the Android technology. 32 00:01:59,240 --> 00:02:02,400 Speaker 1: Why was that the big this mistake when you're in 33 00:02:02,520 --> 00:02:04,720 Speaker 1: a field, you know, we were in the field of 34 00:02:04,760 --> 00:02:07,840 Speaker 1: doing operating systems for personal computers. We knew the mobile 35 00:02:07,840 --> 00:02:12,000 Speaker 1: phone would be very popular, and so we were doing 36 00:02:12,520 --> 00:02:16,760 Speaker 1: what was called Windows Mobile. We missed being the dominant 37 00:02:16,800 --> 00:02:21,000 Speaker 1: mobile operating system by very tiny amount. We were distracted 38 00:02:21,040 --> 00:02:23,960 Speaker 1: during Iron and Trust trial. We didn't assign the best 39 00:02:23,960 --> 00:02:26,959 Speaker 1: people to do the work. So it's the biggest mistake 40 00:02:27,040 --> 00:02:30,239 Speaker 1: I made in terms of something that was clearly within 41 00:02:30,320 --> 00:02:34,040 Speaker 1: our skill set. We were clearly the company that that 42 00:02:34,040 --> 00:02:37,440 Speaker 1: should have achieved that your two million areas of focus 43 00:02:37,440 --> 00:02:40,720 Speaker 1: our cater twelve education, United States, and healthcare in the 44 00:02:41,200 --> 00:02:44,840 Speaker 1: least wealthy parts of the world. Recently, you've decided to 45 00:02:45,040 --> 00:02:48,399 Speaker 1: make another effort, not necessarily through your foundation, but through 46 00:02:48,760 --> 00:02:52,320 Speaker 1: Breakthrough Energy to try to do something about climate change. 47 00:02:52,720 --> 00:02:56,480 Speaker 1: Why are you so worried about climate change? Well, the 48 00:02:56,480 --> 00:02:59,880 Speaker 1: climate change is a problem that gets worse every year, 49 00:03:00,680 --> 00:03:03,560 Speaker 1: and yet what you have to do on a global 50 00:03:03,600 --> 00:03:09,520 Speaker 1: basis is very dramatic and reshaping the entire physical economy 51 00:03:10,160 --> 00:03:13,760 Speaker 1: that we have. And so it's a very complex problem, 52 00:03:14,120 --> 00:03:16,880 Speaker 1: and it's a problem that fits where I see my 53 00:03:17,480 --> 00:03:19,760 Speaker 1: value at it, which is looking at something through the 54 00:03:19,840 --> 00:03:23,160 Speaker 1: lens of innovation, not just the R and D part, 55 00:03:23,320 --> 00:03:26,360 Speaker 1: but the creation of products and the deployment of products, 56 00:03:26,960 --> 00:03:31,160 Speaker 1: and so helping educate people about Okay, what where, what 57 00:03:31,240 --> 00:03:34,800 Speaker 1: are the sources of these greenhouse gases? And how do 58 00:03:34,840 --> 00:03:38,000 Speaker 1: you get on a path of innovation so that you 59 00:03:38,040 --> 00:03:42,360 Speaker 1: can get global adoption and actually bring emissions down dramatically, 60 00:03:42,400 --> 00:03:44,320 Speaker 1: Because that's part of your foundation, or you're doing this 61 00:03:44,400 --> 00:03:49,200 Speaker 1: outside your foundation. Okay, the the part where you mitigate 62 00:03:49,440 --> 00:03:53,120 Speaker 1: and you help the poor countries with better seeds and 63 00:03:53,160 --> 00:03:57,400 Speaker 1: better policies. UH, partly through development aid that is through 64 00:03:57,440 --> 00:04:01,440 Speaker 1: the foundation, that mitigation part the part where you invent 65 00:04:01,520 --> 00:04:07,720 Speaker 1: new ways of making uh fuels, electricity, cement, steal meat 66 00:04:08,360 --> 00:04:13,480 Speaker 1: that is uh done directly by me with a lot 67 00:04:13,520 --> 00:04:16,320 Speaker 1: of investments, including the fund that you mentioned. The so 68 00:04:16,440 --> 00:04:21,240 Speaker 1: called breakto Energy Ventures is a fund that I assembled 69 00:04:21,240 --> 00:04:24,960 Speaker 1: a group of twenty two people uh to put money 70 00:04:25,080 --> 00:04:30,720 Speaker 1: into companies that are trying to commercialize the breakthroughs. Right, 71 00:04:30,720 --> 00:04:33,080 Speaker 1: but that's a fund of one billion dollars, right you 72 00:04:33,120 --> 00:04:37,200 Speaker 1: put in two So can one billion dollars really make 73 00:04:37,240 --> 00:04:42,080 Speaker 1: that much of a difference a billion? Uh? It's actually 74 00:04:42,120 --> 00:04:45,360 Speaker 1: been very catalytic so far. They have twenty investments. UH 75 00:04:45,560 --> 00:04:49,880 Speaker 1: late next year will probably raise another billion to a 76 00:04:49,920 --> 00:04:53,000 Speaker 1: billion and a half. You know, this is all about 77 00:04:53,240 --> 00:04:56,400 Speaker 1: innovation some right now. The premium if you said, okay, 78 00:04:56,680 --> 00:05:00,559 Speaker 1: you have to make steal with no emissions, that steel 79 00:05:00,560 --> 00:05:03,839 Speaker 1: would cost you four times what steel does today. Your 80 00:05:03,960 --> 00:05:08,320 Speaker 1: electric bill would more than double. Uh if if we 81 00:05:08,440 --> 00:05:13,160 Speaker 1: just take the technology we have today. So um Yes, 82 00:05:13,279 --> 00:05:17,200 Speaker 1: supporting those companies and drawing other investors in. One thing 83 00:05:17,240 --> 00:05:19,240 Speaker 1: break Through Energy has done has gotten a lot of 84 00:05:19,240 --> 00:05:22,600 Speaker 1: co investors green investing didn't go very well in the 85 00:05:22,600 --> 00:05:25,640 Speaker 1: first round, and so it looked like a field that 86 00:05:25,839 --> 00:05:30,200 Speaker 1: might UH evaporate to some three because d vs come 87 00:05:30,240 --> 00:05:33,080 Speaker 1: in and being able to bring a depth of understanding 88 00:05:33,080 --> 00:05:35,640 Speaker 1: to these things. Not only have they been able to invest, 89 00:05:35,680 --> 00:05:39,760 Speaker 1: the first billing will be UH fully committed within the 90 00:05:39,839 --> 00:05:43,160 Speaker 1: next year. But we've gotten other investors, so that's gone 91 00:05:44,080 --> 00:05:47,840 Speaker 1: quite well. And in the technology they only invest in 92 00:05:47,960 --> 00:05:52,600 Speaker 1: companies who have a chance of reducing greenhouse gas emissions 93 00:05:52,640 --> 00:05:56,599 Speaker 1: by a half a percent each each company. UH. And 94 00:05:56,760 --> 00:05:58,839 Speaker 1: you know they've found twenty and I'm sure they'll find 95 00:05:58,880 --> 00:06:02,039 Speaker 1: another twenty. I'm the smallest investor in that fund, I think, 96 00:06:02,080 --> 00:06:04,040 Speaker 1: So am I going to get my money back and 97 00:06:04,080 --> 00:06:07,320 Speaker 1: make a return? Or Um? What what do you say? 98 00:06:07,600 --> 00:06:10,880 Speaker 1: I'd say it's of the things you invest in. It's 99 00:06:11,000 --> 00:06:14,160 Speaker 1: probably one of the higher risk things. It is being 100 00:06:14,200 --> 00:06:17,279 Speaker 1: done on a commercial basis. UH. You know, we're likely 101 00:06:17,320 --> 00:06:21,680 Speaker 1: to have a few significant successes, so it's not philanthropic 102 00:06:21,720 --> 00:06:25,320 Speaker 1: in the sense that you can deduct it. UH. But 103 00:06:26,360 --> 00:06:30,039 Speaker 1: the time frame of the returns and the riskiness of 104 00:06:30,040 --> 00:06:33,600 Speaker 1: the turns are fairly high. So we do expect UH 105 00:06:33,720 --> 00:06:36,720 Speaker 1: to make a profit out of that fund. So Why 106 00:06:36,720 --> 00:06:39,640 Speaker 1: do you think some people do not believe that there 107 00:06:39,800 --> 00:06:42,440 Speaker 1: is such a thing as climate change. What is propelling 108 00:06:42,480 --> 00:06:45,960 Speaker 1: them to say there's no climate change? Is it's scientific 109 00:06:46,000 --> 00:06:49,880 Speaker 1: evidence or some other political reason. I won't mention anybody, 110 00:06:49,920 --> 00:06:52,000 Speaker 1: but there are some people who don't think that there 111 00:06:52,080 --> 00:06:56,080 Speaker 1: is climate change. Well, you know, they must not have 112 00:06:56,880 --> 00:07:02,440 Speaker 1: taken enough science courses or something. I don't know it. Uh. 113 00:07:02,800 --> 00:07:10,040 Speaker 1: The climate is a complex issue, and you know, just 114 00:07:10,200 --> 00:07:14,840 Speaker 1: understanding how you do the abatement requires a lot of 115 00:07:14,920 --> 00:07:18,760 Speaker 1: in depth study. In the United States, it's become somewhat 116 00:07:18,760 --> 00:07:24,320 Speaker 1: of a partisan issue, which is unfortunate. Uh, you know, 117 00:07:24,360 --> 00:07:27,360 Speaker 1: it might make it harder to achieve the type of 118 00:07:27,760 --> 00:07:32,720 Speaker 1: agreements we need uh here here in the United States. 119 00:07:32,800 --> 00:07:35,720 Speaker 1: But you know, we have two problems. We have the 120 00:07:35,720 --> 00:07:38,280 Speaker 1: people who deny climate and then we have the people 121 00:07:38,280 --> 00:07:40,880 Speaker 1: who think it's easy to solve, and we need to 122 00:07:40,920 --> 00:07:43,960 Speaker 1: help educate both of those groups. But in the history 123 00:07:44,000 --> 00:07:47,400 Speaker 1: of human civilization, is there any evidence that people will 124 00:07:47,480 --> 00:07:50,640 Speaker 1: do things that will affect their great great grandchildren but 125 00:07:50,680 --> 00:07:54,240 Speaker 1: that they won't see the benefit from. Well, the United States, actually, 126 00:07:54,240 --> 00:07:57,520 Speaker 1: of all governments, has been willing to take on very 127 00:07:57,560 --> 00:08:03,760 Speaker 1: difficult problems like cancer and make gigantic investments knowing that 128 00:08:04,120 --> 00:08:08,960 Speaker 1: the real payoff would be many decades down the road. Uh. 129 00:08:09,120 --> 00:08:12,880 Speaker 1: You know when that was first being pushed. Uh, you know, 130 00:08:13,000 --> 00:08:16,400 Speaker 1: people are saying, hey, this is important. Climate change is 131 00:08:16,480 --> 00:08:21,120 Speaker 1: like that where you've got to take a long term perspective. 132 00:08:21,480 --> 00:08:25,440 Speaker 1: And government at its best is when it's taking that 133 00:08:25,560 --> 00:08:29,040 Speaker 1: long term perspective and funding the basic guaranty and the 134 00:08:29,160 --> 00:08:33,199 Speaker 1: policies that lead to scale deployment. Now, a large part 135 00:08:33,240 --> 00:08:35,080 Speaker 1: of the carbon we have in the atmosphere now is 136 00:08:35,120 --> 00:08:39,240 Speaker 1: caused by the electricity grid, which is about so exactly 137 00:08:39,559 --> 00:08:43,839 Speaker 1: so Um, it comes from agriculture and forestry. Why is 138 00:08:43,880 --> 00:08:48,040 Speaker 1: that causing such a big increasing carbon Well, the mat 139 00:08:48,320 --> 00:08:53,320 Speaker 1: category UH is a variety of things. When you clear land, 140 00:08:53,960 --> 00:08:57,720 Speaker 1: you're taking in the carbon that stored, seeing the trees 141 00:08:57,880 --> 00:09:00,160 Speaker 1: or plants there, and you're releasing all of that like 142 00:09:00,280 --> 00:09:07,720 Speaker 1: burning the land, uh saying Indonesia for um palm oil plantations. 143 00:09:07,760 --> 00:09:12,640 Speaker 1: Another thing is that UH, cows and other grass eating 144 00:09:12,760 --> 00:09:20,320 Speaker 1: species have a digestion system that emits methane, and methane 145 00:09:20,400 --> 00:09:24,319 Speaker 1: is a very powerful greenhouse gas, and so cows alone 146 00:09:25,600 --> 00:09:30,960 Speaker 1: account for about six percent of global emissions. And so 147 00:09:31,000 --> 00:09:35,400 Speaker 1: we need to change cows just cows alone. How are 148 00:09:35,400 --> 00:09:41,240 Speaker 1: we going to do that? Well? Actually, of all the categories, uh, 149 00:09:41,440 --> 00:09:45,000 Speaker 1: the one that has gone better than I would have 150 00:09:45,040 --> 00:09:49,160 Speaker 1: expected five years ago is this work to make what's 151 00:09:49,200 --> 00:09:53,439 Speaker 1: called artificial meat. And so you have people like Impossible 152 00:09:53,559 --> 00:09:57,600 Speaker 1: or Beyond Meat, both which uh I invested in. You 153 00:09:57,640 --> 00:09:59,920 Speaker 1: eat it as well, and you like it, that's absolutely. 154 00:10:00,800 --> 00:10:04,080 Speaker 1: You can go to Burger King and by the Impossible burger. 155 00:10:04,160 --> 00:10:06,280 Speaker 1: All right, is it healthier for you or just healthier 156 00:10:06,280 --> 00:10:09,840 Speaker 1: for the It's slightly healthier for you in terms of 157 00:10:09,920 --> 00:10:16,319 Speaker 1: less cholesterol. It's of course dramatic reduction in methane emissions, 158 00:10:16,840 --> 00:10:21,680 Speaker 1: you know, animal cruelty, manure management, and the pressure that 159 00:10:21,840 --> 00:10:25,160 Speaker 1: meat consumption puts on land use. What about electric cars? 160 00:10:25,200 --> 00:10:30,240 Speaker 1: Do you think that's a solution. Absolutely? They If you 161 00:10:30,280 --> 00:10:37,600 Speaker 1: look at the transport sector, passenger cars with about another 162 00:10:37,760 --> 00:10:41,320 Speaker 1: factor of two to three in battery improvement, which is possible, 163 00:10:42,000 --> 00:10:47,320 Speaker 1: the mainstream for passenger cars can become electric. So you 164 00:10:47,400 --> 00:10:50,920 Speaker 1: have to make that transition. Uh, you've got to scale 165 00:10:50,920 --> 00:10:54,000 Speaker 1: it up. You've got to make sure electricity is zero emission. 166 00:10:54,440 --> 00:10:59,680 Speaker 1: But for trucks and planes, uh, there's almost no chance 167 00:10:59,720 --> 00:11:02,920 Speaker 1: the batteries will be good enough, and so there you'll 168 00:11:02,920 --> 00:11:09,360 Speaker 1: still need to create liquid fuels, either with electricity or biofuels. 169 00:11:10,040 --> 00:11:14,000 Speaker 1: Uh some way. Fuels are amazing. You know, the energy 170 00:11:14,080 --> 00:11:18,280 Speaker 1: density of gasoline is thirty times the energy density of 171 00:11:18,280 --> 00:11:20,800 Speaker 1: the best battery we can make. And so if you 172 00:11:20,840 --> 00:11:25,360 Speaker 1: look at like a a container ship that crosses the ocean, 173 00:11:25,840 --> 00:11:29,440 Speaker 1: having your fuel be thirty times less efficient would mean 174 00:11:29,480 --> 00:11:32,600 Speaker 1: that nine of the weight you're carrying would be the 175 00:11:32,640 --> 00:11:37,400 Speaker 1: batteries instead of the cargo. And so trucks and planes 176 00:11:37,720 --> 00:11:43,280 Speaker 1: and boats, electrification is unlikely to work in those cases. 177 00:11:43,320 --> 00:11:46,520 Speaker 1: So we need ways of making fuels that are are 178 00:11:46,640 --> 00:11:49,280 Speaker 1: zero carbon. When you talk to heads of state about this, 179 00:11:49,400 --> 00:11:51,920 Speaker 1: do they roll their eyes and say, we're happy to me, 180 00:11:52,040 --> 00:11:53,400 Speaker 1: do you can I have a selfie with you? And 181 00:11:53,440 --> 00:11:56,560 Speaker 1: so forth? But do they really do anything? And what 182 00:11:56,600 --> 00:11:59,080 Speaker 1: are you trying to get heads of state to do? Well? 183 00:11:59,120 --> 00:12:04,160 Speaker 1: In the um Paris Climate Conference, one of the things 184 00:12:04,200 --> 00:12:06,280 Speaker 1: that was missing was to focus on R and D 185 00:12:06,920 --> 00:12:11,560 Speaker 1: and uh. So actually France said yes, we want that 186 00:12:11,640 --> 00:12:14,840 Speaker 1: to be for the first time at a COP a 187 00:12:15,000 --> 00:12:18,440 Speaker 1: real issue that gets discussed. And so what was called 188 00:12:18,480 --> 00:12:22,480 Speaker 1: Mission Innovation, which Prime Minister MODI got to pick that name. 189 00:12:23,240 --> 00:12:26,440 Speaker 1: That idea of a commitment of over thirty governments to 190 00:12:26,480 --> 00:12:31,280 Speaker 1: double their energy R and D was a significant milestone 191 00:12:31,800 --> 00:12:35,200 Speaker 1: that came out of that conference. Uh. In order to 192 00:12:35,280 --> 00:12:38,320 Speaker 1: get that commitment, UH I had to make the commitment 193 00:12:38,320 --> 00:12:41,079 Speaker 1: that there would be breakthrough energy that would take things 194 00:12:41,120 --> 00:12:44,280 Speaker 1: out of those labs and help get them into the marketplace. 195 00:12:44,320 --> 00:12:50,480 Speaker 1: So there's been some progress. Climate is complicated enough that UH, 196 00:12:50,640 --> 00:12:53,959 Speaker 1: you know, you don't want you want a broad set 197 00:12:53,960 --> 00:12:58,920 Speaker 1: of people in the government to understand the the complexities 198 00:12:59,559 --> 00:13:01,360 Speaker 1: and in terms of the R and D work that 199 00:13:01,440 --> 00:13:04,839 Speaker 1: needs to be done, unless the US is deeply engaged, 200 00:13:04,960 --> 00:13:08,240 Speaker 1: it's unlikely to happen because so much of the world's 201 00:13:08,280 --> 00:13:12,240 Speaker 1: capacity to do that innovation is here in the United States. 202 00:13:12,440 --> 00:13:15,400 Speaker 1: States pulled out more or less of the Paris Accord, 203 00:13:15,480 --> 00:13:18,360 Speaker 1: though not technically so for another year or so. UM. 204 00:13:18,440 --> 00:13:20,200 Speaker 1: Is that of concern to you? And do you think 205 00:13:20,240 --> 00:13:22,960 Speaker 1: this is going to hurt the effort to change climate 206 00:13:23,040 --> 00:13:26,520 Speaker 1: change around the world. Yeah, it's a huge step backwards. 207 00:13:27,080 --> 00:13:30,960 Speaker 1: Even if you meet all the current commitments in that 208 00:13:31,840 --> 00:13:36,000 Speaker 1: climate accord, you're still way over two degrees of warming 209 00:13:36,640 --> 00:13:41,360 Speaker 1: and most countries are behind the commitments they made. Those 210 00:13:41,360 --> 00:13:44,400 Speaker 1: commitments were a set of reductions where you would compare 211 00:13:44,440 --> 00:13:49,840 Speaker 1: your two emissions to your two thousand five emissions. And 212 00:13:50,760 --> 00:13:52,880 Speaker 1: there's a little bit of that, that's easy. The shift 213 00:13:52,880 --> 00:13:55,520 Speaker 1: from cold and natural gas, which is one kind thing, 214 00:13:55,920 --> 00:13:59,280 Speaker 1: is a lot of that. Uh. And yet the world 215 00:13:59,520 --> 00:14:02,240 Speaker 1: is falling in short and so to him, people like 216 00:14:02,320 --> 00:14:06,160 Speaker 1: the United States say, Okay, that's even that's not important. Uh. 217 00:14:06,200 --> 00:14:08,440 Speaker 1: It just shows how daunting this is going to be. 218 00:14:08,760 --> 00:14:11,120 Speaker 1: There's no way we'll get there without the US coming 219 00:14:11,120 --> 00:14:13,160 Speaker 1: back in in a strong way. You think if you 220 00:14:13,280 --> 00:14:15,920 Speaker 1: met with President Trump you could convince him on Paris 221 00:14:16,360 --> 00:14:18,679 Speaker 1: to maybe get back in. Or is that beyond your 222 00:14:18,720 --> 00:14:25,760 Speaker 1: capabilities to do that? I I someone else should do that. Um. 223 00:14:25,800 --> 00:14:28,560 Speaker 1: All right, now, the largest companies in the world and 224 00:14:28,600 --> 00:14:33,920 Speaker 1: the United States today are technology companies Apple, Facebook, Google, Microsoft, 225 00:14:34,120 --> 00:14:37,360 Speaker 1: and so forth. UM. Do you worry that there's too 226 00:14:37,440 --> 00:14:39,720 Speaker 1: much power and too much data in the hands of 227 00:14:39,760 --> 00:14:42,800 Speaker 1: these technology companies? And are you surprised that government hasn't 228 00:14:42,800 --> 00:14:47,120 Speaker 1: done something more than they've done today about this? Well, 229 00:14:47,120 --> 00:14:51,400 Speaker 1: technology has become so central that government has to think, Okay, 230 00:14:51,400 --> 00:14:53,840 Speaker 1: what does that mean about elections? What does it mean 231 00:14:53,880 --> 00:14:56,960 Speaker 1: about bullying. What does it mean about why are tapping 232 00:14:57,000 --> 00:15:01,160 Speaker 1: authorities that lets you find out uh what's going on 233 00:15:01,320 --> 00:15:06,600 Speaker 1: financially or uh, you know, drug money laundering, things like that. 234 00:15:07,000 --> 00:15:10,640 Speaker 1: So yes, the government needs get involved. I for the 235 00:15:10,680 --> 00:15:13,840 Speaker 1: earliers of Microsoft break two people that I didn't have 236 00:15:13,920 --> 00:15:17,560 Speaker 1: an office in Washington, d C. And eventually I came 237 00:15:17,600 --> 00:15:22,360 Speaker 1: to regret uh that statement because it was kind of 238 00:15:22,360 --> 00:15:27,360 Speaker 1: almost like taunting Washington, d C. Uh. And so now 239 00:15:27,440 --> 00:15:30,960 Speaker 1: the technology companies, partly because of the lesson of Microsoft, 240 00:15:31,680 --> 00:15:34,160 Speaker 1: uh of course, you know, they could have seen that 241 00:15:34,280 --> 00:15:37,120 Speaker 1: lesson through a T and T R IBM or codact 242 00:15:37,160 --> 00:15:40,560 Speaker 1: or a lot of innovators as well. They're very engaged. 243 00:15:40,960 --> 00:15:44,440 Speaker 1: There will be more regulation of the tech sector, things 244 00:15:44,480 --> 00:15:47,960 Speaker 1: like privacy, I'm sure though, and there should be at 245 00:15:48,040 --> 00:15:51,520 Speaker 1: some point federal regulation that relates to that. The fact 246 00:15:51,560 --> 00:15:55,400 Speaker 1: that now this is the way people consume media, uh, 247 00:15:55,480 --> 00:15:59,760 Speaker 1: you know, has really brought it into a realm that uh, 248 00:16:00,080 --> 00:16:01,840 Speaker 1: you know, we need to shape it so that the 249 00:16:01,920 --> 00:16:07,720 Speaker 1: benefits outweigh outweigh the negatives. So if you were twenty 250 00:16:07,840 --> 00:16:09,800 Speaker 1: years old today and you wanted to start a new 251 00:16:09,840 --> 00:16:13,440 Speaker 1: company drop out of Harvard, what company or what area 252 00:16:13,440 --> 00:16:15,560 Speaker 1: would you want to start it in Well, this is 253 00:16:15,600 --> 00:16:18,880 Speaker 1: a great time to be doing innovation because the tools 254 00:16:18,920 --> 00:16:21,600 Speaker 1: of innovation are so much better. There are lots of 255 00:16:21,640 --> 00:16:25,720 Speaker 1: things in biology that are very interesting. Uh, there are 256 00:16:25,840 --> 00:16:29,360 Speaker 1: lots of things in energy that are interesting. Given my background, 257 00:16:29,480 --> 00:16:34,000 Speaker 1: I would start an AI company that whose goal would 258 00:16:34,040 --> 00:16:38,680 Speaker 1: be to uh teach computer's kind of read so that 259 00:16:38,760 --> 00:16:42,400 Speaker 1: they can absorb and understand all the written knowledge of 260 00:16:42,440 --> 00:16:46,000 Speaker 1: the world. That's an area where AI has yet to 261 00:16:46,120 --> 00:16:48,720 Speaker 1: make progress, and it will be quite profound when we 262 00:16:48,760 --> 00:16:51,160 Speaker 1: achieve that goal. So are you worried about the power 263 00:16:51,200 --> 00:16:54,720 Speaker 1: of AI to disrupt our civilization and put people out 264 00:16:54,720 --> 00:16:58,880 Speaker 1: of work? Those kind of things. The increased productivity that 265 00:16:58,920 --> 00:17:04,240 Speaker 1: will come from a I will create dilemmas about what 266 00:17:04,359 --> 00:17:08,200 Speaker 1: should people do with that extra time? And you've got 267 00:17:08,200 --> 00:17:11,720 Speaker 1: to consider that a good thing, even though it will 268 00:17:11,800 --> 00:17:15,800 Speaker 1: be an interesting set of adjustments that have to take place. 269 00:17:16,080 --> 00:17:18,960 Speaker 1: You assess the two most urgent issues were K to 270 00:17:19,040 --> 00:17:21,600 Speaker 1: twelve in the United States and health in the less 271 00:17:21,640 --> 00:17:24,280 Speaker 1: developed areas. How did you pick those two? Any regrets 272 00:17:24,280 --> 00:17:26,600 Speaker 1: about picking those two? And have you made progress on 273 00:17:26,800 --> 00:17:31,440 Speaker 1: either of those two? Well, global health is our biggest area, 274 00:17:31,560 --> 00:17:36,400 Speaker 1: and there the progress has been really unbelievable, not just 275 00:17:36,600 --> 00:17:39,960 Speaker 1: because of our work, but our partners that include the 276 00:17:40,040 --> 00:17:46,080 Speaker 1: US government spending on pet far the European donors who 277 00:17:46,080 --> 00:17:49,040 Speaker 1: have really stepped up on these health issues. One of 278 00:17:49,040 --> 00:17:53,040 Speaker 1: the metrics of importance is the number of children in 279 00:17:53,080 --> 00:17:55,680 Speaker 1: the world who died before the age of five. When 280 00:17:55,720 --> 00:17:58,720 Speaker 1: we got started in the year two thousand, that was 281 00:17:58,760 --> 00:18:02,800 Speaker 1: over ten million a year. Now it's about five million year, 282 00:18:03,800 --> 00:18:08,199 Speaker 1: and so you know, it's just mind blowing, and and 283 00:18:08,359 --> 00:18:10,840 Speaker 1: people aren't that as aware of it. Iss you'd like 284 00:18:10,960 --> 00:18:15,240 Speaker 1: them to be. The those deaths. Because of getting out 285 00:18:16,119 --> 00:18:19,960 Speaker 1: vaccines and understanding a bit more about nutrition, those deaths 286 00:18:19,960 --> 00:18:22,560 Speaker 1: have been cut in half. Now the goal is to 287 00:18:22,600 --> 00:18:29,280 Speaker 1: cut them in half again. Our US education work, uh 288 00:18:29,400 --> 00:18:32,119 Speaker 1: that is not just K through twelve, that includes higher 289 00:18:32,160 --> 00:18:38,640 Speaker 1: it as well. They're the key metrics, uh, dropout rates, UM, 290 00:18:40,200 --> 00:18:45,840 Speaker 1: math and verbal achievement. Those metrics have moved essentially not 291 00:18:46,000 --> 00:18:49,760 Speaker 1: at all. And even as the US is spending more 292 00:18:50,400 --> 00:18:54,480 Speaker 1: resources on education, we spend by far more than any 293 00:18:54,600 --> 00:18:57,720 Speaker 1: any country in the world, and yet our results are 294 00:18:57,840 --> 00:19:02,639 Speaker 1: quite a bit worse. Uh then, Uh, almost all the 295 00:19:02,680 --> 00:19:06,280 Speaker 1: other rich countries and even some middle income countries. You know, 296 00:19:06,359 --> 00:19:10,320 Speaker 1: even Vietnam now is passing us in terms of their 297 00:19:10,400 --> 00:19:15,320 Speaker 1: math results. So the there, the fields as a whole 298 00:19:15,400 --> 00:19:18,160 Speaker 1: in our work has not had the impact we hope for. 299 00:19:18,320 --> 00:19:21,240 Speaker 1: So today people come to you all the time for money. 300 00:19:21,560 --> 00:19:23,840 Speaker 1: I assume everywhere you go people say, by the way, 301 00:19:23,880 --> 00:19:25,720 Speaker 1: I have this thing you should invest in. By I 302 00:19:25,760 --> 00:19:27,879 Speaker 1: have a couple of myself I've mentioned later, No, not 303 00:19:27,920 --> 00:19:29,720 Speaker 1: just gonna do now, a couple of things you should 304 00:19:29,720 --> 00:19:31,439 Speaker 1: invest in or things you should give money to. So 305 00:19:31,480 --> 00:19:33,119 Speaker 1: how do you resist it you have some person who 306 00:19:33,160 --> 00:19:35,480 Speaker 1: says no for you, or how do you do that? 307 00:19:36,000 --> 00:19:39,480 Speaker 1: Let many people? How many people say no? Well, once 308 00:19:39,520 --> 00:19:43,359 Speaker 1: you pick what you care about. If somebody has something 309 00:19:43,400 --> 00:19:46,600 Speaker 1: that can make a difference in global health, we're super interested. 310 00:19:46,920 --> 00:19:50,680 Speaker 1: And you know we have staff people and if it's 311 00:19:50,720 --> 00:19:52,800 Speaker 1: to do with global help, some of those people will 312 00:19:52,800 --> 00:19:55,280 Speaker 1: come out and talk through with you whatever your innovation 313 00:19:55,520 --> 00:19:58,000 Speaker 1: is and how we can partner with you on that. 314 00:19:58,440 --> 00:20:01,359 Speaker 1: You know, so that's clearly in our area. If it's 315 00:20:01,400 --> 00:20:04,640 Speaker 1: something that can substantially improve K through to welve education, 316 00:20:05,200 --> 00:20:07,879 Speaker 1: then we're going to be very interested in it. If 317 00:20:07,880 --> 00:20:10,639 Speaker 1: people are asking outside of those things, then you know, 318 00:20:10,760 --> 00:20:15,040 Speaker 1: fortunately you can say no, because focuses is key to philanthropy. 319 00:20:15,080 --> 00:20:17,919 Speaker 1: So people have recognized over the years that raising children 320 00:20:18,000 --> 00:20:20,919 Speaker 1: is difficult. Jackie Kennedy famously said, if you mess up 321 00:20:21,000 --> 00:20:24,399 Speaker 1: raising your children, nothing else matters. You have three children, 322 00:20:24,400 --> 00:20:26,520 Speaker 1: seem to be well adjusted, and you've kept them out 323 00:20:26,520 --> 00:20:29,080 Speaker 1: of the newspapers and so forth. How do you avoid 324 00:20:29,119 --> 00:20:32,879 Speaker 1: spoiling kids like that? I think that's a huge problem. 325 00:20:32,920 --> 00:20:38,919 Speaker 1: You know. Obviously our kids have benefited from having a 326 00:20:38,960 --> 00:20:43,560 Speaker 1: great education and an opportunity to travel and uh, you know, 327 00:20:43,600 --> 00:20:47,000 Speaker 1: so they're very lucky in that sense. Making sure that 328 00:20:47,040 --> 00:20:50,080 Speaker 1: the visibility or the way people treat them is not unnatural. 329 00:20:51,080 --> 00:20:54,080 Speaker 1: There're some challenges that come with that. So far, they've 330 00:20:54,119 --> 00:20:57,879 Speaker 1: handled it well. You know, Melinda uh is the one 331 00:20:57,880 --> 00:21:02,160 Speaker 1: who deserves any are certainly almost all the credit uh 332 00:21:02,200 --> 00:21:07,000 Speaker 1: for the kids so far doing very well. You know, 333 00:21:07,000 --> 00:21:09,719 Speaker 1: our kids, we've said to them that that, you know, 334 00:21:09,840 --> 00:21:12,600 Speaker 1: the money is going to the foundation, and so they 335 00:21:12,600 --> 00:21:16,800 Speaker 1: don't think of themselves as sort of aristocratic. What do 336 00:21:16,840 --> 00:21:18,520 Speaker 1: they say to tell them that? They say, can you 337 00:21:18,520 --> 00:21:20,120 Speaker 1: give me a little bit or something? Or they don't 338 00:21:20,359 --> 00:21:22,560 Speaker 1: they don't ask for some they'll get a little bit. 339 00:21:22,640 --> 00:21:26,000 Speaker 1: How much money has your Foundation given away today about 340 00:21:26,000 --> 00:21:30,280 Speaker 1: forty billion dollars. Yeah, we're now up to giving six 341 00:21:30,320 --> 00:21:33,840 Speaker 1: billion dollars a year. So finally, UM, if people are 342 00:21:33,840 --> 00:21:35,639 Speaker 1: watching now and they say, all right, I want to 343 00:21:35,640 --> 00:21:38,119 Speaker 1: do something about climate change, but I'm just one person. 344 00:21:38,160 --> 00:21:40,159 Speaker 1: I don't have the resource of Bill Gates says, what 345 00:21:40,280 --> 00:21:43,520 Speaker 1: can any average person do to have some impact on 346 00:21:43,640 --> 00:21:47,440 Speaker 1: climate change? In your view? Well, certainly they can take 347 00:21:48,720 --> 00:21:54,639 Speaker 1: things like, uh, these new meat products or how they 348 00:21:55,560 --> 00:22:00,440 Speaker 1: buy electricity, and they can help, uh draw hive up 349 00:22:00,480 --> 00:22:05,560 Speaker 1: the scale of the green solutions. The most important thing 350 00:22:05,960 --> 00:22:09,520 Speaker 1: at this stage is their political voice. Uh, there's going 351 00:22:09,560 --> 00:22:14,160 Speaker 1: to be a need to put substantial resources into this effort, 352 00:22:14,720 --> 00:22:20,240 Speaker 1: and you know we need will need a bipartisan solution 353 00:22:20,880 --> 00:22:23,840 Speaker 1: and to send the right signal to the market. You 354 00:22:23,920 --> 00:22:26,720 Speaker 1: actually don't if you just win one year and then 355 00:22:26,760 --> 00:22:30,600 Speaker 1: it gets repealed, that doesn't help at all. The key 356 00:22:30,680 --> 00:22:34,720 Speaker 1: is what people see the policies will be over the 357 00:22:34,760 --> 00:22:38,240 Speaker 1: next thirty years on a consistent basis, and that means 358 00:22:38,280 --> 00:22:41,640 Speaker 1: it's a much higher bar than just a one time victory. 359 00:22:42,040 --> 00:22:45,440 Speaker 1: That was Billy Melinda Gates Foundation co chair and Microsoft 360 00:22:45,480 --> 00:22:46,880 Speaker 1: co founder Bill Gates