I’ve been detecting some blowback recently regarding all the attention surrounding emerging AI, it’s near-term effect on jobs, and it’s long-term impact on humanity. Having an anticipatory mindset toward artificial intelligence is just the logical thing to do. As I have said before, designing a car without a braking system would be foolish. Anticipating the eventuality that you might need to slow down or stop the car is just good design. Nevertheless, there are a lot of people, important people in positions of power that think this is a lot of hooey. They must think that human ingenuity will address any unforeseen circumstances, that science is always benevolent, that stuff like AI is “a long way off,” that the benefits outweigh the downsides, and that all people are basically good. Disappointed I am that this includes our Treasury Secretary Steve Mnuchin. WIRED carried the story and so did my go-to futurist Amy Webb. In her newsletter Amy states,
“When asked about the future of artificial intelligence, automation and the workforce at an Axios event, this was Mnuchin’s reply: ‘It’s not even on our radar screen,’ he said, adding that significant workforce disruption due to AI is ‘50 to 100’ years away. ‘I’m not worried at all’”
Sigh! I don’t care what side of the aisle you’re on, that’s just plain naive. Turning a blind eye to potentially transformative technologies is also dangerous. Others are skeptical of any regulation (perhaps rightly so) that stifles innovation and progress. But safeguards and guidelines are not that. They are well-considered recommendations that are designed to protect while facilitating research and exploration. On the other side of the coin, they are also not laws, which means that if you don’t want to or don’t care to, you don’t have to follow them.
Nevertheless, I was pleased to see a relatively comprehensive set of AI principles that emerged from the Asilomar Conference that I blogged about a couple of weeks ago. The 2017 Asilomar conference organized by The Future of Life Institute,
“…brought together an amazing group of AI researchers from academia and industry, and thought leaders in economics, law, ethics, and philosophy for five days dedicated to beneficial AI.”
The gathering generated the Asilomar AI Principles, a remarkable first step on the eve of an awesome technological power. None of these people, from the panel I highlighted in the last blog, are anxious for regulation, but at the same time, they are aware of the enormous potential for bad actors to undermine whatever beneficial aspects of the technology might surface. Despite my misgivings, an AGI is inevitable. Someone is going to build it, and someone else will find a way to misuse it.
There are plenty more technologies that pose questions. One is nanotechnology. Unlike AI, Hollywood doesn’t spend much time painting nanotechnological dystopias, perhaps that along with the fact that they’re invisible to the naked eye, lets the little critters slip under the radar. While researching a paper for another purpose, I decided to look into nanotechnology to see what kinds of safeguards and guidelines are in place to deal with that rapidly emerging technology. There are clearly best practices by reputable researchers, scientists, and R&D departments but it was especially disturbing to find out that none of these are mandates. Especially since there are thousands of consumer products that use nanotechnology including food, cosmetics, clothing, electronics, and more. A nanometer is very small. Nanotech concerns itself with creations that exist in the 100nm range and below, roughly 7,500 times smaller than a human hair. In the Moore’s Law race, nanothings are the next frontier in cramming data onto a computer chip, or implanting them into our brains or living cells. However, due to their size, nanoparticles can also be inhaled, absorbed into the skin, flushed into the water supply and leeched into the soil. We don’t know what happens if we aggregate a large number of nanoparticles or differing combinations of nanoparticles in our body. We don’t even know how to test for it. And, get ready. Currently, there are no regulations. That means manufacturers do not need to disclose it, and there are no laws to protect the people who work with it. Herein, we have a classic example of bad decisions in the present that make for worse futures. Imagine the opposite: Anticipation of what could go wrong and sound industry intervention at a scale that pre-empts government intervention or the dystopian scenarios that the naysayers claim are impossible.