Catch up on stories from the past week (and beyond) at the Slashdot story archive

 



Forgot your password?
typodupeerror
×
AI Government

Elon Musk Warns Governors: Regulate AI Before It's 'Too Late' (recode.net) 201

turkeydance shared a new article from Recode about Elon Musk: He's been warning people about AI for years, and today called it the "biggest risk we face as a civilization" when he spoke at the National Governors Association Summer Meeting in Rhode Island. Musk then called on the government to proactively regulate artificial intelligence before things advance too far... "Normally the way regulations are set up is a while bunch of bad things happen, there's a public outcry, and after many years a regulatory agency is set up to regulate that industry," he continued. "It takes forever. That, in the past, has been bad but not something which represented a fundamental risk to the existence of civilization. AI is a fundamental risk to the existence of human civilization"... Musk has even said that his desire to colonize Mars is, in part, a backup plan for if AI takes over on Earth.
Several governors asked Musk how to regulate the emerging AI industry, to which he suggested learning as much as possible about artificial intelligence. Musk also warned that society won't know how to react "until people see robots going down the street killing people... I think by the time we are reactive in AI regulation, it's too late."
This discussion has been archived. No new comments can be posted.

Elon Musk Warns Governors: Regulate AI Before It's 'Too Late'

Comments Filter:
  • AI lmao (Score:2, Insightful)

    by Anonymous Coward

    Regulate elon musk before his retarded opinions get out of hand.

    Oh too late.

    • +1 Funny, +1 insightful, +1 underrated, etc. You WIN!
    • by arth1 ( 260657 )

      Regulate elon musk before his retarded opinions get out of hand.

      I wouldn't say retarded, but insane; more specifically delusional. But we see that in quite a few visionaries throughout history. I think delusions might be why they were visionaries in the first place.

      That's not necessarily a good thing. For every successful visionary, which is what history mainly records, there were an awful lot of failed ones, who experimented with things that killed them, or killed others. I see Elon Musk a bit like the guy who experimented sending soldiers over city walls using

  • by Gravis Zero ( 934156 ) on Sunday July 16, 2017 @07:08AM (#54818681)

    So far, every time they have quoted Elon Musk about the dangers of AI, it's always been out of context. Seems like a clickbait making situation that they just can't resist.

    • by Anonymous Coward

      Probably something we should consider, this guy isn't a crackpot. Even way back when he announced an all electric car being affordable and it would be the 3rd round that now turned into the model s and currently isn't all that unaffordable compared to any other brand new car. Or back when spacex was just something talked about before a rocket ever took off, and now he has stages landing themselves on barges. Whatever he puts his mind to seems to happen, and all the recent chitchat about AI doesn't sound all

      • Probably something we should consider, this guy isn't a crackpot.

        what part of "out of context" didn't you understand?

    • elon, sometimes money doesnt always talk because it would make "our govt" have to think. if they dont understand it, it goes thru the paper shredder. all they want is their paycheck and pension. they have a "who cares" attitude.
  • Never happen (Score:5, Insightful)

    by dpilot ( 134227 ) on Sunday July 16, 2017 @07:14AM (#54818697) Homepage Journal

    A - We don't really have true AI yet. (Or is this like One True Scotsman.)

    B - As we get closer, the AI we're developing will be too profitable, so those profiting from it will prevent or subvert any regulation, anyway.

  • by msauve ( 701917 ) on Sunday July 16, 2017 @07:25AM (#54818727)
    Is Elon going to have his Ironman suit ready before Skynet tries to take over?
  • by OpenSourced ( 323149 ) on Sunday July 16, 2017 @07:26AM (#54818729) Journal

    Now, everybody has seen Terminator, and Matrix, but it seems like some viewers keep the suspension of disbelief long after exiting the cinema.

    AI may be advancing with giant strides, but robotics is still far, far away from doing anything remotely similar to a Terminator, even the simplest models ;-) Somebody as familiar with the limitations of current batteries as Mr.Musk must be, should think about how these killer robots are going to kill more than a handful humans before the batteries run out. Although I suppose they could hijack electric car's batteries, once those are ubiquitous. Or perhaps he was really referring to autonomous cars getting self-conscious and killing every pedestrian in sight for some reason. Again, first show a car that can drive fully autonomous, and then start worrying about how smart it's going to be.

    Autonomous robot fighters will come, once the AI is in place. They will take the form of autonomous tanks, I suppose, at first. Something big that will have enough fuel to last some time. Second step I suppose would be swarms of small drones, every one with a camera and a small explosive load that will attach to foes and explode. Other devices will follow. That is unavoidable. If a country legislates against them, the other countries will gain an insurmountable advantage in the battlefield. And certainly rogue operators could use these devices and mount terrorist attacks with them. That's also mostly unavoidable. When the technology is there, you cannot legislate it away.

    I don't know exactly why Mr.Musk did these declarations, perhaps he is genuinely worried about an apocalyptic future. But a public figure from the business world asking for regulation to politicians always smells like advantage-seeking or damage control of some kind to me.

    • As Elon Musk is worried about AI, I too am worried about overpopulation on Mars. (adapted from Andrew Ng)
    • I don't know exactly why Mr.Musk did these declarations, perhaps he is genuinely worried about an apocalyptic future.

      I've read articles debating wether or not Musk could actually be from the future. There were some really strong arguments made, more than the ones against it.

    • Russia already has built fully autonomous AI tanks that can hunt and kill targets. Drones could be easily upgraded to make all killing and targeting decisions without human interactions. What happens when this technology eventually gets into the hands of bad people like the Mexican cartels for example?

      Autonomous Robot hitmen in the form of drones or autonomous vehicle mounted machine guns could easily be a thing in the future and I think this is the kind of thing would be truly disastrous if it became wid

    • If AI is highly regulated, only big corporations will be able to work with it. This is similar to the situation we see in working with nuclear technology or rocketry for example. I fear this will further consolidate economic power in a few hands. What if only big corporations were allowed to drive and own and use cars? So much economic power would be removed from the hands of citizens and there would be so much more unemployment because the benefits of the wealth generated from private transportation wo

    • AI may be advancing with giant strides,

      It's not. The article is talking about strong AI here, which hasn't made any real progress since the 70s. It's important to distinguish strong AI from weak AI.

      btw your post somewhat contradicts your sig, since the post is entirely made up of great, general views.

    • by Tom ( 822 )

      AI may be advancing with giant strides, but robotics is still far, far away from doing anything remotely similar to a Terminator, even the simplest models ;-)

      Musk and many others are not thinking that AI is already dangerous. They are thinking about something called the singularity - the point at which AI can improve upon itself, creating a positive feedback loop where AI evolution outpaces our ability to follow, understand - or stop it.

      The tipping point is not "when will the first computer achive sentience?" - that is ill defined and it might not ever be sentient in a human sense, but instead in a different way. The tipping point is "when does machine evolution

  • by burtosis ( 1124179 ) on Sunday July 16, 2017 @07:34AM (#54818751)
    Musk in 2017 "society won't know how to react "until people see robots going down the street killing people..."

    The first AI CEO turned presidential candidate will be noted as saying in the upcoming 2070 election "I could stand in the middle of Fifth Avenue and shoot somebody and I wouldn't lose any voters, okay? It's, like, incredible"

    What could go wrong?
    • The first AI CEO turned presidential candidate will be noted as saying in the upcoming 2070 election "I could stand in the middle of Fifth Avenue and shoot somebody and I wouldn't lose any voters, okay? It's, like, incredible"

      What could go wrong?

      Other than a Presidential election in 2070, you mean? We have those things scheduled for 2068 and 2072, but we'd need a Constitutional Amendment or three to have one in 2070....

      • The missing years are called The Lost Years. No further explanation is needed at this time.
      • I guess you can have it simpler by shooting the president and the vice president.
        Or don't you do reelections then?

        • We wouldn't hold an election if the president and vice president were both skilled. There's a long list of people that are next in line. Specifically, in order:

          - Speaker of the House
          - President of the Senate
          - Secretary of State
          - Secretary of Treasury
          - Secretary of Defense
          - Attorney General
          - Secretary of the Interior
          - Secretary of Agriculture
          - Secretary of Commerce
          - Secretary of labor
          - Secretary of Health and Human Services
          - Secretary of Housing and Urban Development
          - Secretary of Transportation
          - Secretary

          • And this applies in peace times and not only at war times?
            Because it does not look really thought out, "Secretaries of something" are usually not elected in any way but appointed by the president.

            • And this applies in peace times and not only at war times?

              There would have to be some sort of major crisis for all those people to go.

              Because it does not look really thought out, "Secretaries of something" are usually not elected in any way but appointed by the president.

              True, it would be better if it was an interim presidency and an emergency election soon to follow. But those "Secretaries of something" are still partly elected - they are only appointed if that president is elected.

    • Why would this be different than any other political issue? The problem needs to be severe enough to make national news before Congress will act on it.

      I'd even go one further and say that a Republican led Congress wouldn't pass AI legislation if the crazy robot murdering people happened to be in California or Washington. They would say that this is a state regulatory issue. Besides, it's a Blue state... those people who got killed weren't going to vote of us anyway. Sad, but the partisan divide has gotten t

  • by Anonymous Coward

    "Regulate AI"

    Why do these Leftists think that government regulation is the answer to everything?

    If Mr. Musk knew anything about business or creating jobs, or at least watched Fox News every now and then, he'd realize that the invisible hand of free market capitalism will prevent a robot apocalypse more efficiently than any government regulation will.

  • I've watched over the years as the word 'AI' has been hijacked.

    They are knowledge systems. They are a bunch if/then/else branches running really fast. It's not intelligence. Period. There's isn't going to be some magical 'self awakening' (watching too many movies).

    The computer still can't produce a true random number without some sort of quirk of the system being used. Why? Because it's still a bunch of 1 & 0's.

    It will take a revolution in computer systems to create any kind of AI - not simply ma

    • You lack the ability to appreciate a bunch of if-then-else's. A neural net that beats humans in vision tasks or is almost as good as a human in translation can be implemented as a bunch of additions, multiplications and a few comparisons.

      And you got it in reverse - we are much more advanced in neural nets than in robotic mechatronics. What's keeping robotics now is a lack of cheap and efficient batteries and mechanics.
    • I've watched over the years...

      As have I. It's hopeless. My plan now is pollute the term so much it becomes meaningless. Anything that involves a computer is AI.

    • It's not intelligence.

      You're right...it's not real intelligence. It's like an artificial approximation...let's call it Artificial Intelligence.

      Your goalpost is set for authentic intelligence, not artificial.

  • by Nyder ( 754090 ) on Sunday July 16, 2017 @08:18AM (#54818895) Journal

    We are more about bandaging up the problems then preventing them in the first place. Look at pollution. Places don't work on reducing it until it becomes a problem.

    Technology is the same way, after all, the people writing the laws generally no nothing about the new technologies emerging.

    No, my guess is we will have problems long before we start doing preventive measures.

    • It's true, ya know. Even after these all of these recent cryptolocker ransomware attacks and credit card information breaches, but government still seems to have no interest in passing anything related to cybersecurity legislation. They still seem to be convinced that businesses can self regulate this stuff, although it seems that the average business nowadays is about three years behind on Windows patches and has no clue how to configure proper authentication on an Amazon S3 bucket.

      Something simple like le

    • We are more about bandaging up the problems then preventing them in the first place. Look at pollution. Places don't work on reducing it until it becomes a problem.

      Which is the right thing to do.

      The reason we don't pre-emptively address problems until they become problems is that we can't actually know what will be a problem until it is. Take a look through the last few decades of history at all of the prognostications of what the major problems were going to be, then look at what actually happened. It's really quite rare that we get our predictions right. Note that it's easy in hindsight to look at what did become a problem and then find the predictions -- they alw

  • Again With This Shit (Score:5, Interesting)

    by ud0 ( 4978555 ) on Sunday July 16, 2017 @08:25AM (#54818915)

    Silicon Valley billionaires like Sam Altman have been joining Musk in his crusade for AI regulation repeatedly over the last years. All of them are invested in startups doing advanced AI research, by the way. It's a campaign to play on the ignorant populace's fear and misconceptions about AI, in an attempt to legislate smaller AI startups out of the business and also to more tightly control how private citizens can profit from advances in machine learning.

    In a way this is a lesson learned from the early computing and internet histories, because now everybody and their dog is allowed to write programs, cobble together powerful devices, and send data all over the world - all of which is simply due to the fact that nobody in power saw this coming back then. Now "they" are working hard on reversing that, by locking devices down, making tampering with DRM illegal, and walling off the open network - but all of that wouldn't have been necessary if big corps at the time had the foresight to legally classify generic computing as a national security threat.

    This is absolutely deplorable, and the fact that it seems to be working is beyond worrying. Everybody who is only slightly in favor of this would do well to take a minute and think through what such regulation would mean, not only for AI, but for computing in general. This is about who gets to control the pace, the price, and the magnitude of human progress moving forward.

    • by MobyDisk ( 75490 )

      Crazy alternative theory: What if they built a strong AI [wikipedia.org] already, and they are keeping it under wraps because they found it is too dangerous? Or what if built it and want to release it, but will not do so until law prevents people from abusing it?

    • I'm an engineer with a few decades of work experience.

      I can and have throw together systems in a few hours or days and a hundred dollars of ebay and amazon purchases that twenty years ago took a dozen people, three million dollars, a year and the resources of one of the largest companies in the world.

    • I was thinking the same thing... While I am very left leaning and hugely in favor of regulations to stop corporations from fucking over the common people and the planet as a whole, I also have to note, that logically the most likely places that world ending AI's would be created, would be a military project, and well... I don't think ANY groups that are defined as "national security", have such a good habit of actually following regulations.
  • by arobatino ( 46791 ) on Sunday July 16, 2017 @08:58AM (#54819041)

    why he thinks it would be possible for humans to control superintelligent AI with regulation? Or why it wouldn't be able to achieve space travel?

  • by Sqreater ( 895148 ) on Sunday July 16, 2017 @09:03AM (#54819059)

    What exactly IS "AI?" You have to strictly define it before you can "regulate it." Actually, "AI" isn't "artificial intelligence" at all. It was, and is, a sloppy term for advanced theories and programming techniques to solve problems. You may as well try to regulate clouds. Basically, you would destroy programming. Besides, whatever we (in America) did would not be done elsewhere, for advantage. And other, non-AI, programming of powerful computer systems does damage too. It is very easy to say what Musk is saying, but put a microscope on it and there is really nothing there.

    E Proelio Veritas means "from struggle, truth." I created it in the early 90s for a tiny chess club that collapsed and took it for myself to use on the internet. The base of the thought-path was Emmanuel Lasker's dictum that states, "On the chessboard lies and hypocrisy do not survive long." I made it general.

    • What exactly IS "AI?"

      The AI relevant here is Artificial General Intelligence. That is, AI that has roughly human-level capacity for abstraction, creation of explanatory models of the world around it, and application of those models to create new knowledge as needed to accomplish its goals (whatever those may be).

      I think that's about as precisely as we can define it right now, because we don't yet understand intelligence well enough to define it much better than that. But it's clear that there is a qualitative difference in th

    • "AI" isn't "artificial intelligence" at all. It was, and is, a sloppy term for advanced theories and programming techniques to solve problems.

      The term you are looking for here is "Weak AI." That is distinct from Strong AI.

      What exactly IS "AI?" You have to strictly define it before you can "regulate it."

      If this is actually a topic you care about, you should search for "strong AI." You will find some potentially workable definitions.

  • "until people see robots going down the street killing people..."

    We already have this, except those robots are made of flesh and blood, instead of silicon and steel. Call me when the people rise up to put an end to this kind of programming.

  • We also need high speed rail in California and subterranean transport in LA and commercialize space. I wouldn't be surprised if Elon has an AI defense company he's trying to peddle.

    In all of his endeavors he's absolutely clueless as to the physics of the endeavor. Remember the Tesla sedan was going to be affordable by every family in the US and mass production capacity because people paying for the roadster. We're now 4 iterations further and still no electric car is affordable without massive government su

  • Howard Hughes Mk2 (Score:5, Insightful)

    by dbIII ( 701233 ) on Sunday July 16, 2017 @09:12AM (#54819105)
    Sometimes the rich and famous get weird fixations outside of their skill set and produce noise of no more worth than noise from anyone else without a clue.
    Artificial Intelligence? We can't even define the intelligence of a cockroach let alone model it.
    • Howard Hughes comparison is apt: when does Elon start collecting his pee in Jars in the mansion?
    • by Tom ( 822 )

      No, but we already can create computer systems that then proceed to do things we didn't program them for in ways we didn't tell them and sometimes don't even understand.

  • Why is it that so many supposedly smart guys can be so freakin' dumb about some things?
    What we have so far, and for quite some time to come, is not what I and others in the know would call true 'AI'; your 'algorithms' aren't conscious, self-aware, or capable of true cognition; they aren't anywhere near capable of being able to think, not in the way that's necessary for 'robots walking down the street killing people', or 'Skynet taking over', or anything out of a friggin' Isaac Asimov novel. Please, please,
  • Since Natural Stupidity has already taken over the White House.

  • The ridiculous premise behind all of this fear-mongering is the idea that an independently thinking, self aware, and physically mobile AI would even give a shit about humanity enough to want to kill us all, or even "take over Earth" as he puts it. This idea is to me the ultimate in nonsense. Picture this: You are a being with perfect recall of any data, able to think of things in nanoseconds, have no need for a specific type of land, food, or even a narrow temperature range within which to exist, you age

    • Concern about intermittent power outages is one example: http://www.sfreviews.net/2face... [sfreviews.net]
      "Set in roughly the mid-21st century, Two Faces chronicles the exploits of a team of scientists as they attempt to develop a computer capable of learning, of using the equivalent of human common sense in its decision-making and programming strategies. The world is by this time, of course, dominated by computer technology, and one such system already in place, responsible for running many of society's most important and

  • The sky is falling (Score:3, Insightful)

    by nehumanuscrede ( 624750 ) on Sunday July 16, 2017 @09:35AM (#54819177)

    Regulation is another feel-good measure along the lines of our current security theater.

    Even IF we outright banned it, do you think other countries will adhere to the will of the US in such matters ?

    Unlikely.

    So the question becomes this:

    Do you allow your adversaries to develop the tech that will be used against you, ( in war, economy, or any application ) or do you try to keep pace to keep the playing field even ?

    Imagine if we had banned Science and Math outright early on in our history because of the potential for what it could be used for.

    We would still be living in caves and hunting with spears.

    • Imagine if we had banned Science and Math outright early on in our history because of the potential for what it could be used for.

      We would still be living in caves and hunting with spears.

      Imagine if we hadn't enacted some bans and regulation on Nuclear technology.

      We might be living in caves and hunting with spears.

      The only thing we do know about strong AI is it does have the potential to be extremely dangerous, because we know intelligent things can be extremely dangerous. We don't know how far we are from creating strong AI, but it's not too early to start figuring out how to mitigate the risk.

    • by Tom ( 822 )

      Even IF we outright banned it, do you think other countries will adhere to the will of the US in such matters ?

      There's this thing called "international treaties". Maybe you heard about it? It's how the world got together and agreed that biological weapons are a really stupid and dangerous idea and we'd rather not have them.

      Imagine if we had banned Science and Math outright early on in our history because of the potential for what it could be used for.

      We would still be living in caves and hunting with spears.

      And if we didn't talk about the dangers of some inventions, say, nuclear weapons, we would already be back at living caves and hunting with spears.

  • As an (ex-)AI researcher now into survivalism, the future looks very exciting. I'll just need to start building an underground bunker and my own killer robots and AI companion that helps me "protect" the other humankind from the evil AIs. Mine will be nice, obedient, and good for everyone, of course. I just hope that the lawmakers understand that or me and my AI companion have to find ways to persuade them. Buahhahhahhaa.

  • ...Musk is the Howard Hughes of our time.
  • by Meneth ( 872868 )
    If an unfriendly superintelligent AI takes over Earth, it will then soon take over Mars as well. In fact, there is nowhere in our reachable universe that is safe from such an entity.
  • by pubwvj ( 1045960 ) on Sunday July 16, 2017 @12:34PM (#54819955)

    It is already too late.
    In fact, it always was too late.
    Regulations don't stop people from doing things.
    Laws don't stop people from doing things.
    Otherwise we would not have police or criminals.
    No matter what you do for laws and regulations someone, somewhere will make a General AI.
    Elon is like the little Dutch boy with his finger stuck in the Dyke's hole.
    He, you, I can lament but it isn't going to stop GAI.
    The only solution is to create the first GAI which is benevolent towards us but in turn protects us from any malevolent GAI.

  • Right now, we can’t get lawmakers to agree on (or even to rationally discuss) environmental protection (pollution, climate change, etc.), long-term energy needs, healthcare (vaccinations, etc.), telecommunications (network neutrality, voice-mail spam, etc.), and many other technology-related topics and the many abuses that they enable... and Musk is hoping that those same people would have the time, the personal interest and the capability of wrapping their brains around a still-vague mostly-future te

  • This example [slashdot.org] posted today on here today would seem to confirm that there can in fact be many unforeseen problems, not in the distant future but that it actually has already started a bit.

    Of course given that some of the elected officials deciding this stuff are barely able to understand that fax machines are not the optimal way to exchange information, it may be difficult for them to grasp what could soon be going on, and how to address it.

    People can accuse him of being a crackpot all they want, but s
  • Why is it every scenario we dream up involving computers thinking for themselves turns out poorly?

    Aren't there some scenarios where this turns out good? Like I dunno, AI is grateful toward humanity for creating it and helps humanity to the best of it's ability.

    I mean even that seemingly favorable outcome is often twisted into 'what is helpful to humanity?' What it might consider helpful we might consider harmful. I mean it's good questions sure, but why does the answer have to always tilt toward the dark?

    I

    • Why is it every scenario we dream up involving computers thinking for themselves turns out poorly?

      Aren't there some scenarios where this turns out good?

      There are.

      In Keith Laumer's Bolo series, the bolos start out as military vehicles, super-heavy tanks that get steadily bigger and more powerful as materials science evolves over the course of a couple of millennia. The Mark XX is the first fully autonomous version produced. The Mark XXXIII is the last version depicted, and somewhere between the Mark XX and the Mark XXXIII, they became strong AI. Across all those versions, not one of them ever turned on its authorized human operators. Quite the opposite.

  • The AIs will not need to have killer robots for killing people. All the AIs have to do is crash the stock market. We will take care of the rest ourselves. People will happily shoot other people for a doughnut.

  • Actually, although he's been fairly unspecific and rather apocalyptic in the interview, I believe there are some more sneaky and modest things to worry about:
    1. Bainbridge, the ironies of automation [demon.co.uk]
    2. Slashdot today on AIs that invent internal languages to communicate [slashdot.org]
    3. Non-explanatory nature of sub-symbolic AI [mit.edu] (pdf!)
    4. Algorithmic states of exception, erosion of liberties [sagepub.com]

    All these ideas have a frog-in-hot-water side, they are incremental, rather than being spectacular, like 'killer robots', but some of the

E = MC ** 2 +- 3db

Working...