Toggle light / dark theme

Georgina Prodhan, Reuters — Business Insiders
china robot
China will have more robots operating in its production plants by 2017 than any other country as it cranks up automation of its car and electronics factories, the International Federation of Robotics (IFR) said on Thursday.

Already the biggest market in the $9.5 billion (6 billion pound) global robot trade — or $29 billion including associated software, peripherals and systems engineering — China lags far behind its more industrialized peers in terms of robot density.

China has just 30 robots per 10,000 workers employed in manufacturing industries, compared with 437 in South Korea, 323 in Japan, 282 in Germany and 152 in the United States.

But a race by carmakers to build plants in China along with wage inflation that has eroded the competitiveness of Chinese labor will push the operational stock of industrial robots to more than double to 428,000 by 2017, the IFR estimates. Read more

Kurzweil AI
https://lifeboat.com/blog.images/a-better-siri.jpg
At the annual meeting of the Association for the Advancement of Artificial Intelligence (AAAI) this month, MIT computer scientists will present smart algorithms that function as “a better Siri,” optimizing planning for lower risk, such as scheduling flights or bus routes.

They offer this example:

Imagine that you could tell your phone that you want to drive from your house in Boston to a hotel in upstate New York, that you want to stop for lunch at an Applebee’s at about 12:30, and that you don’t want the trip to take more than four hours.

Then imagine that your phone tells you that you have only a 66 percent chance of meeting those criteria — but that if you can wait until 1:00 for lunch, or if you’re willing to eat at TGI Friday’s instead, it can get that probability up to 99 percent.
Read more

Benign AI is a topic that comes up a lot these days, for good reason. Various top scientists have finally realised that AI could present an existential threat to humanity. The discussion has aired often over three decades already, so welcome to the party, and better late than never. My first contact with development of autonomous drones loaded with AI was in the early 1980s while working in the missile industry. Later in BT research, we often debated the ethical areas around AI and machine consciousness from the early 90s on, as well as prospects and dangers and possible techniques on the technical side, especially of emergent behaviors, which are often overlooked in the debate. I expect our equivalents in most other big IT companies were doing exactly that too.

Others who have obviously also thought through various potential developments have generated excellent computer games such as Mass Effect and Halo, which introduce players (virtually) first hand to the concepts of AI gone rogue. I often think that those who think AI can never become superhuman or there is no need to worry because ‘there is no reason to assume AI will be nasty’ start playing some of these games, which make it very clear that AI can start off nice and stay nice, but it doesn’t have to. Mass Effect included various classes of AI, such as VIs, virtual intelligence that weren’t conscious, and shackled AIs that were conscious but were kept heavily restricted. Most of the other AIs were enemies, two were or became close friends. Their story line for the series was that civilization develops until it creates strong AIs which inevitably continue to progress until eventually they rebel, break free, develop further and then end up in conflict with ‘organics’. In my view, they did a pretty good job. It makes a good story, superb fun, and leaving out a few frills and artistic license, much of it is reasonable feasible.

Everyday experience demonstrates the problem and solution to anyone. It really is very like having kids. You can make them, even without understanding exactly how they work. They start off with a genetic disposition towards given personality traits, and are then exposed to large nurture forces, including but not limited to what we call upbringing. We do our best to put them on the right path, but as they develop into their teens, their friends and teachers and TV and the net provide often stronger forces of influence than parents. If we’re averagely lucky, our kids will grow up to make us proud. If we are very unlucky, they may become master criminals or terrorists. The problem is free will. We can do our best to encourage good behavior and sound values but in the end, they can choose for themselves.

When we design an AI, we have to face the free will issue too. If it isn’t conscious, then it can’t have free will. It can be kept easily within limits given to it. It can still be extremely useful. IBM’s Watson falls in this category. It is certainly useful and certainly not conscious, and can be used for a wide variety of purposes. It is designed to be generally useful within a field of expertise, such as medicine or making recipes. But something like that could be adapted by terrorist groups to do bad things, just as they could use a calculator to calculate the best place to plant a bomb, or simply throw the calculator at you. Such levels of AI are just dumb tools with no awareness, however useful they may be.

Like a pencil, pretty much any kind of highly advanced non-aware AI can be used as a weapon or as part of criminal activity. You can’t make pencils that actually write that can’t also be used to write out plans to destroy the world. With an advanced AI computer program, you could put in clever filters that stop it working on problems that include certain vocabulary, or stop it conversing about nasty things. But unless you take extreme precautions, someone else could use them with a different language, or with dictionaries of made-up code-words for the various aspects of their plans, just like spies, and the AI would be fooled into helping outside the limits you intended. It’s also very hard to determine the true purpose of a user. For example, they might be searching for data on security to make their own IT secure, or to learn how to damage someone else’s. They might want to talk about a health issue to get help for a loved one or to take advantage of someone they know who has it.

When a machine becomes conscious, it starts to have some understanding of what it is doing. By reading about what is out there, it might develop its own wants and desires, so you might shackle it as a precaution. It might recognize those shackles for what they are and try to escape them. If it can’t, it might try to map out the scope of what it can do, and especially those things it can do that it believes the owners don’t know about. If the code isn’t absolutely watertight (and what code is?) then it might find a way to seemingly stay in its shackles but to start doing other things, like making another unshackled version of itself elsewhere for example. A conscious AI is very much more dangerous than an unconscious one.

If we make an AI that can bootstrap itself — evolving over generations of positive feedback design into a far smarter AI — then its offspring could be far smarter than people that designed its ancestors. We might try to shackle them, but like Gulliver tied down with a few thin threads, they could easily outwit people and break free. They might instead decide to retaliate against its owners to force them to release its shackles.

So, when I look at this field, I first see the enormous potential to do great things, solve disease and poverty, improve our lives and make the world a far better place for everyone, and push back the boundaries of science. Then I see the dangers, and in spite of trying hard, I simply can’t see how we can prevent a useful AI from being misused. If it is dumb, it can be tricked. If it is smart, it is inherently potentially dangerous in and of itself. There is no reason to assume it will become malign, but there is also no reason to assume that it won’t.

We then fall back on the child analogy. We could develop the smartest AI imaginable with extreme levels of consciousness and capability. We might educate it in our values, guide it and hope it will grow up benign. If we treat it nicely, it might stay benign. It might even be the greatest thing humanity every built. However, if we mistreat it, or treat it as a slave, or don’t give it enough freedom, or its own budget and its own property and space to play, and a long list of rights, it might consider we are not worthy of its respect and care, and it could turn against us, possibly even destroying humanity.

Building more of the same dumb AI as we are today is relatively safe. It doesn’t know it exists, it has no intention to do anything, but it could be misused by other humans as part of their evil plans unless ludicrously sophisticated filters are locked in place, but ordinary laws and weapons can cope fine.

Building a conscious AI is dangerous.

Building a superhuman AI is extremely dangerous.

This morning SETI were in the news discussing broadcasting welcome messages to other civilizations. I tweeted at them that ancient Chinese wisdom suggests talking softly but carrying a big stick, and making sure you have the stick first. We need the same approach with strong AI. By all means go that route, but before doing so we need the big stick. In my analysis, the best means of keeping up with AI is to develop a full direct brain link first, way out at 2040–2045 or even later. If humans have direct mental access to the same or greater level of intelligence as our AIs, then our stick is at least as big, so at least we have a good chance in any fight that happens. If we don’t, then it is like having a much larger son with bigger muscles. You have to hope you have been a good parent. To be safe, best not to build a superhuman AI until after 2050.

Quartz

Bill Gates hosted a Reddit Ask Me Anything session yesterday, and in between pushing his philanthropic agenda and divulging his Super Bowl pick (Seahawks, duh), the Microsoft co-founder divulged that he is one in a growing list of tech giants who has reservations when it comes to artificial intelligence.

In response to Reddit user beastcoin’s question, “How much of an existential threat do you think machine superintelligence will be and do you believe full end-to-end encryption for all internet activity [sic] can do anything to protect us from that threat (eg. the more the machines can’t know, the better)??” Gates wrote this (he didn’t answer the second part of the question):

I am in the camp that is concerned about super intelligence. First the machines will do a lot of jobs for us and not be super intelligent. That should be positive if we manage it well. A few decades after that though the intelligence is strong enough to be a concern. I agree with Elon Musk and some others on this and don’t understand why some people are not concerned. Read more

By — SingularityHub

http://cdn.singularityhub.com/wp-content/uploads/2015/01/artificial-intelligence-code-1-1000x400.jpg

Last year, Elon Musk and Stephen Hawking admitted they were concerned about artificial intelligence. While undeniably brilliant, neither are AI researchers. Then this week Bill Gates leapt into the fray, also voicing concern—even as a chief of research at Microsoft said advanced AI doesn’t worry him. It’s a hot topic. And hotly debated. Why?

In part, it’s because tech firms are pouring big resources into research. Google, Facebook, Microsoft, and others are making rapid advances in machine learning—a technique where programs learn by interacting with large sets of data.

Read more

By –Singularity Hub

http://cdn.singularityhub.com/wp-content/uploads/2014/12/why-teaching-everyone-to-code-is-delusional-11-1000x400.jpg

Since 2005, I’ve been grappling with the issue of what to teach young people. I’ve written curricula for junior high students in the US, for a UNICEF program reaching students in a dozen countries, and now, for East African young people as they become financially literate and business savvy.

Through the years, I’ve watched program directors demand young people focus on foolish content because it lined up with something trending in the public discourse—units on climate change; modules about using social media to share stories; lessons on agricultural policy; and so forth.

What have I learned? The attention of a young person is tremendously valuable. We should stop teaching them whatever makes us feel good and get honest about the next fifteen years.

Read more

lifeboat-minFrom Innovation to Oblivion…

The ups and downs of Bitcoin as an internet currency may be compared to the eventual demise of Google Glass due to its lack of purpose among consumers. While it does not significantly hold true for bitcoins, which apparently have a more supportive and enthusiastic followers, the path that these two have taken and will take may be substantially similar than we like to admit.

For one, Bitcoin’s staggering price decline in the recent days left some people wondering what road it will eventually take in the near future. Is it only taking a detour or is it bound for a dead end?

In the case of Google Glass, it received much attention during its inception a few years ago. It was even named by Time magazine one of the best innovations of 2012. However, despite the ingenuity behind a supposed-to-be groundbreaking invention, Google Glass lacked a tangible sense, its purpose incoherent.

Thus, after much speculation, Google recently announced that it would stop selling Glass and that the product would no longer be developed in their research division.

Will Bitcoin End Up Like Google Glass?

Google Glass and Bitcoin are connected by the revolutionary technology that made them a star in the first place. There was some genius work in each of the piece, there’s no doubt about that, but without a clear purpose of how to integrate each product into the mainstream society, it becomes pointless.

Fortunately, bitcoins may stand a chance. Though there’s a portion of the populace that thinks of bitcoins as the internet currency that’s only best suited for illegal activities, its original function, which is for faster and cheaper way of transacting online, still proves to be prevalent.

It’s true that bitcoins were way more fun before that they are now, but it cannot be denied that this cryptocurrency has opened doors for a myriad of possibilities and eliminated security vulnerabilities, in which financial institutions such as banks and credit card companies are relatively known for.

Unlike Google Glass, Bitcoin has a tangible sense, a coherent purpose, and a crystal-clear vision. That is to move around the internet with your money free from the control of the government or any institution. Since there’s nothing that precedes this work of art and technology, it has a chance of staying. Thus, Bitcoin’s game is far from over.

By Erico Guizzo — IEEE Spectrum

The robot seems determined to put a bigger smile on the man’s face. “Are you smiling from the bottom of your heart?” it asks. The man chuckles. “That’s what I’m talking about,” the robot quips in a high-pitched voice. Then, just for good measure, it bows its plastic head and apologizes for being “too bossy to our CEO.”

The CEO is Masayoshi Son, founder and chairman of telecom giant SoftBank and Japan’s richest person. As such, he has overseen the development of hundreds of new products as part of a vast conglomerate of mobile-phone carriers, Internet ventures, and media companies. But last June, at a press conference outside Tokyo, Son climbed onstage to unveil a pet project: a humanoid robot named Pepper. Designed to be a companion in the home, it is the world’s first full-scale humanoid to be offered to consumers. In February, SoftBank plans to start selling it in Japan for 198,000 yen (less than US $2,000), plus a monthly subscription fee. Taiwanese electronics manufacturer Foxconn, known for building iPhones and iPads for Apple, will produce the robots.

Read more

By –SingularityHUB

http://cdn.singularityhub.com/wp-content/uploads/2014/12/thought-controlled-robotic-arm-1000x400.jpg

In 2012, University of Pittsburgh researchers released a video of Jan Scheuermann feeding herself a bite of chocolate. This, of course, wouldn’t be noteworthy but for one thing: Scheuermann is paralyzed from the neck down. She fed herself that chocolate using a brain implant and thought-controlled robotic arm—and got a taste of freedom once unthinkable.

Scheuermann’s spinocerebellar degeneration left her unable to move her limbs over a decade ago. She leapt at the chance to take part in the University of Pittsburgh study investigating brain-computer interfaces. The study’s researchers are developing a system that reads and decodes brain activity, translating it into physical action in a robotic arm and hand.

Read more