backtop


Print 28 comment(s) - last by Argon18.. on Feb 19 at 11:16 AM

System could be useful for mission critical applications, such as combat robotics

Professor Peter Bentley of the University College of London and his colleague Christos Sakellariou aren't impressed with most everyday computers, which aren't very fault tolerant and can only multitask by flipping their cores between various sequential instruction streams in a program.

He describes in an interview with NewScientist, "Even when it feels like your computer is running all your software at the same time, it is just pretending to do that, flicking its attention very quickly between each program.  Nature isn't like that.  Its processes are distributed, decentralised and probabilistic. And they are fault tolerant, able to heal themselves. A computer should be able to do that."

So the pair set out to make a new hardware and a new operating system, capable of handling tasks differently from most current machines, which even if "parallel" deal with instructions sequentially.

The new machine has instruction set pairs that tell what to do when a certain set of data is encountered.  The instructions-data pairs are then sent to multiple "systems", which are chosen at random to produce results.  Each system has its own redundant stack of instructions, so if one gets corrupted, others can finish up the work.  And each system has its own memory and storage; so "crashes" due to memory/storage errors are eliminated.

Comments Prof. Bentley, "The pool of systems interact in parallel, and randomly, and the result of a computation simply emerges from those interactions."

The results will be presented at an April conference in Singapore. 

The team is currently working on coding the machine so that it can reprogram its own instructions to respond to changes in the environment.  That self-learning, combined with the redundant, pseudorandom nature of the system would make it quite a bit more similar to a human brain than a traditional computer.

Potential applications for such a system include military robotics, swarm robotics, and mission critical servers.  For example, if an unmanned aerial vehicle sustained damage or was hacked, it might be able to reprogram itself and escape errors thanks to the redundancy, allowing it to fly home.

The computer is somewhat similar to so-called "probabilistic" chip designs, which are being researched at other universities.

Source: New Scientist



Comments     Threshold


This article is over a month old, voting and posting comments is disabled

RE: Except
By conquistadorst on 2/18/2013 9:40:02 AM , Rating: 2
Yes, it's entirely the software developers wanting to make things overly complicated. No, no, the users and the stake holders never want anything complicated. They always want simple. The entire universe only gives us simple problems. It's all of these pesky developers' faults for giving them complex solutions to their simple problems.

OR JUST MAYBE, the world has moved beyond solving simple problems to solving complex ones because requirements demand it. Nooooo, that can't be it! I'm sorry for the sarcasm but you just come off as someone with limited exposure and/or experience. Things do get over-engineered for sure but you're painting the world a little simpler than it actually is.


"The whole principle [of censorship] is wrong. It's like demanding that grown men live on skim milk because the baby can't have steak." -- Robert Heinlein














botimage
Copyright 2014 DailyTech LLC. - RSS Feed | Advertise | About Us | Ethics | FAQ | Terms, Conditions & Privacy Information | Kristopher Kubicki