Ok

By continuing your visit to this site, you accept the use of cookies. These ensure the smooth running of our services. Learn more.

Dec 08, 2009

FaceBots

The world's first robot with its own Facebook page (and that can use its information in conversations with "friends") has been developed by Nikolaos Mavridis and collaborators from the Interactive Robots and Media Lab at the United Arab Emirates University.

The main hypothesis of the FaceBots project is that long-term human robot interaction will benefit by reference to "shared memories" and "events relevant to shared friends" in human-robot dialogues.

More to explore:

  • N. Mavridis, W. Kazmi and P. Toulis, "Friends with Faces: How Social Networks Can Enhance Face Recognition and Vice Versa", contributed book chapter to Computational Social Networks Analysis: Trends, Tools and Research Advances, Springer Verlag, 2009. pdf

    N. Mavridis, W. Kazmi, P. Toulis, C. Ben-AbdelKader, "On the synergies between online social networking, Face Recognition, and Interactive Robotics", CaSoN 2009. pdf

    N. Mavridis, C. Datta et al,  "Facebots: Social robots utilizing and publishing social information in Facebook", IEEE HRI 2009. pdf

 

Mar 16, 2008

A second life for AI

Source: Eetimes 

Passing the Turing test - the holy grail of AI (a human conversing with a computer can't tell it's not human) - may now be possible in a limited way with the world's fastest supercomputer (IBM's Blue Gene) and mimicking the behavior of a human-controlled avatar in a virtual world, according to AI experts at Rensselaer Polytechnic Institute. "We are building a knowledge base that corresponds to all of the relevant background for our synthetic character--where he went to school, what his family is like, and so on," said Selmer Bringsjord, head of Rensselaer's Cognitive Science Department and leader of the research project. The researchers plan to engineer, from the start, a full-blown intelligent character and converse with him in an interactive virtual environment, like Second Life.

 

read full article here

Dec 19, 2007

Avatar-controlled robots

Via KurzweilAI.net

Researchers at Korea Advanced Institute of Science and Technology have developed a system for controlling physical robots using software robots, displayed as virtual-reality avatars.

 


Article

 

Dec 04, 2007

Simroid

Via Pink Tentacle 

Simroid --

 

Simroid is a robotic dental patient designed by Kokoro Company Ltd as a training tool for dentists.

The simulated patient can follow spoken instructions, closely monitor a dentist’s performance during mock treatments, and react in a human-like way to mouth pain thanks to mouth sensors.

video

Oct 25, 2007

Japanese android recognizes and uses body language

Via Pink Tentacle

NICT develops robot with nonverbal communication skills --

 

Japan’s National Institute of Information and Communications Technology (NICT) researchers have developed an autonomous humanoid robot that can recognize and use body language. According to the press release, the android can use nonverbal communication skills such as gestures and touch to facilitate natural interaction with humans. NICT researchers envision future applications of this technology in robots that can work in the home or assist with rescue operations when disaster strikes.

 

NICT press release (japanese)

Oct 07, 2007

NIPS 2007 WORKSHOP: Robotics Challenges for Machine Learning

Via Neurobot

Dates: 7-8 December, 2007

Organizers:
Jan Peters (Max Planck Institute for Biological Cybernetics & USC), Marc Toussaint (Technical University of Berlin)

http://www.robot-learning.de
email: nips07@robot-learning.de

Abstract Submission Deadline: October 21, 2007
Acceptance Notification: October 26, 2007

Oct 01, 2007

Artificial brain falls for optical illusions

Via New Scientist

A computer program that emulates the human brain falls for the same optical illusions humans do.

It suggests the illusions are a by-product of the way babies learn to filter their complex surroundings. Researchers say this means future robots must be susceptible to the same tricks as humans are in order to see as well as us.

For some time, scientists have believed one class of optical illusions result from the way the brain tries to disentangle the colour of an object and the way it is lit. An object may appear brighter or darker, either because of the shade of its colour, or because it is in bright light or shadows.

The brain learns how to tackle this through trial and error when we are babies, the theory goes. Mostly it gets it right, but occasionally a scene contradicts our previous experiences. The brain gets it wrong and we perceive an object lighter or darker than it really is – creating an illusion

 

Read full article

 

Sep 20, 2007

Japanese seniors bored with robot

From Robot.net


 

According to a Reuters report, Japanese senior citizens quickly become bored with the simple robots so far introduced into nursing homes:

"The residents liked ifbot for about a month before they lost interest. Stuffed animals are more popular." Ifbot is a small robot that can converse, sing, express emotions, and even present trivia quizzes to senior citizens. According to the article the robot has spent most of the past year sitting in a corner, unused. Another robot, Hopis, that looked like a furry pink dog has gone out of production due to poor sales. Hopis was designed to monitor blood sugar, blood pressue, and body temperature. One problem may be that both robots are little more than advanced toys. Neither can help elderly people with day-to-day problems they face such as getting around their house, reading small print, or taking a bath. The elderly have found utilitarian improvements in existing devices more useful: height-adjustable countertops or extra-big control buttons on household gadgets. Whether seniors will find robots that can help with their utilitarian needs more to their liking than fluffy pupbots that sing remains to be seen"

Sep 16, 2007

EU researchers implanted an artificial cerebellum inside a robot

Roland Piquepaille reports that "An international team of European researchers has implanted an artificial cerebellum - the portion of the brain that controls motor functions - inside a robotic system. This EU-funded project is dubbed SENSOPAC, an acronym for "SENSOrimotor structuring of perception and action for emerging cognition". One of the goals of this project is to design robots able to interact with humans in a natural way. This project, which should be completed at the end of 2009, also wants to produce robots which would act as home-helpers for disabled people, such as persons affected by neurological disorders, such as Parkinson's disease."

SENSOPAC website 

 

Sep 13, 2007

Telepresence robot for interpersonal communication with the elderly

Developing a Telepresence Robot for Interpersonal Communication with the Elderly in a Home Environment.

Telemed J E Health. 2007 Aug;13(4):407-424

Authors: Tsai TC, Hsu YL, Ma AI, King T, Wu CH

"Telepresence" is an interesting field that includes virtual reality implementations with human-system interfaces, communication technologies, and robotics. This paper describes the development of a telepresence robot called Telepresence Robot for Interpersonal Communication (TRIC) for the purpose of interpersonal communication with the elderly in a home environment. The main aim behind TRIC's development is to allow elderly populations to remain in their home environments, while loved ones and caregivers are able to maintain a higher level of communication and monitoring than via traditional methods. TRIC aims to be a low-cost, lightweight robot, which can be easily implemented in the home environment. Under this goal, decisions on the design elements included are discussed. In particular, the implementation of key autonomous behaviors in TRIC to increase the user's capability of projection of self and operation of the telepresence robot, in addition to increasing the interactive capability of the participant as a dialogist are emphasized. The technical development and integration of the modules in TRIC, as well as human factors considerations are then described. Preliminary functional tests show that new users were able to effectively navigate TRIC and easily locate visual targets. Finally the future developments of TRIC, especially the possibility of using TRIC for home tele-health monitoring and tele-homecare visits are discussed.

Jul 26, 2007

Halluc II: 8-legged robot vehicle

Via Pink Tentacle

Halluc II --

 

 

 

 

 

 

 

 

Researchers at the Chiba Institute of Technology have developed a robotic vehicle with eight wheels and legs designed to drive or walk over rugged terrain. The agile robot, which the developers aim to put into practical use within the next five years, can move sideways, turn around in place and drive or walk over a wide range of obstacles.

The researchers hope the robot’s abilities will help out with rescue operations, and they would like to see Halluc II’s technology put to use in transportation for the mobility-impaired.

Here’s a short video of the model in action.

 

Jul 14, 2007

RunBot

Via Medgadget

 

Researchers from Germany and the United Kingdom have developed a bipedal walking robot, capable of self-stabilizing via a highly-developed learning process.

From the study abstract:

In this study we present a planar biped robot, which uses the design principle of nested loops to combine the self-stabilizing properties of its biomechanical design with several levels of neuronal control. Specifically, we show how to adapt control by including online learning mechanisms based on simulated synaptic plasticity. This robot can walk with a high speed (>3.0 leg length/s), self-adapting to minor disturbances, and reacting in a robust way to abruptly induced gait changes. At the same time, it can learn walking on different terrains, requiring only few learning experiences. This study shows that the tight coupling of physical with neuronal control, guided by sensory feedback from the walking pattern itself, combined with synaptic learning may be a way forward to better understand and solve coordination problems in other complex motor tasks.
 

The paper: Adaptive, Fast Walking in a Biped Robot under Neuronal Control and Learning (Manoonpong P, Geng T, Kulvicius T, Porr B, Worgotter F (2007) Adaptive, Fast Walking in a Biped Robot under Neuronal Control and Learning. PLoS Comput Biol 3(7): e134)

Jul 11, 2007

Using a Robot to Teach Human Social Skills

Via KurzweilAI

A humanoid robot designed to teach autistic children social skills has begun testing in British schools. Known as KASPAR (Kinesics and Synchronisation in Personal Assistant Robotics), the $4.33 million bot smiles, simulates surprise and sadness

Read full article

 

Jul 06, 2007

Epigenetic Robotics 2007 (Extended Deadline)

Via NeuroBot

5-7 November 2007, Piscataway, NJ, USA

Seventh International Conference on Epigenetic Robotics: Modeling Cognitive Development in Robotic Systems

http://www.epigenetic-robotics.org


Email: epirob07@epigenetic-robotics.org

Location:
Rutgers, The State University of New Jersey,
Piscataway, NJ, USA

*Extended* Submission Deadline: 1 August 2007

 

Apr 29, 2007

First DARPA Limb Prototype

From the DARPA press release

(via Medgadget

Proto1a.jpg

 

An international team led by the Johns Hopkins University Applied Physics Laboratory (APL) in Laurel, Md., has developed a prototype of the first fully integrated prosthetic arm that can be controlled naturally, provide sensory feedback and allows for eight degrees of freedom--a level of control far beyond the current state of the art for prosthetic limbs. Proto 1, developed for the Defense Advanced Research Projects Agency (DARPA) Revolutionizing Prosthetics Program, is a complete limb system that also includes a virtual environment used for patient training, clinical configuration, and to record limb movements and control signals during clinical investigations.

The DARPA prosthetics program is an ambitious effort to provide the most advanced medical and rehabilitative technologies for military personnel injured in the line of duty. Over the last year, the APL-led Revolutionizing Prosthetics 2009 (RP 2009) team has worked to develop a prosthetic arm that will restore significant function and sensory perception of the natural limb. Proto 1 and its virtual environment system were delivered to DARPA ahead of schedule, and Proto 1 was fitted for clinical evaluations conducted by team partners at the Rehabilitation Institute of Chicago (RIC) in January and February.

"This progress represents the first major step in a very challenging program that spans four years and involves more than 30 partners, including government agencies, universities, and private firms from the United States, Europe, and Canada," says APL's Stuart Harshbarger, who leads the program. "The development of this first prototype within the first year of this program is a remarkable accomplishment by a highly talented and motivated team and serves as validation that we will be able to implement DARPA's vision to provide, by 2009, a mechanical arm that closely mimics the properties and sensory perception of a biological limb."

 

Proto1.jpg

 

APL, which was responsible for much of the design and fabrication of Proto 1, and other team members are already hard at work on a second prototype, expected to be unveiled in late summer. It will have more than 25 degrees of freedom and the strength and speed of movement approaching the capabilities of the human limb, combined with more than 80 individual sensory elements for feedback of touch, temperature, and limb position.

"There is still significant work to be done to determine how best to control this number of degrees of freedom, and ultimately how to incorporate sensory feedback based on these sensory inputs within the human nervous system," Harshbarger says. "The APL team is already driving a virtual model of Proto 2 with data recorded during the clinical evaluation of Proto 1, and the team is working to identify a robust set of grasps that can be controlled by a second patient later this year."

Another exciting development is the functional demonstration of Injectable MyoElectric Sensor (IMES) devices--very small injectable or surgically implantable devices used to measure muscle activity at the source verses surface electrodes on the skin that were used during testing of the first prototype.

 

Apr 27, 2007

Evolution of visually guided behavior in artificial agents

Evolution of visually guided behavior in artificial agents.

Network. 2007 Mar;18(1):11-34

Authors: Boots B, Nundy S, Purves D

Recent work on brightness, color, and form has suggested that human visual percepts represent the probable sources of retinal images rather than stimulus features as such. Here we investigate the plausibility of this empirical concept of vision by allowing autonomous agents to evolve in virtual environments based solely on the relative success of their behavior. The responses of evolved agents to visual stimuli indicate that fitness improves as the neural network control systems gradually incorporate the statistical relationship between projected images and behavior appropriate to the sources of the inherently ambiguous images. These results: (1) demonstrate the merits of a wholly empirical strategy of animal vision as a means of contending with the inverse optics problem; (2) argue that the information incorporated into biological visual processing circuitry is the relationship between images and their probable sources; and (3) suggest why human percepts do not map neatly onto physical reality.

Apr 22, 2007

Seventh International Conference on Epigenetic Robotics

Seventh International Conference on Epigenetic Robotics: Modeling Cognitive Development in Robotic Systems

Call for Papers: Epigenetic Robotics 2007

5-7 November 2007, Piscataway, NJ, USA

Location: Rutgers, The State University of New Jersey, Piscataway, NJ, USA

Apr 11, 2007

MOBI

From Networked Performance

mobi.jpg



MOBI (Mobile Operating Bi-directional Interface), by Graham Smith, is a human sized telepresence robot that users remotely control to move through distant environments, see through its camera eye, talk through its speakers and hear via its microphone ear. Simultaneously a life sized image of themselves is projected onto the robots LCD face, creating a robotic avatar. MOBI allows people to "explore far away art shows, attend distant presentations and make public appearences from anywhere on earth, thus helping to reduce air travel and reduce global warming". MOBI is at DEAF 07.

Graham Smith is a leading expert in the fields of telepresence, virtual reality, videoconferencing and robotics. He has worked with leading Canadian high tech companies for more than 14 years, including Nortel, Vivid Effects, VPL, BNR and IMAX. Graham initiated and headed the Virtual Reality Artist Access Program at the world-renowned McLuhan Program at the University of Toronto, and has lectured internationally. He holds numerous patents in the field of telepresence and panoramic imaging, and was recognized in Macleans magazine as one of the top 100 Canadians to watch
.

Mar 16, 2007

Socially assistive robotics for post-stroke rehabilitation

Journal of NeuroEngineering and Rehabilitation

Maja J Matari

Background: Although there is a great deal of success in rehabilitative robotics applied to patient recovery post stroke, most of the research to date has dealt with providing physical assistance. However, new rehabilitation studies support the theory that not all therapy need be hands-on. We describe a new area, called socially assistive robotics, that focuses on non-contact patient/user assistance. We demonstrate the approach with an implemented and tested post-stroke recovery robot and discuss its potential for effectiveness. Results: We describe a pilot study involving an autonomous assistive mobile robot that aids stroke patient rehabilitation by providing monitoring, encouragement, and reminders. The robot navigates autonomously, monitors the patient's arm activity, and helps the patient remember to follow a rehabilitation program. We also show preliminary results from a follow-up study that focused on the role of robot physical embodiment in a rehabilitation context. Conclusion: We outline and discuss future experimental designs and factors toward the development of effective socially assistive post-stroke rehabilitation robots.

Robot/computer-assisted motivating systems for personalized, home-based, stroke rehabilitation

 
Michelle J Johnson, Xin Feng, Laura M Johnson and Jack M Winters
 
Background: There is a need to improve semi-autonomous stroke therapy in home environments often characterized by low supervision of clinical experts and low extrinsic motivation. Our distributed device approach to this problem consists of an integrated suite of low-cost robotic/computer-assistive technologies driven by a novel universal access software framework called UniTherapy. Our design strategy for personalizing the therapy, providing extrinsic motivation and outcome assessment is presented and evaluated. Methods: Three studies were conducted to evaluate the potential of the suite. A conventional force-reflecting joystick, a modified joystick therapy platform (TheraJoy), and a steering wheel platform (TheraDrive) were tested separately with the UniTherapy software. Stroke subjects with hemiparesis and able-bodied subjects completed tracking activities with the devices in different positions. We quantify motor performance across subject groups and across device platforms and muscle activation across devices at two positions in the arm workspace. Results: Trends in the assessment metrics were consistent across devices with able-bodied and high functioning strokes subjects being significantly more accurate and quicker in their motor performance than low functioning subjects. Muscle activation patterns were different for shoulder and elbow across different devices and locations. Conclusion: The Robot/CAMR suite has potential for stroke rehabilitation. By manipulating hardware and software variables, we can create personalized therapy environments that engage patients, address their therapy need, and track their progress. A larger longitudinal study is still needed to evaluate these systems in under-supervised environments such as the home.

1 2 3 Next