You are here

HUBOT- Bridging the gap between Human and Robot Interaction

A variety of robots are available today and each having their own, unique applications. However achieving an efficient means of communication with these robots is still problematic. It is difficult for a robot to accurately analyse and follow voice commands given by its master. During human-human interactions, multiple sounds are simultaneously processed by our brain and we still segregate the sound containing the information we want. This is not the case with robots; they are unable to differentiate various sounds in its environment. This is a major roadblock in human-robot interactions.

‘HUBOT’- is a term coined by the Speech and Audio Group (SAG) at IISc, to represent any effective interaction which occurs between a Human and a Robot. The researchers at IISc are actively working towards enhancing Hubot communication, in order to obtain better response from robots to given commands.  The techniques used to enable Hubot communication include, voice and keyword recognition, along with the gross localisation of sound. The setup includes placing multiple microphones on the robot which pick up various signals from the environment which are then processed for relevant information. The group is making great strides in its research.

“Humans have only one pair of ears. But a robot can have more! We can place the “ears” anywhere we want - we can have two ears on the robot’s arms and have it wave them around!” says Dr T.V. Sreenivas, Professor at the Department of Electrical Communication Engineering, IISc. He heads the Speech and Audio Group and is leading its research activities.

Through their research these scientists have developed several techniques which work to make Hubot communication more efficient. One of these techniques is the Y-Array technique which works in an indoor environment, for localization and tracking of a moving source. This means that even if the robot and master are both in motion, the robot will still be able to efficiently track the commands given. Another such technique involves the use of multiple rotating microphones on the robot. This produces the same effect as moving your head around to pick up the location of a sound.

Digital Signal Processing techniques and properties of speech signals are studied by the SAG to overcome the existing difficulties in human-robot voice based interactions. A lot of processing steps are involved to make the robot behave in a certain way in the considered environment. First, the robot needs to analyze a real life situation or a physical problem faced by it. It then converts this problem into a mathematical form and written as an algorithm. The algorithm, a set of simple steps, is then converted into a computer program thus making the robot work accordingly. The microcontrollers in the robot are used to process the required computations. If there are larger computations needed, an external server which will be connected to the robot through Wi-Fi to perform them. This avoids unnecessary waiting time and results in the efficient functioning of the robot.

The team is currently working on several challenges that must be addressed before their “Hubot” can be considered perfect. When the source of the sound (master) moves with respect to the robot, the robot cannot follow the commands effectively. Acoustic scene analysis, a technique to process and interpret sound from different perspectives in an environment, is used to solve this. The sound that reaches the many microphones on the robot is detected, classified localized and identified to help robots “hear” better.

“Although we may not achieve a perfect conversation-type interaction with a robot, we can improve on making the robot follow voice commands more efficiently”, says Prof. Sreenivas.  Isn’t it exciting to have a Hubot that listens to you and works for you?


Prof. T.V.Sreenivas can be contacted at +91-97-31-317031. Email: