Robotic sensing

Robotic sensing is a subarea of robotics science intended to give robots sensing capabilities, so that robots are more human-like. Robotic sensing mainly gives robots the ability to see, touch, hear and move and uses algorithms that require environmental feedback.

Vision
Method
The visual sensing system can be based on anything from the traditional camera, sonar, and laser to the new technology radio frequency identification (RFID), which transmits radio signals to a tag on an object that emits back an identification code. All four methods aim for three procedures—sensation, estimation, and matching.

Image processing
Image quality is important in applications that require excellent robotic vision. Algorithm based on wavelet transform for fusing images of different spectra and different foci improves image quality. Robots can gather more accurate information from the resulting improved image.

Usage
Visual sensors help robots to identify the surrounding and take appropriate action. Robots analyze the image of the immediate environment imported from the visual sensor. The result is compared to the ideal intermediate or end image, so that appropriate movement can be determined to reach the intermediate or final goal.

Touch
Signal processing
Touch sensory signals can be generated by the robot’s own movements. It is important to identify only the external tactile signals for accurate operations. Previous solutions employed the Wiener filter, which relies on the prior knowledge of signal statistics that are assumed to be stationary. Recent solution applies an adaptive filter to the robot’s logic. It enables the robot to predict the resulting sensor signals of its internal motions, screening these false signals out. The new method improves contact detection and reduces false interpretation.

Usage
Touch patterns enable robots to interpret human emotions in interactive applications. Four measurable features—force, contact time, repetition, and contact area change—can effectively categorize touch patterns through the temporal decision tree classifier to account for the time delay and associate them to human emotions with up to 83% accuracy. The Consistency Index is applied at the end to evaluate the level of confidence of the system to prevent inconsistent reactions.

Robots use touch signals to map the profile of a surface in hostile environment such as a water pipe. Traditionally, a predetermined path was programmed into the robot. Currently, with the integration of touch sensors, the robots first acquire a random data point; the algorithm of the robot will then determine the ideal position of the next measurement according to a set of predefined geometric primitives. This improves the efficiency by 42%.

In recent years, using touch as a stimulus for interaction has been the subject of much study. In 2010, the robot seal PARO was built, which reacts to many stimuli from human interaction, including touch. The therapeutic benefits of such human-robot interaction is still being studied, but has shown very positive results.

Hearing

Signal processing
Accurate audio sensors require low internal noise contribution. Traditionally, audio sensors combine acoustical arrays and microphones to reduce internal noise level. Recent solutions combine also piezoelectric devices. These passive devices use the piezoelectric effect to transform force to voltage, so that the vibration that is causing the internal noise could be eliminated. On average, internal noise up to about 7dB can be reduced.

Robots may interpret strayed noise as speech instructions. Current voice activity detection (VAD) system uses the complex spectrum circle centroid (CSCC) method and a maximum signal-to-noise ratio (SNR) beamformer. Because humans usually look at their partners when conducting conversations, the VAD system with two microphones enable the robot to locate the instructional speech by comparing the signal strengths of the two microphones. Current system is able to cope with background noise generated by televisions and sounding devices that come from the sides.

Usage
Robots can perceive emotions through the way we talk. Acoustic and linguistic features are generally used to characterize emotions. The combination of seven acoustic features and four linguistic features improves the recognition performance when compared to using only one set of features.

Acoustic feature
Duration
Energy
Pitch
Spectrum
Cepstral
Voice quality
Wavelets

Linguistic feature
Bag of words
Part-of-speech
Higher semantics
Varia

Movement

Usage
Automated robots require a guidance system to determine the ideal path to perform its task. However, in the molecular scale, nano-robots lack such guidance system because individual molecules cannot store complex motions and programs. Therefore, the only way to achieve motion in such environment is to replace sensors with chemical reactions. Currently, a molecular spider that has one streptavidin molecule as an inert body and three catalytic legs is able to start, follow, turn and stop when came across different DNA origami. The DNA-based nano-robots can move over 100 nm with a speed of 3 nm/min.

In a TSI operation, which is an effective way to identify tumors and potentially cancer by measuring the distributed pressure at the sensor’s contacting surface, excessive force may inflict a damage and have the chance of destroying the tissue. The application of robotic control to determine the ideal path of operation can reduce the maximum forces by 35% and gain a 50% increase in accuracy compared to human doctors.

Performance
Efficient robotic exploration saves time and resources. The efficiency is measured by optimality and competitiveness. Optimal boundary exploration is possible only when a robot has square sensing area, starts at the boundary, and uses the Manhattan metric. In complicated geometries and settings, a square sensing area is more efficient and can achieve better competitiveness regardless of the metric and of the starting point.

Source from Wikipedia