Development of a Basic Robot System for Medical and Nursing Care for Patients with Glaucoma

Medical methods to completely treat glaucoma are yet to be developed. Therefore, ophthalmologists manage patients mainly to delay disease progression. Patients with glaucoma are mainly elderly individuals. In elderly people's houses, having an equipment that can provide medical treatment and care can release their family from their care. For elderly people with the glaucoma to live by themselves as much as possible, we developed a support robot having five functions: elderly people care, ophthalmological examination, trip assistance to the neighborhood, medical treatment, and data referral to a hospital. The medical and nursing care robot should approach the visual field that the patients can see at a speed suitable for their eyesight. This is because the robot will be dangerous if it approaches the patients from the visual field that they cannot see. We experimentally developed a robot that brings a white cane to elderly people with glaucoma. The base part of the robot is a carriage, which is a Megarover 1.1, and it has two infrared sensors. The robot moves along a white line on the floor using the infrared sensors and has a special arm, which does not use electricity. The arm can scoop the block attached to the white cane. Next, we also developed a direction detector comprised of a charge-coupled device camera (SVR41ResucueHD; Sun Mechatronics), goggles (MG-277MLF; Midori Anzen Co. Ltd.), and biconvex lenses with a focal length of 25 mm (Edmund Co.). Some young people were photographed using the direction detector, which was put on their faces. Image processing was performed using Scilab 6.1.0 and Image Processing and Computer Vision Toolbox 4.1.2. To measure the people's line of vision, we calculated the iris's center of gravity using five processes: reduction, trimming, binarization or gray scale, edge extraction, and Hough transform. We compared the binarization and gray scale processes in image processing. The binarization process was better than the gray scale process. For edge extraction, we compared five methods: Sobel, Prewitt, Laplacian of Gaussian, fast Fourier transform, and Canny. The Canny method was the optimal extraction method. We performed the Hough transform to search for the main coordinates from the iris's edge, and we found that the Hough transform could calculate the center point of the iris.

Sound Selection for Gesture Sonification and Manipulation of Virtual Objects

New sensors and technologies – such as microphones, touchscreens or infrared sensors – are currently making their appearance in the automotive sector, introducing new kinds of Human-Machine Interfaces (HMIs). The interactions with such tools might be cognitively expensive, thus unsuitable for driving tasks. It could for instance be dangerous to use touchscreens with a visual feedback while driving, as it distracts the driver’s visual attention away from the road. Furthermore, new technologies in car cockpits modify the interactions of the users with the central system. In particular, touchscreens are preferred to arrays of buttons for space improvement and design purposes. However, the buttons’ tactile feedback is no more available to the driver, which makes such interfaces more difficult to manipulate while driving. Gestures combined with an auditory feedback might therefore constitute an interesting alternative to interact with the HMI. Indeed, gestures can be performed without vision, which means that the driver’s visual attention can be totally dedicated to the driving task. In fact, the auditory feedback can both inform the driver with respect to the task performed on the interface and on the performed gesture, which might constitute a possible solution to the lack of tactile information. As audition is a relatively unused sense in automotive contexts, gesture sonification can contribute to reducing the cognitive load thanks to the proposed multisensory exploitation. Our approach consists in using a virtual object (VO) to sonify the consequences of the gesture rather than the gesture itself. This approach is motivated by an ecological point of view: Gestures do not make sound, but their consequences do. In this experiment, the aim was to identify efficient sound strategies, to transmit dynamic information of VOs to users through sound. The swipe gesture was chosen for this purpose, as it is commonly used in current and new interfaces. We chose two VO parameters to sonify, the hand-VO distance and the VO velocity. Two kinds of sound parameters can be chosen to sonify the VO behavior: Spectral or temporal parameters. Pitch and brightness were tested as spectral parameters, and amplitude modulation as a temporal parameter. Performances showed a positive effect of sound compared to a no-sound situation, revealing the usefulness of sounds to accomplish the task.

Design and Fabrication of a Miniature Railway Vehicle

We present design, fabrication, and characterization of a small (12 mm × 12 mm × 8 mm) movable railway vehicle for sensor carrying. The miniature railway vehicle (MRV) was mainly composed of a vibrational structure and three legs. A railway was designed and fabricated to power and guide the MRV. It also transmits the sensed data from the MRV to the signal processing unit. The MRV with legs on the railway was moving due to its high-frequency vibration. A model was derived to describe the motion. Besides, FEM simulations were performed to design the legs. Then, the MRV and the railway were fabricated by precision machining. Finally, an infrared sensor was carried and tested. The result shows that the MRV without loading was moving along the railway and its maximum speed was 12.2 mm/s. Moreover, the testing signal was sensed by the MRV.

Skyline Extraction using a Multistage Edge Filtering

Skyline extraction in mountainous images can be used for navigation of vehicles or UAV(unmanned air vehicles), but it is very hard to extract skyline shape because of clutters like clouds, sea lines and field borders in images. We developed the edge-based skyline extraction algorithm using a proposed multistage edge filtering (MEF) technique. In this method, characteristics of clutters in the image are first defined and then the lines classified as clutters are eliminated by stages using the proposed MEF technique. After this processing, we select the last line using skyline measures among the remained lines. This proposed algorithm is robust under severe environments with clutters and has even good performance for infrared sensor images with a low resolution. We tested this proposed algorithm for images obtained in the field by an infrared camera and confirmed that the proposed algorithm produced a better performance and faster processing time than conventional algorithms.

Using Ultrasonic and Infrared Sensors for Distance Measurement

The amplitude response of infrared (IR) sensors depends on the reflectance properties of the target. Therefore, in order to use IR sensor for measuring distances accurately, prior knowledge of the surface must be known. This paper describes the Phong Illumination Model for determining the properties of a surface and subsequently calculating the distance to the surface. The angular position of the IR sensor is computed as normal to the surface for simplifying the calculation. Ultrasonic (US) sensor can provide the initial information on distance to obtain the parameters for this method. In addition, the experimental results obtained by using LabView are discussed. More care should be taken when placing the objects from the sensors during acquiring data since the small change in angle could show very different distance than the actual one. Since stereo camera vision systems do not perform well under some environmental conditions such as plain wall, glass surfaces, or poor lighting conditions, the IR and US sensors can be used additionally to improve the overall vision systems of mobile robots.

Interactive PTZ Camera Control System Using Wii Remote and Infrared Sensor Bar

This paper proposes an alternative control mechanism for an interactive Pan/Tilt/Zoom (PTZ) camera control system. Instead of using a mouse or a joystick, the proposed mechanism utilizes a Nintendo Wii remote and infrared (IR) sensor bar. The Wii remote has buttons that allows the user to control the movement of a PTZ camera through Bluetooth connectivity. In addition, the Wii remote has a built-in motion sensor that allows the user to give control signals to the PTZ camera through pitch and roll movement. A stationary IR sensor bar, placed at some distance away opposite the Wii remote, enables the detection of yaw movement. In addition, the Wii remote-s built-in IR camera has the ability to detect its spatial position, and thus generates a control signal when the user moves the Wii remote. Some experiments are carried out and their performances are compared with an industry-standard PTZ joystick.

Application of Sensory Thermography as Measuring Method to Study Median Nerve Temperatures

This paper presents an experimental case using sensory thermography to describe temperatures behavior on median nerve once an activity of repetitive motion was done. Thermography is a noninvasive technique without biological hazard and not harm at all times and has been applied in many experiments to seek for temperature patterns that help to understand diseases like cancer and cumulative trauma disorders (CTD’s). An infrared sensory thermography technology was developed to execute this study. Three women in good shape were selected for the repetitive motion tests for 4 days, two right-handed women and 1 left handed woman, two sensory thermographers were put on both median nerve wrists to get measures. The evaluation time was of 3 hours 30 minutes in a controlled temperature, 20 minutes of stabilization time at the beginning and end of the operation. Temperatures distributions are statistically evaluated and showed similar temperature patterns behavior.

Sensor Fusion Based Discrete Kalman Filter for Outdoor Robot Navigation

The objective of the presented work is to implement the Kalman Filter into an application that reduces the influence of the environmental changes over the robot expected to navigate over a terrain of varying friction properties. The Discrete Kalman Filter is used to estimate the robot position, project the estimated current state ahead at time through time update and adjust the projected estimated state by an actual measurement at that time via the measurement update using the data coming from the infrared sensors, ultrasonic sensors and the visual sensor respectively. The navigation test has been performed in a real world environment and has been found to be robust.

Real Time Speed Estimation of Vehicles

this paper gives a novel approach towards real-time speed estimation of multiple traffic vehicles using fuzzy logic and image processing techniques with proper arrangement of camera parameters. The described algorithm consists of several important steps. First, the background is estimated by computing median over time window of specific frames. Second, the foreground is extracted using fuzzy similarity approach (FSA) between estimated background pixels and the current frame pixels containing foreground and background. Third, the traffic lanes are divided into two parts for both direction vehicles for parallel processing. Finally, the speeds of vehicles are estimated by Maximum a Posterior Probability (MAP) estimator. True ground speed is determined by utilizing infrared sensors for three different vehicles and the results are compared to the proposed algorithm with an accuracy of ± 0.74 kmph.