Motional Feedback Systems for Ultrasonic Transducers

  • Journal List
  • Sensors (Basel)
  • v.21(7); 2021 Apr
  • PMC8038776

Sensors (Basel). 2021 Apr; 21(7): 2560.

Construction of Ultrasonic Tactile Force Feedback Model in Teleoperation Robot System

Andrey V. Savkin, Academic Editor

Received 2021 Mar 22; Accepted 2021 Apr 5.

Abstract

The ultrasonic phased array as an emerging interactive tool is increasingly used for aerial tactile interaction. However, there is almost no method to achieve remote variable force feedback through the ultrasonic phased array as far as we know. This article presents a force tactile feedback method for teleoperating robot systems that tracks the five fingers and forms a focus on the fingertips. First, the perceived size of the focus depends on the input parameters. The influence of the parameters on the physical output pressure intensity was obtained through physical test experiments. Then, the absolute threshold and difference threshold of human perception were studied through psychophysical experimental methods. Finally, the input parameters were selected according to the experimental results. According to the collected data, the construction of the force regression model was completed, and different parameters were mapped to the perceived intensity. The contact force generated in the actual operation is fed back to the haptic system, and the constructed model automatically adjusts the control parameters to ensure that the user's hand presents a sensory output corresponding to the intensity change. The entire force feedback system is evaluated, and results show that the system shows good perceptual quality.

Keywords: ultrasound array, tactile interaction, force feedback, teleoperation system, regression model

1. Introduction

Robots are increasingly involved in dangerous tasks that are difficult for personnel to perform in complex environments, such as mine clearance, telemedicine, space exploration, industrial operations, etc. [1,2,3,4,5]. Teleoperation is a new type of robot control technology that allows people to avoid on-site dangers and remotely control robots to perform tasks in a safe environment. Immersion and presence are key factors in the robot teleoperation interactive system. The use of tactile force feedback in remote operation can enable the operator to perceive the remote environment in a dynamic interactive manner and improve the operator's task adjustment ability, thereby enhancing the perception modality [6,7,8]. By using a suitable tactile feedback method, the transparency of the remote operating system can also be improved, and the interactive performance of the system can be improved. The current application of 5G communication technology improves the efficiency of information transmission and greatly reduces system delay. It ensures the stability of dynamic force feedback in teleoperation and provides excellent technical support for the widespread application of teleoperation in the future [9].

Vibration motors are often used in contact tactile interactions, which are convenient, economical, and have significant force feedback effects. Manshad et al. used a single vibration motor to study the role of tactile feedback in mobile payment [10]. Scalera et al. proposed the use of two hand-held joysticks with vibration motors to study tactile stimulation. The combination of vibration motors and other devices brings acceptable tactile perception, but vibration stimulation can only stimulate at a fixed point and cannot reflect the direction of force [11]. González et al. chose the Phantom Omni as the tactile feedback device in the remote operation of industrial robots. It is a small mechanical arm equivalent to a rotary joint. This device can be used in bilateral teleoperation [12]. Unlike point stimulation, such interactive devices can transmit force and torque. The operator perceives force information through contact with the equipment. However, its interaction space is limited, the operator and the device need to be in constant contact, and it is not suitable for simulating the multi-point interaction force transmitted by a multi-finger manipulator. Quek et al. designed a skin deformation tactile feedback device that can provide force information with 3 degrees of freedom [13]. In order to achieve tactile feedback during surgery, the device provides users with force and torque information, simulating normal skin deformation and tangential skin stretching tactile sensation, but its tactile device includes a triangular mechanism and a direct current (DC) motor, which is relatively bulky. Yeh et al. added a piezoelectric actuator to the operating side of the surgical robot to achieve force feedback for remote operation [14]. The piezoelectric actuator has a thin appearance and simple structure and is suitable for haptic feedback in combination with a mechanical structure. In the multi-finger teleoperation robot system, tactile gloves are mainly used for tactile feedback. The three-finger remote control system developed by the German Aerospace Center for the lunar rover prototype and the five-finger grasping system in the virtual reality environment both use the force feedback exoskeleton in the CyberGrasp glove to achieve tactile feedback [15,16,17]. In general, different tactile feedback systems have their own advantages, and different methods are suitable for different application fields and different needs. Vibration motors, mechanical devices, piezoelectric actuators, and haptic gloves are all contact-type tactile interactions. People must contact or wear the device during the interaction. Movement and interaction area in this way will be restricted.

Non-contact tactile interaction breaks the limitations of spatial interaction. Common methods include jet driving force feedback, lasers, and ultrasonic sensor arrays [18,19,20]. Air jets rely on air pulses guided by flexible nozzles to produce a tactile sensation. It is effective in simulating coarse force feedback, but space and time properties are limited. The tactile sensation produced by the laser requires a light-absorbing elastic medium attached to the skin. These two methods are not suitable for multi-finger manipulator force tactile feedback in remote operations.

Ultrasonic phased array formation can form a focal point, and the acoustic radiation pressure generated by it can be effectively sensed by the skin. Compared with other tactile interaction methods, ultrasonic tactile feedback does not require wearing or touching the device. Relevant studies conducted by the University of Tokyo and Johns Hopkins University in Japan have shown that ultrasonic phased arrays have reliable aerial tactile interaction effects [21,22,23]. The ultrasonic phased array can generate multiple movable focal points in space to meet the needs of movement perception and can realize independent force feedback of a single finger. Many laboratories have developed experimental prototypes and semi-commercial products for related research, and researchers have also focused on related human perception and application levels. Sand et al. proposed an aerial tactile feedback system for head-mounted displays, which installed array and hand position sensors on the front surface of the head-mounted virtual reality display [24]. Through the participants completing simple virtual key tapping tasks with or without tactile feedback, the experiment proved that participants had a higher evaluation of self-performance through tactile feedback. Some researchers have also applied ultrasonic tactile feedback in the home environment to simulate the touch perception of a light switch. As far as we know, there is no research on the application of ultrasonic tactile feedback in teleoperation systems.

In the study of a multi-point dynamic force-feedback haptic interaction system, Gonzalo et al. used psychophysical methods in the motor haptic feedback system to study the absolute threshold and the difference threshold. This method can accurately configure the haptic device and create appropriate stimuli to improve the human Machine interactive system [25]. For the five-finger independent force feedback, Yunus et al. developed a new wearable tactile feedback system using a vibration motor. The system can establish a one-to-one mapping between the slave and the master to provide the force perception of a single finger, thereby making the disabled People can recognize the force of each finger [26]. Researchers have proposed a variable motion mapping method by adjusting the motion mapping coefficient to change the feedback stiffness to output accurate force feedback [27]. In the field of ultrasonic haptic feedback, most studies only focus on the impact of rendering parameters on the perceived quality, and no systematic method is proposed for the multi-point dynamic haptic feedback interactive system.

Previous studies mainly focus on the ultrasonic phased array in the field of tactile perception, but there is no research on the changing force feedback of the ultrasonic phased array. This paper proposes a multi-finger changing force tactile feedback method using an ultrasonic phased array during teleoperation. The advantage of this method is that it can approach multi-finger following focus through the ultrasonic sensor array and leap motion. The operator can sense the force transmitted by the manipulator when the arm moves and controls the manipulator. Also, this study develops a force regression model based on control parameters to obtain the feedback and perception of changing force in the interaction process. In the ultrasonic system, the control parameters affecting the output sound pressure intensity are frequency, intensity and focus position. We perform the factor analysis on the parameters through the experiments and develop a force regression model based on the data collected so that the system can adjust the parameters independently to map the changes in the feedback force's value.

2. Description of Teleoperation Robot System

Figure 1 is a conceptual diagram of the system. When the operator remotely controls the robot to perform operations, the manipulator will perform actions such as grasping the target object and obtain contact force information by in-stalling a pressure sensor on the fingertip of the robot. Dynamic interaction force information can be transmitted back to the human–computer interaction terminal through wireless network transmission. The ultrasonic tactile feedback system we propose as suitable for remote operation is the interactive terminal platform, and the transmitted interactive force information is presented through the ultrasonic array. This allows tracking, recognition and positioning of human fingers by using the leap motion. The ultrasonic phased array emits ultrasonic waves and focuses on the positioning point to realize the tactile perception of human fingertips.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g001.jpg

Conceptual diagram of teleoperation robot system.

2.1. Method of Forming Tactile Focus Points

The ultrasonic phased array can form a focal point in the air through-beam focusing, and the acoustic radiation pressure at the focal point can be sensed by human skin. We use the aerial tactile display obtained from Ultrahaptics for secondary development and related research. We need to achieve real-time, dynamic tactile focus points on five fingers. Therefore, we need to use the phased array to generate five focal points and leap motion for hand tracking. The combination of both means that the focal points generated by the phased array are focused on the thumb, index finger, middle finger, ring finger and little finger. Project programming is mainly designed from two aspects: phased array multi-focus generation and leap motion positioning.

The leap motion acquisition process is mainly divided into the following processes: acquisition of a single frame image, image recognition and analysis, hand model reconstruction, and hand information extraction. For hand information, leap motion can identify information including the position, direction, angle, normal vector, and movement speed of the palm, thumb, index finger, middle finger, ring finger, and little finger. Therefore, it is programmed to realize single-frame image collection finger positioning. For the output of finger information, leap motion expresses the finger model in the form of an array. It recognizes the thumb, index finger, middle finger, ring finger and little finger as the five elements in the array. You can select the element position in the array to determine the information of the finger that you need. It is necessary to determine the position of the fingertip and the direction of the finger to realize the display of the focus on the finger and convert the coordinate system to the coordinate system of the phased array to facilitate subsequent operations.

Ultrahaptics' SDK (Software Development Kit) integrates functions such as focus position, controller launch, modulation, etc., to facilitate design and programming. When transmitting ultrasonic waves into space, amplitude modulation API (Application Programming Interface) that uses sinusoidal signals is used to modulate the ultrasonic waves. When using the AM (amplitude modulation) API, you can continuously send control information, including the modulation frequency, focus position, and focus the intensity of the ultrasound. For the generation of focus, the process mainly includes: creating Ultrahaptics transmitter, defining modulation frequency and intensity, creating control point position, and continuously updating control point information. The process sequence is shown in the Figure 2.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g002.jpg

The flow chart of tactile points focusing.

2.2. Integration of the System

The proposed system uses an ultrasonic phased array for dynamic tactile feedback in a robot teleoperation system. The control terminal completes the contact force detection, and the interactive end presents a dynamically changing contact force through an ultrasonic phased array. The PC side completes data processing, transmission and equipment control. The specific technical roadmap is shown in Figure 3. leap motion is needed to determine the focus position, which captures the position of the finger and transmits it to the ultrasonic array control program. The focused pressure is related to the real-time contact force of the control terminal. The pressure information detected by the pressure sensor installed by the manipulator can be transmitted back to the PC by means of wireless network transmission such as an ad hoc network, and the transmitted pressure information needs to be transmitted to the related program that controls the ultrasonic phased array after processing. We realize the transmission of pressure information through network programming. Under normal circumstances, ultrasonic phased array is only used for force sensing. Using the ultrasonic phased array to achieve force feedback requires a new control method. We built a regression model using three parameters: height, distance from centre, and input intensity. The model inputs the received contact force information and can directly output appropriate control parameters to control the ultrasonic array to present a matching tactile focus. The specific construction method of the model will be introduced in detail in the third section.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g003.jpg

The overall flowchart of our proposed system. It includes the human-computer interaction terminal and the control terminal, as well as the method of data transmission and the method of realizing device control.

3. Experiment

The ultrasonic phased array can realize tactile perception, but a single control cannot make the ultrasonic phased array output regular or specific tactile power points. In actual use, it can only be used for perception, not force feedback. Therefore, we built a sound field force feedback regression model to realize that the ultrasonic phased array displays a specific magnitude of force feedback at the focal point in the air. The physical force of the focus depends on the input parameters. According to the physical characteristics of the ultrasonic phased array and the related theory of human skin perception of touch, the parameters that affect the perception effect are command intensity, modulation frequency, distance from the centre, and the height of the focus in the working space. Because there are many parameters that affect feedback perception, it is a relatively complicated situation to obtain stable force feedback under different circumstances. Directly constructing the regression model will make the problem complicated and time-consuming. Therefore, we need to complete the factor analysis through experimental methods. Experiment 1 was a physical test experiment. The results of the experiment showed the influence of the aforementioned parameters on physical output pressure. Experiment 2 was a psychophysical experiment to understand the influence of parameters on human perception. Therefore, we obtained the relationship between output pressure and human skin perception. The constructed model mapped different parameters to the perceived amplitude.

3.1. Measurement Hardware

In order to build a force regression model, we built an experimental platform for testing the sound field to test the characteristics of the sound field under different conditions. An acoustic sensor was used for measurement. Due to the inconvenience of movement when using a balance to measure the force value, only one focus point could be measured at the same time. When the ultrasonic frequency is 40 kHz, the energy loss is only 4%, but if the frequency becomes four times larger, it will lose 50%. At the same time, 40 kHz frequency has a better rendering effect. Therefore, the frequency of the system's ultrasonic transducer was 40 kHz. The selected acoustic sensor was GRAS 46 BE, which could measure high sound pressure levels at high frequencies. Assembling the XYZ workbench works for precise and systematic spatial measurement with acoustic sensors. The sensors could move freely in three-axis directions. The constructed measurement system is shown in Figure 4. Using the measurement system, we changed the commanding strength, height, and distance from the centre to three parameters, collected five samples for each data point for pressure measurement and moved the sensor to the focus for 5 s to measure each data point. The sample average of each data point was calculated as the final result.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g004.jpg

Measurement setup: XYZ workbench, data collector, acoustic sensor, ultrasonic mid-air haptic display.

3.2. Experiment 1

Experiment 1 was a physical test experiment. First we discussed the effect of intensity on rendering pressure. In order to observe the changing trend, we fixed the position of the focus. The height was set at 200 mm, the distance from the centre was 0 mm, and the modulation frequency was 240 Hz. The input intensity was 0 to 1, the step length between two consecutive values was 0.1, and the test records ten values. The recorded data are shown in Figure 5. It can be seen that there is a non-linear relationship between intensity and output pressure, but the overall trend shows that the greater the amplitude is, the greater the pressure becomes.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g005.jpg

The relationship between input command intensity and output intensity.

In order to judge the influence of the distance to the centre on the intensity of the focus, all other parameters remain unchanged, the frequency was 240 HZ, the height was 200 mm, and the input intensity was 1. The distance from the centre to the target was 9 mm and the maximum distance was 81 mm. The distance was the Euclidean distance from the focal point to the centre. The result is shown in Figure 6. It can be seen that as the focal point moves away from the centre point, the pressure intensity decreases monotonously and presents a non-linear relationship.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g006.jpg

The relationship between the distance from centre point and the output intensity.

When studying the influence of the height parameter, a similar experiment was conducted, keeping the frequency at 240 HZ, keeping the distance from the centre at 0 mm, and setting the input intensity to 1. Experiments found that there were many interference points between ultrasonic waves in areas with a height of less than 9 cm, and the perception of the focus with the hand will be greatly interfered with. The spatial perception of heights above 25 cm will be weakened, which is not conducive to tactile perception. The working area should be set between 9–25 cm. It is observed from experimental Figure 7 that the relationship between the two is non-linear, and the output pressure intensity will decrease as the height increases.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g007.jpg

The relationship between height and output intensity.

3.3. Experiment 2

Experiment 2 was a psychophysical experiment. To understand the tactile perception under ultrasound stimulation, we observed the relationship between physical stimuli and user responses. Our experiment mainly determines two values, namely the absolute threshold and the differential threshold. The absolute threshold is the minimum intensity at which the stimulus can be detected, and the differential threshold is the smallest change where the subject can perceive the increase or decrease of the intensity of the stimulus. The experimental methods used to estimate the sensory threshold can usually use psychophysical experimental methods. We chose the adaptive ladder method for evaluation to provide accurate results in less time.

3.3.1. Participants

Ten participants (3 females and 7 males, aged between 23–30) were recruited into the experiment. They were all right-handed and all subjects had no disability and no physical injury affecting palm sensitivity.

3.3.2. Procedure

During the experiment, a piece of foam plastic was placed above the array in a fixed relative position. During the experiment, participants put their palms on the foam with their palms facing down to ensure that the palms were in a fixed position relative to the transducer array and that the palms were as horizontal as possible. The position of the ultrasonic focus was set 20 cm above the upper surface of the array. During the experiment, the ultrasonic transducer array will emit audible sounds. When the amplitude is different, the volume of the sound is also different. We asked participants to wear headphones when listening to the pre-recorded noise of the transducer array and natural rain so that they could not hear the changing sound of the transducer array.

We conducted experiments by adaptive ladder program with variable step length "1up-2down". The experiment was divided into stair case A and stair case B. At the beginning of the experiment, in experiment A, the subjects were provided with higher intensity stimulation, and in experiment B, the subjects were provided with lower stimulation. The duration of the stimulus was about 3 s. The participant was asked if they feel the stimulus. If the participant answered yes twice, the stimulus level would be reduced by one level. If the participant answers no, the stimulus level would be increased in the next test. After each reversal, the stimulus step length was reduced to ensure a more accurate adjustment. The experiment was stopped after eight reversals, and the average value of the stimulus level at the point of the last four reversals was the absolute threshold. The step size is 0.4, 0.3, 0.2, 0.1, 0.05, 0.03 and 0.01.

In the differential threshold experiment, participants were required to compare the two stimulus intensities and answered whether they feel the difference between them. Each test included the first stimulus reference (SR) and comparative stimulus (CS). Each test stayed for 3 s. In the first experiment, the two stimuli were chosen to have larger differences. Participants were asked whether they felt the difference, and the participants answered yes that the stimulation level would decrease in the next stage, and vice versa, the stimulation level would increase. Every time the response result changed, we made the stimulus step smaller to make the adjustment more convenient and accurate. The steps used are 0.5, 0.4, 0.3, 0.2, 0.1, 0.05, 0.03, and 0.02. Similarly, after eight reversals, the differential threshold is calculated based on the average of the last four reversal points.

3.3.3. Results

Figure 8 is an example result of the experiment. The experimental results show that when the input intensity value is 0.39, it is the smallest perceived intensity.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g008.jpg

The result of the absolute threshold measurement. An example: absolute threshold measurement result of two staircase programs.

Figure 9 is an example of a differential threshold experiment. When the input intensity value interval is 0.1, there is the smallest perceptual change.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g009.jpg

Psychophysical adaptive staircase for subject 2 with a reference flow rate of 0.5.

3.4. Regression Model

In the previous test experiments and perception experiments, we learned that the input command strength, height, distance from the centre, and modulation frequency will all affect the subsequent force feedback perception. Our goal is that the system can automatically output the output force corresponding to the change of the input force so that the operator can obtain a stable perception. According to the threshold detection experiment, we can know that under the continuous change of the control intensity, the change of the tactile feedback force generated by the ultrasonic phased array can be perceived by the human finger. The parameter values in all cases are considered, and a regression model is constructed for the mapping between parameters and output. This is the most direct realization method. But this method has some problems. First, we perform a factor analysis. The first is the frequency. Previous studies have shown that when different modulation frequencies are used, the perception of human skin is different and exhibits non-linear changes. As the frequency changes, the measured physical force will not change accordingly. We cannot directly map frequency to output force. The changes in the three parameters of command strength, height, and distance from the centre will not only lead to changes in perception but also changes in physical force, and the direction of change is the same. The different perceptual results of changing the frequency and the perceptual results of changing the commanded intensity will largely overlap. Through analysis, we do not need to set the frequency as the parameter of the regression model. In order to present the maximum tactile perception, we set the modulation frequency of the system to the human body's best perception frequency, 240 Hz. Therefore, our input parameters are set as the input command strength, height, and distance from the centre.

According to our experience when the ultrasonic phased array is arranged in a square, our actual test data do not need to cover the entire space, and we select typical representative points in the horizontal and vertical directions for testing. The tested data points are shown in Figure 10. The three coordinate axes are the input command intensity, height, and distance from the centre. The pressure at each data point is indicated by colour. The test data set is used for multiple regression. Formula 1 represents the regression model, where P is the output sound pressure value, a is the input command intensity, z is the height, and d is the distance from the centre.

P = 98.136 + 840.889 ∗ a − 0.548 ∗ z + 13.588 ∗ d + 266.198 ∗ a2 − 0.003 ∗ z2 − 0.290 ∗ d2

(1)

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g010.jpg

Three-dimensional display of test data points. The x-axis is the intensity of the input command, the y-axis is the height, the z-axis is the distance from the centre point, the test points are characteristic data points of different parameter combinations, and the measured pressure is indicated by color.

The entire sound field force feedback system is shown in Figure 11. The robot end inputs the force information and position information of the interaction end into the force feedback system and then constructs a good regression model to output the predicted value to the ultrasonic tactile feedback system to control the ultrasonic tactile feedback system. The ultrasonic tactile feedback system provides a specific magnitude of force feedback corresponding to the input force for the operator's perception.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g011.jpg

Frame diagram of real-time force feedback system.

4. Results and Discussion

4.1. Measurement

A physical test experiment was designed to check the accuracy of the output pressure after using the regression model. As shown in Figure 12, the interactive part was an ultrasonic phased array combined with leap motion. The sound pressure at the ultrasound focus point was tested using an acoustic sensor GRAS 46 BE (GRAS Sound & Vibration, Holte, Denmark). The 8-channel multifunctional data acquisition module SIRIUS (DEWESoft, Kumberg, Austria) and DEWESoft X (DEWESoft, Kumberg, Austria) were selected for data collection and analysis. A lightweight robotic arm (RobotAnno, Guangzhou, China) was used as a remote control terminal. A pneumatic bionic hand was installed at the end of the robotic arm, and the bionic hand controls the fingers to perform gripping actions through the pipes connected to an air pump. The pressure data of the contact between the finger and the object were measured by the pressure sensor BSLM-2 (BUFSON, Shijiazhuang, China) on the fingers of the bionic hand.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g012.jpg

Test and verification experiment diagram in teleoperation system.

In the experiment, the collected pressure data was directly transmitted back to the PC through wireless ad-hoc network communication. After processing the pressure data into a numerical value corresponding to the sound pressure, this numerical value was used as the predicted value of the experiment. The predicted value was input to the regression model, and then the control parameters were output so that the ultrasonic phased array generated the corresponding focus point at the position where the leap motion detects the fingers. The intensity value of the gathering point was used as the test value in the experiment. Two sets of test experiments were set up. The first group was given a unique predicted value, and the test value was compared with the predicted value at different points on the horizontal plane. The second group set different predicted values at points of different heights and compared the test value with the predicted value. In addition, repeated-measures analyses of variance (ANOVA) were performed to compare the data. All data were assessed for the approximation to a normal distribution and sphericity, and when necessary, degrees of freedom were adjusted using the Greenhouse-Geisser adjustment. All data were performed using SPSS statistical software (V22, Chicago, IL, USA). When the test result showed significance of difference, it was labeled as "p < 0.05" or "p < 0.01", otherwise, "p > 0.05" was labeled.

The height of the selected points in the first group was 250 mm, and the distance from the centre was different. The required output power was set to 500 pa, which was the predicted value. The pressure value generated by the ultrasonic array using the regression model was the test value. Five data points were tested, the reading started from the centre, with a step length of 15 mm, and the experimental results are shown in Figure 13. There is a small error between the predicted value and the test value. The errors of the five points are 0.3%, 2.3%, 3.6%, 5.03%, and 3.49% respectively. The average error is 2.94%. Result of the repeated measures ANOVA shows that there were significant differences between the predicted value and the test value [F (1, 4) = 13.88, p < 0.05].

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g013.jpg

The predicted and measured values at different positions from the centre point on the horizontal plane. The value indicated by the arrow is the relative error.

The second group selected the point at the centre. The height started from 110 mm, the step length was 30 mm, five data points were tested, and the predicted value randomly inputted from small to large. The experimental result is shown in Figure 14. The test value is very close to the predicted value, and there is only a slight error. Result of the repeated measures ANOVA shows that there is no significant difference between the predicted value and the test value [F (1, 4) = 2.10, p > 0.05]. Experiments show that the output of the ultrasonic array is controlled well by the regression model.

An external file that holds a picture, illustration, etc.  Object name is sensors-21-02560-g014.jpg

Test values and predicted values at different heights on the vertical line where the centre point of the array is located.

4.2. Discussion

As mentioned earlier, the tactile points focused by the ultrasonic array were generally used for tactile perception in different situations. The output pressure of the equipment was not usually concerned about this, and changed in an irregular form in the process of movement perception. However, in some cases, we need to perceive changes in contact force through touch. For example, in the robot teleoperation system, the contact points generated by the ultrasonic array were used to express the change of the contact force during the robot's operation, and the required tactile interaction could be obtained more realistically and naturally. Therefore, the force regression model could change the control parameters in real time and output the tactile focus with a specific pressure level. It can be seen from Figure 14 that the ultrasonic array can render more accurate tactile points of a specific size. Although there was a difference between the test value and the predicted value at different points on the horizontal plane, according to human perception characteristics, this degree of error will not cause a significant difference in perception [28,29].

In this study, the method set the best perception frequency to avoid complicated combinations. Therefore, the input command intensity needed to be adjusted in different positions. At the farthest edge point of the array, the maximum output force was when the input command intensity value was the maximum. Due to the characteristics of the ultrasonic array to achieve tactile perception, it may happen that when the contact force was large, the focused ultrasonic pressure cannot be equal to the contact force. The overall output pressure could be scaled down according to needs and effects. In this case, the size of the array became the limiting factor for the size of the focus output pressure, and the expansion of the array could be considered to meet the system requirements.

5. Conclusions

This paper presents a method of using the ultrasonic phased array for non-contact tactile interaction in the teleoperation system. The method enables the ultrasonic phased array to input changing force tactile feedback while following the movement of the finger to focus. The main conclusions were summarized as below:

  1. Leap motion was used to complete the position of the finger during the positioning operation. The ultrasonic phased array focused on adjustable tactile points, which were used to feed back the contact force between the manipulator and the object.

  2. We conducted the physical test experiments and psychological physical experiments to investigate the influence of different parameters on the output intensity of the constructed focus and then completed the factor analysis.

  3. The modulation frequency was set as the human body's best perception frequency, and the three parameters, including height, distance from the centre, and input command strength, were set as system control parameters. The regression model was constructed according to the collected characteristic data points, and the input parameters were mapped to the output intensity.

  4. A physical test experiment was designed to check the accuracy of the output pressure after using the regression model. The results from repeated measures ANOVA showed that there is a significant difference between the test value and the predicted value at different points on the horizontal plane, but the average error is only 2.94%, which will not cause a significant difference in perception. There is no significant difference in test results at different heights.

Author Contributions

X.L., M.L., and J.L. proposed and designed the algorithm and analyzed the data; Z.Z. and H.Z. performed the experiment; J.L. and Y.L. were involved in data collection and drafted the manuscript; Y.L., Z.Z. and H.Z. helped review and edit the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Footnotes

Publisher's Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

1. Shao H., Nonami K., Wojtara T., Yuasa R., Amano S., Waterman D. Neuro-fuzzy position control of demining tele-operation system based on rnn modeling. Robot. Comput. Integr. Manuf. 2006;22:25–32. doi: 10.1016/j.rcim.2005.01.003. [CrossRef] [Google Scholar]

2. Chen Y., Zhang B., Zhou J., Wang K. Real-time 3d unstructured environment reconstruction utilizing vr and kinect-based immersive teleoperation for agricultural field robots. Comput. Electron. Agric. 2020;175:105579. doi: 10.1016/j.compag.2020.105579. [CrossRef] [Google Scholar]

3. Hernansanz A., Casals A., Amat J. A multi-robot cooperation strategy for dexterous task oriented teleoperation. Robot. Auton. Syst. 2015;68:156–172. doi: 10.1016/j.robot.2014.12.007. [CrossRef] [Google Scholar]

4. Fong T., Nourbakhsh I. Interaction challenges in human-robot space exploration. Interactions. 2005;12:42–45. doi: 10.1145/1052438.1052462. [CrossRef] [Google Scholar]

5. Vemula B., Matthias B., Ahmad A. A design metric for safety assessment of industrial robot design suitable for power-and force-limited collaborative operation. Int. J. Intell. Robot. Appl. 2018;2:226–234. doi: 10.1007/s41315-018-0055-9. [PMC free article] [PubMed] [CrossRef] [Google Scholar]

6. Howe R.D., Kontarinis D.A. Task Performance with a Dexterous Teleoperated Hand System. SPIE International Society for Optics and Photonics; Washington, DC, USA: 1993. pp. 199–207. Telemanipulator Technology. [CrossRef] [Google Scholar]

7. van der Putten E., van den Dobbelsteen J.J., Goossens R., Jakimowicz J.J., Dankelman J. The effect of augmented feedback on grasp force in laparoscopic grasp control. IEEE Trans. Haptics. 2010;3:280–291. doi: 10.1109/TOH.2010.23. [PubMed] [CrossRef] [Google Scholar]

8. Estrada E., Yu W., Li X. Stable bilateral teleoperation with phase transition and haptic feedback. J. Frankl. Inst. 2021;358:1940–1956. doi: 10.1016/j.jfranklin.2020.12.027. [CrossRef] [Google Scholar]

9. Jacob E., Astorga J., Jose Unzilla J., Huarte M., Garcia D., Lopez de Lacalle L.N. Towards a 5g compliant and flexible connected manufacturing facility. Dyna. 2018;93:656–662. doi: 10.6036/8831. [CrossRef] [Google Scholar]

10. Manshad M.S., Brannon D. Haptic-payment: Exploring vibration feedback as a means of reducing overspending in mobile payment. J. Bus. Res. 2021;122:88–96. doi: 10.1016/j.jbusres.2020.08.049. [PMC free article] [PubMed] [CrossRef] [Google Scholar]

11. Scalera L., Seriani S., Gallina P., Di Luca M., Gasparetto A. An experimental setup to test dual-joystick directional responses to vibrotactile stimuli. IEEE Trans. Haptics. 2018;11:378–387. doi: 10.1109/TOH.2018.2804391. [PubMed] [CrossRef] [Google Scholar]

12. González C., Solanes J.E., Muñoz A., Gracia L., Girbés-Juan V., Tornero J. Advanced teleoperation and control system for industrial robots based on augmented virtuality and haptic feedback. J. Manuf. Syst. 2021;59:283–298. doi: 10.1016/j.jmsy.2021.02.013. [CrossRef] [Google Scholar]

13. Quek Z.F., Provancher W.R., Okamura A.M. Evaluation of skin deformation tactile feedback for teleoperated surgical tasks. IEEE Trans. Haptics. 2018;12:102–113. doi: 10.1109/TOH.2018.2873398. [PubMed] [CrossRef] [Google Scholar]

14. Yeh C.-H., Su F.-C., Shan Y.-S., Dosaev M., Selyutskiy Y., Goryacheva I., Ju M.-S. Application of piezoelectric actuator to simplified haptic feedback system. Sens. Actuators A Phys. 2020;303:111820. doi: 10.1016/j.sna.2019.111820. [CrossRef] [Google Scholar]

15. Lii N.Y., Chen Z., Pleintinger B., Borst C.H., Hirzinger G., Schiele A. Toward understanding the effects of visual-and force-feedback on robotic hand grasping performance for space teleoperation; Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems; Taipei, Taiwan. 18–22 October 2010; New York, NY, USA: IEEE; 2010. pp. 3745–3752. [Google Scholar]

16. Peer A., Einenkel S., Buss M. Multi-fingered telemanipulation-mapping of a human hand to a three finger gripper; Proceedings of the RO-MAN 2008-The 17th IEEE International Symposium on Robot and Human Interactive Communication; Munich, Germany. 1–3 August 2008; New York, NY, USA: IEEE; 2008. pp. 465–470. [Google Scholar]

17. Magdalon E.C., Michaelsen S.M., Quevedo A.A., Levin M.F. Comparison of grasping movements made by healthy subjects in a 3-dimensional immersive virtual versus physical environment. Acta Psychol. 2011;138:126–134. doi: 10.1016/j.actpsy.2011.05.015. [PubMed] [CrossRef] [Google Scholar]

18. Tsalamlal M.Y., Ouarti N., Ammi M. Psychophysical study of air jet based tactile stimulation; Proceedings of the 2013 World Haptics Conference (WHC); Daejeon, Korea. 14–17 April 2013; New York, NY, USA: IEEE; 2013. pp. 639–644. [Google Scholar]

19. Kim H.-S., Kim J.-S., Jung G.-I., Jun J.-H., Park J.-R., Kim S.-P., Choi S., Park S.-J., Choi M.-H., Chung S.-C. Evaluation of the possibility and response characteristics of laser-induced tactile sensation. Neurosci. Lett. 2015;602:68–72. doi: 10.1016/j.neulet.2015.06.053. [PubMed] [CrossRef] [Google Scholar]

20. Large D.R., Harrington K., Burnett G., Georgiou O. Feel the noise: Mid-air ultrasound haptics as a novel human-vehicle interaction paradigm. Appl. Ergon. 2019;81:102909. doi: 10.1016/j.apergo.2019.102909. [PubMed] [CrossRef] [Google Scholar]

21. Inoue S., Makino Y., Shinoda H. Active touch perception produced by airborne ultrasonic haptic hologram; Proceedings of the 2015 IEEE World Haptics Conference (WHC); Evanston, IL, USA. 22–26 June 2015; New York, NY, USA: IEEE; 2015. pp. 362–367. [Google Scholar]

22. Long B., Seah S.A., Carter T., Subramanian S. Rendering volumetric haptic shapes in mid-air using ultrasound. ACM Trans. Graph. 2014;33:1–10. doi: 10.1145/2661229.2661257. [CrossRef] [Google Scholar]

23. Takahashi R., Hasegawa K., Shinoda H. Tactile stimulation by repetitive lateral movement of midair ultrasound focus. IEEE Trans. Haptics. 2019;13:334–342. doi: 10.1109/TOH.2019.2946136. [PubMed] [CrossRef] [Google Scholar]

24. Sand A., Rakkolainen I., Isokoski P., Kangas J., Raisamo R., Palovuori K. Head-mounted display with mid-air tactile feedback; Proceedings of the 21st ACM Symposium on Virtual Reality Software and Technology; Beijing, China. 13–15 November 2015; pp. 51–58. [Google Scholar]

25. Garcia-Valle G., Arranz-Paraiso S., Serrano-Pedraza I., Ferre M. Estimation of torso vibrotactile thresholds using eccentric rotating mass motors. IEEE Trans. Haptics. 2021 doi: 10.1109/TOH.2020.3048290. [PubMed] [CrossRef] [Google Scholar]

26. Yunus R., Ali S., Ayaz Y., Khan M., Kanwal S., Akhlaque U., Nawaz R. Development and testing of a wearable vibrotactile haptic feedback system for proprioceptive rehabilitation. IEEE Access. 2020;8:35172–35184. doi: 10.1109/ACCESS.2020.2975149. [CrossRef] [Google Scholar]

27. Liu L., Zhang Y., Liu G., Xu W. Variable motion mapping to enhance stiffness discrimination and identification in robot hand teleoperation. Robot. Comput. Integr. Manuf. 2018;51:202–208. doi: 10.1016/j.rcim.2017.12.008. [CrossRef] [Google Scholar]

28. Pang X.-D., Tan H.Z., Durlach N.I. Manual discrimination of force using active finger motion. Percept. Psychophys. 1991;49:531–540. doi: 10.3758/BF03212187. [PubMed] [CrossRef] [Google Scholar]

29. Jones L.A. Matching forces: Constant errors and differential thresholds. Perception. 1989;18:681–687. doi: 10.1068/p180681. [PubMed] [CrossRef] [Google Scholar]


Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)


gregorpaped1989.blogspot.com

Source: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8038776/

0 Response to "Motional Feedback Systems for Ultrasonic Transducers"

Post a Comment

Iklan Atas Artikel

Iklan Tengah Artikel 1

Iklan Tengah Artikel 2

Iklan Bawah Artikel