-
The discussion section compares the three-shape measurements based on their performance on a NIST gauge and applications on real TM samples, followed by discussion of the practicability of the three methods in TM measurements.
Table 1 compares the three shape measurement methods based on the results reported in this paper. The optical setup for the NIST gauge measurements was configured to mimic that used in the TM measurements, which met constraints like magnification requirements, the limited volume for triangulation, and available spatial resolution for the high-speed camera. The computed resolution is better (of lower value) in the NIST gauge measurements compared to the TM measurements because the NIST gauge is more reflective and does not need the higher illumination power required by the TM. Within the NIST gauge measurements, the resolution of the multi-wavelength method is better than the multi-angle, which is better than HSFP.
Fringe projection Multi-angle Multi-wavelength Illumination density High Moderate Low to moderate Acquisition time for the entire measurement (multiple images were taken) 18 ms 26 ms 150 ms Processing time Fast Moderate Moderate Temporal phase unwrapping Not compatible Compatible Compatible Triangulation volume Needed Needed Not needed Resolution TM NIST TM NIST TM NIST 0.11 mm 0.11 mm 0.15 mm 0.04 mm 0.13 mm 0.01 mm Table 1. Comparison of the three high-speed shape measurement methods based on TM and NIST measurement
Performing high-speed imaging on semi-transparent samples such as TM requires a large illumination density, which significantly affects the optical phase quality of the results. The HSFP and multi-wavelength methods use the same 210 mW power laser light source, whereas the tunable laser output is 10 mW. Furthermore, the HSFP focuses all the available optical power on a single fringe, and can achieve the highest illumination power density of the three methods. In contrast, the multi-wavelength method has the lowest optical phase quality due to the tunable laser's lower illumination power. The image acquisition times for the fringe projection and multi-angle method are limited by the scanning speed of the MEMS mirror, and they have a similar measurement time of about 20−30 ms. The measurement time for the multi-wavelength method is limited by the wavelength tuning time of 150 ms. The processing time of the fringe projection method is much less than the multi-angle and multi-wavelength methods because the latter two need more computational power to track the four-step phase-shifted frames at each illumination position or wavelength.
The resolution of the shape measurement method is related to the optical phase quality as well as the sensitivity of the method's configuration. A comparison of factors that affect the sensitivity of the shape measurements is presented in Table 2. The sensitivity of the fringe projection method depends upon the projected fringe density and the triangulation of the illumination and observation direction. The sensitivity of the multi-angle method depends upon the illumination power as well as the illumination and observation direction and the angle difference of the illumination tuning. The multi-wavelength method's sensitivity depends upon the illumination and observation direction and the wavelength change. The fringe projection method has the highest resolution in terms of standard deviation among the TM measurement results and the lowest resolution among the NIST measurements. This may be because the fringe projection method can use the higher available illumination power when measuring the semi-transparent TM, whereas the additional power is not needed when measuring the NIST.
Method Factors that affect the sensitivity Fringe projection method Illumination and observation direction Fringe density Multi-angle method Illumination and observation direction Angle difference Multi-wavelength method Illumination and observation direction Wavelength difference Table 2. Comparison of factors that affect the sensitivity of the three high-speed shape measurement methods
The HSFP method used in this paper is not compatible with temporal phase unwrapping, which requires a series of fringe images, where the fringe width gradually decreases to increase the fringe density. In HSFP the diameter of the laser beam determines the width of the fringe, and it cannot be changed automatically during the measurement in the current configuration. The multi-angle and multi-wavelength methods are suitable for temporal phase unwrapping since the spatial frequency of the optical phase of both methods changes gradually as the tuning of the illumination or the wavelength changes. By introducing temporal phase unwrapping, the resolution of the measurement results can be further improved36.
The HSFP and multi-angle method require triangulation volume. Fringes projected perpendicular to the sample's surface are not sensitive to the shape of the sample along the projection direction. Thus, for both methods, there must be a triangulation volume to allow the scanning of the illumination. The human eardrum is located at the end of an approximately 2.5 centimeter long and 0.7 centimeter diameter ear canal, and the limited space favors the multi-wavelength method. However, the use of a higher power multi-wavelength laser would be advisable. For measurement of the cadaveric human eardrum, the ear canal can be removed to expose widely the TM, so we propose the multi-angle and fringe projection methods using the high-power fixed wavelength laser to measure the deformation and shape of the cadaveric human eardrum.
The HSFP applies a phase-stepping algorithm that utilizes four phase-stepped images. However, the maximum MEMS mirror's scanning speed is 3 kHz which is not fast enough to scan more than three fringes onto the sample during image acquisition by the high-speed camera since the longest exposure time of the camera is 0.001 s (1 kHz). Therefore, the system cannot capture the completed multi-fringe image in one frame. To overcome this limitation, the completed multi-fringe images are reconstructed after the image acquisition by summing up a series of single fringe images. The number of single fringe images needed is related to the width of the projected fringe and the desired fringe density. In this paper, a total of 72 images are captured to measure the shape of the human eardrum. The processing time of these images for a modern computer is less than 1 second allowing rapid examination of measurement quality.
The multi-angle method measures the shape of a sample by scanning the illumination across the sample. A PZT actuator is used as the phase shifter for the four-step phase-shifting algorithm. Locating the correct frames for the four-step phase-shifting method at each scanning position is done by scanning the Pearson correlation computed for each image to determine the shifted phase. This process for a total of 1200 images is more computationally expensive than the process of HSFP. Furthermore, with the multi-angle method, the illumination beam needs to be collimated and expanded to cover an area larger than the sample surface to maintain the illumination during the scanning and lowering the illumination density for high-speed imaging.
The multi-wavelength method utilizes a tunable laser for shape measurement. The wavelength is tuned during the high-speed image acquisition with a PZT actuator for phase shifting. A typical multi-wavelength shape measurement captures 50 high-speed images at 12 wavelengths. Within these 50 frames, four frames with 90-degree phase-shifted are selected via the calculation of the person correlation37−40 to calculate the optical phase at a given wavelength. Recording continuously while the phase shifter is moving linearly minimizes unwanted ringing within the phase shifter improving the accuracy of the phase shifting. It is possible to implement the method with 100 high-speed images at 12 wavelengths to improve the phase-shifting accuracy. The processing time of the multi-wavelength is similar to that of the multi-angle shape measurement. The output power and the tuning speed of the tunable laser limit the illumination power density and the measurement time. In our setup, the tunable laser only provides a minimal illumination power for the shape measurement, with a tuning time of 150 ms. A moving average filter has to be used to extract the optical phase due to the minimal illumination power reducing the spatial resolution of the shape measurement.
In summary, the multi-wavelength method does not require triangulation volume to measure the shape, but the optical power of these lasers are limited and the method generally requires painting of the TM. The Multi-angle method enables the shape measurement on the unpainted TM with more optical power, and the HSFP allows rapid shape measurement of the unpainted TM with the least post-processing time. However, both the Multi-angle and HSFP methods have constraints with the small triangulation volume. Therefore, the ideal approach to measure the TM through the intact ear canal in live ears will be to apply the Multi-wavelength method with a tunable laser with the desired optical power and tuning speed for high-speed imaging.
-
Digital holography (DH) can record and reconstruct an optical wavefield (amplitude and phase) by comparing the optical path difference of the reference and object beams captured by an optical sensor (camera). For a general holographic interference setup, the measured optical phase ϕ is given by the fringe locus function36:
$$ \phi ={\boldsymbol{d}}\cdot {\boldsymbol{e}} $$ (1) where d is the optical path difference vector and e is the sensitivity vector defined as38
$$ {\boldsymbol{e}}=\frac{2\pi }{\lambda }\left[{\boldsymbol{b}}-{\boldsymbol{s}}\right] $$ (2) where λ is the wavelength of the laser, b is the observation unit vector, and s is the illumination unit vector. Generally, the object beam is reflected by a diffusive scattering object, and the measured optical phase contains high spatial varying speckle noise. Therefore, the optical path difference cannot be retrieved from a single optical wavefield. However, the change in the measured optical phase
$ \left(\Delta \phi \right) $ from the two camera exposures is expressed as$$ \Delta\phi ={\boldsymbol{d}}\cdot {{\boldsymbol{e}}}_{2}-{\boldsymbol{d}}\cdot {{\boldsymbol{e}}}_{1} $$ (3) where
$ {{\boldsymbol{e}}}_{1} $ and$ {{\boldsymbol{e}}}_{2} $ are the sensitivity vectors at two camera exposures. The optical phase change$ \left(\Delta \phi \right) $ can be retrieved if the speckle distributions at the two exposures correlate with each other. Eq. 3 is the foundation of digital holographic shape and deformation measurement36,37.For shape measurement, assuming there is no object deformation during the measurement, the height information (shape) of the measured subject can be obtained by modulating either the magnitude or the direction of the sensitivity vector. The magnitude of the sensitivity vector can be changed by tuning the wavelength of the laser source, which is basically the multi-wavelength method. On the other hand, it is possible to modulate the direction of the sensitivity vector by altering the angle of the object beam, which is the multi-angle shape measurement method. The details about the experimental setups for high-speed multi-wavelength shape and multi-angle shape measurement methods can be found in our previous publications29,30,34−36 and are briefly described below.
Gauges traceable to the National Institute of Standards and Technology (NIST) were used to validate the shape measurement capabilities of the methods presented in this paper. We then applied these methods to measure the shape of human TMs and compare the measurement results obtained by the three methods, including illumination density, image acquisition speed, optical phase quality, resolution, processing time, temporal phase unwrapping compatibility, and the need for triangulation volume.
To verify the accuracy of the displacement measurement, a latex drum head subjected to an impulsive sound stimulus was simultaneously measured by the High-speed Digital Holographic system (HDH) and a Laser Doppler Vibrometer (LDV). The acoustically induced displacements of a human TM measured with HDH are presented in the results section to demonstrate the capability of the HDH system.
-
A tunable laser can be used to change the magnitude of the sensitivity vector by changing the wavelength of the laser source, as described in29. The optical phase difference
$ \left(\Delta \phi \right) $ of a holographic measurement of an object with two different wavelengths results in$$ \Delta \phi ={\boldsymbol{d}}\cdot {{\boldsymbol{e}}}_{2}-{\boldsymbol{d}}\cdot {{\boldsymbol{e}}}_{1}={\boldsymbol{d}}\cdot \left[\frac{2\pi }{{\lambda }_{2}}\left({\boldsymbol{b}}-{\boldsymbol{s}}\right)-\frac{2\pi }{{\lambda }_{1}}\left({\boldsymbol{b}}-{\boldsymbol{s}}\right)\right] $$ (4) In Eq. 4, the optical phase (
$ \Delta \phi $ ) is the difference measured from the interference pattern captured by the camera. The wavelengths$ {\lambda }_{1} $ and$ {\lambda }_{2} $ are known experimental parameters for the tunable laser, and b and s are the observation and illumination vectors, respectively, which can be measured directly or obtained by calibration. The optical path difference vector d can be solved to extract shape information. In the simplified 2D case shown in Fig. 9a, P is a random point on the subject, B is the observation point, and S is the illumination point. In this case, the amplitude of the optical path difference vector d can be expressed as:Fig. 9 2D Holographic schematic for shape measurements: a multi-wavelength experimental setup; b multi-angle wavelength experimental setup. The observation unit vectors are
$ {\boldsymbol{P}}{\boldsymbol{B}} $ and the illumination vectors are$ {\boldsymbol{S}}{\boldsymbol{P}} $ for the two methods.$ {S}_{1} $ and$ {S}_{2} $ are the illumination points before and after changing the illumination angle in the multi-angle method.$$\begin{split} \left|{\boldsymbol{d}}\right|=&\left|{\boldsymbol{P}}{\boldsymbol{B}}-{\boldsymbol{S}}{\boldsymbol{P}}\right|=\sqrt{{\left({{\boldsymbol{P}}}_{{\boldsymbol{x}}}-{{\boldsymbol{S}}}_{{\boldsymbol{x}}}\right)}^{2}+{\left({{\boldsymbol{P}}}_{{\boldsymbol{y}}}-{{\boldsymbol{S}}}_{{\boldsymbol{y}}}\right)}^{2}}\\&+\sqrt{{({{\boldsymbol{B}}}_{{\boldsymbol{x}}}-{{\boldsymbol{P}}}_{{\boldsymbol{x}}})}^{2}+{({{\boldsymbol{B}}}_{{\boldsymbol{y}}}-{{\boldsymbol{P}}}_{{\boldsymbol{y}}})}^{2}} \end{split}$$ (5) where
$ {P}_{y} $ is the user-defined pixel coordinates in the y-direction.$ {P}_{x} $ , which contains the shape information in the x-direction, is the only unknown component of Equation (5) but can be solved for. Eqs. 4,5 show that as the wavelength change ($ {\lambda }_{1}-{\lambda }_{2} $ ) increases, the synthetic wavelength ($ {{\lambda }_{1}{\lambda }_{2}}/({{{\lambda }_{1}-\lambda }_{2}})) $ increases, and thus uncertainty of the shape measurement decreases. -
The shape of the TM can also be measured by mechanically changing the direction of the illumination vector s. From Eq. 3,
$$ \Delta \phi ={\boldsymbol{d}}\cdot {{\boldsymbol{e}}}_{2}-{\boldsymbol{d}}\cdot {{\boldsymbol{e}}}_{1}={\boldsymbol{d}}\cdot \left[\frac{2\pi }{\lambda }\left({\boldsymbol{b}}-{{\boldsymbol{s}}}_{2}\right)-\frac{2\pi }{\lambda }\left({\boldsymbol{b}}-{{\boldsymbol{s}}}_{1}\right)\right] $$ (6) where
$ {{\boldsymbol{s}}}_{1} $ and$ {{\boldsymbol{s}}}_{2} $ are the illumination vectors at two camera exposures that can be measured or calibrated30. The shape information is obtained by solving the optical phase difference using Eq. 6.For the experimental arrangement shown in Fig. 9b, by introducing a known angle change (dθ) to the illumination vector, the height in the x direction can be determined by the following relation36:
$$ x=\Delta \phi \left({\frac{2\pi }{\lambda }\sin\theta d\theta}\right)^{-1} $$ (7) The sensitivity of the method increases as the angle change increases; moreover, it also depends on the non-zero illumination angle.
-
In this paper a third shape measurement method called fringe projection method was also implemented in our setup, which has several advantages comparing with the other two methods as being discussed in the Results and Discussion sections. A MEMS mirror (Brand: Mirrorcle) with a cylindrical lens was used to project a single fringe onto the sample, which was set up at its maximum scan rate of 3 kHz, and the frame rate and exposure time of the high-speed camera (Brand: Photron Fastcam SAZ, with a resolution of 512 by 512 pixels) were set at 67, 200 frames per second and
${1}/{67200} $ second, respectively. The camera was configured to be triggered by the position of the MEMS mirror to capture one image of a single fringe at each mirror position. All the single fringe images were summed up numerically to generate a synthesized, completed multi-fringe image after the high-speed acquisition.The experimental setup for the high-speed fringe projection is previously shown in Fig. 2. The light source used is a 532-nm fixed wavelength laser (Oxxius). The laser beam is divided into a reference beam and an illumination beam. Two mirrors direct the reference beam to reach the expansion lens and are combined with the object beam at the beam splitter in front of the High-speed camera. The illumination beam first reaches the MEMS mirror, which has two rotation axes. The vertical axis rotation is used to switch the illumination to the cylindrical lens for fringe projection, and it is also used to switch the expansion lens for displacement measurements. The cylindrical lens is used to generate a single vertical fringe onto the sample. The horizontal rotation of the MEMS mirror is used to scan the single fringe across the surface of the sample during the shape measurement. The mechanical shutter in the reference beam is activated to block the reference beam for fringe projection during the shape measurement. The mechanical shutter is deactivated during the high-speed displacement measurement, and the MEMS mirror fixes the illumination beam to the expansion lens. The reflected light from the sample interferes with the reference beam on the beam splitter in front of the camera. Finally, the resultant speckle pattern is recorded.
The position of the fringes was based on the size of the sample and the configuration of the optical setup. The first single fringe image was taken at the left-most side of the field of view. The second MEMS mirror position was programmed to be at 90° phase-shifted position of the first fringe, the third at the 180° phase-shifted position, and the fourth at the 270° phase-shifted position. (Four-phase shifting was chosen for its balance of shifting speed and overall accuracy; however, any phase shifting strategy can be implemented.) This process was repeated until the entire sample was scanned. The repeatability of the MEMS mirror is examined by investigating the signal fringe image. The MEMS is programmed to project a signal fringe to a fixed position (Fig. 10a) for high-speed imaging and move the fringe out of the field of view back to its zero position. This procedure is repeated five times. A Gaussian curve is horizontally fit to the intensity image row by row, and the peak of the Gaussian profile is used to determine the position of the projected fringe. The difference in fringe position of the five consecutive measurements is 0.63 pixels, and the averaged fringe width is 100 pixels. The error introduced by the MEMS mirror in repositioning the fringe is less than 1% of the fringe width, and is assumed to be negligible.
Fig. 10 The repeatability of the MEMS mirror is examined by checking the repeatability of the fringe position projected by the MEMS mirror during 5 consecutive TM measurements. a One fringe image with the Fringe position determined by averaged pixel position of the intensity image of each row. b the fringe position of 5 consecutive measurements and its averaged standard deviation is ± 0.63 pixels with averaged fringe width of 100 pixels. The reposition error introduced by the MEMS mirror is less the 1% of the fringe width and is assumed to be negligible.
Once all the single fringe images were captured, they were divided into four groups based on the angles of phase-shifting (0°, 90°, 180°, and 270°). The phase-shifting is calibrated based on the optical setup by making the distance between each projected fringe to be
$ {W}/{4} $ , where W is the width of the single fringe. The four-step phase-shifted multi-fringe images represented the numerical summation of each image group. The optical phase of the shape of the sample was calculated using Eq. 8.The fringe projection shape measurement method is based on the projection of well-defined fringe patterns onto the sample. In the high-speed fringe configuration, the high-speed camera captures the 4n single fringe images sequentially,
$ {I}_{1}^{1},{I}_{1}^{2},{I}_{1}^{3},{I}_{1}^{4},{I}_{2}^{1},{I}_{2}^{2},{I}_{2}^{3},{I}_{2}^{4}\dots {I}_{n}^{1},{I}_{n}^{2},{I}_{n}^{3},{I}_{n}^{4} $ , where the superscript indicates the phase-shifting order and the subscript indicates the fringe order. Numerical summations$ ({I}_{1}^{1}+{I}_{2}^{1}\dots {I}_{n}^{1}) $ ,$ \left({I}_{1}^{2}+{I}_{2}^{2}\dots {I}_{n}^{2}\right) $ ,$ ({I}_{1}^{3}+{I}_{2}^{3}\dots {I}_{n}^{3}) $ , and$ \left({I}_{1}^{4}+{I}_{2}^{4}\dots {I}_{n}^{4}\right) $ are calculated to reconstruct the completed multi-fringe images corresponding to the four-step phase-shifting frames. The optical phase φ of the shape of the sample is37$$ \phi =arctan\frac{\left({I}_{1}^{4}+{I}_{2}^{4}\dots {I}_{n}^{4}\right)-\left({I}_{1}^{2}+{I}_{2}^{2}\dots {I}_{n}^{2}\right)}{({I}_{1}^{1}+{I}_{2}^{1}\dots {I}_{n}^{1})-({I}_{1}^{3}+{I}_{2}^{3}\dots {I}_{n}^{3})} $$ (8) -
The displacement measurement method measures the full-field-of-view (> 200,000 points at 67,200 camera frame rate) transient displacements29,31,32−35. The Pearson's correlation coefficients between the deformed frame and two undeformed reference frames with 0 and 90° phase-shifting are calculated to quantify the optical phase of the sample's displacement. Assuming
$ {I}_{r} $ is the intensity of the reference beam and$ {I}_{o} $ is the intensity of the object beam, and φ is the optical phase difference between the object and the reference beams. The interference patterns of the reference (undeformed)$ {I}_{ref} $ and deformed$ {I}_{def} $ states can be expressed as shown in Eqs. 9,1029,38, respectively:$$ {I}_{ref}={I}_{r}+{I}_{o}+2\sqrt{{I}_{r}{I}_{o}}\cos\left(\phi \right) $$ (9) $$ {I}_{def}={I}_{r}+{I}_{o}+2\sqrt{{I}_{r}{I}_{o}}\cos\left(\phi +\Delta \phi \right) $$ (10) where
$ \Delta \phi $ is the optical phase change due to the sample's deformation. To extract$ \Delta \phi $ , we calculate the Pearson's correlation coefficient of a small window (e.g., 3 by 3 pixels) between the deformed and reference frames by using Eq. 1140$$ \rho \left({I}_{ref},{I}_{def}\right)=\frac{\left\langle{{I}_{ref}{I}_{def}}\right\rangle-\left\langle{{I}_{ref}}\right\rangle\left\langle{{I}_{def}}\right\rangle}{{\left[\left\langle{{{I}_{ref}}^{2}}\right\rangle-{\left\langle{{I}_{ref}}\right\rangle}^{2}\right]}^{\frac{1}{2}}{\left[\left\langle{{{I}_{def}}^{2}}\right\rangle-{\left\langle{{I}_{def}}\right\rangle}^{2}\right]}^{\frac{1}{2}}} $$ (11) where operator <> presents the expected value. Since the speckle pattern captured by the camera sensor is randomly distributed and the two same-sized matrices can be correlated, the expected value of the speckle pattern is equivalent to its numerical average39. Assume φ is uniformly distributed and
$ \Delta \phi $ is constant within the small window, then the coefficient can be simplified in the following format by inserting (9) and (10) into (11)32:$$ \rho =\frac{{(r+1)}^{2}+4r\cos\left(\Delta \phi \right)}{{(r+1)}^{2}+4r} $$ (12) where
$ r={{I}_{o}}/{{I}_{r}} $ is the intensity ratio of the object and reference beams. Therefore, the optical phase difference due to the deformation of the sample can be calculated by29$$ \Delta {\rm{\phi} }=arctan\frac{\rho ({I}_{ref},{I}_{def})}{\rho \left({I}_{ref+\pi /2}{,I}_{def}\right)} $$ (13) where
$ \left({I}_{ref+\frac{\pi }{2}}\right) $ is the 90° phase-shifted reference frame of the undeformed frame. Khaleghi et al. have demonstrated that the dominant deformation of the TM is along the surface normal direction40. Razavi40 demonstrated that the sensitivity vector of the holographic system varies when the sample is located close to the imaging lens; however, the magnitude of displacement (S) along the surface normal direction, can be compensated by the sensitivity vector variation K(x,y) and the surface normal vector of the sample n(x,y), can be measured as$ \Delta \phi (x,y,t) $ 34$$ S\left(x,y,t\right)=\frac{\Delta {\rm{\phi} }(x,y,t)}{{\boldsymbol{K}}(x,y)\cdot {\boldsymbol{n}}(x,y)} $$ (14) where x and y are the pixel coordinates of the camera and t is the time.
-
Pearson’s correlation coefficient was used to calculate the optical phase between two camera exposures. The previous algorithm used a sliding window technique on each recorded image. The process consists simply of moving a sliding window (e.g., 3 by 3 pixels) from point to point in the horizontal and vertical directions. Calculating the Pearson’s coefficient of two of 512 by 512-pixel images requires more than 250,000 iterations, which takes about 3.5 seconds on a modern computer. For a typical displacement measurement, more than 1000 images are captured, which takes about one hour to finish the process. The long processing time makes accessing the measurement quality during the measurement impractical. The algorithm is optimized by a vectorization procedure to remove the nested loops to improve the computational performance. Vectorization is the process of converting an algorithm from operating on a single value at a time to operate on a set of vectors or matrices at one time utilizing the vector operation capabilities of modern CPUs.
To generalize the Pearson's correlation coefficient calculation for n by n kernel size (n=3,5,7…), defining
$ {I}^{\text{'}}(i,j) $ are$ {n}^{2} $ translated matrices of the original image matrix I, where i= 1,2,3…n and j = 1,2,3…n. and$ {I}_{x,y} $ is the matrix element at (x,y) coordinate,$ {I}^{\text{'}}\left(i,j\right) $ is obtained by translating$ {I}_{x,y} $ by the following Equation:$$\begin{split}& {I}^{\text{'}}\left(i,j\right)={I}_{\left(x-\frac{n-1}{2}+i-1\right),\left(y-\frac{n-1}{2}+j-1\right)},\\&i= 1,2,3\dots n\;{\rm{and}}\;j= 1, 2,3\dots n \end{split}$$ (15) Both the undeformed frame
$ {I}_{ref} $ and deformed frame$ {I}_{def} $ are translated to$ {{I}_{ref}}^{\text{'}}\left(i,j\right) $ and$ {{I}_{def}}^{\text{'}}\left(i,j\right) $ by Eq. 15. The Pearson's correlation coefficient of$ {I}_{ref} $ and$ {I}_{def} $ can be calculated by Eq.11, and simplified expression is:$$ \rho \left({I}_{ref},{I}_{def}\right)=\frac{\sum _{i}^{n}\sum _{j}^{n}X\cdot Y}{{[\sum _{i}^{n}\sum _{j}^{n}{X}^{2}\cdot \sum _{i}^{n}\sum _{j}^{n}{Y}^{2}]}^{\frac{1}{2}}} $$ (16) where
$ X={{I}_{ref}^{\text{'}}}\left(i,j\right)-\frac{1}{{n}^{2}}\sum _{i}^{n}\sum _{j}^{n}{[{I}_{ref}^{\text{'}}}\left(i,j\right)]\;{\rm{and}}\;Y={{I}_{def}^{\text{'}}}\left(i,j\right)- $ $ \frac{1}{{n}^{2}}\sum _{i}^{n}\sum _{j}^{n}{[{I}_{def}^{\text{'}}}\left(i,j\right)] $ . The fastest calculation is the 3 by 3 pixels kernel version because it requires the minimum number of translated matrices, which takes about 0.15 s to finish on a modern computer. This 3 by 3 pixel kernel of the Pearson’s coefficient is suitable for displaying the results immediately after the measurement to assess the optical phase quality to determine if the measurement needs to be repeated.The high-speed holographic displacement measurement method described in this paper utilizes the optimized Pearson's correlation algorithm to calculate the optical phase change introduced by the surface deformation of the sample. With the optimized algorithm, the calculation becomes 21 times faster using the same computation power. The improved processing speed makes it possible to display the optical phase of the high-speed image immediately after the acquisition for us to examine the measurement quality between the measurements, which streamlines the experiment process and makes measurements on live ears more practicable.
Ultra-high speed holographic shape and displacement measurements in the hearing sciences
- Light: Advanced Manufacturing 3, Article number: (2022)
- Received: 30 September 2021
- Revised: 28 January 2022
- Accepted: 15 February 2022 Published online: 14 March 2022
doi: https://doi.org/10.37188/lam.2022.015
Abstract: The auditory system of mammals enables the perception of sound from our surrounding world. Containing some of the smallest bones in the body, the ear transduces complex acoustic signals with high-temporal sensitivity to complex mechanical vibrations with magnitudes as small as tens of picometers. Measurements of the shape and acoustically induced motions of different components of the ear are essential if we are to expand our understanding of hearing mechanisms, and also provide quantitative information for the development of numerical ear models that can be used to improve hearing protection, clinical diagnosis, and repair of damaged or diseased ears.We are developing digital holographic methods and instrumentation using an ultra-high speed camera to measure shape and acoustically-induced motions in the middle ear. Specifically we study the eardrum, the first structure of the middle ear which initializes the acoustic-mechanical transduction of sound for hearing. Our measurement system is capable of performing holographic measurement at rates up to 2.1 M frames per second. Two shape measurement modalities had previously been implemented into our holographic systems: (1) a multi-wavelength method with a wavelength tunable laser; and (2) a multi-angle illumination method with a single wavelength laser. In this paper, we present a third method using a miniaturized fringe projection system with a microelectromechanical system (MEMS) mirror. Further, we optimize the processing of large data sets of holographic displacement measurements using a vectorized Pearson's correlation algorithm. We validate and compare the shape and displacement measurements of our methodologies using a National Institute of Standards and Technology (NIST) traceable gauge and sound-activated latex membranes and human eardrums.
Research Summary
Ultra-high speed Holographic used to measure the shape and displacement of eardrum for hearing research
We are developing digital holographic methods and instrumentation using an ultra-high speed camera to measure shape and acoustically-induced motions of the eardrum in the middle ear. The eardrum is first structure of the middle ear which enables the transduction of sound for hearing. Our ultra-high speed holographic system can perform holographic measurement at rates up to 2.1M frames per second. We present a novel method using a miniaturized fringe projection system with a microelectromechanical system (MEMS) mirror incorporated into the holographic system. The processing of large data sets of holographic displacement measurements is optimized by a vectorized Pearson's correlation algorithm. the shape and displacement measurements of holographic measurements are validated and compared using a National Institute of Standards and Technology (NIST) traceable gauge and sound-activated latex membranes and human eardrums.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article′s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article′s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.