Developing of Algorithms Monitoring Heartbeat and Respiration Rate of a Seated Person with an FMCW Radar
Lorenz J. Dirksmeyer
Aly Marnach
Daniel Schmiech
Andreas R. Diewald
With a radar working in the 24 GHz ISMband in a frequency modulated continuous wave mode the major vital signs heartbeat and respiration rate are monitored. The observation is hereby contactless with the patient sitting straight up in a distance of 1–2 m to the radar. Radar and sampling platform are components developed internally in the university institution. The communication with the radar is handled with MATLAB via TCP/IP. The signal processing and realtime visualization is developed in MATLAB, too. Cornerstone of this publication are the wavelet packet transformation and a spectral frequency estimation for vital sign calculation. The wavelet transformation allows a fine tuning of frequency subspaces, separating the heartbeat signal from the respiration and more important from noise and other movement. Heartbeat and respiration are monitored independently and compared to parallel recorded ECGdata.
 Article
(2716 KB)  Fulltext XML
 BibTeX
 EndNote
Examining a person's health condition the initial step is the analysis of the vital signs in general. Those are pivotal indicators for any medical condition. Subject literature counts four, some five, vital signs: body temperature, heartbeat rate, respiratory rate, blood pressure and some add the blood's oxygen saturation additionally. The observation is common in hospitals as in daily homecare (BoricLubecke et al., 2016; Schmidt and Thews, 1980).
The electrocardiograph (ECG) is the gold standard in monitoring the heartbeat. It uses the heart's electrical activity causing it to contract to detect its beat. The electric waves pass through the body and can be detected sufficiently with adhesive electrodes on the skin. Two electrodes are used to measure the voltage drop between the muscles on different sides of the heart. The calculated heartbeat from an electrocardiograph is usually the average over a timewindow of 6–12 s. Since adhesive electrodes are stringently needed to tap the electric waves, there is no possibility to apply electrocardiography contactless. This is an issue when patients with skin diseases, skin burns or newborns with extremely sensitive skin are monitored. Beyond, these electrodes must be changed by medical staff regularly (BoricLubecke et al., 2016; Schmidt and Thews, 1980).
The Respiration can be measured with systems which need the tested person to inhale and exhale into a mouthpiece. The respiration is calculated by the air pressure and flow, corresponding measurements are called e.g. spirogramm and pneumotachography. Also possible is the indirect measurement of the respiration rate with a belt around the chest, tracking the expansion (respiratory inductive plethysmography). The chest's stretch is also measurable optoelectronic (optoelectronic plethysmography). Under certain circumstances it is possible to retrieve information about the respiration from ECGdata. The respiration rate is often sporadically observed, no measurement method gained general acceptance among the medical community, as they lack either in precision or practicality or pressure the patient (Neff et al., 2003; Pereira et al., 2017; Van Diest et al., 2014].
The concept of the radar is a promising platform for monitoring two of the most pivotal vital signs: Heartbeat and respiration rate. The characteristics of a radar's working principle allows a continuous noncontact observation even through common clothing.
Another advantage is the easy implementation, a radarsystem just needs a correct alignment to the observed patient unlike to the ECG must no electrodes be placed correctly. A huge advantage has the radar to common respiration measurements, which are usually performed with belts around the chest or require an extra mouthpiece. Both impede the actual respiration as they put extra pressure on the lungs, thus they may distort the measuring. Whereas radarmeasurements puts no extra load on the thorax as it works completely contact free (BoricLubecke et al., 2016; Walterscheid, 2017).
The heartbeat and the respiration, which are focused in this work, cause movement of the upper body. Those movements can basically be understood as periodic oscillations, a condition that enables the usage of radartechnique. The radar is able to track movements by emitting and receiving electromagnetic waves which are harmless to the patient's health. The radar's response provides information about the observed movements in line of sight, yet the information of all movements are superimposed from another. Hence the individual oscillations are not immediately retrievable and must be further processed to determine both vital signs. To separate both movements, an approach with wavelet filter banks and the frequency estimation with the rootMusic algorithm is elaborated in this research. A simulation of upper body movement is assembled providing a huge amount of data simulating different circumstances as different persons height, heartbeat and respiration frequencies and vary over time. The simulation does tremendous work to prove the concept, refine the approach and eradicate teething troubles. Afterwards the working principle is applied to data generated by real persons' body movement. Tested are offline and online approaches of the algorithm. For evaluation of the results, radar data and electrocardiography data are recorded parallel. The electrocardiograph's data is used as reference to calculate statistical errors of the algorithm and validate the successful estimation of respiratory and heartbeat rate (BoricLubecke et al., 2016; Schmidt and Thews, 1980).
Radar is an acronym for RAdio Detection And Ranging. The working principle of a radar is, broadly speaking, emitting RFwaves and collecting the reflected waves. Objects reflect RFwaves. The angle and intensity of the reflected wave is depending on the object's shape and material.
Radar systems are designed for several different frequency bands, in the following are all numbers fitted for the ISMband at 24.00–24.25 GHz and f_{0}=24.00 GHz.
A target moving towards the radar pushes the waves closer to each other increasing the frequency. A receding target flees from the emitted waves, so between two hits from wave fronts the time increases, therefore, the reflected wave's frequency is lowered. Apparently, this frequency shift, the Doppler frequency f_{D}, is only depending on the emitted wave's frequency f_{0}, the target's velocity v and the wave's speed, the speed of the wave in the medium. Assuming the wave travels through air, the wave's speed is approximately the speed of light c. It is assumed the target moves only in radial direction to the radar. A pure lateral motion would not be detected. A motion with both radial and lateral components is only detected by its radial velocity.
To detect the distance of a target to the radar, is commonly an operation mode called frequency modulated continuous wave chosen. Hereby the emitted frequency f is modulated over time. An example is to linear increase the frequency starting at f_{0} until a certain frequency f_{m} is reached, then the frequency drops back to f_{0} and rises again. This is called a continuous sawtooth ramp. The difference between f_{0} and f_{m} is the bandwidth f_{δ}. The time it takes to perform one ramp determines the ramp repetition frequency ${f}_{\mathrm{rr}}={t}_{\mathrm{ramp}}^{\mathrm{1}}$. Depending on the distance d from the target to the radar responds with a certain frequency on its voltage output U(t), the amplitude depends to the target's scattering and distance. However, only the obtained frequency leads to a decisive distance measurement. Is the target in motion there is an additional frequency shift due to the Doppler effect (Mahafza, 2009).
To gather first experiences with thoracic and abdominal movement a simulation in MATLAB is assembled. The simulated person hereby is adult and sits straight up. The simulation features a twobodymodel representing thorax and abdomen respectively (Karahasanovic et al., 2018). Both move due to the respiration and heartbeat. The variables defining the bodies and their movement, which are partly shared by both bodies, alter in a previously set interval to cover a wide amount of possible scenarios. Nevertheless the abdomen and chest movement is a complex and chaotic system, which is not covered in detail with this simplified twobody simulation. Also part of the simulation is a FMCWradar observing the person in a distance of approximately 2 m. The radar's antennas are aligned to the monitored persons' sternum. For later signal analysis the response of a FMCWradar in terms of a voltage over time signal is synthesized working in the 24 GHz ISM band with a ramp repetition frequency of 32 Hz. The twobody model is based on a number of distances, oscillation frequencies and amplitudes. All influencing the movement of the chest and abdomen regarding the heartbeat and respiration motion. The bodies are compromised to a single spot, for the chest (or thorax) this is the sternum, for the abdomen the navel. Both spots are chosen since they appear to have the most significant deviation due to body movement.
Objective of the signal processing is to derive the pulse and the respiration frequency within the dataset. The only input is the radar's voltage respond to objects and their movement in its line of sight. The working principle of a FMCWradar is to detect the distance and velocity of objects with the frequency contained in the radar's voltage output.
Therefore, the radar voltage signal of each frequency ramp is transferred into the frequency domain via FFT for classical range processing. Prior, the signal is windowed with a Hannwindow to suppress leakage and zeropadded to generate a larger amount of supporting points. This is done for all individual ramps separately. By determining the signal's peak in the frequency domain the corresponding frequency can be extracted which belongs to the observed twobody model. Quickly, it is apparent that there is only one peak in the frequency domain and not two, although the twobody model certainly consists of two targets in the radar's line of site. Both targets are so close to each other, their frequency lobes blur into one lobe and are not distinguishable. This is attributable to the resolution of a radar using a bandwidth of f_{δ}=250 MHz. The minimum resolution is Δ_{R min}≈0.6 m, which is far from the two body's distance which is ranging in a single digit centimetre area. Hence, to track the movement due to heartbeat and respiration in the millimetre and centimetre range the pure result of the FFT is not sufficient. When observing the progression of the radar response within consecutive ramps, it is noticeable that those are slightly phase shifted to each other. This phase shift is caused by the slightest movements of the twobody model, this phaseshift is accessible by interpreting the frequency domain signal's phase. From the first ramp, the bin with maximum peak in frequency domain (absolute values) is noted, the value of the phase in the frequency domain at this bin is saved. For the next ramp the phase value at this bin is also saved to this vector as for every following ramp. The phase rises and falls over time corresponding to the targets' movement opening and closing to the radar. This creates a new timedomain signal ϕ(t). The phase progression ϕ(t) is the result of the overlaid distances and movements. Only the respiration movement is vivid in the original not further processed ϕ(t).
The phase data is massively corrupted by fast alterations almost looking like points of discontinuity, making the signal noisy and tough to analyse. These points originate from the superposition of both bodies and their phase shift. An easy way to expose the discontinuities in ϕ(t) is by differentiating the signal with respect to the time. The fast alterations in ϕ(t) are significant peaks in $\dot{\mathit{\varphi}}\left(t\right)$. Yet, the other oscillations keep their frequencies. The noisy peaks are vividly identifiable as not being part of the continuous signal of $\dot{\mathit{\varphi}}\left(t\right)$. The main part of $\dot{\mathit{\varphi}}\left(t\right)$ is located in a limited interval. With the outlier test by Grubbs featuring the generalized extreme Studentized distribution (GESDtest) it is possible to detect data points which appear not to fit to the majority. The GESDtest is an statistical, iteratively applied outlier test for a data series x, based on the Grubbs test. The test detects outlier depending on the tdistribution, also known as Student's tdistribution, and a defined threshold. In advantage to the Grubb's and TietjenMoore test the number of outliers is not known in advance for the GESD test and it performs better when two or more outlier follow each other consecutively (Grubbs, 1950; NIST, 2012).
In detail the GESDtest works as follows: The data vector x is checked for outliers. The vector x with its length N changes in every iteration, that is, it is denoted with the iteration number k. First, the the value R_{k} of the N−klength data vector x_{k} is calculated (Eq. 1), here σ_{x} is the standard deviation and μ_{x} the mean value of x. The data point of x_{k} that maximizes Eq. (1) is the examined data point n_{k} and potential outlier.
Following α is defined to set the significance level. The value of $\mathit{\u03f5}=Nk\mathrm{1}$ varies in each iteration, too. The value of ϵ defines the tdistribution (Eq. 2), whereby ϵ represents the distribution's degree of freedom and Γ the gamma function.
The probability value p is defined as in Eq. (3).
The value of t_{p,ϵ} is the 100p percentage point from the tdistribution from the kth iteration. Eventually, the threshold ξ is calculated (Eq. 4).
If Eq. (5) is true, the data sample n_{k} is classified as outlier and removed. The algorithm restarts with N−k samples respectively. If Eq. (5) is false, n_{k} is not an outlier and there are no (further) outlier in the dataset, the GESD test is completed (Grubbs, 1950; NIST, 2012)
The detected outliers are replaced with interpolated data by using a fitted chirp signal. The processing of the signal is shown in Fig. 1. This signal appears to be pure, as both oscillations due to heartbeat and respiration are already identifiable. The exact angle information gets lost due to the processing. Its frequency spectrum shows a clear peak for the respiration and a several swings in the heartbeat rate region, see Fig. 2.
The wavelet transformation is chosen to separate both vital sign signals in ϕ(t) since it performs signal analysis and synthesis in different frequency ranges simultaneously. In comparison to a common shorttime Fourier analysis the wavelet transformation serves with a better frequency resolution at lower frequencies and further a better time resolution at higher frequencies. The wavelet analysis fits best for signals interrupted by discontinuities, while usually having a smooth progression and for signals that superimpose each other despite being in different frequency ranges, both is the case in this application.
The wavelet transformation is a timefrequency analysis based on filtering. It serves for signal analysis, filtering and synthesis. Its foundation is the use of multiple scaled window functions and time shifting. A wavelet ψ(t) is a zeromean window function with an average frequency f_{ψ}≠0. A not scaled or time shifted wavelet is called mother wavelet. Scaling with a and time shifting with b, as in Eq. (6) with $a,b\in {\mathbb{R}}^{+}$, modifies the wavelet ψ_{a,b} for the later applied wavelet analysis.
Bandwidth Δ_{f} and time duration Δ_{t} of the transform are heavily depending on a. The factor $\frac{\mathrm{1}}{\sqrt{a}}$ secures a constant signal energy of any scaled wavelet. The result of the wavelet transform are the wavelet coefficients. While the average time is τ=b and the average frequency is $f=\frac{{f}_{\mathit{\psi}}}{a}$. A small a leads to fine frequency resolution at high frequencies, large a can perform a fine frequency resolution at low frequencies. Any wavelet function ψ(t) and its corresponding frequency transform Ψ(f) must satisfy (Eq. 7).
Is Eq. (7) true, fulfils the wavelet transform the Parseval energy conservation. The wavelet transform ${W}_{x}^{\mathit{\psi}}(a,b)$ of a signal x(t) is performed with Eq. (8).
A time shift t_{s} in x(t−t_{s}) shifts the wavelet transform by the same time ${W}_{x}^{\mathit{\psi}}(a,b{t}_{\mathrm{s}})$, hence the wavelet transform is invariant for a time shift. Also the wavelet transform is invariant in scaling, a function is scaled with s as $\frac{\mathrm{1}}{\sqrt{\mathrm{}s\mathrm{}}}x\left(\frac{t}{s}\right)$ the transform is scaled, too: ${W}_{x}^{\mathit{\psi}}\left(\frac{a}{s},\frac{b}{s}\right)$ (Percival and Walden, 2000; Kienecke et al., 2008).
To benefit the wavelet transform's potential to resolve good frequency resolutions in high as in low frequencies, the transform is performed multiple times (Ktimes) with differing a and b. The scale factor a is usually chosen dyadic as a_{k}=2^{k} with $k=\mathrm{1},\mathrm{2},\mathrm{\dots}K$. The time scale factor b rises dyadic, too, ${b}_{k,m}={\mathrm{2}}^{k}mT$, that is, the sample distance of the wavelet coefficients are further apart for wavelets with a higher k. The analysing wavelet ψ_{k}(t) is calculated dyadic as in Eq. (9).
The wavelet transform at (m,k) is then ${W}_{x}^{\mathit{\psi}}(m,k)={W}_{x}^{\mathit{\psi}}\left(m{\mathrm{2}}^{k}T,{\mathrm{2}}^{k}\right)$ with ${W}_{x}^{\mathit{\psi}}$ as in Eq. (8).
In general, a wavelet is characterized as bandpassfilter. For a k the wavelet transform projects x(t) in a subspace with the frequencies $]{\mathrm{2}}^{k}{F}_{\mathrm{N}},{\mathrm{2}}^{k+\mathrm{1}}{F}_{\mathrm{N}}]$, whereby F_{N} is the Nyquist frequency. The projection results in a function y_{k}(t). For the original signal x(t) and its projections y_{k}(t) take Eqs. (10) and (11) affect. The signal x_{k}(t) takes the lowpass parts of $x\left(t\right)\phantom{\rule{0.125em}{0ex}}\widehat{=}\phantom{\rule{0.125em}{0ex}}{x}_{k\mathrm{1}}\left(t\right)$ which are below the bandpasssubspace of y_{k}(t). Equations (10) and (11) show the mathematical connection of x(t), x_{k}(t) and y_{k}(t) with $\overline{x}\left(t\right)$ representing x(t) in the relevant frequency domain of [F_{N} 0].
These coherence can also be expressed in their subspaces. For a function x(t)∈L_{2}(ℝ), with L_{2}(ℝ) being the subspace of all squareintegrable functions, is y_{k}(t) the projection in the subspace W_{k} while x_{k}(t) is part of the lowpass subspace V_{k}. The subspaces have the coherences Eqs. (12) and (13).
The transition from V_{k−1} to W_{k} is performed by a bandpass filter G_{BP} with the impulse response g_{BP}(m), and from V_{k−1} to V_{k} by a lowpass filter G_{LP} with the impulse response g_{LP}(m). The subspace V_{k} is spanned by the orthonormal base function ${\mathit{\varphi}}_{m,k}\left(t\right)={\mathrm{2}}^{\frac{k}{\mathrm{2}}}\cdot \mathit{\varphi}({\mathrm{2}}^{k}tmT)$ which is orthogonal to the wavelet ψ_{m,k}(t) for all m,k. The base function ψ is also called scale function and is able to describe x_{k}(t) in V_{k}.
When transitioned into another subspace the impulse responses g(t), wavelet ψ_{k+1}(t) and scale functions ψ_{k}(t) and ψ_{k+1}(t) are related with Eq. (15).
The argument 2l−m needs some closer attention, increasing the index of ψ_{k+1}(t) and ψ_{k+1}(t) by one, increases the index in ψ_{k}(t) by two. Consequently, the subspace V_{k+1} and W_{k+1} are sampled with halve the sample frequency as V_{k}. For the inverse wavelet transformation besides the scale and wavelet function the impulse responses ${h}_{\text{BP}}\left(m\right)={g}_{\text{BP}}^{*}(m)$ and ${h}_{\text{LP}}\left(m\right)={g}_{\text{LP}}^{*}(m)$ and their corresponding filters are needed. Eventually, to determine the filter bank only h_{LP} is needed from the start. The other filter and their responses can be calculated with h_{LP}, the bandpass filter response is ${h}_{\text{BP}}\left(m\right)=(\mathrm{1}{)}^{m}{h}_{\text{BP}}^{*}(\mathrm{1}m)$. The low pass filter H_{LP} and its response h_{LP} are used to calculate the scale function ψ(t) and the wavelet function ϕ(t), too.
With the low pass filter response h_{LP} and the scale filter ϕ(t) the wavelet function ψ is calculated in Eq. (17).
The previously described fully processed ϕ(t) is used for the transformation. Used are the MATLABfunctions modwt
and modwtmra
for transformation and analysis. A Symlet4 is utilized as wavelet and its corresponding scaling function. The wavelet transformation level is 7, following a sample frequency of f_{rr} the signal is divided in frequency intervals as in Table 1 (Percival and Walden, 2000; Kienecke et al., 2008; MathWorks, 2020).
For the respiration frequency intervals VI and VII are to consider, satisfying the respiration rate in a range of 12 to 20 respirations per minute. For the pulse the interval IV is sufficient for the simulation. Although cutting the heartbeat frequency off at 124 bpm which is sufficient for the simulated data. The wavelet coefficients are processed to synthesize the original signal in the corresponding frequency bounds. For the respiration signals the intervals VI and VII are summed up, baseline wander can be removed easily (Kienecke et al., 2008; MathWorks, 2020).
The frequency in the separated signals is estimated with the rootMUSIC. The rootMUSIC is a variant of the MUSIC algorithm, which is a frequency estimation method based on the fundamentals of Pisarenko's harmonic decomposition. It was first presented by Arthur J. Barabell in 1983. The pivotal step is to split the correlation matrix' R_{x} eigenvectors λ in a signal subspace and noise subspace (Manolakis et al., 2005; Barabell, 1983).
For the RootMUSIC algorithm some knowledge about the evaluated signal x(t) is needed: that is the signal length M and more crucial the number of contained harmonic oscillations P. For the rootMUSIC algorithm $M>P+\mathrm{1}$ must be true, that means the evaluated time window of x(t) must be at least contain two more samples than harmonic oscillations. The signal x(t) is assumed to be the sum of the actual signal s(t) and uncorrelated white noise w(t). x(t) can be denoted as Eq. (18), whereby the vector v(f) is the frequency vector in form of a Mlength DFTvector.
Similar to the time signal itself the autocorrelation R_{x} can be split into signal R_{s} and noise correlation matrices R_{s}, assuming independence of signal and noise. The autocorrelation may also be expressed with eigenvalues λ and their respective eigenvectors q, as defined in Eq. (19).
The eigenvalues are sorted in descending order, that ${\mathit{\lambda}}_{\mathrm{1}}\ge {\mathit{\lambda}}_{\mathrm{2}}\ge \mathrm{\dots}\ge {\mathit{\lambda}}_{M}$ is true. For m>P the eigenvalues λ_{k} are caused by noise and can be summarized in ${\mathit{\sigma}}_{w}^{\mathrm{2}}$. For k≤P the eigenvalues λ_{k} are evoked by the actual signal s(t). The subspaces of signal and noise are orthogonal following is the frequency vector v^{H}(f_{p}) orthogonal to Q_{w} with ${\mathbf{\text{Q}}}_{w}=\left[{\mathit{q}}_{P+\mathrm{1}}\mathrm{\dots}{\mathit{q}}_{M}\right]$. The pseudospectrum ${\stackrel{\mathrm{\u203e}}{R}}_{m}$ for one eigenvector q_{k} calculates to:
The denominator of Eq. (20) produces M−1 roots, M−P of them due to the noise subspace and occur on different frequencies without any restrictions. However, P of them are correspond to the P oscillations in s(t). To secure a correct identification of peaks corresponding to the frequencies f_{P}, the MUSICpseudospectrum ${\stackrel{\mathrm{\u203e}}{R}}_{\text{MUSIC}}$ is defined as in Eq. (21).
The peaks of ${\stackrel{\mathrm{\u203e}}{R}}_{\mathrm{MUSIC}}$ represent the frequencies f_{P}. For the rootMUSIC approach the ztransform of the denominator in Eq. (21) is needed, Eq. (22) represents the sum of the noise vector's pseudospectrum in the zdomain.
From Eq. (22) stem M−1 pairs of roots with one inside and one outside the unit circle. Since there is no attenuation on the P oscillations, the Pclosest roots to the unit circle match the oscillations in s(t). For real signals x(t)=Re(x(t)) with k frequencies f_{k}, P=2k roots have to be regarded. Due to the two sided spectrum the roots appear at each f_{k} and at −f_{k} (Manolakis et al., 2005; Barabell, 1983).
The twobody model for the simulation is sufficient for a first testing, yet it is not covering the complex movements of the upper body entirely. Further are the altering variables, such as the heartbeat rate, really complex to model as they follow no known rule and are more like a mathematical chaos. The simulation of their shift is a simple approximation. To proceed with the algorithm development it is tested with real data. The data is measured with a radar working under the same settings as the simulation except for the sampling rate, which is slower, and the ramp repetition frequency, which is faster. Each measurement is 3 min in time. Several test persons were used for the measurement procedure, in some cases they were given special tasks such as holding their breath for a period of time. All have a healthy respiration system. For later comparison the monitored persons are connected to an ECG which tracks the heartbeat data parallel.
Before starting the vital sign calculation, the phase progression ϕ(t) is extracted as described in the previous section. The reconstructed signals for both vital signs there is a noticeable differing in their frequency and amplitude over time which is distinctly higher than in the simulation. The oscillations appear uneven as there are flatter and steeper parts in the phase progression. The signal processing algorithm developed for the simulated signals fails, especially for the heartbeat. Errors between the ECGdata are significant and not tolerable. Hence the algorithm is further developed and the input signal analysed.
In comparison to the simulated twobody model the signal ϕ(t) is derived from the chest movement affected by considerably more radar targets as neither the chest nor the abdomen can be considered as uniformly moving unit. The movement amplitude is differing on the various upper body parts. Additionally, besides the movements due to heartbeat and respiration other irregular motions are to expect, although the test person were told to sit still and not to talk. In sum this certainly leads to a less smooth phase progression as in the simulation with more superimposed intermittent oscillations. Since the wavelet transform's resolution already causes problems in analysis of the simulation an improved version of the wavelet transformation is chosen for the further development: The wavelet packet transform. For most applications in signal processing is it useful to have the targeted frequency parts in preferably few subspaces. Yet, especially in higher frequency ranges the frequency bandwidth can be quite huge, while the short time resolution in those is not essential if the signal is fairly stable in its frequency over time. This issue can be solved by using wavelet packets. In the regular wavelet transformation the subspaces V_{k} and W_{k} are strictly expanded from the previous lowpass subspace V_{k−1}. Thus, the bandpass subspace W_{k} is not further processed. The subspaces in the wavelet packet transformation are expanded in both V_{k} and W_{k}. That is, the signal y_{k}(t) is transformed, too. For k=1 is no difference to the standard wavelet transformation, but if k=2 there are 4 subspaces present. The indice i is introduced along with k:

V_{2,0} the lowpass subspace of V_{1} containing x_{2,0}(t)

W_{2,1} the bandpass subspace of V_{1} containing y_{2,1}(t)

V_{2,2} the lowpass subspace of W_{1} containing x_{2,2}(t)

W_{2,3} the bandpass subspace of W_{1} containing y_{2,3}(t)
For larger k the time resolution becomes worse, yet the frequency resolution improves. By a constant k signals and subspaces with higher i represent higher frequency bandwidths (Kienecke et al., 2008).
This transform is capable to realize an even resolution distribution of the individual frequency intervals. The frequency width of a single interval (or level) is depending on the chosen tree depth. To achieve a desired resolution, the treedepth is chosen to 9 which results in 512 nodes. By a radar ramp repetition frequency of 62.5 Hz the frequency levels have a width of 0.061 Hz (3.662 bpm). MATLAB provides two functions to perform a complete transform with wavelet filter banks: modwpt
ensures an accurate resolution of the energy to the frequency levels, drawback is a delay in the time domain for the wavelet coefficients. This effects arises from the use of filters with a nonlinear phase response. While the other function modwptdetails
is true to the time, which is particularly important for the signal synthesis. In opposite to the other this function utilises zerophase filtering for the transformation, which is mandatory for the application.
The frequency may be estimated from a time domain signal, therefore the wavelet coefficients are regarded, which represent the original signal in the respective frequency level. For this transformation the function modwptdetails
is used. Since the exact value is yet to be determined, the whole pulse related possible frequency range is regarded. That is, the levels 14–49 contain frequencies in the interval of 0.824–3.021 Hz (49.44–181.27 bpm). Therefore, the coefficients of these levels are summed up. This signal equals the original signal in the corresponding frequency bounds.
For heartbeat estimation is the signal cut into 3 s parts. The consequently following data snippet is reconstructed in dependence to the previously detected heartbeat rate. Only adjacent frequency levels of the one containing the heartbeat are regarded, limiting the evaluated frequency interval to 0.31 Hz (18.6 bpm). With the reconstruction being a simple summation of wavelet coefficients, it is easy to handle for each data excerpt.
Since the phase progression appears to be less harmonic and the oscillations' frequency and amplitude differs massively, the estimation with rootMUSIC causes issues. The number of frequencies in a signal processed with rootMUSIC must be known precisely. If the signal or the evaluated signal part is corrupted due to minor body movement the estimated frequency can deviate from the correct heartbeat frequency massively. Hence, the data is searched progressively for erratic oscillations with the PearsonCoefficient. If a fitting part is not detected, it is omitted. Nevertheless, outlier occur those are ascertained and removed by the Grubbs test for outlier and afterwards replaced by interpolation. An example for excluded data is shown in Fig. 3.
For the detection of the respiration movement and consequently the respiration rate the same input data is used as for the heartbeat rate. The processing is similar to the heartbeat detection. The data is differentiated before wavelet transformed, too. The relevant frequency interval is lower and overall smaller as for the heartbeat, yet the level frequency width is chosen the same with 0.061 Hz. That is, the wavelet packet transform must be performed only once to calculate both vital signs. The respiration is assumed in the frequency range of 12 to 20 breaths per minute. Thus, the levels 2 to 6 are used for signal reconstruction, with the signal having frequencies in the bounds of 0.09–0.40 Hz (5.5–23.8 bpm). For the time signal analysis the sliding time window is adjusted to the slower frequencies, too. The window size is 500 sample points (8 s of data), this time window should be sufficient to have one full breathing cycle within, even for slow respiration rates. The time step for each evaluated data excerpt is set to 1.92 s, the alterations in the respiration are expected to be not as sudden as for the heartbeat. The sample frequency of the generated respiration signal following the time step length is roughly 0.5 Hz.
3.1 Results
Having the goldstandard of heartbeat frequency measurement electrocardiography as reference is the best case scenario to evaluate the algorithm results. Observing the error over time m(t) for all measurements are significant deviations partially only for a few seconds noticeable. Looking closely into the data for some parts massive outlier are recognizable, for example in one measurement the ECG calculated heartbeat frequency rises tremendous two times for 2 s peaking at over 220 bpm and falling then back to plausible values. This happens into unnatural deep ranges, too, in another measurement the ECG calculated pulse falls from 90 below 40 within seconds before rising again. Those miscalculations by the ECG may not compromise the quality of the radar based pulse computation. Therefore, the error is calculated once more with the intervals of implausible ECGdata omitted. With this approach the average absolute error over all measurements is $\stackrel{\mathrm{\u203e}}{m}=\mathrm{3.138}$ bpm with a standard deviation of σ=2.839 bpm. The results of one single measurement and comparison to ECGdata is shown in Fig. 4 and in numbers for eight measurements in Table 2.
On the first look, the respiration rate calculated with the radardata is plausible. The respiration has some dynamic over time and yet no discontinues are to find. Only smaller spikes strike arise, these could be avoided by applying a lowpass filter, as a movingaverage filter, with the drawback of more timeinertial. However, within some of the datasets the monitored persons were told to hold the breath for a certain time. These parts of respiratory arrest are not correctly detected at all. The reason for that failure are apparent, the wavelet level of very low frequencies, below 0.09 Hz (5.5 bpm), are excluded to avoid wrong detections at first. In this region is a lot of noise which would disturb the frequency estimation. Additionally forces the frequency estimation with the rootMUSIC algorithm a detection of a single frequency within a signal, although there is none with significant energy. The estimation treats the oscillations equal not regarding their energy. If the monitored person does not move its chest in the observed frequency range, a phase variation signal within the range will be classified as respiration movement and then estimated as respiration frequency. The mean error for the 4 datasets without respiratory standstill is m=1.92 bpm (0.032 Hz) and the standard deviation σ_{m}=1.21 bpm (0.020 Hz). A BlandAltman plot showing the results in comparison to the ECGdata is given in Figs. 5 and 6.
Vital sign measuring is most importantly done to trace the data in a live measurement as the data is needed immediately for further diagnosis. Therefore, the given measurement sets are processed pseudolive, that is, the signal is partially loaded to the algorithm in real time. In opposite to the initial offline processing algorithm the data is completely transformed to the wavelet coefficients at the beginning. Since the single level frequency resolution shall not be decreased, the wavelet packet transformation tree depth is unchanged at 9. For the long offline processed data sequence, this is quite CPUintensive but once done the coefficients can be used in arbitrary ways without extra computations. In the live algorithm the angle data ϕ(t) is only piecewise processable, hence the wavelet packet transform must be performed several times generating an increase in the calculation time. To ensure proper data sampling and enough calculation time the heartbeat rate here is estimated just every 3.6 s. The short data windows causes another hitch. The wavelet transform is based and bound to different filter, those filter appear to have some settling time as the results of the transform are attenuated at the beginning and the end of the signal. The signal snippet is basically split in thirds. The first and the last third are just dummy data for the processing as the reconstructed signal contains corrupted data in these parts caused by the wavelet transform. That is, only the middle section of the signal part is used for frequency estimation. the latest calculated frequency and the ones already present from former iterations in the heartbeat ratevector are checked with the GESD test for outlier. As before, detected outlier are replaced by interpolated values. Although tremendous outliers are removed, the heartbeat frequency over time signal appears erratic also it frames the desired values. These leaps are not explainable by heartbeat variations. When observing the derived ECGsignal it is erratic too, yet not to the degree the signal from the radar derived is. In practice, however, the ECGsignal is always averaged over time before it is displayed. Since the ECG is considered the goldstandard of pulse measuring this is a widely held and proven approach. Following, is the heartbeat over time signal derived from the radar measurement averaged over time, just as the ECGsignal, for a fair comparison. Likewise to the extraction from the respiration in offline processing, the observed time window is enlarged in comparison to the heartbeat detection to fit the appreciable lower frequencies occurring. The lowest in the regarded frequency range frequency has a time period of roughly 10 s, the window size is therefore chosen with a length of 16 s in time.
4.1 Results
Eventually the comparison of live processed smoothed data to ECG's data is closer to realworld data since erratic peaks are usually omitted as those are sleeked for evaluation whereas they are assumed as measurement inaccuracy. In opposite to the offline processed data no extra evaluation, excluding suspected ECGdata measuring errors or disruptive background movement, is done. The time averaged signals are used for comparison, this mitigates the influence of measuring errors to an extend, as single deviations get lost.
The heartbeat rate only suffer slight errors, larger errors only happened only in time intervals of 5 s or less, but observed throughout the measuring period the result is satisfying. The absolute mean error is $\stackrel{\mathrm{\u203e}}{m}=\mathrm{3.311}$ bpm with a standard deviation of σ=3.783 bpm. The pure result might be slightly worse than for the offline processed data but with no probably corrupted ECG failures omitted this is expected and still a fairly good result. Exemplary are the results of one single measurement shown in Fig. 7.
The datasets including a respiratory arrest are excluded from the review, as they are corrupted by measurement failures inevitably. Although the respiration rate's values of both approaches are within the same range and the rate's trend is discernible as similar, the deviations become significant within few intervals. Which is possibly caused by the differing approaches the compared data is acquired.
Overall the mean error is m=1.21 bpm (0.020 Hz) and the standard deviation of the mean error σ_{m}=1.74 bpm (0.029 Hz), the individual numbers are presented for each measurement in Table (3). This result is satisfying, as the radar delivers respiration rates comparable to the ECG. Partly, the radar data is even more plausible, as it omits rapid variation in the respiration rate, which are tracked in the ECGdata. The comparison is visualized in Figs. 7 and 8.
With the lessons learned from the processed simulated data the basics of the algorithm to evaluate data sampled from real humans are quickly established. Yet, other difficulties are encountered as the simulated data lacked in the complexity of the upper body movement and missed common disruptions in the data and measurement noise. The use of correlation, the Pearsoncoefficient and the GESDtest helped to overcome those troubles as corrupt data are detected and excluded from the evaluation. For the heartbeat rate, the offline processing algorithm provided satisfying results overall, when compared to the ECGdata. The estimated respiration rate for all offline processed datasets is plausible in terms of range and dynamic and follow the same trends as the ECGdata. One major issue for the respiration rate estimation is unavoidably the lack of correct respiratory standstill detection.
Since it is a common approach and consequently closer to the reality the calculated live processed data and the corresponding ECGdata time are averaged before comparison. Here too, the results for the heartbeat rate compared to the ECGdata are satisfying. The impacts of disruptive background movements causes inferior errors than in the offline processed data, explainable by the larger observed time window and the time averaging. For all datasets without respiratory standstill the results show the working principle is adequate. The slightly larger regarded data time windows in the calculation have no adverse effect on the respiration rate detection. The retrieved data stands comparison to the ECGdata, similar to the offline processed data. The radar based progression is smoother as it omits rapid variations, making it even more plausible than the ECGdata. The radar trends to detect slightly lower respiration rates, as the deviation between both approaches averages +1.21 bpm (+0.02 Hz).
Due to the fact that the modeling and verification code is under patent application approval, the code cannot be published.
All relevant models and results of the research are contained in the manuscript, the work can be reproduced with the presented models applied on standard raw data form a classical CWradar in the GHz range.
LJD worked on the project as Master student at the Laboratory of Radar Technology and Optical Systems, he modulated the simulation, developed the signal processing and evaluated the results. AM performed the radar and ECG measurements and, along with DS, supervised the signal processing. ARD is the head of the institution, he gave the initial idea and organised the team.
The authors declare that they have no conflict of interest.
Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This article is part of the special issue “Kleinheubacher Berichte 2020”.
Parts of this project are funded by the Federal Ministry of Education and Research within the program KMU innovativ in the project ProDemiis. Grant number: 16SV7933.
This research has been supported by the Bundesministerium für Forschung und Technologie (grant no. 16SV7933).
This paper was edited by Madhu Chandra and reviewed by Dennis Vollbracht and one anonymous referee.
Barabell, A.: Improving the Resolution of Eigenstructurebased DirectionFinding Algorithms, IEEE, Boston, Massachusetts, USA, 1983. a, b
BoricLubecke, O., Lubecke, V., Droitcour, A., Park, B., and Singh, A.: Doppler Radar Physiological Sensing, John Wiley & Sons, Hoboken, NewJersey, USA, 2016. a, b, c, d
Grubbs, F.: Sample Criteria for Testing Outlying Observations, University of Michigan, Ann Arbor, Michigan, USA, 1950. a, b
Karahasanovic, U., Stifter, T., Beise, H., Fox, A., and Tatarinov, D.: Mathematical Modelling and Simulations of Complex Breathing Patterns Detected by RADAR Sensors, 2018 19th International Radar Symposium (IRS), 1–10, https://doi.org/10.23919/IRS.2018.8448045, 2018. a
Kiencke, U., Schwarz, M., and Weickert, T.: Signalverarbeitung – ZeitFrequenzAnalyse und Schätzverfahren, Oldenbourg Verlag, Munich, Germany, 2008. a, b, c, d
Mahafaza, B.: Radar Signal Analysis and Processing Using MATLAB, CRC Press, Boca Raton, Florida, USA, 2009. a
Manolakis, D., Ingle, V., and Kogon, S.: Statistical and Adaptive Signal Processing, Artech House, Norwood, Massachusetts, USA, 2005. a, b
MathWorks Deutschland: MATLAB Documentation, The MathWorks Inc., available at: https://de.mathworks.com/help/matlab/ (last access: 5 October 2021), 2020. a, b
Neff, R., Wang, J., Baxi, S., Evans, C., and Mendelowitz, D.: Respiratory Sinus Arrhythmia, Circ. Res., 93, 565–572, https://doi.org/10.1161/01.RES.0000090361.45027.5B, 2003. a
National Institute of Standards and Technology (NIST): Handbook of Statistical Methods, US Department of Commerce, Gaithersburg, Maryland, USA, 2012. a, b
Percival, D. and Walden, T.: Wavelet Methods for Time Series Analysis, Cambridge University Press, Cambridge, United Kingdom, 2000. a, b
Pereira, M., Porras, D., Lunardi, A., and da Silva, C.: Thoracoabdominal Asynchrony: Two Methods in Healthy, COPD, and Interstitial Lung Disease Patients, PLoS ONE, 12, e0182417, https://doi.org/10.1371/journal.pone.0182417, 2017. a
Schmidt, R. and Thews, G.: Physiologie des Menschen, 20th Edn., Springer, Berlin, Germany, 1980. a, b, c
Van Diest, I., Verstappen, K., Aubert, A., Widjaja, D., Vansteenwegen, D., and Vlemincx, E.: Inhalation/Exhalation Ratio Modulates the Effect of Slow Breathing Heart Rate Variability and Relaxation, Springer Science + Business Media, New York, USA, 2014. a
Walterscheid, I.: Respiration and Heartbeat Using a Distributed Pulsed MIMO Radar, IEEE, Boston, Massachusetts, USA, 2017. a
 Abstract
 Introduction
 Simulation of Thorax and Abdomen Movement
 OffLine Processing for RealWorld Data
 Live Processing for RealWorld Data
 Conclusions
 Code availability
 Data availability
 Author contributions
 Competing interests
 Disclaimer
 Special issue statement
 Acknowledgements
 Financial support
 Review statement
 References
 Abstract
 Introduction
 Simulation of Thorax and Abdomen Movement
 OffLine Processing for RealWorld Data
 Live Processing for RealWorld Data
 Conclusions
 Code availability
 Data availability
 Author contributions
 Competing interests
 Disclaimer
 Special issue statement
 Acknowledgements
 Financial support
 Review statement
 References