http://www.freecsstemplates.org Released for free under a Creative Commons Attribution 3.0 License Name : Unofficial Channels Description: A two-column, fixed-width design with a bright color scheme. Version : 1.0 Released : 20120723 --> Vol. 35 No. 3

 

Instantaneous Position and Orientation of the Body

Segments as an Arbitrary Object in 3D Space by

Merging Gyroscope and Accelerometer Information

J.A. Barraza-Madrigal*
R. Muñoz-Guerrero*
L. Leija-Salas*
R. Ranta**
* Centro de Investigación y de Estudios Avanzados (CINVESTAV-IPN).
** Centre de Recherche en Automatique de Nancy (CRAN-CNRS)
ABSTRACT

This work presents an algorithm to determine instantaneous orientation of an object in 3D space. The orientation was determined by using a Direction Cosine Matrix (DCM), performed by the combination of three consecutive rotations, around each to the main axes of the evaluated system, using quaternions. An inertial measurement unit (IMU), consisting of 3 axes gyroscope and 3 axes accelerometer was used in order to establish 2 coordinate systems; The first one describes object movement, by using gyroscope as a main source of information, relating the angular rate of change along time. The second defines a coordinate reference system, relating the acceleration of gravity to an inertial direction vector. A proportional integral (PI) feedback controller was used, which includes sensors information, eliminating offset, cancelling drift and improving the accuracy of the orientation. The proposed algorithm can be used for assessing both the position and orientation of the body segments which is very important in orthopedic, traumatology and rheumatology important for diagnosis, prognostic, therapeutic, research and as well as the design and fabrication of measuring devices used in surgical instrumentation, prostheses and ortheses. It is important to note that the developed system opens the opportunities to be implemented on ambulatory joint evaluation through a wearable system, due to the dimensions and requirements of the sensors.

Keywords: instantaneous orientation, direction cosine matrix, quaternions, inertial measurement unit, offset and drifting deviation.

Correspondencia:
Barraza-Madrigal J.A
Av. IPN 2508 Zacatenco 07360, México DF
Correo electrónico: jbarraza@cinvestav.mx
Fecha de recepción:
5 de mayo de 2014
Fecha de aceptación:
10 de octubre de 2014

RESUMEN

El presente trabajo presenta un algoritmo para determinar la orientación instantánea de un objeto en el espacio 3D. La orientación fue determinada utilizando una matriz de cosenos directores (DCM) conformada por la combinación de 3 rotaciones consecutivas alrededor de cada uno de los ejes del sistema evaluado, utilizando cuaterniones. Una unidad inercial de medida (IMU) compuesta por un giroscopio de 3 ejes y un acelerómetro de 3 ejes fue utilizada con el objetivo de establecer 2 sistemas coordenados; Un sistema coordenado describiendo el movimiento del objeto, utilizando al giroscopio como fuente principal de información, estableciendo la relación de cambio con respecto al tiempo. Un sistema coordenado de referencia, relacionando la aceleración gravitacional a un vector inercial. Un control por retroalimentación proporcional integral (PI) fue utilizado con el objetivo de combinar la información de los sensores, eliminando las desviaciones por offset y deriva, mejorando la precisión en la orientación. Dadas sus características, el algoritmo propuesto permite su utilización en la evaluación de la posición y la orientación de los segmentos corporales, siendo de suma importancia en ortopedia, traumatología y reumatología para la determinación de diagnósticos, pronósticos terapéuticos e investigación así como el diseño y fabricación de dispositivos de medición, instrumentación quirúrgica, prótesis y ortesis. Cabe destacar que el sistema desarrollado abre oportunidades de ser implementado en el diseño de sistemas ambulatorios de evaluación de las articulaciones, mediante el uso de elementos transportables dadas las reducidas dimensiones y limitaciones de los sensores empleados.

Palabras clave: orientación instantánea, matriz de cosenos directores, cuaterniones, unidad inercial de medida, desviación de offset y deriva.

Introduction

Joints evaluation is very important in orthopedic, traumatology and rheumatology for assessing diagnosis, prognostic, therapeutic and research as well as the design and fabrication of measuring devices used in surgical instrumentation, prostheses and orthesis [1].

The kinematics of joint evaluation requires the determination of both the position and orientation of the body segments involved; It is normally modelled as rigid bodies which is related to an inertial reference system [2]. This is commonly achieved by tracking the movement of a group of reflective point markers attached to the required body segment of interest, this using optical motion systems [3]. Nonetheless, this system is expensive and cannot also be used outside the laboratory, since it requires a specific and controlled environment.

During the last decade, body mounted sensors such as gyroscopes and/or accelerometers also known as inertial sensors have been used to obtain kinematic values, offering the advantage of identifying human motion in a wide variety of environments due their low cost, small size and low power consumption [4]. However, inertial sensors do not measure the position or orientation of the body segment directly but the physical quantities which are related to the motion of the objects where the sensors are fixed [5]. On one hand, gyroscope measures angular velocity which if initial conditions were known, information may be integrated over time to compute the sensors orientation [6] [7]. In theory, both the position and orientation of an arbitrary object or body segment can be estimated by integrating the angular rate data obtained from a gyroscope sensor attached herewith. Although this data can be translated into meaningful three dimensional information, it was reported that the integration of any amount of signal noise in the angular velocity data will cause an accumulative error in the estimated orientation also known as a drifting deviation [8]. On the other hand, the accelerometer measures the earth’s gravitation and provides an absolute reference orientation allowing it to measure the angle of the sensor unit in respect to gravity [9], [10]. This method is appropriate if magnitude of acceleration is neglected with respect to gravity, and less accurate when accelerations due to motion are affected by measured direction of gravity. Furthermore, accelerometer signals do not contain information about the rotation around the vertical axis and therefore do not give a complete description of the orientation [11].

Hence to compute a single estimate of orientation through the optimal fusion of accelerometer and gyroscope sensors is needed [12]. Recently, many studies of human motion based on inertial sensors have proposed different methods to combine gyroscope and accelerometer sensors information for kinematic joint evaluation. Dejnabadi et.al [4] proposed a method of measuring uniaxial joint angles using a combination of accelerometers and gyroscopes. The model was based on the acceleration of the joint center rotation mathematically by shifting the location of the physical sensors. Joint angles were computed by estimating absolute angles without any integration of sensors information. Luinge et.al [13] described a method that used constraints in the elbow to measure the orientation of the lower arm in respect to the upper arm. Accelerometer and gyroscope orientations were achieved by using rotational matrixes. A calibration method to determine the orientation of each sensor was determinant in the accuracy of the orientation measured. Zhang et.al [14] used a particle filter to fuse gyroscope and accelerometer information, compensating drifting deviation and also improving the estimation accuracy. Additionally, quaternions are used to represent orientations of upper limb segments thereby avoiding singularities. Tadano et.al [8] proposed a method of three dimensional gait analysis using wearable sensors and quaternion calculations. Accelerometer information was used to estimate initial orientation of the sensors. Gyroscope information was used to estimate angular displacements during gait, by integrating angular velocity. Orientations of the sensors were converted to the orientations of the body segment by a rotational matrix obtained from a calibration trial. Favre et.al [15] proposed 2 different methods to fuse gyroscope and accelerometer information; Quaternion-based integration of angular velocity and orientation correction using gravity. Both methods are based on recognizing characteristic samples which provide vertical orientation of the sensor.

This work presents an algorithm to determine both the position and orientation of an arbitrary object in 3D space, oriented to be used in kinematic joint evaluation. The proposed algorithm combines both gyroscope and accelerometer information, thereby establishing two suitable coordinate systems describing the object movement in respect to a coordinate reference system. Both the position and orientation were achieved by estimating the spatial relationship between the coordinate systems through a Direction Cosine Matrix (DCM) computed from the combination of three consecutive rotations, around each to the main axis of the evaluated system using quaternions. A PI controller feedback (PI) is used to apply the amount of rotation that is needed to align the estimated orientation by both gyroscope and accelerometer information, forcing them to converge with each other proportionally, eliminating drifting deviation and improving also the accuracy by decreasing noise response.

METHODOLOGY

A combination of 3 axes gyroscope and 3 axes accelerometer was used in order to estimate both the position and orientation of an arbitrary object in 3D space. Gyroscope information (Gx,Gy,Gz) was used to define a coordinate system describing object movement (Omov) through gyroscope angular rate (ωxyz). Accelerometer information (Ax,Ay,Az) was used to establish an absolute coordinate reference (Oref) based on the acceleration of gravity (λ) in order to compensate drifting deviation given by the integration of gyroscope measurement errors over time.

Sensors and data acquisition

Gyroscope and Accelerometer information were obtained from an IMU (Invensense MPU6050) configured at its lowest feature rates [16], a full scale range of ±250o/s and ±2g with a bandwidth of 256Hz and 260Hz respectively. Full scale range parameters and bandwidth were defined considering the common values reported in kinematic joint evaluation; gyroscope from ±200o/s to ±600o/s, accelerometer from ±2g to ±5g and bandwidth from 2Hz to 200Hz [4], [8], [11], [15], [17], [18]. A 16-bit Digital Signal Controller Microchip (Dspic - 30F6014A) was used to perform the optimal fusion of gyroscope and accelerometer information and to implement an algorithm for assessing the orientation of an object in 3D space. Communication between IMU and Dspic was established through the Inter-Integrated Circuit serial protocol (I2C) at 400-kHz which is the standard operating frequency. The response of the implemented algorithm was assessed through a virtual model designed on V-realm builder 2.0. Communication between Dspic and virtual environment was achieved by using MATLAB R2013a version.

Orientation estimate

Orientation estimate problem involves two coordinate systems; one attached to the arbitrary object describing its movement Omov and the other located at some point to the earth but not attached to it establishing a reference frame Oref. Since the coordinates of a vector depend on the frame it is represented in, an arbitrary vector can be represented in the rotation frame. Consequently these two frames can rotate relatively to each other independently [19]. This rotation is estimated by a 4D complex number also called quaternion (Q) which is conformed by an scalar (q0) and a 3D vector (q) with q = [q1,q2,q3], Eq.(1).
Q  = {q ,q}
       0
(1)

The rotational vector Q is associated with the sine (S) and the cosine (C) of the rotation angle σ = [α,ϕ,θ], over each of the main axis (x,y,z) to the evaluated coordinate system in time, Eq.(2).
    1
Q = 2-[C (σ),(q)S (σ)]
(2)

Quaternions can be composed to express several rotations through a quaternion product Qcomp, Eq.(3). Qcomp can also be used to express the rotation on a 3D space, Eq.(4).
Qcomp = Qa Qb = (qa0,qa) (qb0,qb) (3)
= {qa0qb0 - qaqb,qa × qb + qa0qb + qb0qa}
Therefore Qcomp = Qz(θ) Qy(ϕ) Qx(α)
          ⌊| C (θ)C (ϕ) C (α )+ S (θ)S (ϕ )S (α ) ⌋|
Qcomp =  1|| C (θ)C (ϕ) S(α )- S (θ) S(ϕ) C (α ) ||
         2⌈ C (θ)S (ϕ)C (α )+ S (θ) C (ϕ )S (α ) ⌉
            S (θ)C (ϕ)C (α )- C (θ)S (ϕ )S (α )
(4)

To transform vectors from one coordinate system to another using Eq.(4) a Direction Cosine Matrix (DCM) is performed as shown in Eq.(5), which is used to establish the spatial relationship between coordinate systems Omov and Oref which allows estimating the orientation [20], [21].
DCM   = (q20 - q′q)I + 2qq′ - 2q0q*
(5)

When expanded this yields
⌊  q2+ q2-  q2- q2   2(q1q2 + q0q3)   2 (q1q3 - q0q2)  ⌋
|   02(q 1q - 2q q )3  q2- q2 + q2- q2   2 (q q  + qq )  |
⌈      1 2   0 3     0   1    2   3   2  223   201 2 ⌉
    2(q1q3 + q0q2)   2(q2q3 - q0q1)  q0 - q1 - q2 + q3

Direction cosine matrix improvement

To estimate the orientation of an object in 3D space, DCM is updated by incremental changes when either Omov or Oref rotates. These incremental changes can be established by integrating DCM as shown in Eq.(6).
DCM   =  DCMt  -1 * DCMt
(6)

Because DCM is conformed from Qcomp, Eq.(4) it is noted that by expressing elements of Qcomp as approximations rather than identities, Qcomp can be reduced to Qred, Eq.(7).

                         (
                         |{  sin (α)    ⌊ α ⌋
    cos(σ) ≈ 1     -y-ie-lds→    sin (ϕ) ≈  | ϕ |
sin(σ) *sin (σ) ≈ 0       |(            ⌈   ⌉
                            sin (θ)      θ

       1
Qred = 2 [1,α,ϕ,θ ]
(7)

For assessing DCM improvement in order to estimate orientation, incremental changes were estimated from the integration of Qred over time, improving Qred as shown in Eq.(8).
Qimp = Qredt-1 ⊗ Qredt
(8)

When expanded this yields:

        ⌊    ⌋    ⌊                  ⌋
        | q0 |    |  - αq1 - ϕq2 - θq3 |
Q    =  || q1 ||+  1||  αq0 + θq2 - ϕq3 ||
  imp    ⌈ q2 ⌉   2⌈  ϕq0 - θq1 - αq3 ⌉
          q3         θq0 + ϕq1 - αq2

It should be noted that numerical errors in the approximations will reduce orthogonality between rotational axes of the coordinated systems [22]. Consequently, Qimp is normalized before being substituted in Eq.(5) in order to perform an improved DCM, Eq.(9).

Qimp(norm ) =-Qimp--
            ∥Qimp ∥

DCM   =  DCMt
(9)

Coordinate system that describes the object movement (Omov)

Gyroscope output (ωxyz) is used as the main source of information in order to define the coordinate system that describes object movement. Then Omov has its origin in the tri-axial gyroscope and each point along the gyroscope axes (Gx,Gy,Gz). Therefore, by relating the angular rate of change along time (Δt) to its rotation during the movement of the IMU, a single estimation of gyroscope orientation (σgyro) can be determined, Eq.(10).
σgyro = (ω - ωoff)* Δt
(10)

Where ωoff is the intrinsic offset of the gyroscope sensor, estimated only at the beginning of the test by computing the mean value of the total number of the recorded data in one second in static position.

It is noted that σgyro could be used to estimate both the position and orientation of an arbitrary object or body segment by conforming DCM. However, constructing DCM only from σgyro is not only inefficient but also inconvenient, considering that any amount of noise in the angular velocity data will result in integration errors and consequently, an accumulative gyro drift in DCM. Hence, an additional source of information is needed to assess drifting compensation.

Drifting estimation

Accelerometer information (λxyz), defines the coordinate reference system which is used to assess drifting deviation. Therefore, Oref has its origin at the tri-axial accelerometer and each point along accelerometer axes (Ax,Ay,Az). Assuming that the accelerometer is only measuring the acceleration of gravity, then an appropriate convention would be to assume that the direction of gravity ĝ = [0,0,1] defines the vertical axis z [12], [23]. Thus the inertial directional vector ⃗v can be computed from the normalized accelerometer measurements Eq.(11).
⃗v = -λ--
    ∥λ∥
(11)

Given a DCM estimation from gyroscope measurement, a gravity vector is determined as Eq.(12).
^v = ^g *DCM
(12)

Finally considering that the magnitude of cross product between two vectors is proportional to the sine of the angle between them and its direction can be perpendicular to both of them [24], drifting deviation detected by taking cross product between vector ⃗v and ^v , presenting both an axis and an amount of rotation that are needed to rotate ^v to become parallel to the estimated vector σgyro, Eq.(13).
σaccel = ⃗v × ^v
(13)

Data integration

Slow movements are related to low noise levels [25]. Hence, drifting deviation can be corrected by adding the total amount of rotation σaccel to the estimated orientation σgyro, in order to align the coordinate systems that they represent, Eq.(14).
σ =  σgyro + σaccel
(14)

Nonetheless σaccel is susceptible to high noise level due to variations in velocity during movement. Therefore, a half portion of σaccel, computed by multiplying σaccel by a weighed ωσ, is used as a feedback to a Proportional Integral (PI) controller Eq.(15) before being added to σgyro to obtain σ Eq.(14). Hence, it is possible to apply a proportional the amount of rotation needed to align Omov and Oref, instead of applying the total correction. Consequently, both coordinate systems are forced to converge proportionally thereby cancelling drifting.
                             t
                            ∫
σaccel = Kpf (ωσ * σaccel)+ Ki  f(ωσ * σaccel)Δt
                            0
(15)

Proportional and Integral parameters (Kp&Ki) were established based on The Good Gain method for PI (D) controller tuning reported in studies [26], [27].

RESULTS AND DISSCUSION

Proposed algorithms

Gyroscope and accelerometer information were used in order to establish two suitable coordinate systems describing both object movement Omov and reference coordinate system Oref. The description of an object in 3D space was assessed by estimating the spatial relationship between coordinate systems. Commonly, this relationship is given by both the position and orientation of the object in respect to a reference system. However, assuming that both systems overlapped at the origin, it is said that there is no change of position between them, [24]. Therefore, an arbitrary object in 3D space can be described by its orientation. This orientation is determined through a DCM computed from the combination of three consecutive rotations around each of the main axis of the evaluated system using quaternions. Fig. 1, shows a block diagram representing the proposed algorithms.

In Fig. 1, it can be observed that two different solutions proposed to estimate the amount of rotation σ, from which quaternion Q is computed defines the amount of rotation in 3D space. 1.- Estimation of σ by combining σgyro and σaccel without any reduction of noise (σgyro&σaccel). 2.- Estimation of σ by combining σgyro and σaccel information using a PI feedback controller, (σgyro&σaccel)&PI. It is noted that Q is used to perform DCM from which the orientation of an object in 3D space is defined. It is noteworthy that two different solutions were proposed to estimate DCM; 1.- Estimation of DCM from Qcomp which defines the orientation from DCM product. 2.- Estimation of DCM from Qimp which defines the orientation through the quaternion product.

PIC

Fig. 1. Estimation of Q from σ as (σgyro&σaccel) without any reduction of noise., estimation of Q from σ as (σgyro&σaccel)&PI by using a PI feedback controller., estimation of DCM from Qcomp., estimation of DCM from Qimp.

Estimating differences between proposed algorithms

In order to estimate differences between proposed algorithms, the analysis of behavior of the sensor is carried out by estimating the orientation in both static position and during movement. Hence representing the orientation as Euler angles computed from DCM Eq.(16).
α = atan2(DCM23,DCM33)
ϕ = asin(DCM13) (16)
θ = atan2(DCM12,DCM11)

Behavior analysis of the estimated orientation in static position

For analysis of behavior in static position five minutes of the obtained information were recorded, which allows the comparison of drifting deviation over time. It should be noted that to assess the behavior of the proposed algorithms, the estimation of the orientation using either σgyro or σaccel is included to the analysis. A comparison is made by using sensors information separately and also by fusion sensors information through the proposed algorithms.

In figure 2, it can be observed that estimating orientation from σgyro only seems not to be so noisy but drifts over time. On the other hand, estimating orientation from σaccel does not drift over time but seems to be quite noisy especially when compared to σgyro. When analyzing the estimated orientation either through (σgyro&σaccel) or (σgyro&σaccel)&PI, drifting deviation appears to be cancelled and the noise reduced. Despite the fact there seems to be a considerably reduction of noise between the estimated orientation (σgyro&σaccel)&PI in comparison to (σgyro&σaccel), it is noted that drifting deviation is compensated but not eliminated thereby maintaining the behavior of σgyro in z axis. The main reason is that both algorithms depend on the sensors fusion in order to compensate drifting deviation and the accelerometer information can only be used to estimate the orientation of an object in X and Y but not in Z axis because of its dependence on the acceleration of gravity as was reported in [11].

PIC

Fig. 2. Behavior analysis of the estimated orientation in static position in order to establish drifting deviation comparison either in the estimated orientation from σgyro and σaccel independently, as well as through to the proposed algorithms; (σgyro&σaccel) and (σgyro&σaccel)&PI.

Table.1 Drifting deviation and RMSE on static position
X axis
Y axis
Z axis
Drift 5minRMSEDrift 5minRMSEDrift 5minRMSE
σgyro 22.85o 1.3e-4 45.15o 1.4e-4 19.87o 1.8e-4
σaccel 0.34o 0.18 0.17o 0.21 0.002o 0.007
σgyro&σaccel 0.33o 0.18 0.12o 0.21 9.26o 0.007
(σgyro&σaccel)&PI 0.001o 0.04 0.10o 0.05 9.21o 0.007

Consequently, drifting deviation in this axis cannot be completely eliminated. Finally, in order to verify such assumptions; Table. 1 shows a comparison of the amount of drifting deviation in degrees as well as the amount of noise in static position represented by the Root Mean Square Error (RMSE). Similarly in Fig. 3, it can be observed a comparison in the estimation of the orientation in Z axis, in a range of -90 to 90 degrees using only σgyro or σaccel, hence demonstrating the mentioned dependence of the acceleration of gravity.

PIC

Fig. 3 Comparison in the estimated orientation in Z axis.

PIC

Fig. 4. Behavior analysis of the estimated orientation during movement into a range of motion from -90 to 90 degrees, comparing drifting deviation, between σgyro and the proposed algorithm (σgyro&σaccel)&PI, for each axes.

Behavior analysis of the estimated orientation during movement

Although σaccel is susceptible to high noise levels, it can be used to estimate both the position and orientation of an arbitrary object or body segment in X and Y axis which is enough in some applications [2], [17]. Although σaccel does not produce drifting deviation during the static analysis, it was demonstrated that it cannot be used to estimate neither the position nor the orientation in 3D space because of its dependence on the acceleration of gravity. Despite the fact that σgyro is subject to drifting deviation, some authors suggest that it could be useful in the evaluation of body segments due to its characteristic of not being susceptible to noise [2], [17]. Accordingly, due to the analysis of behavior during movement, six repetitions from -90 to 90 degrees for each axis was performed comparing σgyro with (σgyro&σaccel)&PI, which is the proposed algorithm in order to determine both the amount of drifting deviation and noise during movement. However, as it was assumed during static analysis in Fig. 4 X and Y axis, it can be noted that because of drifting deviation in σgyro, the estimated orientation is lost through time during movement even in a short period of time. On the other hand, it can be noted that the estimated orientation using (σgyro&σaccel)&PI, seems to be as noisy as σgyro but as susceptible in drifting as σaccel, therefore demonstrating an optimal fusion sensor. In addition, although Z axis maintains the behavior of σgyro, it was demonstrated that the fusion sensor not only allows to compensate and to eliminate both drifting deviation and noise as can be seen in table 2 but it also allowed estimating the orientation of an arbitrary object or body segment in 3D space successfully.

By representing the orientation of an object in 3D space, DCM was conformed from σ instead of using only Euler angles because of its singularity-free orientation representation. Accordingly, by considering that an arbitrary object in 3D space can be represented as a vector through coordinate triplet Oxyz. The orientation of an object in 3D space will be given by Eq.(17).
Oxyz = Oxyzt- 1 *DCM
(17)

Table.2 Drifting deviation and RMSE during movement
X axis
Y axis
Z axis
Drift (30s)RMSEDrift (30s)RMSEDrift (30s)RMSE
σgyro 48.82o 24.12o 4.39o 11.36o 15.15o 7.14o
(σgyro&σaccel)&PI 2.59o 2.75o 1.32o 2.98o 15.15o 7.14o

PIC

Fig. 5. Virtual representation to the estimated orientation of an object during movement, in a range of motion from 0 to 90 degrees through X axis, based on the sensors orientation.

Finally, Oxyz will define the orientation of an object in 3D space based on the sensors orientation representing it through a virtual environment as a cube while performing a trajectory between 0 to 90 degrees through X axis as shown in Fig. 5.

Conclusion

To estimate the orientation of an arbitrary object or body segment in 3D space cannot be properly achieved using gyroscope or accelerometer information separately. Although, gyroscope angular data rate can be translated into meaningful 3D information, it was demonstrated that the integration of any amount of signal will cause an accumulative drifting deviation over time. On the other hand, accelerometer information provides an absolute reference orientation which allows estimating the orientation in respect to gravity. Nonetheless, it was demonstrated that the accelerometer is not able to give a complete description of the orientation because the accelerometer signals do not contain information about the rotation around the vertical z axis.

Results from the present study shows that the proposed algorithm including PI controller feedback not only eliminates drifting deviation but also reduces noise in both static position and during movement, thereby estimating the orientation of an arbitrary object in 3D space. It should be noted that the proposed algorithm simplifies equations by reducing the operations to be computed and consequently computation cost as it was in contrast to methods based on the usual computation of quaternions and rotation matrixes.

Finally, from the obtained results, considering that parameters of the proposed algorithm were configured according to the kinematic joint evaluation, it was demonstrated that it is possible to estimate the instantaneous position and orientation of the body segments as an arbitrary object in 3D space by merging gyroscope and accelerometer information.

The results obtained as well as low cost, small size and low power consumption of the employed sensors were encouraging for design a wearable system for application in the ambulatory kinematic joint body evaluation as a future work.

ACKNOWLEDGMENT

The authors express thanks to Consejo Nacional de Ciencia y Tecnología (CONACyT, México), for granted scholarship to Barraza-Madrigal J.A and their support to this research. The authors also thank the valuable assistance of Ing. Eladio Cardiel Perez for achieving this study.

References

  1. T. Shimada, “Normal Range of Motion of Joints in Young Japanese People,” Bulletin of allied medical sciences Kobe BAMS (Kobe), 1988; 4:103-109.
  2. D. Giansanti, V. Macellari, G. Maccioni and A. Cappozzo, "Is it Feasible to Reconstruct Body Segment 3-D Position and Orientation Using Accelerometric Data?" IEEE Transactions on Biomedical Engineering, 2003; 50(4):476-483.
  3. Sen Suddhajit, R. Abboud, D. Ming, B. Wan, Y. Liao and Q. Gong, “A motion simulation and biomechanical analysis of the shoulder joint using a whole human model,” in 4th International Conference on Biomechanical Engineering and Informatics (BMEI), 2011; 4:2322-2326.
  4. H. Dejnabadi, B. M. Jolles and K. Aminian, “A New Approach to Accurate Measurement of Uniaxial Joint Angles Based on a Combination of Accelerometers and Gyroscopes,” IEEE Transactions on Biomedical Engineering, 2005; 52(8):1478-1484.
  5. M. A. Sabatini, “Estimating Three-Dimensional Orientation of Human Body Parts by Inertial/Magnetic Sensing,” Sensors, 2011; 11(2):1489-1525.
  6. J. Bortz, “A New Mathematical Formulation for Strapdown Inertial Navigation,” IEEE Transactions on Aerospace and Electronic Systems, 1971; AES-7(1):61-66.
  7. M. Ignangni, “Optimal Strapdown Attitude Integration Algorithms,” Journal of Guidance Control and Dynamics, 1990; 13(2): 363-369.
  8. S. Tadano, R. Takeda and H. Miyagawa, “Three Dimensional Gait Analysis Using Wearable Acceleration and Gyro Sensors Based on Quaternion Calculations,” Sensors, 2013; 13(7):9321-9343.
  9. E. Bernmark and C. Wiktorin, “A triaxial accelerometer for measuring arm movements,” Applied Ergonomics, 2002; 33(6):541-547.
  10. G. Hansson, P. Asterlan, N. Holmer and S. Skerfving, “Validity and reliability of triaxial accelerometers for inclinometry in posture analysis,” Medical and Biological Engineering Computing, 2001; 39(4):405-413.
  11. H. Luinge and P. Veltink, “Measuring orientation of human body segments using miniature gyroscopes and accelerometers,” Medical and Biological Engineering and Computing, 2005; 43(2):273-282.
  12. S. O. Madgwick, A. J. Harrison and R. Vaidyanathan, “Estimation of IMU and Marg orientation using a gradient descent algorithm,” in IEEE International Conference on Rehabilitation Robotics, Zurich-Swizerland, 2011.
  13. J. Luige Henk and H. Peter, “Inclination Measurement of Human Movement Using a 3-D Accelerometer With Autocalibration,” IEEE Transactions on Neural Systems and Rehabilitation Engineering, 2004; 12(1):112-121.
  14. Z. Zhang, Z. Huang and J. Wu, “Hierarchical Information Fusion for Human Upper Limb Motion Capture,” in 12th International Conference on Information Fusion, Seattle-WA-USA, 2009.
  15. J. Favre, B. Jolles, O. Siegrist and K. Aminian, “Quaternion-based fusion of gyroscopes and accelerometers to improve 3D angle measurement,” Electronics Letters, 2006; 42(11):612-614.
  16. InvenSense, MPU-6000 and MPU-6050 Register Map and Descriptions, Revision 4.0, 03/09/2012: 13-15.
  17. E. L. Morin, S. A. Reid and J. M. Stevenson, “Characterization of upper Body Accelerations for task Performance in Humans,” in Proceedings of the 25th Annual Conference of the IEEE EMBS Cancun-México, 2003.
  18. A. Olivares, J. Górriz, J. Ramirez and G. Olivares, “Accurate human limb angle measurement: sensor fusion through Kalman, least mean squares and recursive least-squares adaptive filtering,” Measurement Science and Technology, 2010; 22(2):1-15.
  19. E. Bekir, “Introduction to Modern Navigation Systems,” World Scientific Co. Pte-Ltd, 2007.
  20. H. Schneider and G. Barker, “Matrices and Linear Algebra,” Holt Rinehart and Winston Inc. New York, 1968.
  21. C. Cullen, Matrices and Linear Transformations, Addison-Wesley Co. Reading M.A, 1972.
  22. W. Premerlani and P. Bizard, Direction Cosine Matrix: Theory, 17 May 2009
  23. S. O.H Madgwick, “An efficient orientation filter for inertial and inertial/magnetic sensors arrays,” Technical report, Department of Mechanical Engineering, University of Bristol, 2010.
  24. B. Antonio, Fundamentos de Robótica, McGraw Hill, 1997.
  25. R. Mahony, T. Hamel and J.-M. Pflimlin, “Nonlinear Complementary Filters on the Special Orthogonal Group,” IEEE Transactions on Automatic Control, 2008; 53(5):1203-1218.
  26. F. Haugen, “The Good Gain method for PI(D) controller tuning,” TechTeach, 2010: 1-7.
  27. F. Haugen, “The good Gain method for simple experimental tuning of PI controllers,” Modeling, Identification and Control, 2012; 33(4):141-152.