Fabio Varesano
2011-01-25 16:15:35 UTC
Hi guys,
I finally have a sensor fusion algorithm which gets the values from an
accelerometer, a magnetometer and a gyroscope and computes correctly
the orientation of the sensors in respect to the world.
See
http://www.varesano.net/blog/fabio/initial-implementation-9-domdof-marg-imu-orientation-filter-adxl345-itg3200-and-hmc5843-a
I'm now able to get the values computed by the sensor array on Arduino
into Soya, using threads and pyserial.
Now, the sensor fusion algorithm gives me a quaternion representing
the orientation of the sensor array in respect to the world (eg: so
that the yaw angle computed is the angle between the X axis of the
sensor and the Hearth north and so on..) ..
Now, I'm stuck on how to convert this quaternion to something which I
could use in Soya. I also do have methods to correctly calculate yaw,
pitch and roll (respect the real world) but I've not been successful
using them in Soya.
The idea is to have a body in Soya which should move as the sensors
moves (just like in the processing application linked above) so I
guess I probably have to align the world frame to the Soya scene, an
hint on this would be appreciated.
Thank you,
Fabio Varesano
I finally have a sensor fusion algorithm which gets the values from an
accelerometer, a magnetometer and a gyroscope and computes correctly
the orientation of the sensors in respect to the world.
See
http://www.varesano.net/blog/fabio/initial-implementation-9-domdof-marg-imu-orientation-filter-adxl345-itg3200-and-hmc5843-a
I'm now able to get the values computed by the sensor array on Arduino
into Soya, using threads and pyserial.
Now, the sensor fusion algorithm gives me a quaternion representing
the orientation of the sensor array in respect to the world (eg: so
that the yaw angle computed is the angle between the X axis of the
sensor and the Hearth north and so on..) ..
Now, I'm stuck on how to convert this quaternion to something which I
could use in Soya. I also do have methods to correctly calculate yaw,
pitch and roll (respect the real world) but I've not been successful
using them in Soya.
The idea is to have a body in Soya which should move as the sensors
moves (just like in the processing application linked above) so I
guess I probably have to align the world frame to the Soya scene, an
hint on this would be appreciated.
Thank you,
Fabio Varesano