Final word

Now the project is over and everyone comes back home.
For the final word we wish to thank some particular people :

  • Rudi Giot for his support, his ideas and his help during all the week;
  • Roald Baudoux for his work with Kyma without whom we would not have obtained a sufficient sound result;
  • Gregory Aerts for his work done for catching and analysing the internet radio streaming;
  • Lionel Lawson for his help about OpenInterface Workbench and for the time he has voluntarily granted to us;
  • All the organisation from ISIB and IP for the great job accomplished during all week;
  • And all the people that we would have forgotten without whom our project would not have resulted in a so short time.

Thank you all !

Using the Emotiv Epoc with the Emotiv API

After using OpenVibe to get the signals, we spent time on the Emotiv API.
It proposes directly different functions for three types of behavior:

  1. The emotions;
  2. The cognition;
  3. The expressions.

This last one, is the one wich allows to get the most results in a few time. So, we decided to use it (we don’t have time to explore all the possibility the Epoc offers).
It is also easy to use and, in a short time, we managed to implement it in OpenInterface.

Here is a list of the different expressions that we can recognize (this works for most tests) :

  • The blink of eyes
  • The eyebrow movement
  • The left wink
  • The right Wink
  • The look on the right
  • The look on the left
  • Clench teeth
  • The smile

OpenInterface

The second step of the IP will be to use the data (received from the sensors) into an application (like a multimedia application).
There is most chance that the SDK used to get the data from the sensors and the SDK used to develop the application (or any elements in the chain) doesn’t use the same langage !
For example, we may need to use a c++ SDK to get the data from a sensor, do a signal processing with a java library and use the resulting data in a musical software who uses OSC messages.
All the members of the team will be in charge of one of these processes and it could be difficult to combine all the processes in the end.
It is the reason why we could be able to use OpenInterface.

What is OpenInterface ?

Continue reading

Using OpenVibe with Emotiv Epoc

As said before, the OpenVibe option is more flexible. So we prefer using it in a first time.
We will not explain how OpenVibe works and all its features, all this information are available on the OpenVibe’ website.
We will just explain how we can interface the Emotiv Epoc, getting back the signals and sending them into an other application using a VRPN server.

1) Interfacing the Emotiv Epoc

Continue reading

Interfacing the Emotiv Epoc (OpenVibe + Emotiv SDK)

There are at least 2 ways to interface the Emotiv Epoc:

1) Official Emotiv SDK
Designed especially for the headset, the SDK included many tools to use it.
The Emotiv tools focus on the detection of emotions.
We have to calibrate the headset with the panel control and then, we can get back the information in a c++ application using Emotiv library (all information about using the library are available in the documentation of the SDK).

At the moment, we don’t know yet if it’s possible to get the raw signals with this SDK .

2) OpenVibe
OpenVibe is an OpenSource software designed for the Brain-Computer Interface (BCI).
It can be used to acquire, filter, process, classify and visualize brain signals in real time.
Learn more.

It is possible to interface directly the Emotiv Epoc and getting the raw signals (but it’s necessary to have the Emotiv SDK installed).
Unlike the Emotiv SDK, It’s not directly possible to determine the emotions. However, processing options are more flexible (we can get the raw signals and do all processing we want).