Project overview

SILENSE stands for (ultra)Sound Interfaces and Low Energy iNtegrated SEnsors.

Detailed information about partners is on separate page.


SILENSE researches acoustic technologies and develops concepts to activate and control devices by gesture, data communication, and indoor positioning, exclusively based on these innovative technologies. These concepts can be used in different domains: wearables, automotive and smart home application.

Relevance and Impact

Humans communicate mainly with speech and body gestures. The technologies SILENSE is to develop, take the way we naturally communicate one step further: from communication between humans, to communication between humans and objects. This will become increasingly important in a IoT dominated world. The application of the developed technologies in numerous areas will bring lots of benefits to the general public, ranging from enhanced user experience to improved health and safety. A few specific examples:

  • Intuitive user interface in mobile and wearable devices: user interfaces are nowadays one of the major differentiators in the mobile market. Audio sensing allows the activation and control of mobile devices without touching them. Additionally, the user will no longer suffer from a dirty (touch) screen which leads to an improved user experience and better hygiene.
  • Improved hygiene by touchless control: implementation of technologies enabling touchless control of objects can lead to improved hygiene conditions in buildings, hospitals and home.
  • Enhanced safety by touchless control: sound/voice activation/control of systems in the car (e.g. navigation, entertainment and climate control) and control of machinery in industrial applications.
  • Enhanced security by gestural authentication: in the context of needs for stronger authentication process, gestural identification appears as a new factor, increasing the diversity, hence, the robustness of authentication scenarios.
  • Enhances quality of life for disabled persons or elderly unable to move could also take advantage of both audio and gesture recognition, and the ones suffering from hearing/talking impairment could benefit from sign language interpretation.

Time schedule

Start: 1. 5. 2017
Duration: 36 month