This paper presents ongoing work on the design, deployment and evaluation of a multimodal data acquisition architecture
which utilises minimally invasive motion, head, eye and gaze tracking alongside high-quality audiovisual recording of
human interactions. The different data streams are centrally collected and visualised at a single point and in real time by
means of integration in a virtual reality (VR) environment. The overall aim of this endeavour is the implementation of a
multimodal data acquisition facility for the purpose of studying non-verbal phenomena such as feedback gestures,
hand and pointing gestures and multi-modal alignment. In the first part of this work that is described here, a series of tests
were performed in order to evaluate the feasibility of tracking feedback head gestures using the proposed architecture.