From Aicip
Revision as of 14:11, 26 January 2010 by Hqi (talk | contribs) (New page: CALL FOR PAPERS ACM TOMCCAP Special Issue on *Multimedia* *Sensor* *Fusion* ACM Transactions on Multimedia Computing, Communications and Applications IMPORTANT ...)

(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search
    Special Issue on *Multimedia* *Sensor* *Fusion*
    ACM Transactions on Multimedia Computing, Communications
    and Applications


Submission deadline:

   December 14, 2009

First Notification:

   March 8, 2010:

Final notification:

   April 19, 2010

Camera ready version:

   May 10, 2010

Publication of the special issue:

   Fall 2010


Today's devices like cell-phones, laptops, sensor nodes or even cars
can capture many quantities like temperature, tilt and yaw, audio,
video, GPS coordinates or the proximity of WLAN base stations and
Bluetooth devices. They are true multimedia devices which go beyond
mere audio and video capturing. The joint consideration of all of
these modalities comprises new and exciting possibilities for
multimedia research.


The joint analysis of measurements is often referred to as sensor
fusion. The strength of analyzing data from different domains and/or
multiple sensors lies in its ability
- to find out facts which are not easy to see by analyzing a single
  domain only, and
- to cope with fuzzy information and noise.


Taking different domains into account opens up the possibility of
finding hidden information. For example, a small silhouette of a
person may not reveal whether the back or the front is facing the
camera. However, the additional usage of a passive IR movement sensor
allows the direction of movement to be derived reliably. Another
example is the usage of multiple cameras, jointly focusing on an
object. While the view of each single camera might be blurred by
bushes, fog or other obstacles, an image reconstructed from multiple
cameras with slightly differing locations can result in a clear
picture of the obfuscated object. These are initial examples where
information that can not be easily found from a single source of
information is found by taking additional measurements into
account. With a vast variety of measurements from ubiquitous devices,
there is a yet unexploited potential for deriving high-level
semantics from existing sensor readings.


In this special issue, surveys and original work on the joint
analysis of multiple sensors should be addressed. These may be (but
are not limited to)
- movement, acceleration, vibration or the tilt of a device
- data from force feedback devices, pulse clocks, etc.
- temperature, humidity and brightness
- the presence of other wireless mobile devices or base stations
- information from ZigBee appliances,
- images from multiple or self-propelled cameras, or
- images from camera arrays and infra-red cameras.

GUEST EDITORS Thomas Haenselmann <>, Wolfgang Effelsberg University of Mannheim, Germany

Shao-Yi Chien (簡韶逸) National Taiwan University, Taiwan

Peter Bajcsy University of Illinois at Urbana-Champaign, USA

For your submission via the ACM Manuscript Central, please follow the guidelines for preparation available from Clearly indicate in the cover-letter field that you submit for the special issue 2010.