Jump to: navigation, search
Line 33: Line 33:
 
* [[Special_Sessions#ss7| SS7 Multi-Robot Systems and Mobile Sensor Networks]]  
 
* [[Special_Sessions#ss7| SS7 Multi-Robot Systems and Mobile Sensor Networks]]  
 
* [[Special_Sessions#ss8| SS8 Multisensor Fusion Methods for Radiation Source Localization]]  
 
* [[Special_Sessions#ss8| SS8 Multisensor Fusion Methods for Radiation Source Localization]]  
 +
* [[Special_Sessions#ss9| SS9 Multiple (Extended) Object Tracking]]
 
</div>
 
</div>
 
|}
 
|}
Line 197: Line 198:
  
 
'''Organizers:''' [mailto:john_mattingly@ncsu.edu John Mattingly]
 
'''Organizers:''' [mailto:john_mattingly@ncsu.edu John Mattingly]
 +
 +
|-
 +
|}
 +
| style="border:1px solid transparent;" |<br />
 +
|-
 +
 +
<!--        MFI 2016 Accepted Special Sessions        -->
 +
{| id="mp-upper" style="width: 100%; margin:4px 0 0 0; background:none; border-spacing: 0px;"
 +
<div id="ss9"></div>
 +
<!--        SS9 Multiple (Extended) Object Tracking      -->
 +
| class="MainPageBG" style="width:100%; border:1px solid #bdd6c6; background:#e7f7e7; vertical-align:top; color:#000;" |
 +
{| id="mp-left" style="width:100%; vertical-align:top; background:#e7f7e76;"
 +
| style="padding:2px;" | <h2 id="mp-tfa-h2" style="margin:3px; background:#d6efd6; font-family:inherit; font-size:120%; font-weight:bold; border:1px solid #bdd6c6; text-align:left; color:#000; padding:0.2em 0.4em;">SS9 Multiple (Extended) Object Tracking</h2>
 +
|-
 +
| style="color:#000;" | <div id="mp-tfa" style="padding:2px 5px">
 +
'''Description:''' Autonomous driver safety functions are standard in many modern cars, and semi-automated systems (e.g., traffic jam assist) are becoming more and more common. Construction of a driverless vehicle requires solutions to many different problems, among them multiple object tracking. The multiple object tracking problem is defined as keeping track of an unknown naumber of moving objects, and historically it has been focused on so called point objects which give at most one detection per time step. However, modern sensors have increasingly higher resolution, meaning that it is common to see multiple detections per object. For example, this is the case when automotive radar or lidar sensors are used. In order to be able to use point object algorithms for these sensors, heuristic clustering algorithms are applied to the raw measurements to obtain object hypotheses. In challenging scenarios, the hard decisions of the clustering algorithms affect the performance of the tracking algorithm due to the associated loss of information.
 +
Consequently, so called extended object tracking algorithms which are capable of handling several measurements per object are required. This special session addresses recent results in the area of multiple object tracking for both point objects and especially extended objects.
 +
 +
 +
'''Organizers:''' [mailto:karl.granstrom@chalmers.se Karl Granström], [mailto:stephan.reuter@uni-ulm.de Stephan Reuter], [mailto:marcus.baum@cs.uni-goettingen.de Marcus Baum]
  
 
|-
 
|-

Revision as of 10:58, 23 May 2016

Download Call For Special Sessions

Click here to download the call for special sessions.


List of Special Sessions of MFI 2016


SS1 Multi-Sensor Data Fusion for Autonomous Vehicles

Description: Automotive transportation is currently evolving at an unprecedented rate. Future vehicles operating in a highly assisted and autonomous mode will need the ability to function on any road possible, in a safe, legal and socially acceptable manner and without access to high-precision, precompiled maps. Therefore, at the heart of any autonomous functionality will be the ability of a vehicle to sense its environment, produce a map of static objects in the environment and to track both itself and other dynamic targets within that environment. For autonomous vehicles to be commercially viable it must achieve this high level of situational awareness using only commercially viable sensors. Multi-sensor data fusion offers the ability to greatly reduce the uncertainty of state estimates and estimate physical states which might otherwise be unobservable.

In addition to the challenge of fusing sensors organic to the vehicle, many future infrastructure projects envisage the concept where vehicles will be connected with each other and with sensors in the infrastructure to further improve situational awareness. This presents many challenges where there may be poorly understood correlation, or poor trust in externally shared information, along with constraints in terms of bandwidth and computational capacity. Furthermore, there exist both challenges and opportunities for understanding and characterising the interaction between the driver and the vehicle.

Organizers: Daniel Clarke, Michael Fiegert, Feihu Zhang, Dhiraj Gulati, Benjamin Noack, Florian Faion


SS2 Kalman Filters in Nonlinear State Estimation

Description: Nonlinear state estimation is an important component in navigation, robotics, object tracking, and many other current research fields. Besides popular but computational expensive particle filters, variants of nonlinear Kalman filters or LMMSE estimators are widely used methods for state estimation. Such filters include for example the Unscented Kalman Filter, the Divided Difference Filter, or iterated Kalman filters. This session aims to cover the recent advances in the area of nonlinear Kalman filters with an emphasis on sampling and sigma-point set design, linearization techniques, and applications of Kalman filters in nonlinear state estimation scenarios.

Organizers: Jannik Steinbring, Uwe D. Hanebeck


SS3 Data Fusion Methods for Indoor Localization of People and Objects

Description: Indoor positioning has gained great importance as technology allows for affordable realtime sensing and processing systems. Researchers and developers can take advantage of the pervasiveness of WSNs (e.g., in the form of WLAN) and mobile sensors (such as smartphones) to obtain more accurate results by exploiting already existing infrastructure. Applications for indoor positioning include pedestrian navigation in public buildings and shops, location based services, safety for the elderly and impaired, museum guides, surveillance tasks, and also tracking products in manufacturing, warehousing, etc. Unlike outdoor environments, which are covered by GNSS to a satisfying extent, indoor navigation faces additional challenges depending on the underlying measurement system such as occlusions, reflections and attenuation. While there are a great variety of sensors and measuring principles, in practice every single measuring technique suffers from deficits. While RF and (ultra-)sound are subject to multipath propagation, optical systems are intolerant to NLOS conditions. Some systems require setting up beacons, while others are self-calibrating and easy-to-install. Data fusion can overcome these limitations by combining complementary and redundant sensing techniques, with the application of algorithmic methods such as stochastic filtering.

Organizers: Antonio Zea, Florian Faion, Uwe D. Hanebeck


SS4 Multimodal Image Processing and Fusion

Description: Since the launch of the first version of the Microsoft Kinect in 2010, setting up networks based on multimodal image sensors has become extremely popular. The novelty of these devices includes the availability of not only color information, but also infrared and depth information of a scene, at a price affordable to laymen. The combination of multiple sensors and image modalities has many advantages, such as simultaneous coverage of large environments, increased resolution, redundancy, multimodal scene information, and robustness against occlusion. However, in order to exploit these benefits, multiple challenges also need to be addressed: synchronization, calibration, registration, multi-sensor fusion, large amounts of data, and last but not least, sensor-specific stochastic and set-valued uncertainties. This Special Session addresses fundamental techniques, recent developments and future research directions in the field of multimodal image processing and fusion.

Organizers: Antonio Zea, Florian Faion, Uwe D. Hanebeck


SS5 Homotopy Methods in State Estimation

Description: In Bayesian state estimation, the inherent uncertainties of the underlying system and measurements are represented as probability density functions. Given new measurements, the predicted system state is updated to incorporate the new information. Traditionally, this information is introduced directly, which can lead to problems in recursive applications. In the case of discrete particle representations of the densities for example, the problem of particle degeneration is well known and has to be corrected for. Another approach is to gradually incorporate the new information through homotopy methods, allowing for a smooth transition of the underlying densities.

This special session is concerned with theoretical and practical aspects of homotopy methods in the context of state estimation and all works pertaining to fundamental techniques, recent developments and future research directions in this field are invited.

Organizers: Martin Pander, Uwe D. Hanebeck


SS6 Data Fusion in Sensor-based Sorting

Description: Sensor-based sorting is an established technology for sorting of various products according to quality aspects. Fields of application include food processing, recycling, and industrial mineral processing. Selection of a sensor suitable for material characterization typically depends on the product under inspection as well as the sorting task itself. However, in many cases increased sorting performance is achieved by combining information retrieved from multiple different sensors, for example line-scan and area-scan cameras, near-infrared cameras, X-ray, 3D sensors, or hyperspectral cameras. Hence, sensor data needs to be fused to increase performance by putting information in a temporal and / or spatial context. This applies for systems including several sensors of the same kind as well as heterogeneous combinations. Additionally, sensor-based sorting systems are typically restricted in terms of the time being available to derive a sorting decision. Therefore, real-time capable information fusion methods are required.

Organizers: Georg Maier, Robin Gruna


SS7 Multi-Robot Systems and Mobile Sensor Networks

Description: The objective of the special session is to provide an international forum for the discussion of recent developments and advances in the field of nmlti-robot systems and mobile sensor networks. In-depth discussions of relevant theories and applications related to multi-robot systems and mobile sensor networks are expected, including the presentatiol!l of results of applications to real-world land, sea, underwater, aerial and space multi-vehicle sys1tems, as well as strong theoretical contributions. Additionally, the special session welcomes papers that explore new ways of using visual sensors to solve problems in robotics.

Organizers: Joachim Horn, Hung M.La, Marcus Grooemeyer, Anh Duc Dang


SS8 Multisensor Fusion Methods for Radiation Source Localization

Description: Locating a dangerous source of penetrating (i.e., gamma and neutron) radiation in an urban environment is a critical mission for nuclear counterterrorism and emergency response. Traditional methods for localizing a radiation source have primarily relied on individual, non-networked radiation sensors whose responses are used by loosely coordinated operators to collaboratively locate the source. Recently, significant progress has been made in developing rigorous methods for simultaneously analyzing the response of a network of radiation sensors. This session will present recent work on the analysis of radiation sensor networks to optimize the resources and time required to locate a dangerous radiation source in an urban environment.

Organizers: John Mattingly


SS9 Multiple (Extended) Object Tracking

Description: Autonomous driver safety functions are standard in many modern cars, and semi-automated systems (e.g., traffic jam assist) are becoming more and more common. Construction of a driverless vehicle requires solutions to many different problems, among them multiple object tracking. The multiple object tracking problem is defined as keeping track of an unknown naumber of moving objects, and historically it has been focused on so called point objects which give at most one detection per time step. However, modern sensors have increasingly higher resolution, meaning that it is common to see multiple detections per object. For example, this is the case when automotive radar or lidar sensors are used. In order to be able to use point object algorithms for these sensors, heuristic clustering algorithms are applied to the raw measurements to obtain object hypotheses. In challenging scenarios, the hard decisions of the clustering algorithms affect the performance of the tracking algorithm due to the associated loss of information. Consequently, so called extended object tracking algorithms which are capable of handling several measurements per object are required. This special session addresses recent results in the area of multiple object tracking for both point objects and especially extended objects.


Organizers: Karl Granström, Stephan Reuter, Marcus Baum


Personal tools