Recent IVILAB news

  • February 21, 2019
    Congratulations to Jianbo Shao who successfully defended his dissertation on "Computational methods for improving fiber bundle imaging systems".
  • August 2018
    The IVILAB welcomes four incoming PhD students, Marina Kiseleva, Simon Swenson, Manujinda Wathugala, and Aryan Zarei.
  • July 03, 2018
    Congratulations to Ernesto Brau and Jinyan Guan, Tanya Jeffries, and Kobus Barnard for getting their paper "Multiple-gaze geometry: Inferring novel 3D locations from gazes observed in monocular video," accepted to ECCV 2018.
  • July 11, 2016
    Congratulations to Kyle Simek for getting his paper "Branching Gaussian Processes with Applications to Spatiotemporal Reconstruction of 3D Trees," accepted to ECCV 2016. His coauthors are Ravishankar Palanivelu and Kobus Barnard.
  • April 22, 2016
    IVILA lead, Kobus Barnard, has finally finished his book “Computational methods for integrating vision and language,” (Synthesis Lectures on Computer Vision, April 2016, 227 pages, Morgan & Claypool)
  • See all news

Welcome to IVILAB

IVILAB is led by Kobus Barnard. If you cannot find what you are looking for, which is likely while we are working on this new web presence, try his homepage.

To make sense of the world from data we need to connect it to relevant meaning systems. The IVILAB addresses this directly by working on representations that respect semantics and theory, and linking them to data. We apply this methodology to wide range of fascinating problems. Examples from current work include:
1) learning the structure of common objects;
2) stochastic geometric models for plants and microscopic fungus;
3) representations of neuron form;
4) indoor scene understanding;
5) tracking humans and understanding their activities; and
6) emotion dynamics in close relationships.

For each of these disparate problems, covering entities, environments, and processes, our representations enable making connections to broader endeavors. For example, we seek to extract geometric form of plants in terms of organs such as stems, leaves, and flowers, as quantifying their form and their relationships (e.g., distributions over branching tendency and angles of branches) can be linked to other quantities associates with environmental, molecular, and survival data.

Our models attempt to explain data variability through mechanistic and theoretical considerations. For example, a human tracker based on 2D linkage of observations can be confused by people walking behind others, whereas a 3D tracker expects such occlusion and, in fact, can make use of that information. However, capturing the remaining unexplained variability leads us to statistical characterizations of the observations based on our explanatory models. We use Bayesian statistical methodology to combine all sources of information that link representation to data.

Our approach focuses on models with particular attributes, and we explicitly separate modeling concerns from inference (fitting models to interpret data or learn model parameters). This leads to challenging inference which we handle using various forms of MCMC sampling. For this we have developed significant expertise and software infrastructure over the last decade. Freeing ourselves from inference concerns while modeling enables us to more effectively collaborate with others on model development. In particular, we can work together on translating theoretical ideas into mathematical models, without being influenced by which modelling/inference combination are available in existing software.