no code implementations • 30 Aug 2023 • Jakob Suchan, Jan-Patrick Osterloh
Furthermore we present the ASP approach for interpretation and projection of the driver's situation awareness and its integration within the overall system in the context of a real-world use-case in simulated as well as in real driving.
no code implementations • 28 Dec 2020 • Jakob Suchan, Mehul Bhatt, Srikrishna Varadarajan
The method integrates state of the art in visual computing, and is developed as a modular framework that is generally usable within hybrid architectures for realtime perception and control.
no code implementations • 29 May 2020 • Vasiliki Kondyli, Mehul Bhatt, Jakob Suchan
We develop a human-centred, cognitive model of visuospatial complexity in everyday, naturalistic driving conditions.
no code implementations • 31 May 2019 • Jakob Suchan, Mehul Bhatt, Srikrishna Varadarajan
We demonstrate the need and potential of systematically integrated vision and semantics} solutions for visual sensemaking (in the backdrop of autonomous driving).
no code implementations • 31 May 2018 • Jakob Suchan, Mehul Bhatt, Srikrishna Vardarajan, Seyed Ali Amirshahi, Stella Yu
Key features include a human-centred representation, and a declarative, explainable interpretation model supporting deep semantic question-answering founded on an integration of methods in knowledge representation and deep learning based computer vision.
no code implementations • 17 May 2018 • Carl Schultz, Mehul Bhatt, Jakob Suchan, Przemysław Wałęga
We present ASP Modulo `Space-Time', a declarative representational and computational framework to perform commonsense reasoning about regions with both spatial and temporal components.
no code implementations • 3 Dec 2017 • Jakob Suchan, Mehul Bhatt, Przemysław Wałęga, Carl Schultz
We present the formal framework, its general implementation as a (declarative) method in answer set programming, and an example application and evaluation based on two diverse video datasets: the MOTChallenge benchmark developed by the vision community, and a recently developed Movie Dataset.
no code implementations • 10 Oct 2017 • Jakob Suchan, Mehul Bhatt
We propose a deep semantic characterization of space and motion categorically from the viewpoint of grounding embodied human-object interactions.
no code implementations • 15 Sep 2017 • Jakob Suchan, Mehul Bhatt
We present a commonsense, qualitative model for the semantic grounding of embodied visuo-spatial and locomotive interactions.
no code implementations • 9 Aug 2016 • Jakob Suchan, Mehul Bhatt, Carl Schultz
We present an inductive spatio-temporal learning framework rooted in inductive logic programming.
no code implementations • 26 Jul 2016 • Michael Spranger, Jakob Suchan, Mehul Bhatt, Manfred Eppe
This paper presents a computational model of the processing of dynamic spatial relations occurring in an embodied robotic interaction setup.
no code implementations • 20 Jul 2016 • Michael Spranger, Jakob Suchan, Mehul Bhatt
We present a system for generating and understanding of dynamic and static spatial relations in robotic interaction setups.
no code implementations • 13 Aug 2015 • Jakob Suchan, Mehul Bhatt, Harshita Jhavar
We present a general theory and corresponding declarative model for the embodied grounding and natural language based analytical summarisation of dynamic visuo-spatial imagery.
no code implementations • 22 Jun 2013 • Mehul Bhatt, Jakob Suchan, Carl Schultz
We position a narrative-centred computational model for high-level knowledge representation and reasoning in the context of a range of assistive technologies concerned with "visuo-spatial perception and cognition" tasks.
no code implementations • 5 Jun 2013 • Mehul Bhatt, Jakob Suchan, Christian Freksa
ROTUNDE, an instance of smart meeting cinematography concept, aims to: (a) develop functionality-driven benchmarks with respect to the interpretation and control capabilities of human-cinematographers, real-time video editors, surveillance personnel, and typical human performance in everyday situations; (b) Develop general tools for the commonsense cognitive interpretation of dynamic scenes from the viewpoint of visuo-spatial cognition centred perceptual narrativisation.