Deeply Semantic Inductive Spatio-Temporal Learning

14 downloads 0 Views 10MB Size Report
Aug 9, 2016 - ative Spatial Reasoning; Inductive Logic Programming; AI and Art ... machine interaction and control for commonsense cognitive robotics.
Deeply Semantic Inductive Spatio-Temporal Learning Jakob Suchan1 , Mehul Bhatt1 , and Carl Schultz2 Human-Centred Cognitive Assistance1 hcc.uni-bremen.de and The DesignSpace Group.,1,2 www.design-space.org 2 University of Bremen1 , and University of Munster , GERMANY ¨

arXiv:1608.02693v1 [cs.AI] 9 Aug 2016

Spatial Reasoning., www.spatial-reasoning.com

Abstract. We present an inductive spatio-temporal learning framework rooted in inductive logic programming. With an emphasis on visuo-spatial language, logic, and cognition, the framework supports learning with relational spatio-temporal features identifiable in a range of domains involving the processing and interpretation of dynamic visuo-spatial imagery. We present a prototypical system, and an example application in the domain of computing for visual arts and computational cognitive science. Keywords: Spatio-Temporal Learning; Dynamic Visuo-Spatial Imagery; Declarative Spatial Reasoning; Inductive Logic Programming; AI and Art

1 INTRODUCTION Cognitive assistive technologies and computer-human interaction systems involving an interplay of space, dynamics, and cognition necessitate capabilities for explainable reasoning, learning, and control about space, actions, change, and interaction [1]. Prime application scenarios, for instance, include (A1–A5): (A1). activity grounding from video and point-clouds; (A2). modelling and analysis of environmental processes at the geospatial scale; (A3). medical computing scenarios replete with visuo-spatial imagery; (A4). visuo-locomotive human behavioural data concerning aspects such as mobility or navigation, eye-tracking based visual perception research; (A5). embodied humanmachine interaction and control for commonsense cognitive robotics. A crucial requirement in relevant application contexts (such as A1–A5) pertains to the semantic interpretation of multi-modal human behavioural or socio-environmental data, with objectives ranging from knowledge acquisition (e.g., medical computing, computer-aided learning) and data analyses (e.g., activity intepretation) to hypothesis formation in experimental settings (e.g., empirical visual perception studies). The focus of our research is the processing and interpretation of dynamic visuo-spatial imagery with a particular emphasis on the ability to learn commonsense knowledge that is semantically founded in spatial, temporal, and spatio-temporal relations and patterns. DEEP VISUO-SPATIAL SEMANTICS The high-level semantic interpretation and qualitative analysis of dynamic visuo-spatial imagery requires the representational

2

J. Suchan, M. Bhatt, C. Schultz

and inferential mediation of commonsense abstractions of space, time, action, change, interaction and their mutual interplay thereof. In this backdrop, deep visuo-spatial semantics denotes the existence of declaratively grounded models —e.g., pertaining to space, time, space-time, motion, actions & events, spatio-linguistic conceptual knowledge— and systematic formalisation supporting capabilities such as: (a). mixed quantitative qualitative spatial inference and question answering (e.g., about consistency, qualification and quantification of relational knowledge); (b). non-monotonic spatial reasoning (e.g., for abductive explanation); (c). relational learning of spatio-temporally grounded concepts; (d). integrated inductive-abductive spatio-temporal inference; (e). probabilistic spatio-temporal inference; (f). embodied grounding and simulation from the viewpoint of cognitive linguistics (e.g., for knowledge acquisition and inference based on natural language). Recent perspectives on deep visuo-spatial semantics encompass methods for declarative (spatial) representation and reasoning —e.g., about space and motion— within frameworks such as constraint logic programming (rule-based spatio-temporal inference [4, 24]), answer-set programming (for non-monotonic spatial reasoning [27]), description logics (for spatio-terminological reasoning [3]), inductive logic programming (for inductive-abductive spatio-temporal learning [5, 6]) and other specialised forms of commonsense reasoning based on expressive action description languages for modelling space, events, action, and change [1, 2]. In general, deep visuo-spatial semantics driven by declarative spatial representation and reasoning pertaining to dynamic visuo-spatial imagery is relevant and applicable in a variety of cognitive interaction systems and assistive technologies at the interface of (spatial) language, (spatial) logic, and (visuo-spatial) cognition. INDUCTIVE SPATIO-TEMPORAL LEARNING (WITH DEEP SEMANTICS) This research is motivated by the need to have a systematic inductive logic programming [15] founded spatio-temporal learning framework and corresponding system that: – provides an expressive spatio-linguistically motivated ontology to predicate primi-

tive and complex (domain-independent) relational spatio-temporal features identifiable in a broad range of application domains (e.g., A1–A5) involving the processing and interpretation of dynamic visuo-spatial imagery.

– supports spatio-temporal relations natively such that the semantics of these rela-

tions is directly built into the underlying ILP-based learning framework.

– supports seamless mixing of, and transition between, quantitative and qualitative

spatial data.

We particularly emphasise and ensure compatibility with the general setup of (constraint) logic programming framework such that diverse knowledge sources and reasoning mechanisms outside of inductive learning may be directly interfaced, and reasoning / learning capabilities be combined within large-scale integrated systems for cognitive computing.

Deeply Semantic Spatio-Temporal Learning

3

2 LEARNING FROM RELATIONAL SPATIO-TEMPORAL STRUCTURE: A GENERAL FRAMEWORK AND SYSTEM We present a general framework and working prototype for an inductive spatio-temporal learning system with an elaborate ontology supporting a range of space-time features; we demonstrate the functional capabilities from the viewpoint of AI-based computing for the arts & social sciences, and computational cognitive science. 2.1

THE SPATIO-TEMPORAL DOMAIN OSP , AND QS

The spatio-temporal ontology Osp ≡def is characterised by the basic spatial entities (E) that can be used as abstract representations of domain-objects and the relational spatio-temporal structure (R) that characterises the qualitative spatio-temporal relationships amongst the supported entities in (E). The following primitive spatial entities are sufficient to characterise the learning mechanism and its sample application for this paper: a point is a pair of reals x, y; a vector is a pair of reals vx , vy ; an oriented point consists of a point p and a vector v; a line segment is a pair of end points p1 , p2 (p1 6= p2 ); a rectangle is a point p representing the bottom left corner, a direction vector v defining the orientation of the base of the rectangle, and a real width and height w, h (0 < w, 0 < h); an axis-aligned rectangle is a rectangle with fixed direction vector v = (1, 0); a circle is a centre point p and a real radius r (0 < r); a simple polygon is defined by a list of n vertices (points) p1 , . . . , pn (spatially ordered counter-clockwise) such that the boundary is non-self-intersecting, i.e., there does not exist a polygon boundary edge between vertices pi , pi+1 that intersects some other edge pj , pj+1 for all 1 ≤ i < j < n and i + 1 < j.

Spatio-temporal relationships (R) between the basic entities in E may be characterised with respect to arbitrary spatial and spatio-temporal domains such as mereotopology, orientation, distance, size, motion; Table 1 lists the relevant supported relations from the viewpoint of established spatial abstraction calculi such as the Region Connection Calculus [16], Rectangle Algebra and Block Algebra [7], LR Calculus [20], OrientedPoint Relation Algebra (OPRA) [14], and Space-Time Histories [8, 9]. QS – ANALYTIC SEMANTICS FOR OSP We adopt an analytic approach to spatial reasoning, where the semantics of spatial relations are encoded as polynomial constraints within a (constraint) logic programming setup. The analytic method supports the integration of qualitative and quantitative spatial information, and provides a means for sound, complete and approximate spatial reasoning [4]. For example, let axis-aligned rectangles a, b each be defined by a bottom-left vertex (xi , yi ) and a width and height wi , hi , for i ∈ {a, b} such that xi , yi , wi , hi are reals. The relation that a is a non-tangential proper part of b corresponds to the polynomial constraint: (xb < xa ) ∧ (xa + wa < xb + wb ) ∧ (yb < ya ) ∧ (ya + ha < yb + hb )

Continuing with the example, this is generalised to arbitrarily oriented rectangles. Determining whether a point is inside an arbitrary rectangle is based on vector projection.

4

J. Suchan, M. Bhatt, C. Schultz

S PATIAL D OMAIN (QS) Formalisms Mereotopology

RCC-5, RCC-8 [16]

Spatial Relations (R) disconnected (dc), external contact (ec), partial overlap (po), tangential proper part (tpp), non-tangential proper part (ntpp), proper part (pp), part of (p), discrete (dr), overlap (o), contact (c)

Orientation

Distance, Size Dynamics, Motion

Rectangle & Block algebra [7] LR [20]

proceeds, meets, overlaps, starts, during, finishes, equals

left, right, collinear, front, back, on

OPRA [14]

facing towards, facing away, same direction, opposite direction

QDC [10]

adjacent, near, far, smaller, equi-sized, larger

Space-Time Histories [8, 9]

moving: towards, away, parallel; growing / shrinking: vertically, horizontally; passing: in front, behind; splitting / merging

Entities (E) arbitrary rectangles, circles, polygons, cuboids, spheres axis-aligned rectangles and cuboids 2D point, circle, polygon with 2D line oriented points, 2D/3D vectors rectangles, circles, polygons, cuboids, spheres rectangles, circles, polygons, cuboids, spheres

Table 1. The Spatio-Temporal Domain Osp supported within the Learning Framework

Point p is projected onto vector v by taking the dot product: (xp , yp ) · (xv , yv ) = xp xv + yp yv . With this approach, the task of determining whether a set of spatial relations is consistent then becomes the task of determining whether a system of polynomial constraints is satisfiable. We emphasise that our approach and framework are not limited to the above entities; a wider class of 2D and 3D spatial entities are supported and may be defined as per domain-specific and computational needs [4, 18, 27, 19]. INDUCTIVE LEARNING WITH THE SPATIAL SYSTEM < OSP , QS > Learning is founded on the Aleph ILP system [21]. Learning spatio-temporal structures, is based on integrating the spatial ontology Osp described above, into the basic learning setup of ILP. Given: (1) A set of examples E, consisting of positive and negative examples for the desired spatio-temporal structure, i.e., E = E + ∪E − , where each example is given by a set of spatio-temporal observations in the domain; (2) the (spatio-temporal) background knowledge B. The spatio-temporal learning domain is defined by basic spatial entities (E) constituting the domain objects, the relational spatial structure (R) describing the spatio-temporal configuration of spatial entities in the domain, and rules defining spatio-temporal phenomena and characteristics of the domain. In this context, spatio-temporal facts characterising the learning examples E can be given as, (a) numerical representation of domain objects, (b) qualitative relations between spatial entities, or (c) a mixed qualitativequantitative representations, where the facts are partially grounded in numerical observations. Learning: The learning task is defined as finding hypothesis H consisting of spatiotemporal relations (R) holding between basic spatial entities (E), such that H∪B  E + , and H ∪ B 2 E − . As such, the spatial ontology Osp constitutes an integrated part of the learning setup and spatio-temporal semantics are available throughout the learning process.

Deeply Semantic Spatio-Temporal Learning

Deeply Deeply Semantic Semantic Spatio-Temporal Spatio-TemporalLearning Learning Deeply scene Semantic Spatio-Temporal Learning Fig. 1. Positive examples for symmetric structures at the object level

5

55 5

3 LEARNING CINEMATOGRAPHIC PATTERNS AND THEIR VISUAL RECEPTION: THE CASE OF SYMMETRY Fig. 1. Positive examples for for symmetric symmetric scene scene structures structures at at the theobject objectlevel level

1. Positivefilm examples symmetric scene structures at we the present object level AimedFig. at cognitive studiesfor and visual perception research, a use-case pertaining to the (visual) learning of cinematographic patterns of symmetry and its vi11 of eye-tracking) sual reception (by means of eye-tracking) by by subjects. subjects.11 To To demonstrate demonstrate the the temporal temporal sual reception (by means of eye-tracking) by subjects. To demonstrate temporal framework, we to “axioms ofof aspect of the learning framework, we demonstrate demonstrate the the capability capability to learn learnthe “axioms aspect of the learning framework, we demonstrate the capability to learn “axioms of dynamic eye-tracking visual perception” from dynamic eye-tracking data; data; both both the the chosen chosen films films and and their their visual perception” from dynamic eye-tracking data; both the chosen films and their eye-tracking data data are corresponding eye-tracking are obtained obtained from from aa large-scale large-scale experiment experiment in in visual visual corresponding eye-tracking obtainedexample from a large-scale experiment in visual perception of [23, The presented translates aa variety perception of films films [23, 22]. 22].data Theare presented example translates to to variety of of cases cases perception of films [23, 22]. The presented example translates to a variety of cases involving visual involving visual perception perception and and human human behaviour behaviour studies. studies. involving visual perception and human behaviour studies. Learning Spatial Structures: Structures: Object-Level Object-Level Symmetry Symmetry As As an an example example for for Learning Spatial Learning Spatial Structures: Object-Level Symmetry As an example for learning spatial structures, we consider symmetry in the relative object placement in Learning Spatial Structures: Object-Level Symmetry As an example for learning spatial structures, we consider symmetry in the relative object placement learning spatial structures, we consider symmetry in the relative object placement inin aaalearning movie scene (see Fig. 1). In particular, learning is based on the spatial configuration spatial structures, we consider symmetry in the relative object placement in movie scene (see Fig. 1). In particular, learning is based on the spatial configuration movie scene (see Fig. 1). In particular, learning is based on the spatial configuration of people, faces, andFig. their facing direction, directly obtained from computer visionalala movie scene (see 1).facing In particular, learning isobtained based onfrom the spatial configuration faces, and their direction, directly computer vision of people, faces, and their facing direction, directly obtained from computer vision algorithms [23]. In this context, positive and negative examples, are given of people, faces, and their facing direction, directly obtained from computer vision alas described in [23]. In this context, positive and negative examples, are given gorithms as described in [23]. In this context, positive and negative examples, are given as numerical aboutIndomain domain objects in the the image. image. gorithms as described in [23]. this context, positive and negative examples, are given spatial about objects in as numerical spatial facts facts about domain objects in the image. ... as numerical spatial facts about domain objects in the image. ... detection(id(0), image(3), detection(id(0), image(3), class(person), class(person), rectangle(point(319, rectangle(point(319, 194), 194), 319, 319, 456)). 456)). ... detection(id(1), image(3), class(person), rectangle(point(678, 215), 367, 452)). detection(id(1), detection(id(0), image(3), class(person), rectangle(point(678, rectangle(point(319, 215), 194), 367, 319, 452)). 456)). detection(id(0), image(3), class(face), rectangle(point(438, 246), 86, 86)). detection(id(0), rectangle(point(438, 246), 86, 86)). detection(id(1), image(3), class(face), class(person), rectangle(point(678, 215), 367, 452)). detection(id(1), image(3), class(face), rectangle(point(745, 284), 87, 87)). detection(id(1), class(face), rectangle(point(745, 284), detection(id(0), image(3), rectangle(point(438, 246), 87, 86, 87)). 86)). 2d_facing_dir(id(0), image(3), vector(0.550864, 0.834595), magnitude(6.26042)). 2d_facing_dir(id(0), image(3), vector(0.550864, 0.834595), magnitude(6.26042)). detection(id(1), image(3), class(face), rectangle(point(745, 284), 87, 87)). 2d_facing_dir(id(1), image(3), vector(-0.500519, 0.865726), magnitude(4.82556)). 2d_facing_dir(id(1), 0.865726),magnitude(6.26042)). magnitude(4.82556)). 2d_facing_dir(id(0), image(3), vector(-0.500519, vector(0.550864, 0.834595), ... ... 2d_facing_dir(id(1), image(3), vector(-0.500519, 0.865726), magnitude(4.82556)). ...

We define define representations We representations of of domain domain objects objects linking linking the the numerical numerical description description of of obobWe define of domain objects linkingdifferent the numerical of objects in the the representations image entities describing aspects of jects in image to to basic basic spatial spatial entities describing different aspectsdescription of these theseobjects, objects, jectsthe in bounding the imagebox to basic spatial entities describing different e.g. the bounding (rectangles), or (points). e.g. box (rectangles), or the the center-point center-point (points).aspects of these objects, e.g. the bounding box (rectangles), or the center-point (points). entity(center(person(P)), point(X, Y), image(Img)) :entity(center(person(P)), point(X, Y), image(Img)) :detection(_, image(Img), class(person), rectangle(point(Xr, Yr), W, H)), detection(_, image(Img), class(person), rectangle(point(Xr, Yr), W, H)), entity(center(person(P)), Y), image(Img)) :X is Xr + W/2, Y is Yr+point(X, H/2. X is Xr + W/2, Y is Yr+ H/2. detection(_, image(Img), class(person), rectangle(point(Xr, Yr), W, H)), X is Xr + W/2, Y is Yr+ H/2. In addition to the detected domain objects, we define abstract geometric objects needed

In addition to the detected domain objects, we define abstract geometric objects needed In addition symmetry, to the detected objects, wein abstract geometric to describe describe the axis the of image. to symmetry, e.g. e.g. domain the symmetry symmetry axis indefine the center center of the the image. objects needed toentity(symmetry_obj(center_axis), describe symmetry, e.g. the symmetry axis0,in X, theY), center of the image. line(X, image(Img)) :entity(symmetry_obj(center_axis), line(X, 0, X, Y), image(Img))image(Img)), :img(image(Img)), media_size(size(MediaWidth, MediaHeight), img(image(Img)), media_size(size(MediaWidth, MediaHeight), entity(symmetry_obj(center_axis), line(X, 0, X, Y), image(Img))image(Img)), :X is MediaWidth/2, Y is MediaHeight. X is MediaWidth/2, Y is MediaHeight. img(image(Img)), media_size(size(MediaWidth, MediaHeight), image(Img)), X is MediaWidth/2, Y is MediaHeight.

Learning: We Learning: We learn learn the the relational relational spatial spatial structure structure consisting consisting of of qualitative qualitative spatial spatial 1 relationships characterising symmetry in the configuration of the spatial entities ininviewthe Learning: We learn the relational spatial structure consisting of qualitative spatial Our case-study is motivated by a broader multi-level interpretation of symmetry from the relationships characterising symmetry in the configuration of the spatial entities the image, i.e. we consider relations of topology, orientation, distance, and size. relationships symmetry in the ofofthe entities on in one the point i.e. of film cinematography [25];ofhowever, theconfiguration specific example thisspatial paper focusses image, wecharacterising consider relations topology, orientation, distance, and size. image, i.e. relations of topology, orientation, distance, and size. of we thisconsider multi-level symmetry characterisation involving relative object placement in a :-aspect modeh(1,symmetric(+img)). :- modeb( *,entity(#ent,-obj,+img)). ::- modeh(1,symmetric(+img)). modeb(*,topology(rcc8(#rel),+obj,+obj)). movie scene. :,topology(rcc8(#rel),+obj,+obj)). modeh(1,symmetric(+img)). *,orientation(#rel,+obj,+obj)). :- modeb( modeb( * :- modeb(*,orientation(#rel,+obj,+obj)). ,topology(rcc8(#rel),+obj,+obj)). :- modeb(*,orientation(#rel,+obj,+obj)). 1 1 1

::::::-

modeb(**,distance(#rel,+obj,+obj)). ,entity(#ent,-obj,+img)). modeb( modeb(**,size(#rel,+obj,+obj)). ,distance(#rel,+obj,+obj)). ,entity(#ent,-obj,+img)). modeb( modeb(*,size(#rel,+obj,+obj)). ,distance(#rel,+obj,+obj)). modeb(*,size(#rel,+obj,+obj)).

Our case-study is motivated by a broader multi-level interpretation of symmetry from the viewOur case-study is motivated by a broader multi-level interpretation of symmetry from the viewpoint of film cinematography [25]; however, the specific example ofofthis paper focusses onviewone Our case-study is motivated by a broader multi-level interpretation symmetry from the point of film cinematography [25]; however, the specific example of this paper focusses on one aspect offilm thiscinematography multi-level symmetry characterisation involving relative object placementoninone a point of [25]; however, the specific example of this paper focusses aspect of this multi-level symmetry characterisation involving relative object placement in a movie scene. aspect of this multi-level symmetry characterisation involving relative object placement in a movie scene. movie scene.

We define representations of domain objects linking the numerical description of objects in the image to basic spatial entities describing different aspects of these objects, e.g. the bounding box (rectangles), or the center-point (points). entity(center(person(P)), point(X, Y), image(Img)) :detection(_, image(Img), class(person), rectangle(point(Xr, Yr), W, H)), X is Xr + W/2, Y is Yr+ H/2.

In addition to the detected domain objects, we define abstract geometric objects needed to describe symmetry, e.g. the symmetry axis in the center of the image. entity(symmetry_obj(center_axis), line(X, 0, X, Y), image(Img)) :J. Suchan, M. Bhatt, C. Schultz img(image(Img)), media_size(size(MediaWidth, MediaHeight), image(Img)), X is MediaWidth/2, Y is MediaHeight.

6

Learning: We learn the relational spatial structure consisting of qualitative spatial relationships characterising symmetry in the configuration of the spatial entities in the image, i.e. we consider relations of topology, orientation, distance, and size. :- modeh(1,symmetric(+img)). :- modeb(*,entity(#ent,-obj,+img)). modeb( ,topology(rcc8(#rel),+obj,+obj)). :- modeb(*,distance(#rel,+obj,+obj)). *Suchan, J. M. Bhatt, C. Schultz J. Suchan, M. Bhatt, C. Schultz modeb(*,orientation(#rel,+obj,+obj)). :- modeb(*,size(#rel,+obj,+obj)). J. Suchan, M. Bhatt, C. Schultz

66:6:6

J. Suchan, M. Bhatt, C. Schultz

spatial learned by include following. 1 Exemplary symmetrical include the the following. Exemplary symmetrical spatial structures, learnedinterpretation by the the system system include the following. Our case-study is motivated by a structures, broader multi-level of symmetry from the viewExemplary symmetrical spatial structures, learned by the system include the following. symmetric(A) :entity(center(person(0)),B,A), entity(center(person(1)),C,A), point of film cinematography [25]; however, the specific example of this paper focusses on one Exemplary symmetrical spatial structures, learned entity(center(person(1)),C,A), by the system include the following. symmetric(A) :- entity(center(person(0)),B,A), entity(symmetry_object(center_axis),D,A), distance(equidistant,D,C,B). symmetric(A) entity(center(person(0)),B,A), entity(center(person(1)),C,A), entity(symmetry_object(center_axis),D,A), distance(equidistant,D,C,B). aspect of this:multi-level symmetry characterisation involving relative object placement symmetric(A) :- entity(center(person(0)),B,A), entity(center(person(1)),C,A), entity(symmetry_object(center_axis),D,A), distance(equidistant,D,C,B). symmetric(A) entity(person(1),C,A), size(same,C,B). movie scene. :entity(symmetry_object(center_axis),D,A), distance(equidistant,D,C,B). symmetric(A) :- entity(person(0),B,A), entity(person(0),B,A), entity(person(1),C,A), size(same,C,B). symmetric(A) :- entity(person(0),B,A), entity(person(1),C,A), size(same,C,B). symmetric(A) :- entity(person(0),B,A), entity(person(1),C,A), size(same,C,B).

in a

Learning Dynamics: Axioms of of Perception Perception We illustrate illustrate Axioms Perception: Learning Spatio-Temporal Dynamics We LearningofSpatio-Temporal Spatio-Temporal Dynamics: Axioms Learning Spatio-Temporal Dynamics: Axioms of Perception We illustrate learning of spatio-temporal perception, by learning dynamics in the context of visual perception, by learning learning of spatio-temporal Learning Spatio-Temporal Dynamics: Axioms of Perception We illustrate learning of patterns spatio-temporal dynamics in the context of visual perception, by learning perceptual movie scene. As an an from eye-tracking data and people tracks in a movie scene. As perceptual patterns learning of spatio-temporal dynamics in the context of visual perception, by learning perceptual patterns from eye-tracking data and people tracks inindividual a movie to scene. As an example we focus another. on attention of a person switching from one individual to another. example we focus perceptual patterns from eye-tracking data and people tracks in a movie scene. As an example we focus on attention of a person switching from one individual to another. detection(id(0), frame(426), class(person), rectangle(point(385,66),244,271)). example we focus on attention of a person switching from one individual to another. detection(id(0), frame(426), class(person), rectangle(point(385,66),244,271)). detection(id(1), class(person), rectangle(point(111,68),332,276)). detection(id(0), detection(id(1), frame(426), frame(426), class(person), class(person), rectangle(point(385,66),244,271)). rectangle(point(111,68),332,276)). gazepoint(frame(426), point(859,212)). detection(id(0), rectangle(point(385,66),244,271)). detection(id(1), frame(426), class(person), rectangle(point(111,68),332,276)). gazepoint(frame(426), point(859,212)). detection(id(1), frame(426), class(person), rectangle(point(111,68),332,276)). gazepoint(frame(426), point(859,212)). gazepoint(frame(426), Learning: We adapt point(859,212)). the general learning setup of the example above, for learning

Learning: We adapt the general learning setup of the example above, for learning learning Learning: Wedynamics adapt thebygeneral learning of the example toabove, spatio-temporal introducing the setup predicate holds-in/2 denotefor that spaspatio-temporal introducing the predicate holds-in/2 denote that aa spaLearning: Wedynamics adapt thebygeneral learning setup of the exampleto above, for learning denote that a spaspatio-temporal dynamics by introducing the predicate holds-in/2 to tial relation holds between two entities at a time point. tial relation holds betweenby two entities at the a time point. holds-in/2 to denote that a spaspatio-temporal dynamics introducing predicate tial relation holds between two entities at a time point. ... ...relation holds between two entities at a time point. tial :modeb( , holds_in(topology(#rel, +ent, +ent), +time)). ... :- modeb(* , holds_in(topology(#rel, +ent, +ent), +time)). * time(#rel, ... :+ent, +ent), +time)). :- modeb( modeb(**,, holds_in(topology(#rel, time(#rel, +time, +time, +time)). +time)). ... holds_in(topology(#rel, +ent, +ent), +time)). :modeb( , time(#rel, +time, +time)). * ... :- modeb(*, time(#rel, +time, +time)). ... ... Spatio-temporal dynamics switches include include the the following. following. Spatio-temporal dynamics constituting constituting attention attention switches

Spatio-temporal dynamics constituting attention switches include the following. Spatio-temporal dynamics constituting attention switches includeA), the following. att_switch(B) holds_in(topology(inside, gaze, person(1)), A), att_switch(B) ::holds_in(topology(inside, gaze, person(1)), holds_in(topology(inside, gaze, B), time(consecutive, att_switch(B) :- holds_in(topology(inside, gaze, B), person(1)), A), holds_in(topology(inside, gaze, person(2)), person(2)), time(consecutive, att_switch(B) :- holds_in(topology(inside, gaze,B), person(1)), A), holds_in(topology(inside, gaze, person(2)), time(consecutive, holds_in(topology(inside, gaze, person(2)), B), time(consecutive,

A, A, A, A,

B). B). B). B).

4 DISCUSSION AND OUTLOOK 4 DISCUSSION AND OUTLOOK 44 DISCUSSION AND OUTLOOK DISCUSSION AND OUTLOOK Directly comparable to this research is the line of work on integrated inductive-abductive Directly to this research is therelational line of work on integrated inductive-abductive reasoningcomparable for learning spatio-temporal models from video in [5, 6]; here, Directly comparable to this research is the line of work on integrated inductive-abductive reasoning for learning spatio-temporal relational models from video infor[5,the6];case here, Directly comparable to this research is the line of work on integrated inductive-abductive spatio-temporal learning in the context of ILP has only been addressed of reasoning for learning spatio-temporal relational models from video in [5, 6]; here, spatio-temporal learning in the context of ILP has only been addressed for the case of reasoning for learning spatio-temporal relational models from video in [5, 6]; here, topological relations. Furthermore, the ILP learning framework does not have built-in spatio-temporal learning in the context of ILP has only been addressed for the case of topological relations. Furthermore, the Aside ILP learning framework does not have built-in spatio-temporal in the context of ILP has this, only been addressed the built-in case of semantics for thelearning topological relations. from learning relational spatial structopological relations. Furthermore, the ILP learning framework does notforhave semantics for the topological relations. Aside from this, learning relational spatial structopological relations. Furthermore, the ILP learning framework does not have built-in tures was investigated in the context of learning spatial relations from language[12], and semantics for the topological relations. Aside from this, learning relational spatial structures was in the context of learning spatial relations from language[12], and semantics for the topological relations. Aside from this,Logic learning relational spatial strucwithin theinvestigated geospatial domain [13, 26]. Probabilistic Programming frameworks tures was investigated in the context of learning spatial relations from language[12], and within the geospatial domain [13, 26]. Probabilistic Logic Programming frameworks tures was investigated in theProbLog context of learning spatial relations from language[12], and such asthe PRISM [17] and [11] have been used forProgramming learning parameters, and within geospatial domain [13, 26]. Probabilistic Logic frameworks such asthe PRISM [17] and ProbLog [11] have been used forProgramming learning spatial parameters, and within geospatial domain [13, 26]. Probabilistic Logic frameworks the structure, of probabilistic logic programs, although (qualitative) reasoning such as PRISM [17] and ProbLog [11] have been used for learning parameters, and the of [17] probabilistic logic programs, although spatial reasoning suchstructure, as been PRISM and ProbLog been used(qualitative) for learning parameters, and has not directly addressed. The[11] mainhave point-of-departure of this paper with respect the structure, of probabilistic logic programs, although (qualitative) spatial reasoning has not been directly addressed. The main point-of-departure of this paper with respect the structure, of probabilistic logic programs, although (qualitative) spatial reasoning to the state of the art in (qualitative) spatial learning is that the semantics of spatial, has not been directly addressed. The main point-of-departure of this paper with respect to ofdirectly the artaddressed. in (qualitative) spatial learning is that thethe semantics of spatial, hasthe notstate been The main of paper with respect temporal, and relations arepoint-of-departure directly built within inductive to the state of spatio-temporal the art in (qualitative) spatial learning is that thethis semantics of learning spatial, temporal, and spatio-temporal relations are directly built within the inductive learning to the state of the art in (qualitative) spatial learning is that the semantics of spatial, framework of ILP. Pragmatically, what this implies is that it is possible to seamlessly detemporal, and spatio-temporal relations are directly built within the inductive learning framework of ILP. Pragmatically, what this implies is built that itwithin is possible to seamlessly detemporal, and spatio-temporal relations are directly the inductive learning cribe a learning problem using a generic relational spatio-temporal ontology directly as framework of ILP. Pragmatically, what this implies is that it is possible to seamlessly decribe a alearning problem using a generic relational spatio-temporal ontology directlydeas framework of ILP. Pragmatically, what this implies is that it is possible to seamlessly part of logic programming based learning environment. To the best of our knowledge, cribe a learning problem using a generic relational spatio-temporal ontology directly as part logic programming based learning environment. To the bestontology of our knowledge, cribeofaa alearning problem using alearning generic relational spatio-temporal directly as such spatio-temporal framework withTo built semantics for mixed part of ageneral logic programming based learning environment. thein best of our knowledge, such a general spatio-temporal learning framework with built in semantics for mixed qualitative-quantitative spatio-temporal hassemantics not beenfor available such a general spatio-temporal learning reasoning frameworkcapabilities with built in mixed qualitative-quantitative spatio-temporal reasoning capabilities has not been available before. Furthermore, the ontology of space-time features supported in our framework qualitative-quantitative spatio-temporal reasoning capabilities has not been available before. Furthermore, the ontology of space-time features supported in our framework goes much beyond topological relations addressing orientation, distance, size. Fubefore. Furthermore, the ontology of space-time features supported in ourand framework goes much beyond topological relations addressing orientation, distance, and size. Future research will focus on enhancing the expressivity of the spatio-temporal relations goes much beyond topological relations addressing orientation, distance, and size. Future research willrange focusof ondomain-independent enhancing the expressivity of the spatio-temporal relations to cover a wider features characterising spatio-temporal ture research will focus on enhancing the expressivity of the spatio-temporal relations to cover a wider range of domain-independent features characterising spatio-temporal dynamics. to cover a wider range of domain-independent features characterising spatio-temporal dynamics. dynamics.

Deeply Semantic Spatio-Temporal Learning

7

part of a logic programming based learning environment. To the best of our knowledge, such a general spatio-temporal learning framework with built in semantics for mixed qualitative-quantitative spatio-temporal reasoning capabilities has not been available before. Furthermore, the ontology of space-time features supported in our framework goes much beyond topological relations addressing orientation, distance, and size. Future research will focus on enhancing the expressivity of the spatio-temporal relations to cover a wider range of domain-independent features characterising spatio-temporal dynamics.

Bibliography

[1] M. Bhatt. Reasoning about Space, Actions and Change: A Paradigm for Applications of Spatial Reasoning. In Qualitative Spatial Representation and Reasoning: Trends and Future Directions. IGI Global, USA, 2012. ISBN ISBN13: 9781616928681. [2] M. Bhatt and S. Loke. Modelling Dynamic Spatial Systems in the Situation Calculus. Spatial Cognition and Computation, 8(1):86–130, 2008. ISSN 1387-5868. [3] M. Bhatt, F. Dylla, and J. Hois. Spatio-terminological inference for the design of ambient environments. In Spatial Information Theory, 9th International Conference, COSIT 2009, Aber Wrac’h, France, September 21-25, 2009, Proceedings, volume 5756 of Lecture Notes in Computer Science, pages 371–391. Springer, 2009. ISBN 978-3-642-03831-0. doi: 10.1007/978-3-642-03832-7 23. [4] M. Bhatt, J. H. Lee, and C. P. L. Schultz. CLP(QS): A declarative spatial reasoning framework. In Spatial Information Theory - 10th International Conference, COSIT 2011, Belfast, ME, USA, September 12-16, 2011. Proceedings, pages 210–230, 2011. doi: 10.1007/978-3-642-23196-4 12. [5] K. S. R. Dubba, M. Bhatt, F. Dylla, D. C. Hogg, and A. G. Cohn. Interleaved inductiveabductive reasoning for learning complex event models. In S. Muggleton, A. TamaddoniNezhad, and F. A. Lisi, editors, Inductive Logic Programming - 21st International Conference, ILP 2011, Windsor Great Park, UK, July 31 - August 3, 2011, Revised Selected Papers, volume 7207 of Lecture Notes in Computer Science, pages 113–129. Springer, 2011. ISBN 978-3-642-31950-1. doi: 10.1007/978-3-642-31951-8 14. [6] K. S. R. Dubba, A. G. Cohn, D. C. Hogg, M. Bhatt, and F. Dylla. Learning relational event models from video. J. Artif. Intell. Res. (JAIR), 53:41–90, 2015. doi: 10.1613/jair.4395. [7] H. W. Guesgen. Spatial reasoning based on Allen’s temporal logic. Technical Report TR89-049. International Computer Science Institute Berkeley, 1989. [8] P. J. Hayes. Naive physics I: ontology for liquids. In J. R. Hubbs and R. C. Moore, editors, Formal Theories of the Commonsense World. Ablex Publishing Corporation, Norwood, NJ, 1985. [9] S. M. Hazarika. Qualitative spatial change: space-time histories and continuity. PhD thesis, The University of Leeds, 2005. [10] D. Hern´andez, E. Clementini, and P. Di Felice. Qualitative distances. Springer, 1995. [11] A. Kimmig, L. De Raedt, and H. Toivonen. Probabilistic explanation based learning. In European Conference on Machine Learning, pages 176–187. Springer, 2007. [12] P. Kordjamshidi, P. Frasconi, M. van Otterlo, M. Moens, and L. D. Raedt. Relational learning for spatial relation extraction from natural language. In S. Muggleton, A. TamaddoniNezhad, and F. A. Lisi, editors, Inductive Logic Programming - 21st International Conference, ILP 2011, Windsor Great Park, UK, July 31 - August 3, 2011, Revised Selected Papers, volume 7207 of Lecture Notes in Computer Science, pages 204–220. Springer, 2011. doi: 10.1007/978-3-642-31951-8 20. [13] D. Malerba and F. A. Lisi. Discovering associations between spatial objects: An ilp application. In Proceedings of the 11th International Conference on Inductive Logic Programming, ILP ’01, pages 156–163, London, UK, UK, 2001. Springer-Verlag. ISBN 3-540-42538-1. [14] R. Moratz. Representing relative direction as a binary relation of oriented points. In ECAI, pages 407–411, 2006. [15] S. Muggleton and L. D. Raedt. Inductive logic programming: Theory and methods. JOURNAL OF LOGIC PROGRAMMING, 19(20):629–679, 1994.

Deeply Semantic Spatio-Temporal Learning

9

[16] D. A. Randell, Z. Cui, and A. G. Cohn. A spatial logic based on regions and connection. KR, 92:165–176, 1992. [17] T. Sato and Y. Kameya. New advances in logic-based probabilistic modeling by prism. In Probabilistic inductive logic programming, pages 118–155. Springer, 2008. [18] C. Schultz and M. Bhatt. A numerical optimisation based characterisation of spatial reasoning. In Rule Technologies. Research, Tools, and Applications: 10th International Symposium, RuleML 2016, Stony Brook, NY, USA, July 6-9, 2016. Proceedings, pages 199–207. Springer International Publishing, 2016. ISBN 978-3-319-42019-6. [19] C. Schultz, M. Bhatt, and J. Suchan. Probabilistic spatial reasoning in constraint logic programming. In Tenth International Conference on Scalable Uncertainty Management (SUM 2016) (to appear), 2016. [20] A. Scivos and B. Nebel. The Finest of its Class: The Natural, Point-Based Ternary Calculus LR for Qualitative Spatial Reasoning. In C. Freksa et al. (2005), Spatial Cognition IV. Reasoning, Action, Interaction: International Conference Spatial Cognition. Lecture Notes in Computer Science Vol. 3343, Springer, Berlin Heidelberg, volume 3343, pages 283–303, 2004. [21] A. Srinivasan. The Aleph Manual, 2001. URL http://web.comlab.ox.ac.uk/ oucl/research/areas/machlearn/Aleph/. [22] J. Suchan and M. Bhatt. Semantic question-answering with video and eye- tracking data – ai foundations for human visual perception driven cognitive film studies. In IJCAI 2016: 25th International Joint Conference on Artificial Intelligence, New York City, USA, July 2016. [23] J. Suchan and M. Bhatt. The geometry of a scene: On deep semantics for visual perception driven cognitive film, studies. In 2016 IEEE Winter Conference on Applications of Computer Vision, WACV 2016, Lake Placid, NY, USA, March 7-10, 2016, pages 1–9. IEEE Computer Society, 2016. ISBN 978-1-5090-0641-0. doi: 10.1109/WACV.2016.7477712. [24] J. Suchan, M. Bhatt, and P. E. Santos. Perceptual narratives of space and motion for semantic interpretation of visual data. In Computer Vision - ECCV 2014 Workshops - Zurich, Switzerland, September 6-7 and 12, 2014, Proceedings, Part II, pages 339–354, 2014. doi: 10.1007/978-3-319-16181-5 24. [25] J. Suchan, M. Bhatt, and S. Yu. The perception of symmetry in the moving image: multilevel computational analysis of cinematographic scene structure and its visual reception. In E. Jain and S. J¨org, editors, Proceedings of the ACM Symposium on Applied Perception, SAP 2016, Anaheim, California, USA, July 22-23, 2016, page 142. ACM, 2016. ISBN 978-1-4503-4383-1. doi: 10.1145/2931002.2948721. [26] D. Vaz, V. S. Costa, and M. Ferreira. Fire! firing inductive rules from economic geography for fire risk detection. In P. Frasconi and F. A. Lisi, editors, Inductive Logic Programming 20th International Conference, ILP 2010, Florence, Italy, June 27-30, 2010. Revised Papers, volume 6489 of Lecture Notes in Computer Science, pages 238–252. Springer, 2010. doi: 10.1007/978-3-642-21295-6 27. [27] P. A. Walega, M. Bhatt, and C. P. L. Schultz. ASPMT(QS): Non-Monotonic Spatial Reasoning with Answer Set Programming Modulo Theories. In Logic Programming and Nonmonotonic Reasoning - 13th International Conference, LPNMR 2015, Lexington, KY, USA, September 27-30, 2015. Proceedings, pages 488–501, 2015. doi: 10.1007/ 978-3-319-23264-5 41.

Suggest Documents