Detection of Obstacles in Monocular Image

0 downloads 0 Views 5MB Size Report
Nov 26, 1997 - of detection of objects in monocular ... motion-based segmentation algorithm based on the planar motion .... the pilot has to depend on an SVS which derives information ..... In this analysis,weprojectedthe imageplaneonto the ground,and ...... A circular maskis shownat sevendifferentpositionsin the image.
Detection

of Obstacles

in Monocular

Image

Sequences

..a_

Final Technical

Report

"A Vision-Based

for NASA Co-Operative Number NCC 2-916

Obstacle

Period

Detection

of Grant:

System

August

Research

for Aircraft

Agreement Navigation"

1, 1995 to July 31, 1997

i

Submitted NASA .2,.

Ames

Technical Flight

Research

Officer:

Management

to Center

Albert

Ahumada

and Human

Factors

Division

Mail Stop 262-2 Moffett Field, CA 94035 I

i by Rangachar

Kasturi Principal

Departments

of Computer

Science

and Octavia

Camps

Investigators and Engineering

and

Electrical

The Pennsylvania State University University Park, PA 16802 Tel: (814) E-Mail:

863-4254

{kasturi,camps}

i _z._

i

(814)

865-3176

@cse.psu.edu

Graduate

Students:

Sadashiva Tarak

Devadiga Gandhi

November i

Fax:

26, 1997

Engineering

Detection

of Obstacles

in Monocular

Image

Sequences

Preface

This research

was initiated

development

of a High

Center.

Primary

to detect from

two

camera.

the

developing runway

Speed

goal of the research of sensors,

report a

sequence

into

approach of

images

aircraft

in monocular

Millimeter two

parts.

for detecting obtained

Wave The

features

corresponding

to the

at NASA

runway

markers.

histogram-based thresholding is used to detect second part addresses the problem of detection obtained from an on-board video camera.

image

(PMMW) part

runways/taxiways from

a

(SVS)

image

first

approximate runway model and the position information Global Positioning System (GPS) is used to define the image

System

here is to develop

obstacles

a Passive

Vision

(HSCT)

reported

and

is divided

a model-based from

of the Synthetic

Civil Transport

runways/taxiways

types The

as a part

moving

project Ames

Research

analysis

algorithms

sequences sensor

of this

obtained

and

report

and PMMW

for the

objects

a video aims

at

on the

sensor.

An

of the camera provided by the region of interest to search for Once

the

runway

is identified,

a

obstacles on and outside the runway. The of objects in monocular image sequences A recursive motion-based segmentation

algorithm identified

based on planar motion model is used. The background motion due to camera is as the dominant motion and the outliers which do not follow this motion are

identified

as obstacles.

ii

Abstract

The ability on-board

to detect sensors

takeoff,

and

different

weather

modalities.

and locate

is an essential

taxiing

phase

and lighting

mounted

such as detection and landing

In this report, sequences sensor their the

and a video

camera

same,

type.

resolution, different

These

goal

moving poor,

for the image identified,

outside

a single

real PMMW

alternative

on the

camera

runway

landing,

functions

under

sensors

of different

sensors

of different

systems

in functions

surface

navigation,

the sensors

differ

using

these

depending

regions

is tested

sensors

is not

on the

sensor

a method

of images

for

camera

provided

Once obstacles

using image

plane

the runway on

the

sequences

a

is very

We use

in the image

markers.

from

quality

runways/taxiways. of the

detecting

obtained

and the image

of interest

to detect

in

of this report.

is low

information

to the runway

This algorithm

Since

obstacles

for detecting

thresholding

(PMMW)

aircraft. obtained

image

Wave

is to develop

resolution

in monocular

Millimeter

in a sequence

position

to define

corresponding

airport

of objects

in two parts

report

approach

(GPS)

the runway.

a landing

for detecting

the sensor

use histogram-based

regions

The

System

collisions,

of the images

of the

and the

with

ground-based

a Passive

separately

part

Since

model

features

we

first

a model-based

runway

Positioning

on-board

are used

objects

sensor.

we propose

approximate Global

and

PMMW

mounted

(SVS),

of detection

of sensors,

are described

of the

runways/taxiways

the problem types

approaches

runway

flight,

of these

by using

using

conditions.

and the quality

approaches

The

two

System

captured

of low-altitude

be facilitated

the current

of potential

we address from

spatial

complements

in images

Automation

Vision

in all weather

obtained

can

Synthetic

and prevention

and takeoff

navigation.

situations,

on-board,

and obstacles

first step in the automation

of aircraft

An aircraft-based

modalities

runways/taxiways

the

by the to search region

runway

simulated

is and

from

image.

position

is to estimate

information

the camera

provided

position

using

iii

by

the

image-based

GPS

is not features,

accurate. such

An

as points

and

lines.

sensor

The

of sensor model

resolution

for three

due

required

The

aircraft.

plane

part

image

using a corner

planar

and propose model

recovered

the dominant

actual

flight

In this report

position

tested

a video

problem camera

at coruer-like

using

this

are presented.

is useful

We

camera

in deciding

background

both synthetic

mounted

the

points

is either

regions. due

planar

as due to the obstacles.

on the

or is piecewise

based Model

in

a landing

detected

to camera

and real image

of objects

on-board

algorithm

motion

are identified

of detection

feature

that the runway

into separate

are considered

we have plane

tested

using

plane

containing

plane)

for a given

explored

parameter

and velocity

the plane

the

flow vectors

using

obtained

in terms

in the estimated

analysis

segmentation

and the outliers

results

the error

to compute

for accuracy

of the sensor

motion-based

The

model

of the

of SVS for navigation.

addresses from

resolution

on the planar parameters

are

is identified Various

sequences

as

stages

obtained

from

test.

Line features

the runway

positions

theoretical

is computed

of the region.

were

Analytical

for computing

We assume

for segmenting

of the algorithm

in 3-D.

Such

obtained

detector.

motion

position

approach

a recursive

for each

equations

and develop

at six different

flow

on the

position

of this report

the optical

depends

an analytical

for use in the design

sequences

image

however,

in 3-D. We propose

quantization.

resolution

second

First,

motion

sensors

a new algorithmic

to image

monocular

camera

and the sensor

different

camera

estimates,

of the sensor

of the estimated

also propose

could

of such

and the position

the accuracy

pose

accuracy

using

sequences. runway

line segment

parameter

not be properly

could

runway

Since

marker

markers, obstacles

the

in two frames

and hence

and using

Though good

not be obtained.

iv

object

features.

(1) estimating

and

estimating

the

lines. The

algorithm

was

the projecting line

planes

segment

small to resolve,

line features

estimate

than point

purposes:

(2) tracking

corresponding was too

reliable

for two

the angle between

and

not be obtained.

located

and more

the use of line features

of the ground-based

real image the

to be prominent

were

for the motion

good tracked,

(i.e., the

in the

image

estimate

for

end points

of the object

could

Table

of Contents

ii

Preface

.°°

Abstract

Ul

1

1. Introduction

3

1.1 Overview 1.2 Review

of Literature

on Obstacle

Detection

Systems

for Navigation

9

1.3 Organization

2. Sensor

Sensitivity

2.1 Sensor

Evaluation Positional

Sensitivity

Imaging

2.1.3

Quantitative

11

Evaluation

15

Geometry Results

Calibration

2.2.1 Error

11

and Calibration

2.1.1

2.2 Camera

due to Image of Image

2.2.3 Experimental

21

and Discussion Features

24

Quantization

25

from Image-based

2.2.2 Analysis

Plane Plane

27

Quantization

38

Results

2.2.3.1

Data

2.2.3.2

Results

38

Simulation

40

and Discussion

45

2.3 Summary

3. Detection

of Objects

3.1 Model

in Passive

Millimeter

Wave

47

Images

48

Transformation

3.1.1 Defining

Regions

3.1.2 Defining

Search

3.2 Feature

6

Localization

3.2.1

Runway

3.2.2

Object

3.3 Experimental

of Interest Space

and Object Localization

Detection

Results

for Runway

for Horizon Detection

Line

Edges

49 51 52 52 52 53

3.3 ExperimentalResults

53

3.4 Summary

54

4. Multiple Motion Segmentation andEstimation 4.1 Methodologiesfor Motion DetectionandEstimation

58 59

4.1.1Feature-based Motion DetectionandEstimation

60

4.1.2OpticalFlow-basedMotion Estimation

62

4.1.3Multiple Motion SegmentationandEstimation

64

4.2 ProblemStatementandthe ProposedApproach

66

4.3 OpticalFlow Computation

68

4.3.1FeatureDetection

75

4.3.2HierarchicalApproachfor OpticalFlow Computation

77 85

4.4 Motion-basedSegmentation 4.4.1PlanarMotion Model

86

4.4.2RecoveringModel Parameterfrom OpticalFlow

91

4.4.2.1LeastSquareModel Fitting

91

4.4.2.2DetectingOutliers

93

4.4.2.3Algorithm for Recoveryof SingleMotion Model

94

4.4.3SegmentationandEstimationof Multiple Motion

95

4.4.3.1Split andMergeAlgorithm

97

4.4.3.2ExperimentalResults

99

4.5 PerformanceCharacterization

100

4.6 Summary

106

5 StructureandMotion Estimationfrom Line Features 5.1 Estimationof PlaneParameters

115 116

5.1.1Analysis

116

5.1.2FeatureDetectionandMatching

120

5.1.3Discussion

125

5.2Estimationof 3D PositionandVelocity

vi

128

5.2.1Analysis

130

5.2.2ExperimentalResults

134

5.3 Summary

136

6 Conclusions

137

6.1PrimaryContribution

138

6.2 FutureResearch

140

References

143

vii

1

Introduction

The

ability

on-board

sensors

takeoff

and

igates the

to detect

the

(ILS), the

phase

aircraft and

up the

Microwave

human

pilot

about

landing

if the

visibility

information darkness, based

from

navigation,

such

and

Implementing IIIa 1 or better, Another

capability

step

and

in the

height

of the

ground

(MLS)

etc.)

assumes

control

of the

aircraft.

Crew

and

System

detection

landing an SVS which

motivating

are

in order bad

would

factor

Hence,

the

equipped

the

landing

with

development

runway

CAT

System

which

decisions

fusion

of

conditions

of

for an aircraft-

airport

in various surface

[17, 107].

of over

1100

II 2 or CAT

of a SVS

and

systems

collisions,

on

or delayed

under

ground-based

conditions

capabilities

aborted

need

nav-

after

makes

combination

is a strong

the

Landing

pilot

mission

landing,

installed

landing

is often

their

of potential

in all weather

for the

point,

there

to complement

upgrade

currently

The

during

using

system

system

as Instrument

this

to accomplish

prevention

takeoff

At

flight,

autopilot

autopilot

the runway

captured

altitude

the

on a sophisticated

weather.

(SVS)

and

and

rely

such

above

visibility.

members

sensors

visibility Vision

external

present

on the

systems

in images

of low

At

(depending

System

on the

obstacles

automation

of aircrafts.

Landing

multiple

as

first

to a certain

is poor.

Synthetic

runways/taxiways

of navigation

depending

reduced

functions

locate

is an essential

taxiing

aircraft

and

is the

US airports

13 capability design

to CAT [24,

107].

of a supersonic

1The minimum vertical visibility ceiling is 12-35 feet with runway visual range of 100-300 meters. A fully automatic landing system with automatic flare and a failure survival autopilot system with a probability of catastrophic failure leas than 10 -T per hour is required. Pilot assumes control at touchdown. 2The minimum required vertical visibility ceiling is 100 feet and runway visual range is 400 meters. A fail passive autopilot is required. Pilot takes over landing at a height of 100 feet. 3There is sufficient vertical visibility at a height of 200 feet with a runway visual range of at least 800 meters for the pilot to carry out a safe landing under manual control. A simplex autopilot system is acceptable. Pilot assumes control at a height of 200 feet.

commercial

aircraft.

a Concorde-like weight[107,

It is argued

drooped

116].

nose

Without

the pilot has to depend

and

locate

the

sound

cockpit

aircraft

sensor's

information

(HSCT) in the

sensor

to see through

external

System

(GPS)

to detect

determine

visibility

is designed

of sensors. sensor

to be operated

Outputs

information

of these

(yaw, pitch, Navigation

the runways/taxiways conflicts,

and

due to fog is known

and the Inertial

potential

takeoff

Wave (PMMW)

fog [132].

and orientation

without

gross

from combination

Milli-Meter

A PMMW

position

database

the airport,

we address

for obstacle

the problem

obtained

detection

navigation,

use high resolution

images

are often used to estimate

are developed

and

vehicle

and high quality

with having accurately.

of ground-based

and

issue

and

System obstacles,

advisories,

Most of the algorithms evaluated

and

systems

obtained

to an obstacle.

such a system there

in monoc-

available

for applications

The use of stereo

Hence,

obstacles

for use in an SVS. Any object

and highway

images

the distance

and poor quality.

for SVS due to the difficulty the 3D position

sensors

to be an obstacle.

and intelligent

ever, axe of low resolution

of detection

from on-board

cations

to estimate

reduction has limited

with a Passive

Positioning

geometry

within

sequences

vehicle

derives

the ability

with aircraft

field of view is considered

literature

the cockpit

Transport

alarms.

In this research,

road

thus providing

an airport

image

nose,

by the Global

(INS),

ular

a drooped

Civil

in significant

and video cameras.

will be integrated

roll) provided

result

Speed

(e.g., 94 GHz), at which point the energy attenuation

to be at a minimum, sensors

could

to be equipped

to the infra-red

at lower frequencies

a High

on an SVS which

The SVS is envisioned in addition

that

PMMW

with a large

is a strong

enough

These

camera.

sensor

is not currently

in the

in robotics,

(IVHS).

using a video

in the

images,

appliStereo how-

recommended base distance

need to design and evaluate

computervision algorithmsfor usein SVSfor obstacledetectionusing monocularimage sequences.

1.1

Overview

This report

is about

moving

sensor.

analyze

image

on-board

The initial sequences

the

PMMW

detection

aircraft

sensors

obtained to detect

orientation

of aircraft,

The position is provided accuracy. such

data

in the

to guide

of the aircraft

by the INS. For example,

updates

could

a few hundred

is the key to detecting in this information

obtained

form of the our system

could

vision

and

instruments

runway.

sensors

is poor.

approximate

are known

position

of the camera

motion

information

in an error

Although contrast However,

position

position

of the aircraft

only up to a certain

data

position

a few

to be as much from

and velocity.

in the estimated

and

image.

and it is likely that

camera

to

mounted

in the low resolution

once every second

their

system

and radiometric

causing

estimating

from a

system

from the GPS, and the orientation

potentially

result

using these

objects

are updated

and

on the

resolution

to locate

GPS data

objects

obtained

a computer

objects

model

from these

Knowledge

sequences

Wave imaging

airport

The data

feet off.

and

in fog, their spatial

is available

be missed,

Millimeter

runways/taxiways

of the images

image

was to design

from a Passive

have good response

additional

in monocular

goal of this research

are very low, and the quality we have

of objects

the

as

GPS/INS

Any inaccuracy

and

velocity

of the

objects. An alternate

approach

objects

with

known

world

sections

of runway/taxiways,

to obtain coordinates corners

an improved and

their

of buildings,

estimate position etc.).

of camera in the

image

This requires

position plane

is to use (e.g.,

an analytical

interstudy

of the

relationship

among

distances

between

accuracy

of the estimated

sensor

the aircraft

resolution

for computing

the camera

and

and objects.

camera

sensor

position

are useful

resolution

equations

in 3-D. We also propose

camera

images,

model

and

for accuracy

quantization.

resolution

the

to compute

the

in terms

a new algorithmic

pose due to plane

the required

of the

an analytical

and develop

camera

in deciding

the

We propose

position

error in the estimated

ical analysis

parameters,

of

approach

Such theoret-

for use in the design

of

SVS for navigation. Initial camera

experiments located

at a fixed

50 x 150 pixel sequence our

image.

not be continued

an aircraft,

• Camera

Hence, objects that

position

obtained

mechanically information

were encouraging,

scanning and

Although

further

from

the

pixel

it to obtain airport

results

research

a single

using

model,

obtained

a a

using

PMMW

sensor

reasons:

of pixels

could

not be developed,

such a sensor could

be safely

mounted

on-board

and

from image-based

was not better

in monocular

We describe

image

for this experiment.

at this time whether

the remaining

the objects

a test

and then

camera

with an array

pose estimation

sor images

in space

the

images

using

for the following

camera

• It is not clear

point

was simulated

on these

• A practical

conducted

Using

of 30 frames

algorithms

could

were

part

images

were moving

a new recursive

than

the position

of this research obtained

features

information

is focused

by a moving

on a background motion-based

which

video

provided

on detecting camera.

is either

segmentation 4

using low resolution

sen-

by the GPS.

independently In this work,

planar

algorithm

PMMW

moving

we assume

or is piecewise for segmenting

planar. images

into regionscorresponding to

independently

moving

objects

and

estimating

their

motion

parameters. Apparent first.

motion

Since the

the image. grouped

camera

Planar

square

model

as outliers

hypothesis

surface

motion

Line features types board

a landing

to the 3-D objects least

the estimated in motion features

and estimate

This

solved tration requires

parameters,

the

problem

using

digital

knowledge

to be moving

from the runway

a single

for the runway

violating

planar

is computed

the planar

motion

and more reliable

sequences

plane Using

obtained

in are

surface

in

using a least

model

are detected

filter-based

and velocity

and

are computed

features.

corresponding

Two

mounted

on-

line features

due

by matching

motion

approach

of the object

point

from a camera

camera

in the image

Kalman

than

markers

parameters the known

line features

the position

information

at and

to the 3°D object

is used to track

in 3-D assuming

that

the

line

the object

runway.

addresses

segmentation,

from

flow, is computed

appears

resulting

due to the runway

Runway

A recursive

on the planar research

- - line features

in two frames.

are tracked.

is moving

based

plane

are resulting

parameters

in image

in the scene.

two line features

flow vectors

to be prominent

are detected

aircraft

the optical

as obstacles.

are considered

of line features

they

Flow vectors

and are identified

called

(i.e., the runway)

motion,

that

model

approach.

in images,

background

of camera

on the

fitting

patterns

is moving,

Using knowledge

based

motion

of brightness

feature

many

tracking,

of motion-based warping

difficult

dynamic

scene

segmentation

and other

of the warping

problems

statistical

parameters.

5

in computer

vision,

analysis,

In the past,

etc.

using

the Hough

methods.

The

It is mainly

method,

digital

used

such

as motionresearchers image

warping

to compensate

regis-

approach for the

dominantmotion; all featuresviolatingthe dominantmotionare detectedas outliers or obstacles.This resultsin warpingnoisein additionto the digitization noise.The Hough methodcanbeusedto segment andestimatemultiplemotionssimultaneously, althoughit is computationallyexpensive andits success andaccuracyaremainlydependenton knowledge of the rangeandresolutionof the parameterspace. Unlike the Hough-based approach,our approachdoesnot requireprior knowledgeof the rangeof parametervaluesfor the motionmodel.Instead,parametervaluesaxedirectly estimated.This estimationis refinedby removingthe outliersat everyiteration. Usingour algorithm,basicsegmentationcanbe donewithout knowledgeof the motionparameters, as opposedto the warpingtechnique,which requiresthe cameramotion parametersto computethe warpingmatrix. Todealwith the computationof multiplemotions,wepropose incorporationof split and mergetechniqueto the motion-basedsegmentation algorithm. The algorithmsdevelopedin this work havebeentestedusingboth real andsynthetic images.The syntheticimagesweregeneratedby the programdevelopedfor this report, aswell as a simulationsoftwaredevelopedat the NASA AmesResearchCenter. Several real imagesequences usedin this workwereobtainedby sensorsmountedon-boarda landing aircraft, and wereprovidedby the NASA AmesResearchCenterand NASA Langley ResearchCenter.

1.2

Review

of the

Literature

on

Obstacle

Detection

Systems

for Naviga-

tion

In recent

years, considerable

vision algorithms

effort has been put into evaluating

for these tasks.

This section

the feasibility

gives a brief overview

of research

of computer work done

at NASA The

and other detection

to the runway the runway

institutions of runway

where

the camera

an image

model

database.

during

flight,

but

they

from the captured

defining

difference model

image

A cost function minimized

reasonably

good

the runway

range

models

accurate

the

map

image

acquired

using

the sensor

buildings,

etc.).

landing

the position the position

- - namely

is expected

sensor

sensor

input.

of of a

by the

input

and

The choice

and speed of convergence edge-based

the viewed

of

model,

scene

image

of such an

area-based

model,

with an expected with the camera

scene. image

is

position.

detection

is supplied

with a moving

for all points

is segmented A Kalman

that

from the actual

of the expected

camera

of obstacle

if either

estimation

scene, as well as the choice of a cost function

to the accuracy

the matching

the problem

runway

automatic

to determine

and the expected

- - are used to compare

more

position

of

obtained

airport

are used to determine

is computed

the actual

different

quantifying

on a model image

the

to permit

have attempted

of the sensor

between

solved

sky, runways,

and

specific

[4, 61, 69].

of fit, axe crucial

model

to obtain

In principle,

camera

enough

lines, etc.,

of authors

for generating

In [102] three

and texture-based

accurate

runway

have been based on some model of environment

the goodness

algorithm.

of the

properties

in [30] is based

with the expected

position

such as points,

[5, 60, 73]. The position

a reasonable

is compared

for navigation.

in [55, 77] using

Work described

axe not considered

in [4, 61, 70]. A number

some perceived

lines).

the known

detection

attempted

INS and GPS can provide

features

Most approaches sensor

image

using

On-board

[101]. Image-based

sensor

lines has been

(e.g., the use of parallel

by generating

a sensor

in the area of obstacle

and

in front

recognized

filter-based

camera

recursive

can be completely

of the

camera,

into constituent estimation

or the

parts procedure

(i.e., for

estimatingrangeto featurepointsin the imageis explainedin [95,108,110,111]. Image regionsof size11 x 11pixelswith high variancearedetectedasfeatures.The initial range of thesefeaturesis computedusingmotionstereo.The featurepositionsin the subsequent framesare predicted,and Kalman

filter.

and images

The

estimation

least squares state

method

parameters.

[120] an optical Image

regions

background optical are

camera flow-based

over

an extended

and velocity

obstacle

runway

is a planar

motion

using

the camera

moving originally

few frames using

of the moving

incremental

tracking

in the world

weighted

using known analysis

linear.

moving

by Nelson These

in In

objects. from the

[87] on the

detected

algorithm

coordinate

camera

described

are segmented

proposed

to

in [109].

independently objects

images

are assumed

to be piecewise

of the sequence.

a model-based

objects

image

motion

indoor

a

in [97]. Motion-based

objects

plane

using

regions

described

is estimated

in

using

filter.

detection

image

of stationary

filtering

from the initial

images

are compared

is used for detecting

the constraint

frames

stereo

the epipolar

by assuming

are refined

for both

in [118] uses a simple

to the independently

subsequent

Kalman

extends

are reported

estimation

the position

approach

corresponding

[56]. Position

helicopter

motion

range

points

In this work, all obstacles

to handle

described

for estimating

flow computed

The

for passive procedure

by applying

tracked

and the results

was modified

This algorithm

[14, 8] to general

to the feature

from the actual flight test.

algorithms

range

of ranges

is tested

The algorithm

and stereo-based The

algorithm

acquired

be stationary.

the estimates

surface

algorithm

described

and computes

warping.

The

in [115, 44] assumes

the residual

algorithm

flow by compensating

is tested

flight.

8

that

using

images

the obstacle-free for the camera obtained

during

a

1.3

Organization

This section in this

provided

report.

Even

in monocular

image

a brief introduction

to the research

though

is based

this report

sequences,

the

total

from two different

sensors.

Detecting

modalities

different

processing

require

using PMMW

sensor

images.

on the main

research

obstacles

dealt

in images

techniques.

Subsequent

problem

theme

obtained

describe

addressed

of object

with processing

The next

sections

and objectives

detection

images

from sensors

two sections

obtained of different

explain

our work

the work done using video

image

sequences. Section

2 describes

orientation

of the

described

camera

parameters. - - the

Passive

(FLIR)

sensor,

and

quantitative

in Section

array

approach

Section into regions

for error

in camera

Wave

Definition

error

features.

of its position

for detecting

were obtained

by the NASA

was not completely

using this sensor

image-based

the

in the

An analytical

and orientation,

_parameters

(PMMW)

sensor,

Television

(HDTV)

position

and

model other

sensor

Looking

is

sensor

are also derived.

the Forward

and

Three Infrared

- - are evaluated

and

are presented.

3. Results

provided

High

using

for computing

in terms

Millimeter

the

results

A model-based

sensor

Equations

sensors

model

estimated

for an on-board

internal

camera

the analytical

modality.

4 describes corresponding

our

in a PMMW

using an image

Langley

developed

objects

Research

by NASA's

The remaining motion-based to different

sequence

Center. commercial

sections

obtained

and

image

partner,

algorithm estimating

is described

from a single

Since a practical

of this report

segmentation motions

sensor

PMMW

we could

use video

sensor

not continue

images.

for segmenting their

pixel

motion

images

parameters.

An algorithmfor computingthe optical flow from imagesequences is described.A planar motionmodelanda linearalgorithmfor recoveringthe modelparametersfrom vectors

are presented.

dent motions

A recursive

is described.

algorithm

The results

for estimating

obtained

using

single

both

and

synthetic

optical

multiple

and

flow

indepen-

real images

are

presented. Algorithm 5.

Details

image runway.

are given

sequence

future

the plane parameters

for a Kalman

and estimating

The feasibility

axe presented suggests

for estimating

the position

of these methods

in this section. research

filter-based

Section

using line features approach

and velocity

was tested 6 contains

topics.

10

is described

for tracking of 3-D objects

using real image a summary

Of the

in Section

line features

in the

moving

on a planar

sequences;

the results

entire

research

and

2

Sensor

Sensitivity

A Synthetic

Vision

craft during

landing

acquired

System

sensors

navigation (INS)

instruments

are known

used),

are more camera

accurate

position.

landmarks sponding requires (i.e.

2.1

An alternative

positions

in the

and orientation

the sensor

Sensor

computers.

Positional

represent

data

data

from these

a better

from

on-board

relationship

among

instruments

information

of the

their

corre-

position.

This

positional

and

accurate of certain

and

camera

the camera

characteristics,

navigation

more

of buildings

estimate

Naviga-

the type of the instru-

for obtaining

corners

cameras.

by the Inertial

upon

they are useful

in 3-D), the sensor

Sensitivity

of mapping in both

The

resolution

limited

the scene

The

the air-

and images

and low light TV

is provided

Since

database

is to use the 3-D location

to obtain

is quantized

to an M × N array

infrared

and control

parameters

the relative

distance

and objects.

is a process

plane

second.

data,

airport

(depending

of runways/taxiways,

of the

to navigate

(GPS).

accuracy

approach

image

study

wave,

System

every

the GPS-based

such as intersection

position

Imaging image

than

once

ability

by the system

only to a certain

an analytical

between

needed

Positioning

axe updated

Calibration

uses an on-board

such as millimeter

and Global

and

and

the pilot's

operations

information

tion System

ments

for enhancing

and taxiing

by external

Additional

Evaluation

of points introduces

Evaluation

a large spatial

called

3-D

scene

directions

of the sensor pixels.

significant

Such

amount

11

onto a small

to facilitate requires

image

being

plane.

processing

that the entire

an array of error

2-D

scene

too small

into computations

This

2-D

by digital be mapped

to adequately involving

the

locationsof imagepointsandfeatures.Anothernatural outcomeof the imagingprocessis the lossof depthinformationresultingfrom mappingof an infinite numberof pointson a line-of-sightontoa singlepoint in the imageplane.As a result,reconstructionof a sceneby computingthe depthof scenepointsusingimage-based featuresis a challengingproblem. The inverseprocess,knownas the cameracalibrationproblem,is to computethe camera positionby triangulationbetweenknownscenepointsin the world andtheir corresponding positionsin the imageplane. Nearly all of the methodsfor solvingthis problemrequire only onemonocularimagewith specialmarkswhichcanbeman-madepoint featuresor line features[42,46, 54,64,69,113].Theaccuracyof suchcomputationalapproaches, however, dependsuponthe camerageometryandthe quantizationof the imageplane. Most analysisfor obtainingthree-dimensional positioninformationof the scenepoints are basedon the triangulationsystem[13,76]. To computethe depth to a scenepoint, triangulation must be performedusing its projectiononto two images,captured either simultaneously,by two camerasseparatedby a basedistance(asin the caseof stereo),

or separately,

stereo). other the

This requires image.

the spatial

computation

Assuming

3-D information

correct

using

quantization

Equations vision

by a single camera

established

percentage

error

occupying

the

shift

a correspondence

in the

the next step is to recover

of such a computation

depends

on

plane.

in [76] present

between

(as in the case of motion

and

been found,

The accuracy

error in distance

in distance

positions

in one image

have already

triangulation.

for calculating

system

of features

matches

of the image

at two different

binocular

measurement

measurement

a worst

case error

is inversely

the two images

and

12

between

analysis,

proportional

directly

an object and

and a stereo show

to the number

proportional

to the

that

the

of pixels object

dis-

tance. BlostienandHuang[13]havedeveloped equationsto determinethe probabilitythat a certainestimateis within a specifiedtolerancegiventhe camerageometryof a stereo setup. Error equationsdevelopedin [31,32] for determiningthe optimumline width for visualnavigationof anautonomous mobilerobot givethe percentage oferror for anysensor geometry,line width, and error conditions. Even thoughtheir analysislookssomewhat similar to our work,the basisfor their analysiswasmeasurement of line width. In this research, weestimatedthe aircraftpositionby trackingknownstaticscenepoints in the imageplane. Sincethe accuracyof this estimationdependson varioussensorparameters,wewantedto developan analyticalmodelthat could relatethe accuracyof the estimatedsensorpositionto the sensorpositionalparametersandattributesofthe captured image.Thesensorpositionalparametersincluderange,crossrange,altitude andpitch, roll, and yawangles.Sensorimagingattributesincludethe numberof pixelsin the imageand the optical angularview (measuredin degrees). Weassumea pin-holecameraand,therefore,ignorecameralensdistortion andoptical non-linearities. We alsoassumethat the problemof featurecorrespondence was solved usingsomecorrespondence algorithmandthat the correspondence wasaccurateto a single pixel. Hence,the analysisdoesnot dependon the type of algorithmusedfor establishing featurecorrespondence. Any errorin computationwasbasicallydueto spatialquantization of the imageplane. In this analysis,weprojectedthe imageplaneonto the ground,and the pixel (p,q) area

represented

the minimum next

in the image

plane

by the pixel). change

pixel in the image

in a camera

was modeled

We defined parameter

as a patch

the sensitivity that

plane.

13

would

on the ground or the error

plane

(ground

in computation

move a fixed ground

point

as to the

s

i

,/

Z

Ceater of gravity

¥ s

I/

I

/J

i i

/s

¢ PX Lateral World

Coordinate

System

pitching

Figure

2.1.1

Imaging

Throughout

center

analysis,

scribe

gravity. dicular

of reference

The position coordinate

axes

(pitch,

of the aircraft

with pitch

axes

we assume

that

axis

roiling

an imaging

the

sensor

sensor

position

or

axis

and aircraft

of notations

in Fig. 1. The figure shows

roll, and yaw) is defined

passing

with respect

and horizontal

is located

at the position

of the earth.

the basis of the system

axes X, Y, and Z. The image

axes of the airplane,

Fig. 2 shows

forms

is shown

to the rolling axis, with its vertical pitching

Longitudinal or yawing

body

the effect of curvature

axis that

of the sensor

perpendicular

world

1: Airplane

Hence, we can use the terms

We also neglect

the position

mutually

Vertical

for convenience,

of gravity.

interchangeably. The system

_

axis

Geometry

the

aircraft's

or

through

an aircraft

with three

the aircraft's

center

to the three

plane is assumed

axes coinciding

used to de-

mutually

of

perpen-

to be perpendicular

with the yawing

and

the

respectively. situation

during

landing,

angle 8, zero yaw, and zero roll angle.

14

where

the aircraft

is at (Xc, Yc, Zc),

Let c_ = 90 - 8. The

field of view of the

camerais determinedby two viewingangles:Aa

defined

right

extent

angles

extent.)

to Aa.

Even

(Aa

though

determines

the

image

area

captured

by the sensor

Aa,

A/3, and

various

pixel

in the image

pixel-patch

2.1.2

the vertical

obtained

sensor

Consider

a point

coordinates

whose

a rectangle, and

orientation,

on the ground

plane.

the ground

area etc.

depend Note

on

that

a

We refer to this as a

Analysis

feature

which

of this feature

has been detected

in its position

feature

at the same pixel (p, q). Hence,

by a certain

will always

give the

same

the camera

position

is large enough

We define

this minimum

Note that

this is a measure

position,

image

(p, q) in the image Let Nx and

at some pixel (p, q). Let the actual

be (P, Q, 0). Since a pixel represents

could change

camera

change

amount

a camera

pose for nearby for the feature

in camera

of camera

size in number

of pixels,

unless

of the

triangulation the change

in the neighboring

as the sensitivity

the

in

pixel.

of the camera.

estimate

and is a function

resolution,

and the pixel

of the location

plane.

Ny represent pixels

the number

are numbered

of pixels

N___ 1 in the horizontal 2

direction.

the bottom

of the patch

right corner

--_,...

in the vertical

-_

pass through

positions,

position angular

the image

by passive

to be observed

The

0,

camera

world

on the ground,

while still retaining

displacement

of accuracy

a patch

pose estimation

respectively.

' ......

and Aft its horizontal

side length

like position,

to a patch

camera

2

is always

as 0, and Af_ at

(See Fig. 3).

Sensitivity

camera

ABCD

parameters

plane corresponds

of the image

by the sensor

is a trapezoid

other

in the same plane

0,...

15

-_ -

The rolling

and

horizontal

directions,

1 in the vertical

direction

axis of the plane

is assumed

on the ground

plane,

which

and

corresponds

to

Z

C

X

A

D

Figure

2: Image

obtained

by the sensor

16

is projected

towards

the

ground

,Ep

(X3. r3)

r.q

.41.w

Figure 3: Ground area covered in the actual image

to the

center

pixel

The coordinates estimated

in the

by the sensor.

image

of the reference

by the following

plane. corner

Other

X

=

Xc + Zctan(a

Y

=

Yc+

plane are obtained

Since a pixel-patch corners

become

the

pixels

corresponds

are referenced area

covered

to a pixel

in a similar by a pixel

manner.

(p, q) can be

relations:

Z_

rolling angle ¢, the ground

(p, q) in the image

small trapezoid

of the ground

cos(_

For a non-zero

Each

is referenced reference

+p-_-) hc,

+ p_)

coordinates

with replacing

=

pcos¢-qsin¢

q'

=

psin¢+qcos¢

of its three

(X', yI)

which

correspond

to a pixel

by (pr, q_), where

(2)

right corner

neighboring

17

(1)

(p, q) in the equation

pr

by the bottom

A_

tan(q-KT--)

pixel

of the pixel, the other patches,

as shown

three

in Fig. 4.

p

(X3 '. 1"3)

T

(._'. r2'_

(p+ l. q+ l)

I

(p+ l, tl)

(XI" YI') (X_', r_ ')

(p,0)

(p, q+ l)

Figure 4: A pixel

Thus,

the

1, where

four corners of this pixel-patch (p, q) are replaced

by (p',q'),

p_ !

qi

where (Pl,ql)= Eq.

1 gives

ground

point

(P, q), (P2,q2)

corresponding

of the

parameter

that

imagery

parameter.

(X_, Y/),

the ground

i = 1, 2, 3, 4 are obtained

=

pi cos ¢ - qi sin ¢

=

pi sin ¢ + qi cos ¢

between to a pixel

sensor.

the partial

by using

Eq.

such that

the camera (p, q).

This

would move a fixed ground

this by taking

towards

(3)

--- (P+ 1,q), (P3, q3) -- (P+ 1,q+

the relationship

sensitivity

obtain

(p, q) projected

derivative

parameters

We are

is defined point

1), and

--- (P,q+

1).

(Xc, Yc, Zc, O, ¢) and

the

now interested

as the

minimum

(P4,q4)

in computing change

in a camera

to the next pixel in the image

of X{ and Y{ with respect

the

plane.

We

to the corresponding

For example, Dxx

= OX_ y OX'----_' Dx-

18

OY_ = OXc

(4)

This derivationis an approximationto the amountof Xc.

Thus,

we estimate

to Y_ (which

that

the amount

define the corners

of change

of adjacent

pixels)

Note that

SY xc = oc, as expected.

in a similar

manner.

in the direction image

sensitivity

sensor

plane

horizontal

function

is a function

is inclined

directions,

Equivalently,

in Table

positional

the

in

X{ to X_ or YI'

(5)

--Dx_ to other

1. In general,

parameter

ground

area

accuracy

of estimation

represented

of estimation

angle to the number

plane,

parameters

S_ stands

j computed

the border

parameters

is defined

for sensitivity

at pixel

and, hence,

position

parameters.

pixels.

of the sensor

is a function

of sensor

varies

is a function using

For a given

at the top half of the sensor by these

and sensor

the sensitivity

of sensor

as well as other

are observed

as we move towards

sensor

the sensor plane

accuracy

that

of various

to the ground

along

of pixel position

using features large

i due to the sensor

to change

change

as

with reference

are summarized

in X_ for unit

(p, q) in the

plane.

Sensor the

These

Sensitivity

in Xc in order

Sxo-

'

change

of pixels in the image.

19

in the vertical

ground

truth

range,

are less accurate

in the horizontal

Since

of pixel number

Also, for a given p, the

characteristic

attitudes.

data

and (p, q). is a

the estimation because

accuracy

of the

decreases

direction.

In summary,

the

and the ratio

of the sensor

view

Sensor

Sensitivity 2Zc sin(cos

cos(2a+

_

at (p, q)

Sensitivity

_b_)

((2p-{- 1) cos _b-2q

at

sin(_) sin ¢))+1

2Zc

cos(2a+

_

sin(2a+

_)

)+ i

CO oo

O0

S Y

Y_

z_

S xZc SY Zc

Table

1: Sensor

positional

2O

sensitivity

equations

(0, 0) ¢ = 0

2.1.3

Quantitative

The

sensitivity

Results

analysis

sensors

at six different

tivities

SxXc, S_,

are plotted

proaching feature Thus,

6, and

section

for landing).

range

applied

to three

different

are given in Table

2. Sensi-

(i.e., p = 0, q = 0) for various In all of the

is 0 feet (typical

Note that

horizontal

was

of the sensors

7, respectively.

is 0 °, and cross

the

velocity

stationary The

sensitivity

also improves that

for the objects

that

objects above

than

are located are located

values

sXc is larger

pixel before

above

than

when

sensor cases,

positions

pitch

angle

an aircraft

is ap-

Szr at (0, 0) and, hence,

it moves

to the next

GPS

number

In addition

the accuracy correspond

of camera

for the

sensor

is moved

unless

that that

vertical

a

pixel.

to the

highest

frames,

estimation

pixel

ground.

resolution.

It becomes

poor

axis (top of the sensor As expected,

accurately

that

of camera

sensor

by knowing

or by using more

i.e.,

the position the position

21

estimation Note

by motion

feature

this analysis

by any algorithm

scene points.

state

is employed.

can be obtained

to being useful in sensor design,

to known

the

of

are closer to the aircraft.

the accuracy

improvements

state

with

closer

more

a high resolution

of image

sensor

at the far end of the vertical

can be computed

indicate

potential

is best

at the far end of the runway).

on the ground

results

the

a large

minimum.

as the

of the aircraft

do not consider

that

previous

Characteristics

move to the next

for the features

using

in the

only S Zc x is important.

Sensitivity

better

5,

the runway would

Discussion

SzX, at the aim point

in Figures

As expected,

and

described positions.

and

is -3.0 °, roll angle

and

that

points

that

would these

stereo than

be no results

techniques the required

will also help us evaluate uses image-based

features

Sensor

Positional

Parameter

Sensor

Location

Range

Altitude

Threshold CAT II-DH

in ft. 0.0 908.1

in ft. 50.0

HDTV

100.0 200.0

FLIR MMW

CAT I-DH Middle Marker 1000' Altitude Outer Marker Table

2816.2 4500.0 18081.1 29040.1

Sensor

Tgpe

Characteristic Pixel

Field

(H x V) 1920 x 1035 512x512

(H x V)deg 30x24 28x21 27x22

80x64

288.2 1000.0 1574.3

2: Sensor positional

parameters

and sensor

characteristics

100f

5OO

0.05

Figure

0.1

, 0.15 Reso_u_en

5: Sensitivity

0.2 in degre4Vl_xe4

0.2S

in the direction

22

i 0.3

0.35

of range

of View

180 t60

120

i: 80 4C 2O

o, O._

Figure

6: Sensitivity

0.15 0,2 Reso_utt_ in _xal

0-245

0.3 0.35

in the direction

of cross range

20O

..E

L

0_

Figure

0,(_

0.1

7: Sensitivity

0.15 Ruo_on in _

o2

0.25

in the direction

23

0.3 0_._

of altitude

2.2

Camera

Camera and

Calibration

calibration

orientation

extrinsic tation include

They

is the problem

- from

parameters,

can be divided

provide

These

and

placed

[130],

scene

points

camera

Fischler

and

of rectangles

of pixels

and

pose

position

intrinsic

in the image are

location

into two categories:

the camera

and

for the purpose

and

and orien-

parameters,

which

plane.

available

in the

literature.

Tsai

their

[39], and

of the rotation [23] used house

of unknown using

parallelepiped

in [113]. For calibration

image

the scene points, research

and translation corners

as marks

with known dimensions mounted

24

or from certain

landmarks

points,

triangulation

or by developing

Liu et al.

vector

of the

object,

Haralick

as the calibration

points, by Wolf that

are separable.

[46] used the

whereas

and

[70] showed

camera

view angle

on an autonomous

image

was done

[43].

the camera

as a calibration

of a camera

man-made

the corresponding

and

and

from certain

on this approach

Ganapathy

size to determine a cube

arise either

is done either by applying

corresponding

Earlier

pose has been to use point

of calibration

Estimation

that relate

Bolles

the camera

point or line features

parameters.

et al.[21] proposed rectangular

system,

the camera

exist in the 3D scene.

the computation Chou

in units

- including

can be grouped

regarding

to estimating

in the scene

a set of equations

the

information

parameters

into two classes:

line correspondences.

solving

parameters

for estimating

approach

to known

camera

world coordinate

most common

that already

Features

of determining

and scale factors

of methods

objects

Image-based

Camera

to a reference

focal length

• The

images.

which

with respect

A number

from

corners

parameters. a method object

Chen that

uses

is proposed

land vehicle

running

on an outdoor road,Liu andDeng[66]usedroad boundariesas calibrationobjects. Fukuchi[42]usedspecialman-mademarkswith certainconstraintsfor determining the positionof a robot usingstandardshapedmarks. • Objectswith curveshavebeenusedfor cameracalibration,suchasa planewith conic or polygonarcsas in [48]and semi-circlein [65]. Mageeand Aggarwalproposeda methodthat usesa calibratedsphere[75]. All ofthe aboveprocedures involvetwo steps.Thefirst stepis to locatefeaturepointson the imageplanethat correspond to the known3-Dpoints,andthe secondstepis to formulate and solvea set of equationsthat relatethe scenepoints and the imagepoints, thereby satisfyingcertainconstraints. Thesecomputationalapproachesassumean ideal pinhole cameraandmodelthe l_ixelsaspointsof insignificantdimension,largelyignoringthe error introducedby the imageplanequantization.Moreaccuratecalibrationof the camera,or a priori knowledgeabouttheerrorintroducedby theimageplanequantization,isquitecritical in variousaspectsof computervision like objectrecognition,scenereconstruction,robot navigation,etc. A brief reviewof someof the pastwork in error analysisof triangulation to imageplanequantizationis givenin the next section.

2.2.1

Error

Quantization outcome problems sensors.

due

to Image

of the image

of various

impact

Quantization

plane as well as the intensity

computational

are particularly The

Plane

approaches

levels have significant

to solving

important

when

the

of quantization

error

on computer

25

images

computer are

captured

vision

vision using

impact

on the

problems.

The

low resolution

was addressed

as early

as

1969.A report by Hart on stereo-scopic calculationsdiscusses sensitivityof triangulation processto pan, tilt, and quantizationerrors[49]. McVeyand Leedevelopedequationsfor measuringthe worstcaseerror in calculatingthe distancebetweenan object anda stereo vision system[76]. The navigationsystemreportedin [31,32]usesa singlecontinuousstrip paintedon the floor markingthe robot'sroute. Forsuccessful navigationof the robot, an importantdesign considerationwasthe width of conditions

on the

line width

on the floor.

The

effect

stereo

set-up

equally

three

was

in [13, 85].

method,

in this section,

points

and the corresponding

consider

four

scene

points

focal

point

These

the effects

plane,

works

of object

assumed

formed

the probability

of various

to determine

determination

the volume

a new approach

is based on the fact that

size of the

image

pixel,

lines-of-sight

pixels

that

by the

the

distribution

error

the optimal

location scene

lateral

using

point

was

[13] or axial of the

through

and their

corresponding

image

is expected

to lie. The

pose.

of proper

We propose

for determining

errors

[85] in all

the error in the camera

two or more lines-of-sight meet at a single point

for each scene

passing

in the camera selection

on the

derived

we describe

Our approach

to finite

and

analyzed

directions.

state.

Due

within

study

in the image

quantization

to be everywhere

component

Later

plane

analyzed

triangulation

That

of the line, as seen

of image

likely

stereo

width

the line.

volume

to minimize

point

the four pixels

and

scene points - and also the number

26

pixel, Two

within

is proportional

by considering

the scene

as the focal point.

of each pixel.

in a polyhedron

of this polyhedron this volume

known

the corresponding

corners

result

connecting

we can or more

which

to the error

more points.

of scene points - reduces

the

the region

The of

uncertaintyin the determinationof the camerapose.This analysiscanbeusedto determine goodcalibratingpoints,their distribution,andthe numberof scenepointsrequiredto stay within the allowedrangeofcomputationalerrorfor aparticularvisionproblem.In situations whereit is not possibleto determinethe calibratingpointsin advance,this analysiscanbe usedto computethe error in the estimatedcameraposition. Our analysiscan be usedto designalgorithmto selectgoodfeaturepointsfor calibrationfrom availablepool of feature points dynamically.

2.2.2

Analysis

Determination minimum

of Image

of camera

number

(I1, I2, I3,..., requires

Under

ideal

Basic

all the

finite

scene points a polyhedron

as shown

for camera

state

world

coordinate

effects

number

within

in Fig. 8. The

parameters system

the

in the

within

which approach

its orientation

focal

image image

coordinate

presented

with respect

image lens.

and

lens

there

can

Due

to the

onto the image any two such

lines of sight form

of the camera

of the center

points,

When

of these

here computes

points

camera

plane

system.

the finite sized pixel. the intersection

image

of the

can be projected

the focal point

27

of the

a certain

corresponding

point

corresponding world

requires

corresponding and the

of quantization

(i.e., the 3-D coordinates and

points

through

onto the image plane,

volume

scene

a given scene point

of points

features

Sn) and their

and their

for the camera

pixel, however,

are projected

anywhere,

the

set of scene points

an infinite

of finite

of these to pass

Error

using point-based

($1, $2, $3,...,

triangulation

position

size of an image

plane through

parameters

(ignoring

for a given

be only one unique

Quantization

lines-of-sight

conditions

distortion),

external

of scene points

In).

points

Plane

is expected the range

of the image

to the reference

to lie

of values

plane in the

axes that

would

w v

SL &?: $¢¢_t Pinata & tZ' t.mte._,.ts Xc , Y¢.Z¢ : I_

plar_ t_a tstv

itc

Figure

satisfy

8: Lines of sight

the constraint

As explained of known volume

scene within

extrinsic

point

points which

is located

image

the

Hence,

concept,

The problem of error

there

the camera

position

and

distance and that are ranges

image

system.

f (focal length)

quantization

of values

of values

for each camera

as a result,

28

where

plane

that

parameter problem

a stereo the region

system

this

the focal along

point

the

Ii in the

would

the mathematical

from the well known

and,

camera's

that

to image

parameter

of finite

the

clear

of the image

we present

using

Since

it is known

Si corresponds

of the objects

a polyhedron

it is not

However,

for each camera

In the next two sections

basic triangulation

to lie.

in front

the scene point

is fixed and known

forms

are uncertain,

coordinate

here is different

points

lens is expected

orientation)

the 3-D position

plane

due to image plane quantization,

of the

to find the range

addressed

in computing

point

due to image

of the focal point).

corresponding

in the world

the above two constraints. of the above

focal

(position

axis of the camera, plane.

section,

.and their

is at a perpendicular

optical

a polyhedron

for the valid location

in the previous

parameters

polyhedron

forming

satisfy

treatment

separately. of determination [13, 85], where

of uncertainty

for the

object locationis well definedby the camerageometry.However,in our problemscenario the location of itself

is uncertain.

problem

being

approach

Fig.

9

systems

addressed

shows

roll angles

problem

their

discussed

analysis

not make

geometry

the coordinate

a priori,

any

since the camera

in [31, 32] is somewhat

is limited

used

to a particular

assumptions

about

of the

camera.

position"

is in front

the coordinate

application. the

kind

system

in the image plane.

the position

image

plane.

given

of the center

In the above

numbered

Z = 0 to be the

Let ] be the focal length

the u and v directions

as (0, 0). We make

• This work analyzes scene points

of the image

geometry,

• A pinhole

the following

camera

of error

model

plane,

assumptions

in obtaining

on a single quantized

image

plane,

"camera and the

the gimbal

the height

of the

respectively,

M × N be the number

respectively,

with the

center

of

pixel

in this analysis:

Image

pose by triangulation

plane

of the camera

[59], thereby

29

and

and

the tilt, yaw,

we assume

Therefore,

the camera

plane.

in computation

is assumed

plane.

of the camera

of the image

the accuracy

to be the only source

ground

coordinate

in [45]: the term

is Zc. Let A u and A v be the size of a pixel along the u and v directions,

along

of outdoor

system,

and (8, ¢, ¢) are, respectively,

sensor

plane.

The

(Xc, Yc, Zc) is the

offset

pixels

to be zero and

system

coordinate

center

on the image

position

close to the

Two separate

the world

We follow the convention means

of the

in this analysis.

(X, Y, Z) represents

in the world coordinate

or "sensor

focal point

system

(u, v) represents

position

position"

here,

is not defined

for calibration.

the

are shown:

camera

the

in this work does

available

the coordinate

of uncertainty

Although

taken

scene points

and

the region

ignoring

quantization

of 3D

is assumed

position.

camera

lens distortion

and

¥

Zv

% /°

Yc

X

Figure

other

• The

optical

The

of detecting

to have been

calibration

process

focal point

Assuming

formed

Analysis

solved.

- especially

• Assume

the

plane

calibration

when

corresponds

coordinates

to a given

scene

of the scene points

plane,

point

is

used in the

as is the case in any practical

to be in front of the lens center introduce

significant

the scene points

position

coordinates

that

to be accurate.

out in the following

camera

for the world

The world

is in front of the image

since this might

is carried

the image point

are assumed

the image

analysis,

geometry

non-linearities.

problem

assumed

9: Camera

vector

error

is not appropriate

in the volume

camera. for this

of the polyhedron

are close to the camera.

four steps:

to be (Xc,Yc, Zc, 8, ¢, ¢), and

of the corners

of the image

process.

3O

pixels that

develop

equations

are to be used in the

* Formthe planeequationsthat containthe scenepoint andthe two adjacentcorners of the pixel corresponding to the scenepoint. • Computethe coordinatesof the focal point in the world coordinatesystemfor the assumedcamerapositionvector. Verify that the focalpoint lieswithin the polyhedronformedby the intersectionof the aboveplanes. The mathematicaltreatmentof the abovefour stepsaredescribedin detail below.

A.

Computing

We assume

the world coordinates

the

image

plane

of the corners

to be at the origin

of image

of the world

tilt, roll, and yaw, as shown in Fig. 10. In this situation, XZ When

plane

and the world

the camera

coordinates

is rotated,

tem can be computed

of a pixel

the coordinates

by applying

necessary

(i,j)

pizels coordinate

the image

in the image

of the pixel

system,

plane coincides plane

are (iAu,

(i, j) in the world

point transformations

with zero with the O, jay).

coordinate

[45] to the original

syspoint

as given below. Xij

iAu

o

=R

(6)

jay k

1

J

k

31

1

¥

P

x

Figure

10: A point

(i,j)

in the image

plane

rotated

about

(X, Y, Z)

where Too

rOl

to2

to3

rio

rll

r12

r13

T20

r21

r22

r23

r30

r31

r32

r33

(7)

R = RcP_R_ =

The

individual

following

three

elements rotation

of the

rotation

vector

where

=

by premultiplying

/cos0//loo0//c o0/

C stands

0

0

0

S_

0

C_

0

0

-Se

Ce

0

0

0

0

1

0

0

0

1

plane,

plane by applying

S#

1

for cosine

- S_

=

R_

and S for sine of the respective with computing

we can think proper

Re

Ce

0

Since we are concerned in the image

be computed

the

vectors.

0

R_

can

rotation

the world

of the imaging vectors,

process

32

0

0

0

0

1

0

0

0

0

1

(8)

angles.

coordinates as rotating

with the center

C_

=

of the corners every

point

of the image plane

of any pixel in the image

as the origin

of

the world

coordinate

system

coordinate

system

by applying

the camera

position

[45], and then translating

vector,

proper

translation

the world coordinate

them

vector.

to a new position

Therefore,

of a pixel

in the world

for nonzero

(i, j) in the image

values

plane

of

is given

by Xij

iAu

., c _C

=R

(9)

+

1

1

where

-

The

3D coordinates

coordinate

system

_ dt are

considered

to be outliers,

where

dt is a threshold

value

decided

experimentally.

4.4.5

Algorithm

From

the above

for Recovery

analysis,

from a single planar the motion

model

model

parameter

small,

then

If there

surface

accuracy

outliers,

need

the motion

points

1. Compute

iterative

3. Compute

data

consider

at that

point,

the

the point

points

These

on the least

_ using N data

pixel using

points

the

theoretical

distance a threshold

for refining

flow the recovered

in the optical

flow is

of the parameter

noise

data

points,

before

reliable

square

fit method

vector. effect

called

estimates

in the least square (x, y) and

the

can be

for recovering

in the following

plane

optical

result

can have significant

is described

the gradient-based

in the previous

If d is above

these

of a pixel in the image

compute

Mahalanobis

If noise

systematically

based

optical

a good estimate

then

all flow vectors

can be easily used to recover

parameters.

parameters.

removed

algorithm

_ computed

flow.

errors,

model

the coordinate

(up, vp) computed

parameter

fit approach

can provide

with large

the model parameter

where

is no noise in the estimated

estimate

and

Model

ideal situations

from a set of N noisy flow vectors

include

2. For each

under

to the actual

of the recovered

model

model

If there

to be detected

Motion

a least square

will be identical

A simple

points

motion,

parameter.

are few data

obtained.

it is clear that

the least square

on the

of Single

four steps.

sense.

the optical

approach

described

flow (ut, vt)

using

Data flow

earlier.

the

model

stage.

d between

the practical

dr, consider

the estimate

94

in future

the data iterations.

flow and point

the theoretical

as an outlier,

else

.

If there

are no outliers

to converge parameter

and

the

data

4.4.6

these

which

and

described The

data

the

obtained

surface

least square

in this iteration

motion.

points

algorithm is the

If one or more outliers

and

repeat

steps

is said

best

model

are detected

1 - 4 for the

fit algorithm

breakdown

moving

in different

directions.

the optical

flow vectors

not follow the planar In [20], the plane containing

point

of flow vectors,

of these vectors

from a differently

of Multiple

remaining

could

object.

fails if more than

2) The

computed

using

that

an optical

are in

This can happen

different

although

by matching

a portion

in

resulting

objects

may not be perfectly

for

surface

to the few flow vectors

flow algorithm,

were first computed

was used to warp

the second

image to the first.

using the warped

image to detect

flow at other

points

points

from a single planar

come from

the scene

region

flow was computed.

points

of good

moving

planar; correct,

thus, may

model.

parameters

camera

in addition

data

the planar

Using the known

method.

axe resulting

be in error

number

50% of the data

for a least square

though

It is also possible

surface

Motion

can be used only when a sufficient

least square

1) Majority

many

Estimation

above

is the actual

two reasons:

image

parameter

the planar

ignore

Segmentation

are available.

motion,

in this iteration,

points.

The algorithm

error,

model

describing

in this iteration,

detected

motion

obstacles.

image.

Model

The residual

parameter

optical

In [115], the runway

and the plane parameters,

Large residual

pixels was assumed

to the current

flow vectors to be caused

95

by errors

thus recovered

flow was then computed

was assumed

images

due to obstacles

of a previous

to be planar.

were warped

were removed.

in the initial

model

and residual The residual

parameter

and

was used to improve the known moving

the model's

camera

average

motion

and

of the optical

from the available

computing earlier. and

that

a more general

mainly

problem.

motion

due

were computed

were tracked

We computed

the outliers

to obstacles.

Our

parameter

in the set of data

was due to the

model,

planar

algorithm

a

as obstacles.

the model

motion

using

by computing

flow were detected

flow from the estimated

the dominant

were

Features

and identified

of the optical

that

the outliers

parameters.

flow vectors,

the deviation We assumed

plane

In [44], warping parameters

flow. Pixels with large residual

In this work, we considered vector

accuracy.

points

by

as explained

surface

in motion,

can be used

for two

applications:

• By computing addition

the dominant

to detecting

for the planar

• Planar

outliers,

surface

motion

to describe Each

motion

optical

[53] and values expected

to vote

the object

if the range

approaches

in which

uses Hough

votes

between

optical

with the into the

bin.

space

resulting

from

the background

assumption

can be used.

uses the afflne motion the

six-parameter using

in the

multiple Hough

an optical

defined

a single

in clusters

96

model parameter

however,

to detect

for the model

All flow vectors

same bin,

planar

flow value computed

flow expected

In

to the scene is large compared

[1], his study

to a set of bins in the

the optical

the motion

the ground,

case piecewise

at any

and

in the form of outliers.

is moving.

objective

point,

are detected

also provides

work had a similar

the theoretical

associated

on which

planar,

flow vector

on the difference

our algorithm

As the camera

may not be perfectly

Adiv's

obstacles

model is a good assumption

to the focal length.

Although

motion,

rigid

model

motions.

space

based

flow algorithm

by the

six parameter

object

in motion

Hough

space,

where

are each

clustercorresponds to separatemovingobjects. knowledge

about

the range

parameter

space.

The higher

computationally

is then

4.4.7

Split

We assume

initially, derived

and

that

to be very

Merge

the runway

or the terrain

planar.

We started

applied

motion

model

failed then models diagram

we hypothesized

The

starts

model

A higher

the camera the runway

of the although

Hough

approach

resolution

Hough

space.

is moving

The outline

is either

to be perfectly

flow vectors.

to be piecewise

with the hypothesis

parameter

parameter The

the algorithm

converges

the deviation

algorithm

the ego-motion

1_, and k can be computed objects.

is computed

is refined

flow vectors.

moving

in which

to all of the available the runway

by computing

model

represents

resolution

If the

planar,

of the algorithm

planar,

initial

and applied

planar and

hypothesis

planar

is shown

motion

in the block

in Fig. 32.

are detected The

is used.

in the first Hough

by hypothesizing

to each of the pieces separately.

The algorithm surface.

detected

on the

the segmentation,

In [122], an adaptive space

depends

Algorithm

or is piecewise planar

Hough

the clusters

and the

the more accurate

expensive.

a low resolution around

of this method

for each of the parameters

the resolution,

it is going

is used where, space

of values

The success

that using

of each

by iteratively

of the camera from _. The

If the outliers

the least

then

and the structure outliers

are distributed

97

and

outliers

computed

model.

from

the initial

outliers

plane

of the background.

they

due to noise

set of

are detected.

the computed

are either

randomly,

from the

no more

to a single planar

approach,

the outliers

when

50% outliers,

belong

square

of the points

removing

is said to converge

with less than

all flow vectors

If

parameter

Note

that

54

or independently

can be simply

ignored.

If the

Feature detection

Optical

flow

computation

square

I

fit

Repeat for every region

Reject outliers 0

Split region No No For outliers

Merge

Figure

regions

32: Split and merge

98

algorithm

I

optical

flow vectors

object;

a separate

If the lution

least

are close together, motion

square

is completely

Piecewise

entire

image

The regions This

model

distributed

The

based

in this study plane

because

and,

we split

the image

to be piecewise

constraint

local regions

into four parts. region.

recursively

until

square

algorithm

merged

motion-based

the least

if the error after

Experimental

the

so-

planar.

local regions.

and the merging

error.

points

are

may not be well defined

merging

The

Where succeeds

is below

image segmentation.

iterative necessary,

least the

square regions

for a given

region.

In this algorithm were

split

Regions

are

a threshold.

Results

segmentation

algorithm

for detection

for the background/runway

is tested

3 of an image

then

for each of these

used for intensity

to each

technique

50% outliers,

it is not clear how the feature

therefore,

independently

model

sets of outliers.

local regions of size N × N over the

on the spatial

is applied

motion

to a single

distributed.

application,

4.4.8

than

can be computed

2. We follow the split and merge algorithm

then

from these

is assumed

to non-overlapping

parameters

on the image

or evenly

more

to belong

in two ways.

can be then merged

is not practical

with

The background/runway

can be applied and

can be considered

can be recovered

converges

can be applied

algorithm

they

parameters

fit algorithm

dropped.

planarity

1. The

model

then

using both

sequence

simulated

obtained

using

with

the

incorporation

and real image the

NASA

99

of obstacles

sequences.

simulation

based of the

on the split

Fig. 33(a)

software.

Fig.

and

planar merge

shows frame 33(b)

shows

the featuresdetectedin frame 3, and Fig. 33(c)showsthe optical flow computedfor the featurepointsusing our hierarchicalframework. Fig. 33(d) showsflow vectorsviolating the planarmotionmodelconstraint,whichhavebeenidentifiedaspotentialobstacles.The planecrossingthe runwayis very clearlydetectedas an obstaclein the images.(Fewfalse alarmsaredueto noisein the estimatedoptical flow.) Fig.34(a)showsframe50ofthe realimagesequence runway_crossing_new a camera

mounted

in Fig. 34(b). and

34(d)

4.5

on-board

Fig

shows the set of feature

In this section, that

algorithms. then

image

optical

flow vectors

the approach of the

in this image

are shown

flow algorithm,

as obstacles.

and

function

for which

it is being

the optical

planar

surface

of this motion

model

using

recovery

in motion.

from a sequence

The

the planar

estimation

contains

optical

motion

algorithm

algorithm,

proposed

of images

the gradient-based

are used to compute

system two sub-

flow method,

model

requires

as-

parameters.

determining

the

sub-aigorithms.

six components

propagating,

of our motion

parameters

described

dating,

by applying

detected

flow using our optical

identified

flow is computed

of each of these

Unlike

a single

model

of the performance

performance

optical

the performance

contains

the motion

optical

Evaluation

in terms

we evaluate

First,

these

vectors

Features

from

Characterization

the

for recovering

aircraft.

34(c) shows the computed

Performance

suming

a landing

obtained

in [47] for characterizing of protocol

optimizing), used.

flow constraint

the computer

(i.e., modeling,

we evaluate For example, equation

i00

the

within

annotating,

algorithm

the optical

vision

estimating,

in terms

flow algorithms

a local region.

algorithms

Thus,

of the

valigeneral

are developed the ideal input

(a)

(b)

•,-_--- ..... _"_--- --_,-- ."..'_'.,'V.'.:':;:', __.'_.-:_'_.:.-::._,:'_'= :__':11 .-_

'

II (c)

(d)

Figure 33: Results obtained for a simulated landing sequence Landing_normal_32L (b)Detected Features flow vectors detected as due to obstacles

101

image sequence: (a)Frame 3 of image (c)Computed optical flow (d)Optical

.-

(a)

:

...

:.

-

(b)

i -i-

!

/ (c)

(d)

Figure 34: Results obtained for a real image sequence: way_crossing_new (b)Detected Features (c)Computed detected as due to obstacle

102

(a)Frame 60 of image sequence runoptical flow (d)Optical flow vectors

to this algorithmrequiresgenerationof data accordingto the followingconstraintequation within a localneighborhood. ,rxu+

where

Ix, I v, and

images.

The output

the local noise

It are the spatial

region.

models

for error

parameters

of the algorithm. flow algorithms

in [47], partly the

noise

because

functions

or gathering for optical

a representative

data

to the 2-D projection

characterize input

motion

of brightness

flow algorithms

and standard

and output

datasets.

computation velocity

pattern

five or more

data

model

with proper computation are the tuning

a protocol

as suggested

or proper

modeling

or perhaps

in annotating

underlying

the algorithm

smoothness,

single

in real imaging

in the image

within

and

of

motion,

constant

situations.

Optical

is only approximately

of 3-D velocity. were evaluated,

deviations

in [47] for characterizing the algorithms

The

ideal

for the

is applied,

computation,

to a large extent,

equivalent

described

the

values

masks

by designing

are violated,

apparent

mean,

in creating

set.

are gradient

the constraint

about

flow is the

density,

which

using

to the above equation

The gradient

assumptions

etc.,

optical

inputs

with the derivative

illumination,

Various

size within

difficulty

associated

which

computation.

computed

solution

have never been evaluated

of the

flow involves

gradients

to [47], model

in gradient

and the local region

(53)

is the least square

according

of gradient,

Optical

and temporal

of the algorithm Hence,

= -xt

in error

proper

quantitative

were reported

the computer

by selecting

and

in [10]. Unlike

vision algorithms, model for input,

They used a set of image sequences

103

measures

Barton

output,

such as flow the approach

et al. [10] do not

and perturbation

to compare

the performance,

in

and used angular performance Our

deviation

the computed

and the true

has been

together,

a given

sequence

images

which

to evaluate

in terms

of images.

are input

the

of their Hence,

optical

ability

to recover

the input

to the optical

flow algorithm

flow algorithm.

the estimated

parameter

vector

was moving

the known camera model

the image.

motion

The optical optical

We evaluated motion.

vector

this weighting

flow and the estimated

to the

of a planar

algorithm surface.

we computed

computed

the optical

is the sequence

of

is motion

we assumed

and square

the actual

optical

from

ap and the covariance

sequences,

with sinusoidal

stage,

flow algorithm

using

of the estimated

optical

we feel that

function.

that

texture.

(theoretical)

of the

Using planar

flow (ut, vt) at every point

on

flow (up, vp) and the covariance

optical

quantitative

square

(Up -

7zt) _1

sequences and

evaluation distance measure,

(Up -

of planar

surfaces

of the optical between which

texture

in

function

flow should

the true optical

is given

by

(54)

l_t) T

obtained

sinusoidal

104

images

flow will be used as weighting

flow as the performance

will be image We used

synthetic

We use the Mahalanobis

E u =

Input

surface

parameter

flow _u.

Since the covariance

consider

as the

estimation

of this system

vector

image

at and the theoretical

flow algorithm

our optical

in the segmentation

the

and plane parameter,

parameter

of the estimated

To create

in front of a planar

model

The output

in the form of an eight-parameter

motion

vector

motion

to this whole vision system

parameters

Ea.

and

the motion

and structure

camera

displacement

metric.

objective

algorithms

between

using a camera for the

plane.

moving

in front

A sequence

of

15 imageswasobtainedfor differentvelocitiesof the camera. Figures frames

1, 5, 10, and

velocity

15 of a 15-frame

image

of (15, 0, 0) in front of a planar

feet from

the planar

These

image

model

and

sequences known

this parametric optical

surface

form,

geometry,

optical

optical

form for optical

plane. motion

where

distance

a

of 100

the planar

at pixels

flow at pixel

with

flow was derived.

The Mahalanobis

optical

moving

to the ground

Using

flow was computed

36 show

was at a height

respect

flow algorithm.

was available.

flow and the practical

a camera

the camera

of 60 ° with

a parametric

the theoretical

simulating

where

to the optical

flow using the flow algorithm

the theoretical

surface,

with an inclination

were input

camera

sequence

35 and

Using

practical

Eu between

(x, y) on the image

plane

was computed. Figures

37(a)

and

as a percentage

38(a)

of the

show the probability

total

number

of available

continuous

function,

the Eu axis has been

and

show

cumulative

38(b)

the

Suggest Documents