This report is on the Lehigh/ColumbiaMURI contract. While the original focus was on sensors for manufacturing, the natural evolution of our basic researchhas led us to more general problemsin more generic settings. As a multifaculty multi-disciplinaryprojectmuchof the work is naturally done in smaller subgroup. The major resultsover the past year were on 3D modeling/sensorplanning,omni-directionalimaging,and reflectance/image/noise modeling.Thereweremore focusedresultsin deformablemodels,featuredetection, appearancematching,andvideosegmentation. Notethatsomeprojects(e.g.,theoutdoor3D model building) bring togethermany of the above topics. This reportprovidesshortsummariesof our significant contributions with citationsto relatedpapers. Lengthof presentationhereindoesnot reflect level of effort nor our view of its significance– many of the most importantareashave paperselsewherein theseproceedings. 1 AutomatedSiteModeling 3-D modelsof outdoorenvironments,known assite models,areusedin many differentapplicationsthat includecity planning,urbandesign,fire andpolice planning,surveillanceandvirtual reality modeling. Creatingsite modelsfor urbansceneswhich contain large structures(i.e., buildings) thatencompass a wide rangeof geometricshapesandcontainhigh occlusionareasis quitechallenging. Thesemodels are typically createdby hand in a painstakinganderror proneprocess.We arebuilding a systemto automatethis procedurethat extendsour previous work in 3-D model acquisition This work supported by ONR/ARPA MURI program ONR N00014-95-1-0601.Several other agenciesand companieshave also supportedpartsof this research. Figure1: Mobile robotfor automatedsitemodeling. usingrangedata[Reed-1998, ReedandAllen-1998, Reedet al.-1997, Allen and Yang-1998]. This is an incrementalvolumetricmethodthat canacquire andmodelmultiple objectsin a sceneandcorrectly merge models from different views of the scene. Modelsarebuilt from successi ve sensingoperations that aremergedwith the currentmodelbeingbuilt, called the compositemodel. The merging is performedusingaregularizedsetintersectionoperation. Thesepartial,compositemodelscanserveasinputto oursensorplanningsystemthatcanreducethenumber of views neededto fully acquirea scene.Planningeachsensingoperationreducesthedatasetsizes andacquisitiontime for a complex model.Most existing systemsdo not useplanning,but rely on humaninteractionor theneedfor largeoverlapsin imagesto assureadequatecoverageof thescene.Our plannercanincorporatedifferentconstraintsincluding visibility, field-of-view andsensorplacemento find the next viewing position that will reducethe model’s uncertainty . Thesystemhasbeentestedon Figure2: Recovered3-D modelsof a blocksworld city. All 3 objectswererecoveredat once. Sensor planningalgorithmswere usedto reducethe numberof rangescansandfind unoccludedviewing positions. indoor modelsandwe arenow extendingit to outdoorsceneswith multipleobjects. At eachstepof theprocess,a partialsolid modelof the sceneis created. The facesof this model consistof correctlyimagedfacesandfacesthatareocclusionartifacts. We can label thesefacesas “imaged”or “unimaged”andpropagate/update theselabelsasnew imagesareintegratedinto thecomposite model. The faceslabeled“unimaged”are thenthe focusof thesensorplanningsystemwhichwill try to positionthesensorto allow these“unimaged”faces to be scanned.The set intersectionoperationmust be ableto correctlypropagatethe surface-typetags fromsurfacesin themodelsthroughto thecomposite model.Retainingthesetagsaftermergingoperations allows viewpoint planningfor unimagedsurfacesto proceed. Figure2 is a recovered3-D solid modelof a simulatedcity scenemadeupof 3 toy buildingsplacedon our laserscannerturntable.Four initial views were takenat intervals,andthismergedpartialmodel wasusedby theview plannerto choosethenext view to reducethemodel’s uncertainty . This processof a partial modeldriving the plannerfor the next view wasuseduntil 8 moreimagesweretaken, reducing the model’s uncertaintyto a small threshold. The modelis accurateandthemethodcanrecover structurethatis occluded. For automatingthis task outdoors,we are equipping a mobile vehiclewith sensorsand algorithms to accomplishthis task. A pictureof the vehicle is shown in figure1. Theequipmentconsistsof anRWI ATRV mobile robot base,a spot rangescanner(80 meterrangespotscannerwith 2-DOFscanningmirrors for acquiringa whole rangeimage,not shown in image),centimeteraccuracy on-boardGPS,color camerasfor obtainingphotometryof thescene,and mobilewirelesscommunicationsfor transmissionof dataandhigh level controlfunctions. Theplanningalgorithmscanbeusedto navigatethe mobilerobotscannersystemto a positionfor a new scanthat will reducethe uncertaintyin the scene. Given a partial modelof the scene,we cantag the surfacesacquiredby the scannerin sucha way as to know whatregionsareimagedandwhich areocclusion surfaces. We can thenusetheseocclusion surfacesto find unobstructedviewpoints. Oncewe computetheseviewpoints, we can then command themobilesystemto navigateinsidea visibility volumeandtake a new scan,thuscompletinga partial modeland“filling in the blanks” to build the completemodel.Detailscanbefoundin thisproceedings [Allen et al.-1998]. 2 Interacti veSensorPlanning Theautomatedsitemodelingsystemwill bemerging therangedatawith 2-D imageryto enhancethemodels. This alsorequiresa view planningcomponent. In clutteredandcomplex environmentssuchasurban scenes, it canbevery difficult to determinewherea camerashouldbeplacedto view multipleobjectsand regionsof interest.We have built aninteracti ve sensor planningsystem[Stamosand Allen-1998] that canbe usedto selectviewpointssubjectto camera visibility, field of view and task constraints. This work buildsuponourearlierwork in sensorplanning [Abrams-1997 ]. Application areasfor this method includesurveillanceplanning,safetymonitoring,architecturalsite designplanning,andautomatedsite modeling. Givena descriptionof thesensor’ s characteristics,theobjectsin the3-D scene,andthetargetsto be viewed, our algorithmscomputethe set of admissibleview pointsthatsatisfytheconstraints. The systemfirst builds topologically correct solid modelsof the scenefrom a variety of datasources. Viewing targetsarethenselected,andvisibility volumesandfield of view conesarecomputedandintersectedto createviewing volumeswherecameras can be placed. The usercan interacti vely manipulate the sceneandselectmultiple target featuresto beviewedby acamera.Theusercanalsoselectcan-
[1]
Shree K. Nayar,et al.
Reflectance and texture of real-world surfaces
,
1997,
Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[2]
Shree K. Nayar,et al.
A theory of catadioptric image formation
,
1998,
Sixth International Conference on Computer Vision (IEEE Cat. No.98CH36271).
[3]
Rangasami L. Kashyap,et al.
Characterization and estimation of two-dimensional ARMA models
,
1984,
IEEE Trans. Inf. Theory.
[4]
John R. Kender,et al.
Video scene segmentation via continuous video coherence
,
1998,
Proceedings. 1998 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No.98CB36231).
[5]
Glenn Healey,et al.
Illumination and geometry invariant recognition of texture in color images
,
1996,
Proceedings CVPR IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[6]
Hiroshi Murase,et al.
Real-time 100 object recognition system
,
1996,
Proceedings of IEEE International Conference on Robotics and Automation.
[7]
Rick S. Blum,et al.
On estimating the quality of noisy images
,
1998,
Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '98 (Cat. No.98CH36181).
[8]
J. Koenderink,et al.
Surface bidirectional reflection distribution function and the texture of bricks and tiles.
,
1997,
Applied optics.
[9]
Ioannis Stamos,et al.
Automated model acquisition from range images with view planning
,
1997,
Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[10]
Manfred H. Hueckel.
An Operator Which Locates Edges in Digitized Pictures
,
1971,
J. ACM.
[11]
Ina Ruck,et al.
USA
,
1969,
The Lancet.
[12]
Alex Pentland,et al.
Probabilistic visual learning for object detection
,
1995,
Proceedings of IEEE International Conference on Computer Vision.
[13]
Terrance E. Boult,et al.
Remote Reality via omni-directional imaging
,
1998,
SIGGRAPH '98.
[14]
Thomas O. Binford,et al.
On Detecting Edges
,
1986,
IEEE Transactions on Pattern Analysis and Machine Intelligence.
[15]
Frank O'Gorman,et al.
Edge Detection Using Walsh Functions
,
1976,
Artif. Intell..
[16]
Takeo Kanade,et al.
Surface Reflection: Physical and Geometrical Perspectives
,
1989,
IEEE Trans. Pattern Anal. Mach. Intell..
[17]
H. Hotelling.
Some Improvements in Weighing and Other Experimental Techniques
,
1944
.
[18]
Steven Abrams.
Sensor planning in an active robot work-cell
,
1997
.
[19]
S. Nayar,et al.
Optimal Weighting Functions for Feature Detection
,
1998
.
[20]
Terrance E. Boult.
Remote Reality Demonstration
,
1998,
Proceedings. 1998 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No.98CB36231).
[21]
A. Ardeshir Goshtasby,et al.
Design of a single-lens stereo camera system
,
1993,
Pattern Recognit..
[22]
Andrea J. van Doorn,et al.
Bidirectional Reflection Distribution Function Expressed in Terms of Surface Scattering Modes
,
1996,
ECCV.
[23]
Ioannis Stamos,et al.
Interactive sensor planning
,
1998,
Proceedings. 1998 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No.98CB36231).
[24]
Ioannis Stamos,et al.
3-D modeling from range imagery: an incremental method with a planning component
,
1997,
Proceedings. International Conference on Recent Advances in 3-D Digital Imaging and Modeling (Cat. No.97TB100134).
[25]
David G. Morgenthaler.
A new hybrid edge detector
,
1981
.
[26]
Rick S. Blum,et al.
Multisensor image fusion using a region-based wavelet transform approach
,
1997
.
[27]
Ruigang Yang,et al.
Registering, integrating, and building CAD models from range data
,
1998,
Proceedings. 1998 IEEE International Conference on Robotics and Automation (Cat. No.98CH36146).
[28]
Shree K. Nayar,et al.
Catadioptric omnidirectional camera
,
1997,
Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[29]
Shree K. Nayar,et al.
Generation of Perspective and Panoramic Video from Omnidirectional Video
,
1997
.
[30]
Andrea J. van Doorn,et al.
Illuminance texture due to surface mesostructure
,
1996
.
[31]
S. Nayar,et al.
Appearance Matching with Partial Data ∗
,
1998
.
[32]
Manfred H. Hueckel.
A Local Visual Operator Which Recognizes Edges and Lines
,
1973,
JACM.
[33]
Shree K. Nayar,et al.
Reflectance and texture of real-world surfaces
,
1999,
TOGS.
[34]
Ioannis Stamos,et al.
View Planning for Site Modeling
,
2000
.
[35]
John Krumm,et al.
Eigenfeatures for planar pose measurement of partially occluded objects
,
1996,
Proceedings CVPR IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[36]
H.D. Tagare,et al.
A framework for the construction of reflectance maps for machine vision
,
1993
.
[37]
Terrance E. Boult,et al.
Imaging-consistent warping and super-resolution
,
1998
.
[38]
Hiroshi Murase,et al.
Software Library for Appearance Matching (SLAM)
,
1994
.
[39]
Michal Irani,et al.
Motion Analysis for Image Enhancement: Resolution, Occlusion, and Transparency
,
1993,
J. Vis. Commun. Image Represent..
[40]
Michal Irani,et al.
Improving resolution by image registration
,
1991,
CVGIP Graph. Model. Image Process..
[41]
Jitendra Malik,et al.
On perpendicular texture or: Why do we see more flowers in the distance?
,
1997,
Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[42]
Horst Bischof,et al.
Dealing with occlusions in the eigenspace approach
,
1996,
Proceedings CVPR IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[43]
Edward M. Riseman,et al.
Image-based homing
,
1991,
IEEE Control Systems.
[44]
Terrance E. Boult,et al.
Applications of omnidirectional imaging: multi-body tracking and remote reality
,
1998,
Proceedings Fourth IEEE Workshop on Applications of Computer Vision. WACV'98 (Cat. No.98EX201).