422 research outputs found
Invariant Categorisation of Polygonal Objects using Multi-resolution Signatures
With the increasing use of 3D objects and models, mining of 3D databases is becoming an important issue. However, 3D object recognition is very time consuming because of variations due to position, rotation, size and mesh resolution. A fast categorisation can be used to discard non-similar objects, such that only few objects need to be compared in full detail. We present a simple method for characterising 3D objects with the goal of performing a fast similarity search in a set of polygonal mesh models. The method constructs, for each object, two sets of multi-scale signatures: (a) the progression of deformation due to iterative mesh smoothing and, similarly, (b) the influence of mesh dilation and erosion using a sphere with increasing radius. The signatures are invariant to 3D translation, rotation and scaling, also to mesh resolution because of proper normalisation. The method was validated on a set of 31 complex objects, each object being represented with three mesh resolutions. The results were measured in terms of Euclidian distance for ranking all objects, with an overall average ranking rate of 1.29
Multi-scale keypoints in V1 and face detection
End-stopped cells in cortical area V1, which combine out-
puts of complex cells tuned to different orientations, serve to detect line
and edge crossings (junctions) and points with a large curvature. In this
paper we study the importance of the multi-scale keypoint representa-
tion, i.e. retinotopic keypoint maps which are tuned to different spatial
frequencies (scale or Level-of-Detail). We show that this representation
provides important information for Focus-of-Attention (FoA) and object
detection. In particular, we show that hierarchically-structured saliency
maps for FoA can be obtained, and that combinations over scales in
conjunction with spatial symmetries can lead to face detection through
grouping operators that deal with keypoints at the eyes, nose and mouth,
especially when non-classical receptive field inhibition is employed. Al-
though a face detector can be based on feedforward and feedback loops
within area V1, such an operator must be embedded into dorsal and
ventral data streams to and from higher areas for obtaining translation-,
rotation- and scale-invariant face (object) detection
Segmentação de imagem em três dimensões
Tese mestr., Engenharia de Sistemas e Computação, 1998, Universidade do Algarv
Image morphology: from perception to rendering
Complete image ontology can be obtained by formalising a top-down meta-language wich must address all possibilities, from global message and composition to objects and local surface properties
Multi-scale lines and edges in V1 and beyond: brightness, object categorization and recognition, and consciousness
In this paper we present an improved model for line and edge detection in cortical area V1. This model is based on responses of simple and complex cells, and it is multi-scale with no free parameters. We illustrate the use of the multi-scale line/edge representation in different processes: visual reconstruction or brightness perception, automatic scale selection and object segregation. A two-level object categorization scenario is tested in which pre-categorization is based on coarse scales only and final categorization on coarse plus fine scales. We also present a multi-scale object and face recognition model. Processing schemes are discussed in the framework of a complete cortical architecture. The fact that brightness perception and object recognition may be based on the same symbolic image representation is an indication that the entire (visual) cortex is involved in consciousness
Multi-scale cortical keypoint representation for attention and object detection
Keypoints (junctions) provide important information for
focus-of-attention (FoA) and object categorization/recognition. In this
paper we analyze the multi-scale keypoint representation, obtained by
applying a linear and quasi-continuous scaling to an optimized model of
cortical end-stopped cells, in order to study its importance and possibilities
for developing a visual, cortical architecture.We show that keypoints,
especially those which are stable over larger scale intervals, can provide
a hierarchically structured saliency map for FoA and object recognition.
In addition, the application of non-classical receptive field inhibition to
keypoint detection allows to distinguish contour keypoints from texture
(surface) keypoints
Cortical object segregation and categorization by multi-scale line and edge coding
In this paper we present an improved scheme for line and edge detection in cortical area V1, based on responses
of simple and complex cells, truly multi-scale with no free parameters. We illustrate the multi-scale
representation for visual reconstruction, and show how object segregation can be achieved with coarse-to-finescale
groupings. A two-level object categorization scenario is tested in which pre-categorization is based on
coarse scales only, and final categorization on coarse plus fine scales. Processing schemes are discussed in the
framework of a complete cortical architecture
Visual cortex frontend: integrating lines, edges, keypoints and disparaty
We present a 3D representation that is based on the pro-
cessing in the visual cortex by simple, complex and end-stopped cells.
We improved multiscale methods for line/edge and keypoint detection,
including a method for obtaining vertex structure (i.e. T, L, K etc). We
also describe a new disparity model. The latter allows to attribute depth
to detected lines, edges and keypoints, i.e., the integration results in a
3D \wire-frame" representation suitable for object recognition
A vision fronted with a new disparity model
We are developing a frontend that is based on the image representation in the
visual cortex and plausible processing schemes. This frontend consists of multiscale
line/edge and keypoint (vertex) detection, using models of simple, complex
and end-stopped cells. This frontend is being extended by a new disparity model.
Assuming that there is no neural inverse tangent operator, we do not exploit Gabor
phase information. Instead, we directly use simple cell (Gabor) responses at
positions where lines and edges are detected
- …
