30 research outputs found
Catastrophic forgetting: still a problem for DNNs
We investigate the performance of DNNs when trained on class-incremental
visual problems consisting of initial training, followed by retraining with
added visual classes. Catastrophic forgetting (CF) behavior is measured using a
new evaluation procedure that aims at an application-oriented view of
incremental learning. In particular, it imposes that model selection must be
performed on the initial dataset alone, as well as demanding that retraining
control be performed only using the retraining dataset, as initial dataset is
usually too large to be kept. Experiments are conducted on class-incremental
problems derived from MNIST, using a variety of different DNN models, some of
them recently proposed to avoid catastrophic forgetting. When comparing our new
evaluation procedure to previous approaches for assessing CF, we find their
findings are completely negated, and that none of the tested methods can avoid
CF in all experiments. This stresses the importance of a realistic empirical
measurement procedure for catastrophic forgetting, and the need for further
research in incremental learning for DNNs.Comment: 10 pages, 11 figures, Artificial Neural Networks and Machine Learning
- ICANN 201
Modelling and Understanding of Chatter
Recent analysis in chatter modelling of BTA deep-hole drilling consisted in phenomenological modelisation of relationships between the observed time series and appearance of chatter during the process. Using the newly developed MEWMA control chart [4, 5], it has even been possible to predict the occurence of chatter about 30 to 50 mm in advance (i.e. up to one minute before the chatter starts). Unfortunately, no relationships between the machine and model parameters have been detected. Therefore, in this paper a physical model of the boring bar is taken into account. Simulation studies of the regenerative process are performed. These simulated time series show the same characteristics as the data recorded during the drilling process and thus support the validity of our model. By running such simulations, we intend to find strategies for chatter prevention in future work. --
Modelling and Understanding of Chatter
Recent analysis in chatter modelling of BTA deep-hole drilling consisted in phenomenological modelisation of relationships between the observed time series and appearance of chatter during the process. Using the newly developed MEWMA control chart [4, 5], it has even been possible to predict the occurence of chatter about 30 to 50 mm in advance (i.e. up to one minute before the chatter starts). Unfortunately, no relationships between the machine and model parameters have been detected. Therefore, in this paper a physical model of the boring bar is taken into account. Simulation studies of the regenerative process are performed. These simulated time series show the same characteristics as the data recorded during the drilling process and thus support the validity of our model. By running such simulations, we intend to find strategies for chatter prevention in future work
Modelling and understanding of chatter
Recent analysis in chatter modelling of BTA deephole drilling consisted in phenomenological modelisation of relationships between the ob
served time series and appearance of chatter during the process. Using the newly developed MEWMA control chart [4, 5], it has even been possible to predict the occurence of chatter about 30 to 50 mm in advance (i.e. up to
one minute before the chatter starts).
Unfortunately, no relationships between the machine and model parameters have been detected. Therefore, in this paper a physical model of the
boring bar is taken into account. Simulation studies of the regenerative process are performed. These simulated time series show the same characteristics as the data recorded during the drilling process and thus support the validity of our model. By running such simulations, we intend to find strategies for
chatter prevention in future work
Neural Network Fusion of Color, Depth and Location for Object Instance Recognition on a Mobile Robot
International audienceThe development of mobile robots for domestic assistance re-quires solving problems integrating ideas from different fields of research like computer vision, robotic manipulation, localization and mapping. Semantic mapping, that is, the enrichment a map with high-level infor-mation like room and object identities, is an example of such a complex robotic task. Solving this task requires taking into account hard software and hardware constraints brought by the context of autonomous mobile robots, where short processing times and low energy consumption are mandatory. We present a light-weight scene segmentation and object in-stance recognition algorithm using an RGB-D camera and demonstrate it in a semantic mapping experiment. Our method uses a feed-forward neural network to fuse texture, color and depth information. Running at 3 Hz on a single laptop computer, our algorithm achieves a recognition rate of 97% in a controlled environment, and 87% in the adversarial con-ditions of a real robotic task. Our results demonstrate that state of the art recognition rates on a database does not guarantee performance in a real world experiment. We also show the benefit in these conditions of fusing several recognition decisions and data from different sources. The database we compiled for the purpose of this study is publicly available
Biased Competition in Visual Processing Hierarchies: A Learning Approach Using Multiple Cues
In this contribution, we present a large-scale hierarchical system for object detection fusing bottom-up (signal-driven) processing results with top-down (model or task-driven) attentional modulation. Specifically, we focus on the question of how the autonomous learning of invariant models can be embedded into a performing system and how such models can be used to define object-specific attentional modulation signals. Our system implements bi-directional data flow in a processing hierarchy. The bottom-up data flow proceeds from a preprocessing level to the hypothesis level where object hypotheses created by exhaustive object detection algorithms are represented in a roughly retinotopic way. A competitive selection mechanism is used to determine the most confident hypotheses, which are used on the system level to train multimodal models that link object identity to invariant hypothesis properties. The top-down data flow originates at the system level, where the trained multimodal models are used to obtain space- and feature-based attentional modulation signals, providing biases for the competitive selection process at the hypothesis level. This results in object-specific hypothesis facilitation/suppression in certain image regions which we show to be applicable to different object detection mechanisms. In order to demonstrate the benefits of this approach, we apply the system to the detection of cars in a variety of challenging traffic videos. Evaluating our approach on a publicly available dataset containing approximately 3,500 annotated video images from more than 1 h of driving, we can show strong increases in performance and generalization when compared to object detection in isolation. Furthermore, we compare our results to a late hypothesis rejection approach, showing that early coupling of top-down and bottom-up information is a favorable approach especially when processing resources are constrained
