5,350 research outputs found
SINVAD: Search-based Image Space Navigation for DNN Image Classifier Test Input Generation
The testing of Deep Neural Networks (DNNs) has become increasingly important
as DNNs are widely adopted by safety critical systems. While many test adequacy
criteria have been suggested, automated test input generation for many types of
DNNs remains a challenge because the raw input space is too large to randomly
sample or to navigate and search for plausible inputs. Consequently, current
testing techniques for DNNs depend on small local perturbations to existing
inputs, based on the metamorphic testing principle. We propose new ways to
search not over the entire image space, but rather over a plausible input space
that resembles the true training distribution. This space is constructed using
Variational Autoencoders (VAEs), and navigated through their latent vector
space. We show that this space helps efficiently produce test inputs that can
reveal information about the robustness of DNNs when dealing with realistic
tests, opening the field to meaningful exploration through the space of highly
structured images
Learning of Image Dehazing Models for Segmentation Tasks
To evaluate their performance, existing dehazing approaches generally rely on
distance measures between the generated image and its corresponding ground
truth. Despite its ability to produce visually good images, using pixel-based
or even perceptual metrics do not guarantee, in general, that the produced
image is fit for being used as input for low-level computer vision tasks such
as segmentation. To overcome this weakness, we are proposing a novel end-to-end
approach for image dehazing, fit for being used as input to an image
segmentation procedure, while maintaining the visual quality of the generated
images. Inspired by the success of Generative Adversarial Networks (GAN), we
propose to optimize the generator by introducing a discriminator network and a
loss function that evaluates segmentation quality of dehazed images. In
addition, we make use of a supplementary loss function that verifies that the
visual and the perceptual quality of the generated image are preserved in hazy
conditions. Results obtained using the proposed technique are appealing, with a
favorable comparison to state-of-the-art approaches when considering the
performance of segmentation algorithms on the hazy images.Comment: Accepted in EUSIPCO 201
EmbraceNet for Activity: A Deep Multimodal Fusion Architecture for Activity Recognition
Human activity recognition using multiple sensors is a challenging but
promising task in recent decades. In this paper, we propose a deep multimodal
fusion model for activity recognition based on the recently proposed feature
fusion architecture named EmbraceNet. Our model processes each sensor data
independently, combines the features with the EmbraceNet architecture, and
post-processes the fused feature to predict the activity. In addition, we
propose additional processes to boost the performance of our model. We submit
the results obtained from our proposed model to the SHL recognition challenge
with the team name "Yonsei-MCML."Comment: Accepted in HASCA at ACM UbiComp/ISWC 2019, won the 2nd place in the
SHL Recognition Challenge 201
- …
