6,245 research outputs found
Of Wines and Reviews: Measuring and Modeling the Vivino Wine Social Network
This paper presents an analysis of social experiences around wine consumption
through the lens of Vivino, a social network for wine enthusiasts with over 26
million users worldwide. We compare users' perceptions of various wine types
and regional styles across both New and Old World wines, examining them across
price ranges, vintages, regions, varietals, and blends. Among other things, we
find that ratings provided by Vivino users are not biased by cost. We then
study how wine characteristics, language in wine reviews, and the distribution
of wine ratings can be combined to develop prediction models. More
specifically, we model user behavior to develop a regression model for
predicting wine ratings, and a classifier for determining user review
preferences.Comment: A preliminary version of this paper appears in the Proceedings of the
IEEE/ACM International Conference on Advances in Social Networks Analysis and
Mining (ASONAM 2018). This is the full versio
Extended Capabilities for a Privacy-Enhanced Participatory Sensing Infrastructure (PEPSI)
Participatory sensing is emerging as an innovative computing paradigm that
targets the ubiquity of always-connected mobile phones and their sensing
capabilities. In this context, a multitude of pioneering applications
increasingly carry out pervasive collection and dissemination of information
and environmental data, such as, traffic conditions, pollution, temperature,
etc. Participants collect and report measurements from their mobile devices and
entrust them to the cloud to be made available to applications and users.
Naturally, due to the personal information associated to the reports (e.g.,
location, movements, etc.), a number of privacy concerns need to be taken into
account prior to a large-scale deployment of these applications. Motivated by
the need for privacy protection in Participatory Sensing, this work presents
PEPSI: a Privacy-Enhanced Participatory Sensing Infrastructure. We explore
realistic architectural assumptions and a minimal set of formal requirements
aiming at protecting privacy of both data producers and consumers. We propose
two instantiations that attain privacy guarantees with provable security at
very low additional computational cost and almost no extra communication
overhead.Comment: A preliminary version of this article, titled "PEPSI:
Privacy-Enhanced Participatory Sensing Infrastructure", appears in the
Proceedings of the 4th ACM Conference on Wireless Security (WiSec 2011). This
is the extended (journal) version, currently in revisio
Measuring Membership Privacy on Aggregate Location Time-Series
While location data is extremely valuable for various applications,
disclosing it prompts serious threats to individuals' privacy. To limit such
concerns, organizations often provide analysts with aggregate time-series that
indicate, e.g., how many people are in a location at a time interval, rather
than raw individual traces. In this paper, we perform a measurement study to
understand Membership Inference Attacks (MIAs) on aggregate location
time-series, where an adversary tries to infer whether a specific user
contributed to the aggregates.
We find that the volume of contributed data, as well as the regularity and
particularity of users' mobility patterns, play a crucial role in the attack's
success. We experiment with a wide range of defenses based on generalization,
hiding, and perturbation, and evaluate their ability to thwart the attack
vis-a-vis the utility loss they introduce for various mobility analytics tasks.
Our results show that some defenses fail across the board, while others work
for specific tasks on aggregate location time-series. For instance, suppressing
small counts can be used for ranking hotspots, data generalization for
forecasting traffic, hotspot discovery, and map inference, while sampling is
effective for location labeling and anomaly detection when the dataset is
sparse. Differentially private techniques provide reasonable accuracy only in
very specific settings, e.g., discovering hotspots and forecasting their
traffic, and more so when using weaker privacy notions like crowd-blending
privacy. Overall, our measurements show that there does not exist a unique
generic defense that can preserve the utility of the analytics for arbitrary
applications, and provide useful insights regarding the disclosure of sanitized
aggregate location time-series
On Collaborative Predictive Blacklisting
Collaborative predictive blacklisting (CPB) allows to forecast future attack
sources based on logs and alerts contributed by multiple organizations.
Unfortunately, however, research on CPB has only focused on increasing the
number of predicted attacks but has not considered the impact on false
positives and false negatives. Moreover, sharing alerts is often hindered by
confidentiality, trust, and liability issues, which motivates the need for
privacy-preserving approaches to the problem. In this paper, we present a
measurement study of state-of-the-art CPB techniques, aiming to shed light on
the actual impact of collaboration. To this end, we reproduce and measure two
systems: a non privacy-friendly one that uses a trusted coordinating party with
access to all alerts (Soldo et al., 2010) and a peer-to-peer one using
privacy-preserving data sharing (Freudiger et al., 2015). We show that, while
collaboration boosts the number of predicted attacks, it also yields high false
positives, ultimately leading to poor accuracy. This motivates us to present a
hybrid approach, using a semi-trusted central entity, aiming to increase
utility from collaboration while, at the same time, limiting information
disclosure and false positives. This leads to a better trade-off of true and
false positive rates, while at the same time addressing privacy concerns.Comment: A preliminary version of this paper appears in ACM SIGCOMM's Computer
Communication Review (Volume 48 Issue 5, October 2018). This is the full
versio
Systematizing Genome Privacy Research: A Privacy-Enhancing Technologies Perspective
Rapid advances in human genomics are enabling researchers to gain a better
understanding of the role of the genome in our health and well-being,
stimulating hope for more effective and cost efficient healthcare. However,
this also prompts a number of security and privacy concerns stemming from the
distinctive characteristics of genomic data. To address them, a new research
community has emerged and produced a large number of publications and
initiatives.
In this paper, we rely on a structured methodology to contextualize and
provide a critical analysis of the current knowledge on privacy-enhancing
technologies used for testing, storing, and sharing genomic data, using a
representative sample of the work published in the past decade. We identify and
discuss limitations, technical challenges, and issues faced by the community,
focusing in particular on those that are inherently tied to the nature of the
problem and are harder for the community alone to address. Finally, we report
on the importance and difficulty of the identified challenges based on an
online survey of genome data privacy expertsComment: To appear in the Proceedings on Privacy Enhancing Technologies
(PoPETs), Vol. 2019, Issue
- …
