2,168 research outputs found
Why Does Democracy Need Education?
Across countries, education and democracy are highly correlated. We motivate empirically and then model a causal mechanism explaining this correlation. In our model, schooling teaches people to interact with others and raises the benefits of civic participation, including voting and organizing. In the battle between democracy and dictatorship, democracy has a wide potential base of support but offers weak incentives to its defenders. Dictatorship provides stronger incentives to a narrower base. As education raises the benefits of civic participation, it raises the support for more democratic regimes relative to dictatorships. This increases the likelihood of democratic revolutions against dictatorships, and reduces that of successful anti-democratic coups.
Did the Death of Distance Hurt Detroit and Help New York?
Urban proximity can reduce the costs of shipping goods and speed the flow of ideas. Improvements in communication technology might erode these advantages and allow people and firms to decentralize. However, improvements in transportation and communication technology can also increase the returns to new ideas, by allowing those ideas to be used throughout the world. This paper presents a model that illustrates these two rival effects that technological progress can have on cities. We then present some evidence suggesting that the model can help us to understand why the past thirty-five years have been kind to idea-producing places, like New York and Boston, and devastating to goods-producing cities, like Cleveland and Detroit.
Enriching Frame Representations with Distributionally Induced Senses
We introduce a new lexical resource that enriches the Framester knowledge
graph, which links Framnet, WordNet, VerbNet and other resources, with semantic
features from text corpora. These features are extracted from distributionally
induced sense inventories and subsequently linked to the manually-constructed
frame representations to boost the performance of frame disambiguation in
context. Since Framester is a frame-based knowledge graph, which enables
full-fledged OWL querying and reasoning, our resource paves the way for the
development of novel, deeper semantic-aware applications that could benefit
from the combination of knowledge from text and complex symbolic
representations of events and participants. Together with the resource we also
provide the software we developed for the evaluation in the task of Word Frame
Disambiguation (WFD).Comment: In Proceedings of the 11th Conference on Language Resources and
Evaluation (LREC 2018). Miyazaki, Japan. ELR
Efficient pruning of large knowledge graphs
In this paper we present an efficient and highly accurate algorithm to prune noisy or over-ambiguous knowledge graphs given as input an extensional definition of a domain of interest, namely as a set
of instances or concepts. Our method climbs the graph in a bottom-up fashion, iteratively layering
the graph and pruning nodes and edges in each layer while not compromising the connectivity of the set of input nodes. Iterative layering and protection of pre-defined nodes allow to extract semantically coherent DAG structures from noisy or over-ambiguous cyclic graphs, without loss of information and without incurring in computational bottlenecks, which are the main problem of stateof- the-art methods for cleaning large, i.e., Webscale,
knowledge graphs. We apply our algorithm to the tasks of pruning automatically acquired taxonomies using benchmarking data from a SemEval evaluation exercise, as well as the extraction of a domain-adapted taxonomy from theWikipedia category hierarchy. The results show the superiority of our approach over state-of-art algorithms in terms of both output quality and computational efficiency
Improving Hypernymy Extraction with Distributional Semantic Classes
In this paper, we show how distributionally-induced semantic classes can be
helpful for extracting hypernyms. We present methods for inducing sense-aware
semantic classes using distributional semantics and using these induced
semantic classes for filtering noisy hypernymy relations. Denoising of
hypernyms is performed by labeling each semantic class with its hypernyms. On
the one hand, this allows us to filter out wrong extractions using the global
structure of distributionally similar senses. On the other hand, we infer
missing hypernyms via label propagation to cluster terms. We conduct a
large-scale crowdsourcing study showing that processing of automatically
extracted hypernyms using our approach improves the quality of the hypernymy
extraction in terms of both precision and recall. Furthermore, we show the
utility of our method in the domain taxonomy induction task, achieving the
state-of-the-art results on a SemEval'16 task on taxonomy induction.Comment: In Proceedings of the 11th Conference on Language Resources and
Evaluation (LREC 2018). Miyazaki, Japa
Knowledge-rich Image Gist Understanding Beyond Literal Meaning
We investigate the problem of understanding the message (gist) conveyed by
images and their captions as found, for instance, on websites or news articles.
To this end, we propose a methodology to capture the meaning of image-caption
pairs on the basis of large amounts of machine-readable knowledge that has
previously been shown to be highly effective for text understanding. Our method
identifies the connotation of objects beyond their denotation: where most
approaches to image understanding focus on the denotation of objects, i.e.,
their literal meaning, our work addresses the identification of connotations,
i.e., iconic meanings of objects, to understand the message of images. We view
image understanding as the task of representing an image-caption pair on the
basis of a wide-coverage vocabulary of concepts such as the one provided by
Wikipedia, and cast gist detection as a concept-ranking problem with
image-caption pairs as queries. To enable a thorough investigation of the
problem of gist understanding, we produce a gold standard of over 300
image-caption pairs and over 8,000 gist annotations covering a wide variety of
topics at different levels of abstraction. We use this dataset to
experimentally benchmark the contribution of signals from heterogeneous
sources, namely image and text. The best result with a Mean Average Precision
(MAP) of 0.69 indicate that by combining both dimensions we are able to better
understand the meaning of our image-caption pairs than when using language or
vision information alone. We test the robustness of our gist detection approach
when receiving automatically generated input, i.e., using automatically
generated image tags or generated captions, and prove the feasibility of an
end-to-end automated process
Event-based Access to Historical Italian War Memoirs
The progressive digitization of historical archives provides new, often
domain specific, textual resources that report on facts and events which have
happened in the past; among these, memoirs are a very common type of primary
source. In this paper, we present an approach for extracting information from
Italian historical war memoirs and turning it into structured knowledge. This
is based on the semantic notions of events, participants and roles. We evaluate
quantitatively each of the key-steps of our approach and provide a graph-based
representation of the extracted knowledge, which allows to move between a Close
and a Distant Reading of the collection.Comment: 23 pages, 6 figure
- …
