1,992 research outputs found

    Binary extensions and choice theory

    Get PDF

    The computational complexity of rationalizing Pareto optimal choice behavior.

    Get PDF
    We consider a setting where a coalition of individuals chooses one or several alternatives from each set in a collection of choice sets. We examine the computational complexity of Pareto rationalizability. Pareto rationalizability requires that we can endow each individual in the coalition with a preference relation such that the observed choices are Pareto efficient. We differentiate between the situation where the choice function is considered to select all Pareto optimal alternatives from a choice set and the situation where it only contains one or several Pareto optimal alternatives. In the former case we find that Pareto rationalizability is an NP-complete problem. For the latter case we demonstrate that, if we have no additional information on the individual preference relations, then all choice behavior is Pareto rationalizable. However, if we have such additional information, then Pareto rationalizability is again NP-complete. Our results are valid for any coalition of size greater or equal than two.

    The computational complexity of boundedly rational choice behavior

    Get PDF
    This research examines the computational complexity of two boundedly rational choice models that use multiple rationales to explain observed choice behavior. First, we show that the notion of rationalizability by K rationales as introduced by Kalai, Rubinstein, and Spiegler (2002) is NP-complete for K greater or equal to two. Second, we show that the question of sequential rationalizability by K rationales, introduced by Manzini and Mariotti (2007), is NP-complete for K greater or equal to three if choices are single valued, and that it is NP-complete for K greater or equal to one if we allow for multi-valued choice correspondences. Motivated by these results, we present two binary integer feasibility programs that characterize the two boundedly rational choice models and we compute their power. Finally, by using results from descriptive complexity theory, we explain why it has been so difficult to obtain `nice' characterizations for these models.

    Porting concepts from DNNs back to GMMs

    Get PDF
    Deep neural networks (DNNs) have been shown to outperform Gaussian Mixture Models (GMM) on a variety of speech recognition benchmarks. In this paper we analyze the differences between the DNN and GMM modeling techniques and port the best ideas from the DNN-based modeling to a GMM-based system. By going both deep (multiple layers) and wide (multiple parallel sub-models) and by sharing model parameters, we are able to close the gap between the two modeling techniques on the TIMIT database. Since the 'deep' GMMs retain the maximum-likelihood trained Gaussians as first layer, advanced techniques such as speaker adaptation and model-based noise robustness can be readily incorporated. Regardless of their similarities, the DNNs and the deep GMMs still show a sufficient amount of complementarity to allow effective system combination

    The computational complexity of rationalizing Pareto optimal choice behavior

    Get PDF
    We consider a setting where a coalition of individuals chooses one or several alternatives from each set in a collection of choice sets. We examine the computational complexity of Pareto rationalizability. Pareto rationalizability requires that we can endow each individual in the coalition with a preference relation such that the observed choices are Pareto efficient. We differentiate between the situation where the choice function is considered to select all Pareto optimal alternatives from a choice set and the situation where it only contains one or several Pareto optimal alternatives. In the former case we find that Pareto rationalizability is an NP-complete problem. For the latter case we demonstrate that, if we have no additional information on the individual preference relations, then all choice behavior is Pareto rationalizable. However, if we have such additional information, then Pareto rationalizability is again NP-complete. Our results are valid for any coalition of size greater or equal than two.

    ECAPA-TDNN: Emphasized Channel Attention, Propagation and Aggregation in TDNN Based Speaker Verification

    Get PDF
    Current speaker verification techniques rely on a neural network to extract speaker representations. The successful x-vector architecture is a Time Delay Neural Network (TDNN) that applies statistics pooling to project variable-length utterances into fixed-length speaker characterizing embeddings. In this paper, we propose multiple enhancements to this architecture based on recent trends in the related fields of face verification and computer vision. Firstly, the initial frame layers can be restructured into 1-dimensional Res2Net modules with impactful skip connections. Similarly to SE-ResNet, we introduce Squeeze-and-Excitation blocks in these modules to explicitly model channel interdependencies. The SE block expands the temporal context of the frame layer by rescaling the channels according to global properties of the recording. Secondly, neural networks are known to learn hierarchical features, with each layer operating on a different level of complexity. To leverage this complementary information, we aggregate and propagate features of different hierarchical levels. Finally, we improve the statistics pooling module with channel-dependent frame attention. This enables the network to focus on different subsets of frames during each of the channel's statistics estimation. The proposed ECAPA-TDNN architecture significantly outperforms state-of-the-art TDNN based systems on the VoxCeleb test sets and the 2019 VoxCeleb Speaker Recognition Challenge.Comment: proceedings of INTERSPEECH 202

    Robust language recognition via adaptive language factor extraction

    Get PDF
    This paper presents a technique to adapt an acoustically based language classifier to the background conditions and speaker accents. This adaptation improves language classification on a broad spectrum of TV broadcasts. The core of the system consists of an iVector-based setup in which language and channel variabilities are modeled separately. The subsequent language classifier (the backend) operates on the language factors, i.e. those features in the extracted iVectors that explain the observed language variability. The proposed technique adapts the language variability model to the background conditions and to the speaker accents present in the audio. The effect of the adaptation is evaluated on a 28 hours corpus composed of documentaries and monolingual as well as multilingual broadcast news shows. Consistent improvements in the automatic identification of Flemish (Belgian Dutch), English and French are demonstrated for all broadcast types

    Noncooperative household consumption with caring

    Get PDF
    We present a household consumption model that accounts for caring household members, while allowing for noncooperative behavior in decisions on public goods. The intrahousehold consumption outcome critically depends on the degree of caring between the household members. By varying the degree of intrahousehold caring, the model encompasses a whole continuum of household consumption models that are situated between the fully cooperative model and the noncooperative model without caring. This feature is used to define a measure for the degree of cooperation within the household. We also establish a dual characterization of our noncooperative model with caring preferences: we show that the model is dually equivalent to a noncooperative model with non-caring preferences that is characterized by intrahousehold transfers. Finally, following a revealed preference approach, we derive testable implications of the model for empirical data. We demonstrate the practical usefulness of the model through an illustrative application.

    Testable implications of general equilibrium models: an integer programming approach

    Get PDF
    Focusing on the testable implications on the equilibrium manifold, we show that the rationalizability problem is NP-complete. Subsequently, we present an integer programming (IP) approach to characterizing general equilibrium models. This approach avoids the use of the Tarski-Seidenberg algorithm for quantifier elimination that is commonly used in the literature. The IP approach naturally applies to settings with any number of observations, which is attractive for empirical applications. In addition, it can easily be adjusted to analyze the testable implications of alternative general equilibrium models (that include, e.g., public goods, externalities and/or production). Further, we show that the IP framework can easily address recoverability questions (pertaining to the structural model that underlies the observed equilibrium behavior), and account for empirical issues when bringing the IP methodology to the data (such as goodness-of-fit and power). Finally, we show how to develop easy-to-implement heuristics that give a quick (but possibly inconclusive) answer to whether or not the data satisfy the general equilibrium models.General equilibrium, equilibrium manifold, exchange economies, production economies, NP-completeness, nonparametric restrictions, GARP, integer programming.
    corecore