24 research outputs found
The Beaker phenomenon and the genomic transformation of northwest Europe
From around 2750 to 2500 bc, Bell Beaker pottery became widespread across western and central Europe, before it disappeared between 2200 and 1800 bc. The forces that propelled its expansion are a matter of long-standing debate, and there is support for both cultural diffusion and migration having a role in this process. Here we present genome-wide data from 400 Neolithic, Copper Age and Bronze Age Europeans, including 226 individuals associated with Beaker-complex artefacts. We detected limited genetic affinity between Beaker-complex-associated individuals from Iberia and central Europe, and thus exclude migration as an important mechanism of spread between these two regions. However, migration had a key role in the further dissemination of the Beaker complex. We document this phenomenon most clearly in Britain, where the spread of the Beaker complex introduced high levels of steppe-related ancestry and was associated with the replacement of approximately 90% of Britain’s gene pool within a few hundred years, continuing the east-to-west expansion that had brought steppe-related ancestry into central and northern Europe over the previous centuries
Delete-m jackknife for unequal m
In this paper, the delete-mj jackknife estimator is proposed. This estimator is based on samples obtained from the original sample by successively removing mutually exclusive groups of unequal size. In a Monte Carlo simulation study, a hierarchical linear model was used to evaluate the role of nonnormal residuals and sample size on bias and efficiency of this estimator. It is shown that bias is reduced in exchange for a minor reduction in efficiency. The accompanying jackknife variance estimator even improves on both bias and efficiency, and, moreover, this estimator is mean-squared-error consistent, whereas the maximum likelihood equivalents are not
Interpreting degenerate solutions in unfolding by use of the vector model and the compensatory distance model
Interpreting degenerate solutions in unfolding by use of the vector model and the compensatory distance model
catholic university of leuven In this paper, we reconsider the merits of unfolding solutions based on loss functions involving a normalization on the variance per subject. In the literature, solutions based on Stress-2 are often diagnosed to be degenerate in the majority of cases. Here, the focus lies on two frequently occurring types of degen-eracies. The first type typically locates some subject points far away from a compact cluster of the other points. In the second type of solution, the object points lie on a circle. In this paper, we argue that these degenerate solutions are well fitting and informative. To reveal the information, we introduce mixtures of plots based on the ideal point model of unfolding, the vector model, and on the signed distance model. In addition to a different representation, we provide a new iterative majorization algorithm to optimize the average squared correlation between the distances in the configuration and the transformed data per individual. It is shown that this approach is equivalent to minimizing Kruskal’s Stress-2
Unfolding Incomplete Data: Guidelines for Unfolding Row-Conditional Rank Order Data with Random Missings
Unfolding, Incomplete data, Missing data, BIBD, PREFSCAL,
Comparing methods of analysing datasets with small clusters-case studies using four paediatric datasets
Studies of prematurely born infants contain a relatively large percentage of multiple births, so the resulting data have a hierarchical structure with small clusters of size 1, 2 or 3. Ignoring the clustering may lead to incorrect inferences. The aim of this study was to compare statistical methods which can be used to analyse such data: generalised estimating equations, multilevel models, multiple linear regression and logistic regression. Four datasets which differed in total size and in percentage of multiple births (n = 254, multiple 18%; n = 176, multiple 9%; n = 10 098, multiple 3%; n = 1585, multiple 8%) were analysed. With the continuous outcome, two-level models produced similar results in the larger dataset, while generalised least squares multilevel modelling (ML GLS 'xtreg' in Stata) and maximum likelihood multilevel modelling (ML MLE 'xtmixed' in Stata) produced divergent estimates using the smaller dataset. For the dichotomous outcome, most methods, except generalised least squares multilevel modelling (ML GH 'xtlogit' in Stata) gave similar odds ratios and 95% confidence intervals within datasets. For the continuous outcome, our results suggest using multilevel modelling. We conclude that generalised least squares multilevel modelling (ML GLS 'xtreg' in Stata) and maximum likelihood multilevel modelling (ML MLE 'xtmixed' in Stata) should be used with caution when the dataset is small. Where the outcome is dichotomous and there is a relatively large percentage of non-independent data, it is recommended that these are accounted for in analyses using logistic regression with adjusted standard errors or multilevel modelling. If, however, the dataset has a small percentage of clusters greater than size 1 (e.g. a population dataset of children where there are few multiples) there appears to be less need to adjust for clustering
Contingency theory to study leadership styles of small businesses owner-managers at Santo Domingo, Ecuador
A recursive partitioning method for the prediction of preference rankings based upon Kemeny distances
Multivariate analysis of psychological dat
