344 research outputs found

    Stochastic Growth Equations and Reparametrization Invariance

    Full text link
    It is shown that, by imposing reparametrization invariance, one may derive a variety of stochastic equations describing the dynamics of surface growth and identify the physical processes responsible for the various terms. This approach provides a particularly transparent way to obtain continuum growth equations for interfaces. It is straightforward to derive equations which describe the coarse grained evolution of discrete lattice models and analyze their small gradient expansion. In this way, the authors identify the basic mechanisms which lead to the most commonly used growth equations. The advantages of this formulation of growth processes is that it allows one to go beyond the frequently used no-overhang approximation. The reparametrization invariant form also displays explicitly the conservation laws for the specific process and all the symmetries with respect to space-time transformations which are usually lost in the small gradient expansion. Finally, it is observed, that the knowledge of the full equation of motion, beyond the lowest order gradient expansion, might be relevant in problems where the usual perturbative renormalization methods fail.Comment: 42 pages, Revtex, no figures. To appear in Rev. of Mod. Phy

    Longitudinal chirality, enhanced non-reciprocity, and nano-scale planar one-way guiding

    Full text link
    When a linear chain of plasmonic nano-particles is exposed to a transverse DC magnetic field, the chain modes are elliptically polarized, in a single plane parallel to the chain axis; hence, a novel longitudinal plasmon-rotation is created. If, in addition, the chain geometry possesses longitudinal rotation, e.g. by using ellipsoidal particles that rotate in the same plane as the plasmon rotation, strong non-reciprocity is created. The structure possesses a new kind of chirality--the longitudinal chirality--and supports one-way guiding. Since all particles rotate in the same plane, the geometry is planar and can be fabricated by printing leaf-like patches on a single plane. Furthermore, the magnetic field is significantly weaker than in previously reported one-way guiding structures. These properties are examined for ideal (lossless) and for lossy chains.Comment: to appear in PR

    Subspace Projection Approaches to Classification and Visualization of Neural Network-Level Encoding Patterns

    Get PDF
    Recent advances in large-scale ensemble recordings allow monitoring of activity patterns of several hundreds of neurons in freely behaving animals. The emergence of such high-dimensional datasets poses challenges for the identification and analysis of dynamical network patterns. While several types of multivariate statistical methods have been used for integrating responses from multiple neurons, their effectiveness in pattern classification and predictive power has not been compared in a direct and systematic manner. Here we systematically employed a series of projection methods, such as Multiple Discriminant Analysis (MDA), Principal Components Analysis (PCA) and Artificial Neural Networks (ANN), and compared them with non-projection multivariate statistical methods such as Multivariate Gaussian Distributions (MGD). Our analyses of hippocampal data recorded during episodic memory events and cortical data simulated during face perception or arm movements illustrate how low-dimensional encoding subspaces can reveal the existence of network-level ensemble representations. We show how the use of regularization methods can prevent these statistical methods from over-fitting of training data sets when the trial numbers are much smaller than the number of recorded units. Moreover, we investigated the extent to which the computations implemented by the projection methods reflect the underlying hierarchical properties of the neural populations. Based on their ability to extract the essential features for pattern classification, we conclude that the typical performance ranking of these methods on under-sampled neural data of large dimension is MDA>PCA>ANN>MGD
    corecore