1,465 research outputs found

    Vector particles tunneling from BTZ black holes

    Full text link
    In this paper we investigate vector particles' Hawking radiation from a BTZ black hole. By applying the WKB approximation and the Hamilton-Jacobi Ansatz to the Proca equation, we obtain the tunneling spectrum of vector particles. The expected Hawking temperature is recovered

    Stacking-symmetry governed second harmonic generation in graphene trilayers

    Full text link
    Crystal symmetry plays a central role in governing a wide range of fundamental physical phenomena. One example is the nonlinear optical second harmonic generation (SHG), which requires inversion symmetry breaking. Here we report a unique stacking-induced SHG in trilayer graphene, whose individual monolayer sheet is centrosymmetric. Depending on layer stacking sequence, we observe a strong optical SHG in Bernal (ABA) stacked non-centrosymmetric trilayer, while it vanishes in rhombohedral (ABC) stacked one which preserves inversion symmetry. This highly contrasting SHG due to the distinct stacking symmetry enables us to map out the ABA and ABC crystal domains in otherwise homogeneous graphene trilayer. The extracted second order nonlinear susceptibility of the ABA trilayer is surprisingly large, comparable to the best known 2D semiconductors enhanced by excitonic resonance. Our results reveal a novel stacking order induced nonlinear optical effect, as well as unleash the opportunity for studying intriguing physical phenomena predicted for stacking-dependent ABA and ABC graphene trilayers.Comment: To appear in Science Advance

    Correlation effects in the ground state of trapped atomic Bose gases

    Full text link
    We study the effects of many-body correlations in trapped ultracold atomic Bose gases. We calculate the ground state of the gas using a ground-state auxiliary-field quantum Monte Carlo (QMC) method [Phys. Rev. E 70, 056702 (2004)]. We examine the properties of the gas, such as the energetics, condensate fraction, real-space density, and momentum distribution, as a function of the number of particles and the scattering length. We find that the mean-field Gross-Pitaevskii (GP) approach gives qualitatively incorrect result of the kinetic energy as a function of the scattering length. We present detailed QMC data for the various quantities, and discuss the behavior of GP, modified GP, and the Bogoliubov method under a local density approximation.Comment: 11 pages, 12 figures, as typeset using REVTEX4. Submitted to Phys. Rev.

    Pairing, Charge, and Spin Correlations in the Three-Band Hubbard Model

    Full text link
    Using the Constrained Path Monte Carlo (CPMC) method, we simulated the two-dimensional, three-band Hubbard model to study pairing, charge, and spin correlations as a function of electron and hole doping and the Coulomb repulsion VpdV_{pd} between charges on neighboring Cu and O lattice sites. As a function of distance, both the dx2y2d_{x^2 - y^2}-wave and extended s-wave pairing correlations decayed quickly. In the charge-transfer regime, increasing VpdV_{pd} decreased the long-range part of the correlation functions in both channels, while in the mixed-valent regime, it increased the long-range part of the s-wave behavior but decreased that of the d-wave behavior. Still the d-wave behavior dominated. At a given doping, increasing VpdV_{pd} increased the spin-spin correlations in the charge-transfer regime but decreased them in the mixed-valent regime. Also increasing VpdV_{pd} suppressed the charge-charge correlations between neighboring Cu and O sites. Electron and hole doping away from half-filling was accompanied by a rapid suppression of anti-ferromagnetic correlations.Comment: Revtex, 8 pages with 15 figure

    The Counterattack of CNNs in Self-Supervised Learning: Larger Kernel Size might be All You Need

    Full text link
    Vision Transformers have been rapidly uprising in computer vision thanks to their outstanding scaling trends, and gradually replacing convolutional neural networks (CNNs). Recent works on self-supervised learning (SSL) introduce siamese pre-training tasks, on which Transformer backbones continue to demonstrate ever stronger results than CNNs. People come to believe that Transformers or self-attention modules are inherently more suitable than CNNs in the context of SSL. However, it is noteworthy that most if not all prior arts of SSL with CNNs chose the standard ResNets as their backbones, whose architecture effectiveness is known to already lag behind advanced Vision Transformers. Therefore, it remains unclear whether the self-attention operation is crucial for the recent advances in SSL - or CNNs can deliver the same excellence with more advanced designs, too? Can we close the SSL performance gap between Transformers and CNNs? To answer these intriguing questions, we apply self-supervised pre-training to the recently proposed, stronger lager-kernel CNN architecture and conduct an apple-to-apple comparison with Transformers, in their SSL performance. Our results show that we are able to build pure CNN SSL architectures that perform on par with or better than the best SSL-trained Transformers, by just scaling up convolutional kernel sizes besides other small tweaks. Impressively, when transferring to the downstream tasks \texttt{MS COCO} detection and segmentation, our SSL pre-trained CNN model (trained in 100 epochs) achieves the same good performance as the 300-epoch pre-trained Transformer counterpart. We hope this work can help to better understand what is essential (or not) for self-supervised learning backbones

    Spatially and Spectrally Consistent Deep Functional Maps

    Full text link
    Cycle consistency has long been exploited as a powerful prior for jointly optimizing maps within a collection of shapes. In this paper, we investigate its utility in the approaches of Deep Functional Maps, which are considered state-of-the-art in non-rigid shape matching. We first justify that under certain conditions, the learned maps, when represented in the spectral domain, are already cycle consistent. Furthermore, we identify the discrepancy that spectrally consistent maps are not necessarily spatially, or point-wise, consistent. In light of this, we present a novel design of unsupervised Deep Functional Maps, which effectively enforces the harmony of learned maps under the spectral and the point-wise representation. By taking advantage of cycle consistency, our framework produces state-of-the-art results in mapping shapes even under significant distortions. Beyond that, by independently estimating maps in both spectral and spatial domains, our method naturally alleviates over-fitting in network training, yielding superior generalization performance and accuracy within an array of challenging tests for both near-isometric and non-isometric datasets. Codes are available at https://github.com/rqhuang88/Spatiallyand-Spectrally-Consistent-Deep-Functional-Maps.Comment: Accepted by ICCV202

    Efficient IoT Inference via Context-Awareness

    Full text link
    While existing strategies to execute deep learning-based classification on low-power platforms assume the models are trained on all classes of interest, this paper posits that adopting context-awareness i.e. narrowing down a classification task to the current deployment context consisting of only recent inference queries can substantially enhance performance in resource-constrained environments. We propose a new paradigm, CACTUS, for scalable and efficient context-aware classification where a micro-classifier recognizes a small set of classes relevant to the current context and, when context change happens (e.g., a new class comes into the scene), rapidly switches to another suitable micro-classifier. CACTUS features several innovations, including optimizing the training cost of context-aware classifiers, enabling on-the-fly context-aware switching between classifiers, and balancing context switching costs and performance gains via simple yet effective switching policies. We show that CACTUS achieves significant benefits in accuracy, latency, and compute budget across a range of datasets and IoT platforms.Comment: 12 pages, 8 figure
    corecore