4,537 research outputs found

    Fundamental results from microgravity cell experiments with possible commericial applications

    Get PDF
    Some of the major milestones are presented for studies in cell biology that were conducted by the Soviet Union and the United States in the upper layers of the atmosphere and in outer space for more than thirty-five years. The goals have changed as new knowledge is acquired and the priorities for the use of microgravity have shifted toward basic research and commercial applications. Certain details concerning the impact of microgravity on cell systems is presented. However, it needs to be emphasized that in planning and conducting microgravity experiments, there are some important prerequisites not normally taken into account. Apart from the required background knowledge of previous microgravity and ground-based experiments, the investigator should have the understanding of the hardware as a physical unit, the complete knowledge of its operation, the range of its capabilities and the anticipation of problems that may occur. Moreover, if the production of commercial products in space is to be manifested, data obtained from previous microgravity experiments must be used to optimize the design of flight hardware

    Cells in Space

    Get PDF
    Discussions and presentations addressed three aspects of cell research in space: the suitability of the cell as a subject in microgravity experiments, the requirements for generic flight hardware to support cell research, and the potential for collaboration between academia, industry, and government to develop these studies in space. Synopses are given for the presentations and follow-on discussions at the conference and papers are presented from which the presentations were based. An Executive Summary outlines the recommendations and conclusions generated at the conference

    Cost-Effective Control of Infectious Disease Outbreaks Accounting for Societal Reaction

    Get PDF
    Background Studies of cost-effective disease prevention have typically focused on the tradeoff between the cost of disease transmission and the cost of applying control measures. We present a novel approach that also accounts for the cost of social disruptions resulting from the spread of disease. These disruptions, which we call social response, can include heightened anxiety, strain on healthcare infrastructure, economic losses, or violence. Methodology The spread of disease and social response are simulated under several different intervention strategies. The modeled social response depends upon the perceived risk of the disease, the extent of disease spread, and the media involvement. Using Monte Carlo simulation, we estimate the total number of infections and total social response for each strategy. We then identify the strategy that minimizes the expected total cost of the disease, which includes the cost of the disease itself, the cost of control measures, and the cost of social response. Conclusions The model-based simulations suggest that the least-cost disease control strategy depends upon the perceived risk of the disease, as well as media intervention. The most cost-effective solution for diseases with low perceived risk was to implement moderate control measures. For diseases with higher perceived severity, such as SARS or Ebola, the most cost-effective strategy shifted toward intervening earlier in the outbreak, with greater resources. When intervention elicited increased media involvement, it remained important to control high severity diseases quickly. For moderate severity diseases, however, it became most cost-effective to implement no intervention and allow the disease to run its course. Our simulation results imply that, when diseases are perceived as severe, the costs of social response have a significant influence on selecting the most cost-effective strategy.United States. Defense Threat Reduction Agency (Contract HDTRA1-12-C-0061

    The performance of RAMS in representing the convective boundary layer structure in a very steep valley

    Get PDF
    Data from a comprehensive field study in the Riviera Valley of Southern Switzerland are used to investigate convective boundary layer structure in a steep valley and to evaluate wind and temperature fields, convective boundary layer height, and surface sensible heat fluxes as predicted by the mesoscale model RAMS. Current parameterizations of surface and boundary layer processes in RAMS, as well as in other mesoscale models, are based on scaling laws strictly valid only for flat topography and uniform land cover. Model evaluation is required to investigate whether this limits the applicability of RAMS in steep, inhomogeneous terrain. One clear-sky day with light synoptic winds is selected from the field study. Observed temperature structure across and along the valley is nearly homogeneous while wind structure is complex with a wind speed maximum on one side of the valley. Upvalley flows are not purely thermally driven and mechanical effects near the valley entrance also affect the wind structure. RAMS captured many of the observed boundary layer characteristics within the steep valley. The wind field, temperature structure, and convective boundary layer height in the valley are qualitatively simulated by RAMS, but the horizontal temperature structure across and along the valley is less homogeneous in the model than in the observations. The model reproduced the observed net radiation, except around sunset and sunrise when RAMS does not take into account the shadows cast by the surrounding topography. The observed sensible heat fluxes fall within the range of simulated values at grid points surrounding the measurement sites. Some of the scatter between observed and simulated turbulent sensible heat fluxes are due to sub-grid scale effects related to local topograph

    CoGeNT: A Search for Low-Mass Dark Matter using p-type Point Contact Germanium Detectors

    Full text link
    CoGeNT employs p-type point-contact (PPC) germanium detectors to search for Weakly Interacting Massive Particles (WIMPs). By virtue of its low energy threshold and ability to reject surface backgrounds, this type of device allows an emphasis on low-mass dark matter candidates (wimp mass of about 10 GeV/c2). We report on the characteristics of the PPC detector presently taking data at the Soudan Underground Laboratory, elaborating on aspects of shielding, data acquisition, instrumental stability, data analysis, and background estimation. A detailed background model is used to investigate the low energy excess of events previously reported, and to assess the possibility of temporal modulations in the low-energy event rate. Extensive simulations of all presently known backgrounds do not provide a viable background explanation for the excess of low-energy events in the CoGeNT data, or the previously observed temporal variation in the event rate. Also reported on for the first time is a determination of the surface (slow pulse rise time) event contamination in the data as a function of energy. We conclude that the CoGeNT detector technology is well suited to search for the annual modulation signature expected from dark matter particle interactions in the region of WIMP mass and coupling favored by the DAMA/LIBRA resultsComment: 20 pages, 31 figures. Several figures have been added, including an updated allowed region (both 90% and 99% confidence level contours) based on this analysis. There is also the addition of a Pb-210 background estimat

    Quantifying black carbon deposition over the Greenland ice sheet from forest fires in Canada

    Get PDF
    Black carbon (BC) concentrations observed in 22 snowpits sampled in the northwest sector of the Greenland ice sheet in April 2014 have allowed us to identify a strong and widespread BC aerosol deposition event, which was dated to have accumulated in the pits from two snow storms between 27 July and 2 August 2013. This event comprises a significant portion (57% on average across all pits) of total BC deposition over 10 months (July 2013 to April 2014). Here we link this deposition event to forest fires burning in Canada during summer 2013 using modeling and remote sensing tools. Aerosols were detected by both the Cloud‐Aerosol Lidar with Orthogonal Polarization (on board CALIPSO) and Moderate Resolution Imaging Spectroradiometer (Aqua) instruments during transport between Canada and Greenland. We use high‐resolution regional chemical transport modeling (WRF‐Chem) combined with high‐resolution fire emissions (FINNv1.5) to study aerosol emissions, transport, and deposition during this event. The model captures the timing of the BC deposition event and shows that fires in Canada were the main source of deposited BC. However, the model underpredicts BC deposition compared to measurements at all sites by a factor of 2–100. Underprediction of modeled BC deposition originates from uncertainties in fire emissions and model treatment of wet removal of aerosols. Improvements in model descriptions of precipitation scavenging and emissions from wildfires are needed to correctly predict deposition, which is critical for determining the climate impacts of aerosols that originate from fires

    Active cooling control of the CLEO detector using a hydrocarbon coolant farm

    Full text link
    We describe a novel approach to particle-detector cooling in which a modular farm of active coolant-control platforms provides independent and regulated heat removal from four recently upgraded subsystems of the CLEO detector: the ring-imaging Cherenkov detector, the drift chamber, the silicon vertex detector, and the beryllium beam pipe. We report on several aspects of the system: the suitability of using the aliphatic-hydrocarbon solvent PF(TM)-200IG as a heat-transfer fluid, the sensor elements and the mechanical design of the farm platforms, a control system that is founded upon a commercial programmable logic controller employed in industrial process-control applications, and a diagnostic system based on virtual instrumentation. We summarize the system's performance and point out the potential application of the design to future high-energy physics apparatus.Comment: 21 pages, LaTeX, 5 PostScript figures; version accepted for publication in Nuclear Instruments and Methods in Physics Research

    Modeling organic aerosols in a megacity: potential contribution of semi-volatile and intermediate volatility primary organic compounds to secondary organic aerosol formation

    Get PDF
    It has been established that observed local and regional levels of secondary organic aerosols (SOA) in polluted areas cannot be explained by the oxidation and partitioning of anthropogenic and biogenic VOC precursors, at least using current mechanisms and parameterizations. In this study, the 3-D regional air quality model CHIMERE is applied to estimate the potential contribution to SOA formation of recently identified semi-volatile and intermediate volatility organic precursors (S/IVOC) in and around Mexico City for the MILAGRO field experiment during March 2006. The model has been updated to include explicitly the volatility distribution of primary organic aerosols (POA), their gas-particle partitioning and the gas-phase oxidation of the vapors. Two recently proposed parameterizations, those of Robinson et al. (2007) ("ROB") and Grieshop et al. (2009) ("GRI") are compared and evaluated against surface and aircraft measurements. The 3-D model results are assessed by comparing with the concentrations of OA components from Positive Matrix Factorization of Aerosol Mass Spectrometer (AMS) data, and for the first time also with oxygen-to-carbon ratios derived from high-resolution AMS measurements. The results show a substantial enhancement in predicted SOA concentrations (2–4 times) with respect to the previously published base case without S/IVOCs (Hodzic et al., 2009), both within and downwind of the city leading to much reduced discrepancies with the total OA measurements. Model improvements in OA predictions are associated with the better-captured SOA magnitude and diurnal variability. The predicted production from anthropogenic and biomass burning S/IVOC represents 40–60% of the total measured SOA at the surface during the day and is somewhat larger than that from commonly measured aromatic VOCs, especially at the T1 site at the edge of the city. The SOA production from the continued multi-generation S/IVOC oxidation products continues actively downwind. Similar to aircraft observations, the predicted OA/ΔCO ratio for the ROB case increases from 20–30 μg sm<sup>−3</sup> ppm<sup>−1</sup> up to 60–70 μg sm<sup>−3</sup> ppm<sup>−1</sup> between a fresh and 1-day aged air mass, while the GRI case produces a 30% higher OA growth than observed. The predicted average O/C ratio of total OA for the ROB case is 0.16 at T0, substantially below observed value of 0.5. A much better agreement for O/C ratios and temporal variability (<i>R</i><sup>2</sup>=0.63) is achieved with the updated GRI treatment. Both treatments show a deficiency in regard to POA ageing with a tendency to over-evaporate POA upon dilution of the urban plume suggesting that atmospheric HOA may be less volatile than assumed in these parameterizations. This study highlights the important potential role of S/IVOC chemistry in the SOA budget in this region, and highlights the need for further improvements in available parameterizations. The agreement observed in this study is not sufficient evidence to conclude that S/IVOC are the major missing SOA source in megacity environments. The model is still very underconstrained, and other possible pathways such as formation from very volatile species like glyoxal may explain some of the mass and especially increase the O/C ratio
    corecore