Rowlands D.J.,University of Oxford |
Rowlands D.J.,Center for the Analysis of Time Series |
Frame D.J.,University of Oxford |
Frame D.J.,Victoria University of Wellington |
And 33 more authors.
Nature Geoscience | Year: 2012
Incomplete understanding of three aspects of the climate system-equilibrium climate sensitivity, rate of ocean heat uptake and historical aerosol forcing-and the physical processes underlying them lead to uncertainties in our assessment of the global-mean temperature evolution in the twenty-first century 1,2. Explorations of these uncertainties have so far relied on scaling approaches 3,4, large ensembles of simplified climate models 1,2, or small ensembles of complex coupled atmosphere-ocean general circulation models 5,6 which under-represent uncertainties in key climate system properties derived from independent sources 7,9. Here we present results from a multi-thousand-member perturbed-physics ensemble of transient coupled atmosphere-ocean general circulation model simulations. We find that model versions that reproduce observed surface temperature changes over the past 50 years show global-mean temperature increases of 1.4-3 K by 2050, relative to 1961-1990, under a mid-range forcing scenario. This range of warming is broadly consistent with the expert assessment provided by the Intergovernmental Panel on Climate Change Fourth Assessment Report, but extends towards larger warming than observed in ensembles-of-opportunity 5 typically used for climate impact assessments. From our simulations, we conclude that warming by the middle of the twenty-first century that is stronger than earlier estimates is consistent with recent observed temperature changes and a mid-range 'no mitigation' scenario for greenhouse-gas emissions. © 2012 Macmillan Publishers Limited. All rights reserved.
De La Flor G.,University of Oxford |
Jirotka M.,Oxford search Center |
Luff P.,King's College London |
Pybus J.,Oxford search Center |
Kirkham R.,Oxford search Center
Computer Supported Cooperative Work | Year: 2010
e-Research and Cyberinfrastructure programmes actively promote the development of new forms of scientific practice and collaboration through the implementation of tools and technologies that support distributed collaborative work across geographically dispersed research institutes and laboratories. Whilst originating in scientific domains, we have more recently seen a turn to the design of systems that support research practices in the social sciences and the arts and humanities. Attempts to embed large-scale infrastructures into research settings has brought to the fore the necessity of understanding the knowledge, skills and practices of researchers within a variety of disciplines that might use these technologies. In this paper, we consider an approach to gathering requirements through the introduction of various technical interventions for relatively short term periods so that we may come to an understanding their impact on routine work practices. Drawing upon an analysis of the detailed ways in which classicists work with digital images, we discuss the requirements for systems that support them as they collaborate in the interpretation of particular types of images. We discuss implications for the development of infrastructures to support research collaboration in this area and conclude with reflections upon the experiences gained from conducting naturalistic studies in parallel with design interventions. © 2010 Springer Science+Business Media B.V.
Harvey P.,Durham University |
Kuprov I.,Oxford search Center |
Parker D.,Durham University
European Journal of Inorganic Chemistry | Year: 2012
The physicochemical basis of probe design for 19F MRS and MRI applications is reviewed. Complexes that give a single major resonance in solution are described, in which the Ln 3+ ion is about 6 É from the 19F label. Sensitivity improvements of 15-fold are reported in both imaging and spectroscopy based on longitudinal relaxation enhancement. The pseudocontact shift allows an amplification of chemical shift non-equivalence in responsive 19F probes, e.g. for monitoring pH in the range between 5 and 8. Sensitivity in 19F magnetic resonance spectroscopy and imaging is enhanced by placing a paramagnetic lanthanide within 7 E of the spin label. Faster relaxation allows more rapid data acquisition for systems generating one main resonance, and the proximate lanthanide ion amplifies the chemical shift non-equivalence in responsive probes. Copyright © 2012 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Dasgupta A.,UNC Charlotte |
Chen M.,Oxford search Center |
Kosara R.,UNC Charlotte
Computer Graphics Forum | Year: 2012
Uncertainty is an intrinsic part of any visual representation in visualization, no matter how precise the input data. Existing research on uncertainty in visualization mainly focuses on depicting data-space uncertainty in a visual form. Uncertainty is thus often seen as a problem to deal with, in the data, and something to be avoided if possible. In this paper, we highlight the need for analyzing visual uncertainty in order to design more effective visual representations. We study various forms of uncertainty in the visual representation of parallel coordinates and propose a taxonomy for categorizing them. By building a taxonomy, we aim to identify different sources of uncertainty in the screen space and relate them to different effects of uncertainty upon the user. We examine the literature on parallel coordinates and apply our taxonomy to categorize various techniques for reducing uncertainty. In addition, we consider uncertainty from a different perspective by identifying cases where increasing certain forms of uncertainty may even be useful, with respect to task, data type and analysis scenario. This work suggests that uncertainty is a feature that can be both useful and problematic in visualization, and it is beneficial to augment an information visualization pipeline with a facility for visual uncertainty analysis. Computer Graphics Forum © 2012 The Eurographics Association and Blackwell Publishing Ltd.
Salvini S.,Oxford search Center |
Wijnholds S.J.,Netherlands Institute for Radio Astronomy
2014 31th URSI General Assembly and Scientific Symposium, URSI GASS 2014 | Year: 2014
Alternating Direction Implicit (ADI) methods provide a computationally efficient way to solve for antenna based gains in full polarization. In this paper, we analyze the convergence of such methods in simulations. We show that convergence of a basic implementation can be quite slow and we propose two forms of relaxation to improve convergence behavior. The algorithm can be shown to be statistically efficient in low-SNR scenarios, which makes this approach particularly suitable for calibration of large radio astronomical arrays like the Square Kilometre Array (SKA). This also led to the name Statistically Efficient and Fast Calibration (StEFCal). The algorithm was implemented in several stages of the Low Frequency Array (LOFAR) calibration pipeline. We report on calibration performance improvement achieved with StEFCal in this pipeline. © 2014 IEEE.
Timlin C.,Particle Therapy Cancer Research Institute |
Houston M.,Particle Therapy Cancer Research Institute |
Houston M.,Oxford search Center |
Jones B.,Particle Therapy Cancer Research Institute |
Jones B.,University of Oxford
British Journal of Radiology | Year: 2011
The aim of this study was to display malignant induction probability (MIP) maps alongside dose distribution maps for radiotherapy using X-ray and charged particles such as protons. Dose distributions for X-rays and protons are used in an interactive MATLAB® program (MathWorks, Natick, MA). The MIP is calculated using a published linear quadratic model, which incorporates fractionation effects, cell killing and cancer induction as a function of dose, as well as relative biological effect. Two virtual situations are modelled: (a) a tumour placed centrally in a cubic volume of normal tissue and (b) the same tumour placed closer to the skin surface. The MIP is calculated for a variety of treatment field options. The results show that, for protons, the MIP increases with field numbers. In such cases, proton MIP can be higher than that for X-rays. Protons produce the lowest MIPs for superficial targets because of the lack of exit dose. The addition of a dose bath to all normal tissues increases the MIP by up to an order of magnitude. This exploratory study shows that it is possible to achieve three-dimensional displays of carcinogenesis risk. The importance of treatment geometry, including the length and volume of tissue traversed by each beam, can all influence MIP. Reducing the volume of tissue irradiated is advantageous, as reducing the number of cells at risk reduces the total MIP. This finding lends further support to the use of treatment gantries as well as the use of simpler field arrangements for particle therapy provided normal tissue tolerances are respected. © 2011 The British Institute of Radiology.
Warren A.J.,Diamond Light Source |
Armour W.,Diamond Light Source |
Armour W.,Oxford search Center |
Axford D.,Diamond Light Source |
And 9 more authors.
Acta Crystallographica Section D: Biological Crystallography | Year: 2013
The focus in macromolecular crystallography is moving towards even more challenging target proteins that often crystallize on much smaller scales and are frequently mounted in opaque or highly refractive materials. It is therefore essential that X-ray beamline technology develops in parallel to accommodate such difficult samples. In this paper, the use of X-ray microradiography and microtomography is reported as a tool for crystal visualization, location and characterization on the macromolecular crystallography beamlines at the Diamond Light Source. The technique is particularly useful for microcrystals and for crystals mounted in opaque materials such as lipid cubic phase. X-ray diffraction raster scanning can be used in combination with radiography to allow informed decision-making at the beamline prior to diffraction data collection. It is demonstrated that the X-ray dose required for a full tomography measurement is similar to that for a diffraction grid-scan, but for sample location and shape estimation alone just a few radiographic projections may be required.
Obara B.,Oxford search Center |
Fricker M.,University of Oxford |
Gavaghan D.,Oxford Center for Integrative Systems Biology |
Grau V.,University of Oxford
IEEE Transactions on Image Processing | Year: 2012
Many biomedical applications require detection of curvilinear structures in images and would benefit from automatic or semiautomatic segmentation to allow high-throughput measurements. Here, we propose a contrast-independent approach to identify curvilinear structures based on oriented phase congruency, i.e., the phase congruency tensor (PCT). We show that the proposed method is largely insensitive to intensity variations along the curve and provides successful detection within noisy regions. The performance of the PCT is evaluated by comparing it with state-of-the-art intensity-based approaches on both synthetic and real biological images. © 1992-2012 IEEE.
Darch P.,University of Oxford |
Carusi A.,Oxford search Center
Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences | Year: 2010
Volunteer computing projects (VCPs) have been set up by groups of scientists to recruit members of the public who are asked to donate spare capacity on their personal computers to the processing of scientific data or computationally intensive models. VCPs serve two purposes: to acquire significant computing capacity and to educate the public about science. A particular challenge for these scientists is the retention of volunteers as there is a very high drop-out rate. This paper develops recommendations for scientists and software engineers setting up or running VCPs regarding which strategies to pursue in order to improve volunteer retention rates. These recommendations are based on a qualitative study of volunteers in a VCP (climateprediction.net). A typology of volunteers has been developed, and three particularly important classes of volunteers are presented in this paper: for each type of volunteer, the particular benefits they offer to a project are described, and their motivations for continued participation in a VCP are identified and linked to particular strategies. In this way, those setting up a VCP can identify which types of volunteers they should be particularly keen to retain, and can then find recommendations to increase the retention rates of their target volunteers. © 2010 The Royal Society.
Olteanu D.,University of Oxford |
Van Schaik S.J.,Oxford search Center
Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining | Year: 2012
DAGger is a clustering algorithm for uncertain data. In contrast to prior work, DAGger can work on arbitrarily correlated data and can compute both exact and approximate clusterings with error guarantees. We demonstrate DAGger using a real-world scenario in which partial discharge data from UK Power Networks is clustered to predict asset failure in the energy network. © 2012 ACM.