College of William and Mary | Date: 2015-09-22
Gesture-enabled remote control is implemented using a portable device having motion sensors and a wireless transmitter. An activation signal is received when the motion sensors detect a first prescribed motion of the portable device. A neutral orientation is then assigned to the portable device. The neutral orientation is defined by a position of the portable device at the time the activation signal is received. A control signal is generated when the motion sensors detect one of a plurality of prescribed movements of the portable device occurring within a prescribed window of time. Each prescribed movement includes movement of the portable device away from the neutral orientation and then back to the neutral orientation. The control signal is formatted for wireless transmission to an electronic device for control thereof.
Nanjing University, College of William and Mary | Date: 2015-11-10
Systems, methods and techniques are provided for interacting with mobile devices using a camera-based keyboard. The system comprises a processor system including at last one processor. The processor system is configured to at least capture a plurality of images in connection with the keyboard and at least one hand typing on the keyboard via the camera. Based on the plurality of captured images, the processor system is further configured to locate the keyboard, extract at least a portion of the keys on the keyboard, extract a hand, and detect a fingertip of the extracted hand. After that, a keystroke may be detected and localized through tracking the detected fingertip in at least one of the plurality of captured images, and a character corresponding to the localized keystroke may be determined.
College of William and Mary | Date: 2016-12-01
Methods and kits are described for accelerating an aged appearance in indido-dyed denim garments. A high concentration of cellulase enzymes is used to treat denim garments such as blue jeans, followed by a washing step and an abrasion step, followed by repetition of the cycle. This process can be performed at home in standard washing machines, and provides consumers with an alternative approach to customizing their denim garments.
College of William and Mary | Date: 2016-12-15
Herein we describe shotgun shells each having a degradable wad comprising a polyhydroxyalkanoate polymer, as well as methods for ensuring that such wad degrades quickly after landing in an aquatic environment. Plastic, aquatic debris has a negative economic and ecological impact, and thus it is advantageous to use implements that will degrade into environmentally benign compounds. Incorporating a degradable wad into shotgun shells provides an effective, economical solution. Suitable degradable shotgun wads and methods are described herein.
Carone C.D.,College of William and Mary
Physics Letters, Section B: Nuclear, Elementary Particle and High-Energy Physics | Year: 2013
The possibility of explaining the current value of the muon anomalous magnetic moment in models with an additional U(1) gauge symmetry that has kinetic mixing with hypercharge are increasingly constrained by dark photon searches at electron accelerators. Here we present a scenario in which the couplings of new, light gauge bosons to standard model leptons are naturally weak and flavor nonuniversal. A vector-like state that mixes with standard model leptons serves as a portal between the dark and standard model sectors. The flavor symmetry of the model assures that the induced couplings of the new gauge sector to leptons of the first generation are very small and lepton-flavor-violating processes are adequately suppressed. The model provides a framework for constructing ultraviolet complete theories in which new, light gauge fields couple weakly and dominantly to leptons of the second or third generations. © 2013 Elsevier B.V.
Carlson C.E.,College of William and Mary
Progress in Particle and Nuclear Physics | Year: 2015
The proton size, specifically its charge radius, was thought known to about 1% accuracy. Now a new method probing the proton with muons instead of electrons finds a radius about 4% smaller, and to boot gives an uncertainty limit of about 0.1%. We review the different measurements, some of the calculations that underlie them, some of the suggestions that have been made to resolve the conflict, and give a brief overview new related experimental initiatives. At present, however, the resolution to the problem remains unknown. © 2015 Elsevier B.V. All rights reserved.
Agency: NSF | Branch: Standard Grant | Program: | Phase: SEES Coastal | Award Amount: 407.67K | Year: 2016
This research examines the potential for achieving sustainability in coastal systems where natural resources are impacted by both climate change and human responses to climate change. Chesapeake Bay shorescapes (a shoreline zone which includes riparian, intertidal, and near-shore shallow water areas) are used as a test bed because sea level rise creates risks for shoreline property owners and shoreline marshes. Property owner perception and response to the risk varies and their choice of a shoreline protection approach (armoring, living shoreline, do nothing) has consequences for the capacity of shoreline marshes to continue to provide ecosystem services. Impacts on marsh services that include supporting fisheries and improving water quality affect the larger estuarine system, with consequences for the many users of that system. This leads to regulation by government officials, who have their own perception of all these issues. Modeling the decision-making process of shoreline property owners, the ecological consequences of shoreline management decisions, and the perceptions of officials developing or implementing natural resource policies will discover opportunities and options for managing the combined human and natural system to desired outcomes.
The goal of the project is to discover the elements of the shorescape social-ecological system that have the greatest influence on attainment of sustainable outcomes. The research describes the trajectory of Chesapeake Bay shorescapes in terms of changes in amount, distribution, and character of shoreline types. The primary drivers of these changes are rising sea levels and actions of shoreline property owners to combat erosion. An analysis of existing information on shoreline conditions, property owner characteristics and property management decisions is used to model future shoreline management choices. A series of field investigations comparatively quantify multiple ecosystem functions (habitat provision, primary production, nutrient and carbon storage) of living shorelines and natural marshes along a continuum of estuarine shorescape settings and project future shifts in function under varying management scenarios. This information is compiled in a marsh function model that can use input from the shoreline management model. Future outcomes then are forecast under scenarios with alternative sea level rise and management conditions. Surveys of government officials responsible for policy development and implementation document the operative feedbacks from the ecological consequences of property owner decisions. Synthesis of this information identifies the characteristics of shorescape socio-ecological systems that enhance or detract from their ability to achieve sustainable outcomes. Formal guidance for coastal managers and policy makers will be developed using the results of this integrative approach to sustainability in Chesapeake Bay shorescapes.
Agency: NSF | Branch: Standard Grant | Program: | Phase: MAJOR RESEARCH INSTRUMENTATION | Award Amount: 300.00K | Year: 2016
Quantum Chromodynamics (QCD) is the theory of strong interactions, the fundamental force in Nature that binds quarks and gluons into subatomic particles (hadrons), such as protons and neutrons. The study of QCD properties is one of the central scientific challenges in nuclear physics. Nuclear physics scientists use computational methods and large-scale simulations to further the detailed understanding of hadronic structure. These simulations make use of state-of-the-art computational resources and researchers continue to improve their high-performance computing (HPC) methods to take advantage of the latest breakthrough in computer technology. This project will allow the acquisition of a parallel computer cluster incorporating the next-generation Intel multi-core processors (Knights Landing) and a high-performance communication network. The Knights Landing architecture has the potential to revolutionize scientific computing, by providing large computing resources at low cost and using significantly less power than conventional computer clusters. This computer cluster will be housed in a state-of-the-art computing center in the new Integrated Science Center at the College of William & Mary and will form the centerpiece of high-performance computing on the W&M campus.
The primary scientific driver for the system is to use lattice QCD methods to build a solid foundation for nuclear physics based on the Standard Model of elementary particle physics. With increased computational power and recent algorithmic developments, the research carried out with this system will advance the Lattice QCD program. This system will be used as a test bed for the development of new algorithms for lattice QCD calculations. These algorithms will be designed in collaboration with computer scientists specifically for the modern multi-core architectures that are currently the trent in high-performance computing. Furthermore, this computer cluster will serve as the center of interdisciplinary research between computer scientists and physicists and will catalyze the integration of research and education through the training of students and postdoctoral researchers in the groups of the senior investigators and in the wider academic community at William and Mary.
This project advances the objectives of the National Strategic Computing Initiative (NSCI), an effort aimed at sustaining and enhancing the U.S. scientific, technological, and economic leadership position in High-Performance Computing (HPC) research, development, and deployment.
Agency: NSF | Branch: Standard Grant | Program: | Phase: COMPUTER SYSTEMS | Award Amount: 299.99K | Year: 2016
Large, distributed systems are nowadays ubiquitous and part of sustainable IT solutions to a broad range of customers and applications. Data centers in the private or public cloud and high performance computing systems are two examples of complex, highly distributed systems: the former are used by almost everyone on a daily basis, the latter are used by computational scientists for advancing science and engineering. High availability and reliability of these complex systems are important for the quality of user experience. Efficient management of such systems contributes to their availability and reliability, and relies on a priori knowledge of the timing of the collective demands of users and a priori knowledge of certain performance measures (e.g., usage, temperature, power) of various systems components.
This project aims to provide a systematic methodology to improve the operational efficiency of complex, distributed systems by developing neural networks that can efficiently and accurately predict the incoming workload within fine and coarse time scales. Such workload prediction can dramatically improve the operational efficiency of data centers and high performance systems by driving proactive management strategies that specifically aim to enhance reliability. For datacenters, the focus is on actively reducing performance tickets that are automatically triggered by pro-actively managing virtual machine resizing and migration. For high performance computing systems the focus is on predicting hardware faults to autonomically improve the schedulers efficiency, direct cooling, and improve performance and memory bandwidth.
Agency: NSF | Branch: Standard Grant | Program: | Phase: CRII CISE Research Initiation | Award Amount: 175.00K | Year: 2017
Graphics Processing Units (GPUs) are becoming an inevitable part of every computing system because of their ability to enable orders of magnitude faster and energy-efficient execution. However, the necessary and continuous scaling of GPUs in terms of performance and energy efficiency will not be an easy task. Prior works have shown that two biggest impediments towards this scaling are the limited memory bandwidth and the excessive data movement across different levels of the memory hierarchy. In order to alleviate these two issues, die-stacking technology is gaining momentum in the realm of high-performance energy-efficient GPU computing. This technology not only enables very high memory bandwidth for better performance but also provides support for processing-near-memory (PNM) to reduce data movement, access latencies, and energy consumption. Although these technologies seem promising, the architectural support and execution models for PNM-based GPUs and their implications on the entire system design have largely been unexplored. This project takes a fresh look at the design and execution model of a PNM-enabled GPU, which consists of multiple memory stacks and each memory stack incorporates a 3D-stacked logic layer that can consist of multiple PNM GPU cores and other uncore components. Considering that GPUs are becoming an inevitable part of every computing system ranging from warehouse-scale computers to wearable devices, the insights resulting from this research can have a long-term positive impact on the GPU-based computing. The findings of this research will be incorporated to existing and new undergraduate and graduate courses, which will directly help in educating and training students, including women and students from diverse backgrounds and minority groups.
First, a detailed design space exploration will be performed, which will involve the study of the impact and interactions of different design choices related to PNM cores (e.g., register file, SIMD width, pipeline components, warp occupancy), uncore components at the logic layer (e.g., caches) and stacked memory (e.g., number of stacked memories). Second, a computation distribution framework (CDF) will be developed that will answer: a) when is it preferable to map computations to PNM cores, b) which PNM cores and computations they should be?, and c) how can we effectively take advantage of both PNM and regular GPU cores? The CDF will leverage different static and runtime strategies to address many of such similar questions to push the envelopes of energy efficiency and performance even further. The proposed research components will be evaluated via a wide-range of GPGPU applications. If successful, the findings of this research would better equip PNM-enabled GPUs to effectively alleviate the two major bottlenecks: memory bandwidth and energy.