Jay Parker

Learn More
We propose to define, design, develop, deploy, and test a data semantics based system to provide interoperability for heterogeneous data in the earthquake science domain. We focus on the database management aspects of the work, including modeling the meaning of the data, providing for web service based access to heterogeneous data sources to scientists to(More)
The QuakeSim Problem Solving Environment uses a web-services approach to unify and deploy diverse remote data sources and processing services within a browser environment. Here we focus on the high-performance crustal modelling applications that will be included in this set of remote but in-teroperable applications. PARK is a model for unstable slip on a(More)
—QuakeSim is a project to develop a modeling environment for studying earthquake processes using a web services environment. In order to model interseismic processes multiple data types must be ingested including spaceborne GPS and InSAR data, geological fault data, and seismicity data. QuakeSim federates data from these multiple sources and integrates the(More)
We are using Web (Grid) service technology to demonstrate the assimilation of multiple distributed data sources (a typical data grid problem) into a major parallel high-performance computing earthquake forecasting code. Such a linkage of Geoinformatics with Geocomplexity demonstrates the value of the Solid Earth Research Virtual Observatory (SERVO) Grid(More)
In order to develop a solid earth science framework for understanding and studying of active tectonic and earthquake processes, this task develops simulation and analysis tools to study the physics of earthquakes using state-of-the-art modeling, data manipulation, and pattern recognition technologies. We develop clearly defined accessible data formats and(More)
—QuakeSim is problem-solving environment for understanding earthquake processes through the integration of multiscale models and data. The goal of QuakeSim is to substantially improve earthquake forecasts, which will ultimately lead to mitigation of damage from this natural hazard. Improved earthquake forecasting is dependent on measurement of surface(More)
Computer simulations will be key to substantial gains in understanding the earthquake process. Emerging information technologies make possible a major change in the way computers are used and data is accessed. An outline of a real-izable computational infrastructure includes standardization of data accessibility , harnessing high-performance computing(More)
—Advances in understanding earthquakes increasingly requires the integration of models and multiple distributed data products. Increasingly, data are acquired through large investments and utilizing their full potential requires a coordinated effort by many groups and experts who are often distributed both geographically and by expertise.