112 documents found in 204ms
# 101
Parton, Graham A. • Pepler, Sam J.
Abstract: The 2007 British Atmospheric Data Centre (BADC) Users Survey examined the skill base of the BADC’s user community. Results indicated a large proportion of users who were familiar with data held in ASCII formats such as comma-separated variables (csv) and there was a high degree of familiarity with spreadsheet programmes (e.g. Excel) for data analysis purposes. These result, combined with the experiences of the BADC staff dealing with user enquiries and assisting data suppliers in preparing data for submission, and the metadata requirements of the BADC, highlighted the need for a new ASCII format to be generated. The BADC-CSV format adheres to metadata conventions covered by the NASAAmes and NetCDF formats, the CF, and Dublin Core metadata conventions, the ISO19115 standard and the metadata requirements of the BADC and its sister data centres within the Natural Environment Research Council (NERC). The format meets end user and data supplier requirements by being a native format for spreadsheet software as well as other commonly used data production and analysis tools (e.g. IDL, MatLab). This paper presents the requirements for the format resulting from the 2007 user survey and data centre requirements, describes the structure of the format and demonstrates the format through short examples. Finally, ongoing work to further develop the format is discussed.
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 107-111
# 102
Schmidt, Patrick • Märker, Michael
Abstract: In this paper we present a concept for the storage, exchange, presentation, and analysis of geospatial, environmental, and archaeological data with the aim to study and assess the dynamics of early human expansion. We present geographic information systems (GIS), database solutions and web based technologies for handling and processing archaeological and physiographic information. The project deals with a range of variables and file formats from the fields of geology, geomorphology, paleontology, and archaeology presented in vector, raster, and text formats. In order to demonstrate the possibilities of the ROCEEH Out of Africa Database (ROAD), this study presents a unique set of archaeological information sampled from a period between 3Ma and 20Ka covering the region of Southern Africa. The data includes site locations, their geology, and most important, the archaeological, paleo-biological, and climatic data that are accessed for these different locations. To help explaining the spatial distribution of archaeological and paleontological sites, detailed topographic information was provided at scales of 5m, 10m, 25m, 90m, and 1km based on aerial photography, ASTER, and SRTM elevation data. Moreover, bathymetry was provided at a resolution of 1km. Combining these data sets, we show the first spatial and conceptual analysis of the data using ROAD. The results of different research questions show the potential of these methods and give some perspectives for future applications.
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 113-117
# 103
Stockhause, Martina • Höck, Heinke
Abstract: At the KlimaCampus (http://www.klimacampus.de), Cluster of Excellence at the University of Hamburg, an Integrated Climate Data Center (ICDC: http://www.icdc.zmaw.de) is established, suitable for data during the scientific project phase as well as storing long-term archive data. ICDC aims to make data out of different internal and external archives easily accessible for the daily work of the KlimaCampus scientists. It extends the existing services by the announcement of data during the scientific project phase, a data portal and collaboration services. Therein, ICDC utilizes the available infrastructure at the WDC Climate by using it for metadata storage and as a long-term archive. The concept of ICDC, its functionality, its implementation status, and future perspectives are presented.
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 127-135
# 104
Fleury, Laurence • Boichard, Jean-Luc • Brissebrat, Guillaume • Cloché, Sophie • Eymard, Laurence • (et. al.)
Abstract: AMMA program includes in situ measurements at many locations of West Africa and in the Gulf of Guinea, an intensive use of satellite data, diverse modelling studies, as well as human sciences field surveys and value-added products processing. Therefore, AMMA database aims at storing a great amount and a large variety of data, and at providing the data as rapidly and safely as possible to the AMMA research community. In order to stimulate the exchange of information and collaboration between researchers from different disciplines or using different tools, the database provides a detailed description of the products and uses standardized formats. AMMA database and the associated online tools have been fully developed and are managed by two teams in France (IPSL Data Centre, Palaiseau and OMP Data Centre, Toulouse). Datasets are stored in one or the other centre depending on their types, but all of them can be accessed through a single and friendly data request user interface. The complete system has been duplicated at AGHRYMET Regional Centre (CRA) in Niamey, Niger and is operational there since January 2009.
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 45-51
# 105
Effertz, Eva
Abstract: The German Research Foundation (Deutsche Forschungsgemeinschaft, DFG) is well aware of the importance of systematic and integrated data management in research funding. At first the DFG, its aims and expenditure, will be introduced. It then will be explained why scientific infrastructure is funded by the DFG and how this is performed with the funding of specific information infrastructure through the Department of Scientific Library Services
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 35-38
# 106
Bolten, Andreas • Hoffmeister, Dirk • Willmes, Christian • Bubenzer, Olaf • Bareth, Georg
Abstract: The data management concept of the CRC/TR 32 'SVA-Patterns' (Curdt et al. 2008, 2010) offers a perfect tool to store and manage the data of an interdisciplinary research project. The focus here is on data management of soil (S), vegetation (V) and atmosphere (A) data for SVA-modelling purpose for a distinct watershed, the Rur-Catchment in Western Germany. The presented paper reflects the possibility and the adapting steps to transfer the concept to another interdisciplinary project. However, the technical transfer of the framework of the database is evident but the system needs to be adapted to the requirements of the benefit project. The Collaborative Research Centre (CRC) 806 'Our Way to Europe' performs a totally different investigation question and area, to capture the major intercontinental and transcontinental events of dispersal of modern man from Africa to central Europe. Therefore, other data formats and data amount occurs and other demands on the data management system are made. The paper identifies three spheres of action to adapt the concept to the CRC 806. (i) The integration of external databases to afford both, the transfer of data to other data management systems and to provide an interface to allow requests of external databases. (ii) The visualization of 3D-data in a web mapping application. (iii) The possibility for external and internal users to use geoprocessing tools to perform analysis based on the geodata in a web mapping application.
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 7-11
# 107
Curdt, Constanze • Hoffmeister, Dirk • Jekel, Christian • Udelhoven, Krischan • Waldhoff, Guido • (et. al.)
Abstract: Centralized storage and exchange of data is a key issue of interdisciplinary research projects particularly if they focus on environmental field studies and regional modelling. Furthermore, the overall success of such projects depends on a well organized data management and data exchange between all involved disciplines. This includes the organization of data, the implementation of a database and the maintenance of such a system for intensive data exchange between the project sections. This paper presents the implementation of a centralized data management system for the Transregional Collaborative Research Centre 32: 'Patterns in Soil-Vegetation-Atmosphere Systems: Monitoring, Modelling, and Data Assimilation' funded by the German Research Foundation. The design is a combination of file management, database, and online-interface including web mapping functionalities.
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 27-33
# 108
Eichler, Mareike • Francke, Till • Kneis, David • Reusser, Dominik E.
Abstract: Monitoring and modeling projects usually involves time series data originating from different sources. Often, file formats, temporal resolution, and metadata documentation rarely adhere to a common standard. As a result, much effort is spent on converting, harmonizing, merging, checking, resampling, and reformatting these data. Moreover, in work groups or during the course of time, these tasks tend to be carried out redundantly and repeatedly, especially when new data becomes available. The resulting duplication of data in various formats strains additional resources. We propose a database structure and complementary scripts for facilitating these tasks. The GOLM- (General Observation and Location Management) framework allows import and storage of time series data of different types while assisting metadata documentation, plausibility checking and harmonization. The imported data can be visually inspected and its coverage among locations and variables may be visualized. Supplementing scripts provides options for data export of selected stations and variables and resampling of the data to the desired temporal resolution. These tools can, for example, be used for generating model input files or reports. Since GOLM fully supports network access, the system can be used efficiently by distributed working groups accessing the same data over the internet. GOLM’s database structure and the complementary scripts can easily be customized to specific needs. Any involved software such as MySQL, R, PHP, OpenOffice as well as the scripts for building and using the data base, including documentation, are free for download. GOLM was developed out of the practical requirements of the OPAQUE-project. It has been tested and further refined in the ERANET-CRUE and SESAM projects, which all used GOLM to manage meteorological, hydrological and/or water quality data.
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 39-44
# 109
Fuhrmann, Patrick
Abstract: With the ignition of the Large Hadron Collider at CERN, End of November 2009, the largest physics experiment on earth will start to generate a stream of data, accumulating to about 15 Petabytes per year. Following a Tier model, this data will be replicated to various places around the globe. dCache is the storage system in charge of managing the largest share of this data outside CERN. It’s a data management technology designed for storing, retrieving, and managing huge amounts of data, distributed among a large number of heterogeneous server nodes, under a single virtual file system. The largest dCache installations are storing more than three Petabytes of data in a single instance on disk and on tape and they are expected to grow into the tenth Petabytes area by 2012. While dCache has been designed to fit the needs of High Energy Physics experiments, new challenges are waiting for large-scale storage solutions. Experiments in astronomy and in the area of extreme light sources will increase the required data rates and storage capacities by at least a factor of 10 and similarly important, they will force community storage systems to become compliant with industry standards. This publication will briefly describe how dCache fulfils the needs of LHC storage grid and is preparing itself for the upcoming challenges in other scientific disciplines.
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 53-57
# 110
Baru, Chaitanya K.
Abstract: This paper provides an overview of the range of techniques available for integration of heterogeneous data. These range from the wrapper-mediator architecture for integration of structured and semistructured databases: semantic mediation, which involves mapping schema elements and data values to ontologies: to ad hoc, vertical data integration where the user is in the loop of the integration. Every integration technique requires an expert in the loop—at different points in time and at different places in the system, depending on the integration technique employed. Future directions include provision of provenance and social networking information corresponding to the integrated data result.
Proceedings of the Data Management Workshop, 29-30 October 2009, University of Cologne, Germany, Kölner Geographische Arbeiten, 90, pp. 1-6
spinning wheel Loading next page