
NSF Org: |
RISE Integrative and Collaborative Education and Research (ICER) |
Recipient: |
|
Initial Amendment Date: | September 16, 2016 |
Latest Amendment Date: | July 9, 2020 |
Award Number: | 1639759 |
Award Instrument: | Standard Grant |
Program Manager: |
Eva Zanzerkia
RISE Integrative and Collaborative Education and Research (ICER) GEO Directorate for Geosciences |
Start Date: | September 1, 2016 |
End Date: | August 31, 2021 (Estimated) |
Total Intended Award Amount: | $783,999.00 |
Total Awarded Amount to Date: | $783,999.00 |
Funds Obligated to Date: |
|
History of Investigator: |
|
Recipient Sponsored Research Office: |
1 E JACKSON BLVD CHICAGO IL US 60604-2287 (312)362-7388 |
Sponsor Congressional District: |
|
Primary Place of Performance: |
243 South Wabash Avenue Chicago IL US 60604-2287 |
Primary Place of
Performance Congressional District: |
|
Unique Entity Identifier (UEI): |
|
Parent UEI: |
|
NSF Program(s): | EarthCube |
Primary Program Source: |
|
Program Reference Code(s): |
|
Program Element Code(s): |
|
Award Agency Code: | 4900 |
Fund Agency Code: | 4900 |
Assistance Listing Number(s): | 47.050 |
ABSTRACT
Scientific reproducibility -- the ability to independently verify the work of other scientists -- continues to be a critical barrier towards achieving the vision of cross-disciplinary science. Federal agencies and publishers increasingly mandate and incentivize scientists to, at a minimum, establish computational reproducibility of scientific experiments. To comply scientists must connect descriptions of scientific experiments in scholarly publications with the underlying data and code used to produce the published results and findings. However, in practice, computational reproducibility is hard to achieve since it entails isolating necessary and sufficient computational artifacts and then preserving those artifacts in a standard way for later re-execution. Both isolation and preservation present challenges in large part due to the complexity of existing software and systems as well as the implicit dependencies, resource distribution, and shifting compatibility of systems that evolve over time -- all of which conspire to break the reproducibility of an experiment. The goal of the GeoTrust project is to understand the research lifecycle of scientific experiments from conception to publication and establish a framework that will improve their reproducibility.
GeoTrust will develop sandboxing-based systems and tools that help scientists effectively isolate computational artifacts associated with an experiment, use languages and semantics to preserve artifacts, and re-execute /reproduce experiments by deploying the artifacts, changing datasets, algorithms, models, environments, etc. This reproducible framework will be adopted by and integrated within community infrastructures of three geoscience sub-disciplines viz. Hydrology, Solid Earth, and Space Science. Using cross-disciplinary science uses cases from these sub-disciplines, and engaging independent evaluators, we will assess the effectiveness of the framework in achieving reproducibility of computational experiments. Finally, verified results will be associated with ?stamps of reproducibility?, establishing community recognition of computational experiments. The framework will be developed as an EarthCube capability, with software developed and released as per EarthCube requirements. Early adopters across other geoscience sub-disciplines will be continually sought.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH
Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external
site maintained by the publisher. Some full text articles may not yet be available without a
charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from
this site.
PROJECT OUTCOMES REPORT
Disclaimer
This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.
The key outcomes of this project are as follows:
1. Established Sciunit, a novel containerization mechanism for establishing computational reproducibility. Sciunit combines automatic encapsulation with provenance and namespace isolation to achieve different levels of computational reproducibility. Production-ready Sciunit is available via http://sciunit.run
2. Demonstrated the effectiveness of this mechanism in a wide range of geoscience disciplines: solid earth, space science, and hydrology. Lead to 4 journal papers (3 accepted and 1 pending) in geoscience and computational domains, 6 conference papers, and 10 AGU and EGU presentations, along with numerous presentations and EarthCube All Hands Meetings and NSF-funded workshops.
3. Increased awareness and investigation of the issues affecting computational reproducibility. The PI's work lead to additional funding from the EarthCube program for deeper integration of Sciunit within community cyberinfrastructure (CI), and from the prestigious NSF CAREER program for developing formal and systematic approaches for guaranteeing reproducibility in complex computational and data science applications.
4. The project provided the funds to support and train 1 postdoc (minority), 4 research engineers (3 minority), and 8 Master's-level graduate students over a period of five years.
5. All primary and secondary repositories created for this project are open-source and accessible via https://bitbucket.org/geotrust/. The primary project is available at: https://bitbucket.org/geotrust/sciunit2/src/master/
6. The PI served on the EarthCube Nominations Committee for 2 years and participated in several EarthCube governance meetings, and every year at the EarthCube All Hands meeting.
Last Modified: 12/30/2021
Modified by: Tanu Malik
Please report errors in award information by writing to: awardsearch@nsf.gov.