1. Motivation: The Importance of Flux Data Standardization and Reproducibility
Collecting and handling eddy covariance data presents challenges on many front, including site and instrument selection, high frequency data processing, data post-processing, and QA/QC. While there are many resources online resources that can help with all of these different challenges of eddy covariance measurements (see examples below), here we focus on the approach that the EcoFlux Lab uses for flux data post-processing and QA/QC. Here we provide a detailed outline of our procedures for data post-processing QA/QC, gap-filling, and CO2 flux partitioning. We also provide data visualization tools. We draw on expertise from regional networks and FLUXNET, and leverage widely used tools in the field such as REddyProc and other recently developed machine learning-based gap-filling resources . By following these procedures, you will be able to develop a robust data pipeline to ensure high quality data that you can submit to your regional flux networks. Data standardization and reproducibility helps minimize errors, enhances data reliability, and enables the integration of data sets from diverse ecosystems into global networks like FLUXNET.
Note that in this pipeline, we are focused on data post-processing, rather than the processing of high frequency (e.g., 20 Hz) data. While we do provide resources on high frequency flux processing, this is not within the main scope of this data pipeline. For more information on site set-up, instrument selection, and processing of high frequency data, please contact Xiangmin Sun from the FLUXNET CH4 and N2O processing committee.
While the data pipeline described here requires some coding knowledge, the whole pipeline can be run without having to modify any existing code. You only need to edit some initialization files, and then execute the code.
Additional resources: