Raw data for entity 1 was run through a locally developed Python program QAQC procedure (hja_hobo_clean) by study originator. There is a version of entity 1 data offline, but available, that have had a light-adjustment applied based on the illumination values in order to account for bias in the temperature measurements created by light hitting the sensor shield. The light adjustment is based on paired light and temperature measurements, from the Discovery Tree (see document hobo_correction_181012.pdf). It is Corrected air T = 0.2267*(lux/1000) + 0.04955 or y = 0.1987x + 0.0726. This light adjustment was applied to the cleaned version of the dataset.
Raw data for entity 2 was flagged and not filled using the GCE Toolbox workflow and based on the flagging algorithms in the hja_hobo_clean workflow. Researchers provide a folder of HOBO data files with a predefined naming convention and structure that can be processed by Andrews Forest System Administrator. The files are run through GCE Data Toolbox, which is a comprehensive software framework for metadata-based analysis, quality control, transformation and management of ecological data sets. https://gce-lter.marsci.uga.edu/public/im/tools/data_toolbox.htm. The GCE workflow was developed based on the original checks from the hja_hobo_clean Python program.
Raw data for entity 1 was run through a locally developed Python program QAQC procedure (hja_hobo_clean). These data have only been flagged and include extreme values, gaps, spikes and possible snow indication. Some visual inspection of the data using a visualization tool occurred. The Python program can create cleaned and filled versions of the data based on specified parameters. More information on this is found in the supplementary documents (Python workflow, Python flag definitions).
Raw data for entity 2 was flagged and not filled using the GCE Toolbox workflow and based on the flagging algorithms in the hja_hobo_clean workflow. Like entity 1, these data were checked for burial by snow, detection of extreme values and jumps, as well as the influence of high/extreme light intensity. Rather than having multiple data columns for each flag, as in entity 1, these data use an aggregated flagging system for each data variable (temperature and light).