-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Portage data workflow #11
Conversation
Files from compasstools's `portage` branch
Added what the SERC folks call the "normalization" step: matching loggernet variables with their Example html output: https://rpubs.com/bpbond/971152
|
@stephpenn1 FYI in e86ca27 I added a skeleton |
Visualization of files moving through the processing pipeline:
|
* Name all chunks * Add GitHub Action and badge
* Move units and OOB checks from L1a to L1_normalize * Add missing design_link data * Change Quarto file params to relative to DATA_ROOT * Clean up L1a functionality; move timestamp rounding to L1b
* Remove other new variables code * L1a don't use templates; written to site-year-month folders * Rearrange columns * Disable L1b
* Fix L0 data-writing location problem * Update test data: spans August and September, adds CheckTable * Use ALL Portage data for testing * Update design table to be consistent with Google Sheet * Update metadata_vars to be consistent with google sheet * readr/vroom error on GA; try write.csv
* Improve handling and logging of unexpected errors * Quarto files now write error info to driver log
* L0 pivots data to long and assigns unique observation IDs
* MInor fixes to read_csv_group * Initial pass at writing L1 metadata * Update helpers.R * OK to have missing year-month combinations; remove dupes in L0 * Print Git commit in reproducibility section
* Drop site column per @stephpenn1 #49 * Update L1_metadata_template.txt
Guarantee files not overwritten by adding a short hash
* Remove dplyr from L1_normalize * Update check-workflow.yaml
Superseded by #57 |
Sample workflow for processing raw data to L0. This includes
*.dat
files in the raw folder, reads them in, parses out the Logger and Table info, and writes them to the L0 folder. It then generates a summary table.Raw/
data (see below)Sample html documentation file: https://rpubs.com/bpbond/970986
Raw folder before processing:
L0 folder after processing:
This is very much a first pass—feedback welcome @stephpenn1