Rachel Wegener , Jacob O. Wenegrat, Veronica Lance, Skylar Lama
Use poetry to install the dependencies. This can be done by running the following command in the root directory of the repository:
poetry install
A virtual environment will be created at this root of this repo and the dependencies will be installed.
Activate the environment by running:
poetry shell
Install the IPFS desktop app or Kubo CLI client as this will will allow you to run a local IPFS node on your machine.
Overview of Steps:
- Download the datasets: geopolar, nasa mur, Chesepeake Bay program in situ data from IPFS using the scripts provided. There should then be three datasets in the
data/raw
folder after data has been processed. - Run the notebooks in the
notebooks
folder. This generates the figures.
Folders created to hold data throughout the processing pipeline, from raw data to fully processed data. The folders are empty on github, but the data can be processed running the following scripts:
00b1_download_from_ipfs.py
- Downloads the data from IPFS and saves it to the directory./data/scratch
- Specify the collection to be downloaded by changing the
DATASET
variable in the script. Valid options for DATASET aregeopolar
,mur
, orcbp
- Specify the collection to be downloaded by changing the
00b2_crop_satellite_data.py
- Crops the satellite data to the Chesapeake Bay region and saves it to the directory./data/interim
. The images are then merged into a single.netcdf
file and saved to./data/raw
.- Specify the collection to be downloaded by changing the
DATASET
variable in the script. Valid options for DATASET aregeopolar
ormur
.
- Specify the collection to be downloaded by changing the
Contains the final figures, as generated in the data analysis notebooks
Jupyter notebooks written in Python with instructions for processing the data from the raw
format through the generation of relevant figures
.