Shared repository for code related to the fossils project in Thomas Serre's lab at Brown University. Full project repo containing sub-packages for database, dataloaders, models, training, and analysis.
To install, follow procedure below:
1. Clone git repo:
>> git clone https://github.com/JacobARose/pyleaves.git
2. Navigate to pyleaves root directory:
>> cd pyleaves
3. Make sure to create/activate environment:
-- Implement "ENVIRONMENT MANAGEMENT" section below
4. Install pyleaves with pip, include -e option to make sure it's an editable installation:
>> pip install -e .
-pyleaves.yml -> Tensorflow 1.15 version
-pyleaves2.yml -> Tensorflow 2.1 version (experimental)
This can be used as a base reference, but individual packages may need to be installed in a largely trial and error fashion atm. Creating a minimum environment specification is TBD.
1. Locate up-to-date environment spec file 'pyleaves.yml', usually in root pyleaves directory
2. Create conda environment from spec file:
>> conda env create -f pyleaves.yml -n pyleaves
3. Activate conda environment:
>> conda activate pyleaves
4. Install latest version of pyleaves:
>> git pull
5. To update your existing environment when pulling new changes to the repo:
>> conda env update -n pyleaves --file environment.yml --prune
Note, where it says "-n pyleaves' above, you can replace 'pyleaves" with your preferred choice of env name.
Only relevant when making changes to repo that require additional dependencies
5a. Export conda environment spec file (Only do this if making major changes to requirements or working on new git branch):
>> conda env export --no-builds > pyleaves.yml
5b. Export conda environment spec file using env history rather than list of specific versions
>> conda env export --from-history > pyleaves_from_history.yml
-- Useful option for when working on new machine and experience compatibility issues
GETTING STARTED: Take a look at pyleaves_demo.ipynb in the root directory for an example of how to interact with the database and query data. For further reference, see documentation for SQLAlchemy.
The most up-to-date json master record of datasets should always be found in './pyleaves/leavesdb/resources' under the filename 'full_dataset_frozen.json'. This what's used to create the SQLite database that enables efficient data management on a per-experiment basis.
The format is standardized to allow easy human readability, in addition to flexible modularity for adding new datasets or new metadata related to current entries.
When pulling from the latest repo, one should be sure to recreate the db from the corresponding master json. This can be done using the default args in create_db.py, located in the root directory of pyleaves.
(a) Main Usage: Run from the cmd line with defaults:
>> python create_db.py
(b) Custom Usage: Run from the cmd line with custom json location or db location (db file is auto-named leavesdb.db in create_db.py, its parent dir is customizeable):
>> python create_db.py --json_path 'PATH/TO/SOURCE/FILE.json' --output_folder 'PATH/TO/RESOURCES/DIR'
To modify and push changes to git:
1. Navigate to root directory of package (/pyleaves)
2. git add [list of files to commit] or git add --all
3. git commit -m "message describing changes"
4. git push
5. '[Enter GitHub login credentials]'
Major updates and plans that may break functionality
3/30/2020: Beginning to refactor project format to more closely align with the template found here: https://github.com/Mrgemy95/Tensorflow-Project-Template Key differences include: -Separating train and trainer scripts contained in pyleaves/train into 2 separate directories: pyleaves/trainers and pyleaves/mains -Better isolating data_loader functionality from trainer classes