|
1 | | -Azavea data analytics team python project template |
| 1 | +Azavea Data Analytics team python project template |
2 | 2 | ============================== |
3 | 3 |
|
4 | | -template for Azavea Data Analytics team Python projects |
| 4 | +A file structure template, development environment and rule set for python data analytics projects on the data analytics team |
| 5 | + |
| 6 | +Getting Started |
| 7 | +------------ |
| 8 | +From within the root directory, first remove git tracking from the project |
| 9 | + |
| 10 | +`rm -rf .git` |
| 11 | + |
| 12 | +If you have not already done so, build the Docker image (you will only need to do this once) |
| 13 | + |
| 14 | +`docker build -t da-project-template .` |
| 15 | + |
| 16 | +Run a Docker container |
| 17 | + |
| 18 | +`./scripts/container.sh .' |
| 19 | + |
| 20 | +This will open a bash shell within the Docker container. Within the container the 'project' directory on the host machine (as specified as a parameter of `container.sh` above) will map to `/project` within the container. You can now access the full file structure of this template from within the container. |
| 21 | + |
| 22 | +To exit |
| 23 | + |
| 24 | +`exit` |
5 | 25 |
|
6 | 26 | Project Organization |
7 | 27 | ------------ |
8 | 28 |
|
9 | | - ├── LICENSE |
10 | | - ├── Makefile <- Makefile with commands like `make data` or `make train` |
11 | 29 | ├── README.md <- The top-level README for developers using this project. |
12 | 30 | ├── data |
13 | | - │ ├── external <- Data from third party sources. |
14 | | - │ ├── interim <- Intermediate data that has been transformed. |
15 | | - │ ├── processed <- The final, canonical data sets for modeling. |
16 | | - │ └── raw <- The original, immutable data dump. |
| 31 | + │ ├── interm <- Intermediate data that has been transformed |
| 32 | + │ ├── organized <- Raw datasets that have been renamed or reorganized into a new folder structure but have not been changed at all |
| 33 | + │ ├── processed <- The final, canonical data sets for modeling |
| 34 | + │ └── raw <- The original, immutable data dump |
| 35 | + │ |
| 36 | + ├── docs <- A default Sphinx project; see sphinx-doc.org for details (currently not configured) |
17 | 37 | │ |
18 | | - ├── docs <- A default Sphinx project; see sphinx-doc.org for details |
| 38 | + ├── guide <- A set of markdown files with documented best practices, guidelines and rools for collaborative projects |
19 | 39 | │ |
20 | 40 | ├── models <- Trained and serialized models, model predictions, or model summaries |
21 | 41 | │ |
22 | 42 | ├── notebooks <- Jupyter notebooks. Naming convention is a number (for ordering), |
23 | | - │ the creator's initials, and a short `-` delimited description, e.g. |
24 | | - │ `1.0-jqp-initial-data-exploration`. |
| 43 | + │ the creator's initials, and a short `-` delimited description, e.g |
| 44 | + │ `1.0-jqp-initial-data-exploration` |
25 | 45 | │ |
26 | 46 | ├── references <- Data dictionaries, manuals, and all other explanatory materials. |
27 | 47 | │ |
28 | 48 | ├── reports <- Generated analysis as HTML, PDF, LaTeX, etc. |
29 | 49 | │ └── figures <- Generated graphics and figures to be used in reporting |
30 | 50 | │ |
31 | | - ├── requirements.txt <- The requirements file for reproducing the analysis environment, e.g. |
32 | | - │ generated with `pip freeze > requirements.txt` |
33 | | - │ |
34 | | - ├── src <- Source code for use in this project. |
35 | | - │ ├── __init__.py <- Makes src a Python module |
36 | | - │ │ |
37 | | - │ ├── data <- Scripts to download or generate data |
38 | | - │ │ └── make_dataset.py |
39 | | - │ │ |
40 | | - │ ├── features <- Scripts to turn raw data into features for modeling |
41 | | - │ │ └── build_features.py |
42 | | - │ │ |
43 | | - │ ├── models <- Scripts to train models and then use trained models to make |
44 | | - │ │ │ predictions |
45 | | - │ │ ├── predict_model.py |
46 | | - │ │ └── train_model.py |
47 | | - │ │ |
48 | | - │ └── visualization <- Scripts to create exploratory and results oriented visualizations |
49 | | - │ └── visualize.py |
50 | | - │ |
51 | | - └── tox.ini <- tox file with settings for running tox; see tox.testrun.org |
| 51 | + ├── requirements.txt <- The requirements file for reproducing the analysis environment |
| 52 | + │ |
| 53 | + └── src <- Source code for use in this project. |
| 54 | + │ |
| 55 | + ├── data <- Scripts to download or generate data |
| 56 | + │ └── make_dataset.py |
| 57 | + │ |
| 58 | + ├── features <- Scripts to turn raw data into features for modeling |
| 59 | + │ └── build_features.py |
| 60 | + │ |
| 61 | + ├── models <- Scripts to train models and then use trained models to make |
| 62 | + │ │ predictions |
| 63 | + │ ├── predict_model.py |
| 64 | + │ └── train_model.py |
| 65 | + │ |
| 66 | + └── visualization <- Scripts to create exploratory and results oriented visualizations |
| 67 | + └── visualize.py |
| 68 | + |
52 | 69 |
|
53 | 70 |
|
54 | 71 | -------- |
|
0 commit comments