documentation
This commit is contained in:
parent
140a4c4573
commit
6918f80eb4
25
README.md
25
README.md
|
@ -5,3 +5,28 @@ This project is intended to compute an estimated value of risk for a given datab
|
|||
1. Pull meta data of the database and create a dataset via joins
|
||||
2. Generate the dataset with random selection of features
|
||||
3. Compute risk via SQL using group by
|
||||
## Python environment
|
||||
|
||||
The following are the dependencies needed to run the code:
|
||||
|
||||
pandas
|
||||
numpy
|
||||
pandas-gbq
|
||||
google-cloud-bigquery
|
||||
|
||||
|
||||
## Usage
|
||||
|
||||
*Generate The merged dataset
|
||||
|
||||
python risk.py create --i_dataset <in dataset|schema> --o_dataset <out dataset|schema> --table <name> --path <bigquery-key-file> --key <patient-id-field-name> [--file ]
|
||||
|
||||
* Cmpute risk
|
||||
|
||||
python risk.py compute --i_dataset <dataset> --table <name> --path <bigquery-key-file> --key <patient-id-field-name>
|
||||
## Limitations
|
||||
- It works against bigquery for now
|
||||
@TODO:
|
||||
- Need to write a transport layer (database interface)
|
||||
- Support for referential integrity, so one table can be selected and a dataset derived given referential integrity
|
||||
- Add support for journalist risk
|
Loading…
Reference in New Issue