Skip to content

leouchoa/udacity_mlops_nycAirbnb_short_term_rental

Repository files navigation

Build an ML Pipeline for Short-Term Rental Prices in NYC

You are working for a property management company renting rooms and properties for short periods of time on various rental platforms. You need to estimate the typical price for a given property based on the price of similar properties. Your company receives new data in bulk every week. The model needs to be retrained with the same cadence, necessitating an end-to-end pipeline that can be reused.

The goal of this project is to make the end-to-end machine learning pipeline in a production-ready manner, which means that the object to be deployed is ready to be consumed by customers.

Setup

This pipeline can be almost entirely run on it's own, you'll just need to setup 4 things:

Please install python3, conda and git by following the instruction in the bullet links. Next we will setup the conda development environment and your Weights and Biases authorization.

Cloning the Repo

You can get a local copy of the repository by using the following command:

git clone https://github.com/leouchoa/udacity_mlops_nycAirbnb_short_term_rental.git

After that step into the udacity_mlops_nycAirbnb_short_term_rental folder.

Conda Environment Creation

Make sure to have conda installed and ready, then create the development environment using the environment.yml file provided in the root of the repository and activate it:

conda env create -f environment.yml
conda activate nyc_airbnb_dev

Get API key for Weights and Biases

Let's make sure we are logged in to Weights & Biases. Get your API key from W&B by going to https://wandb.ai/authorize and click on the + icon (copy to clipboard), then paste your key into this command:

> wandb login [your API key]

You should see a message similar to:

wandb: Appending key for api.wandb.ai to your netrc file: /home/[your username]/.netrc

Pipeline Snapshot

The pipeline constructed will work in a similar fashion to the one describe in the picture bellow and it can be described in 7 steps:

  1. Fetch data
  2. Pre-processing
  3. Data checks
  4. Data segregation (train/validation/test split)
  5. Model training and selection
  6. Model testing
  7. Deployment

Pipeline snapshot

How to Use

The 7 pipeline steps are present in the src and components folders. The src is has 4 major steps:

a. eda: the exploratory data analysis used to understand data and discover important feature engineering strategies. b. basic_cleaning: data treatment step necessary to fix or remove problematic data points. c. data_check: testing of the data treatment step, necessary to ensure that the basic cleaning step went as expected. d. train_random_forest: encompasses data segregation and model training.

while the components folder has some subtasks of train_random_forest and also the data fetching step.

To make the entire pipeline run automatically you can use:

mlflow run .

but you can also call individual steps with

mlflow run . -P steps="your_desired_step_here"

which are "download", "basic_cleaning", "data_check", "data_split", "train_random_forest" and "test_regression_model" (described in main.py). So for example to run the "download" step use:

mlflow run . -P steps="download"

and to run multiple steps you can, for example run:

mlflow run . -P steps="download,basic_cleaning,data_check"

Customizing a Step

You can also customize each step call. Steps are controlled by the hydra package, which is a framework to configure complex applications. The customizable pipeline parameters are contained in the config.yaml file.

So for example if you want to run the entire pipeline with default parameters, but also want to change the random forest n_estimators and the etl step parameter min_price, you can run:

mlflow run . -P hydra_options="modeling.random_forest.n_estimators=10 etl.min_price=50"

according to the way specified in the config.yaml file. Another example is:

mlflow run . \
  -P steps=download,basic_cleaning,data_check \
  -P hydra_options="data_check.kl_threshold=0.01 etl.min_price=50"

where we change the etl step parameter min_price and the data_check param kl_threshold.

Multiple Runs

Another feature of this implementation is that you can make multiple runs. Example:

mlflow run . \
  -P steps=train_random_forest \
  -P hydra_options="modeling.random_forest.max_depth=10,50,100 modeling.random_forest.n_estimators=100,200,500 -m"

For even more features, like bayesian optimization and parallel runs, please refer to the hydra documentation.

Running the Release From Github

In this repository there's also a release version of the pipeline. You can can use the release to automatically run the training pipeline without cloning the repo, just by having followed the instructions in the setup.

To run the latest release at thist points version 1.0.1) from github use in a different sample dataset use:

mlflow run -v 1.0.1 https://github.com/leouchoa/udacity_mlops_nycAirbnb_short_term_rental.git -P hydra_options="etl.sample='sample2.csv'"

Selecting Production Model and Test Set Scoring with the test_regression_model Step

This step is use to test your production model against the test set and does not occur automatically: it must be manually triggered. The reason for this is that it requires a production-tagged weights and biases run.

After the train_random_forest (specially if you use multiple runs, as described in the Multiple Runs section), you must go to weights and biases and tag a run. In the figure bellow you can see that the output of the vivid-sun-39 run has a "prod" tag, along with the "v15" tag. So if you run the pipeline you also must to go your Weights and Biases project, select the run that provides the best evaluation metric (you can find helpful statistics under the table tab in the left side) and mark is as "prod".

After that you can use

mlflow run . -P steps="test_regression_model"

to score you final model against the test set. Now if you go to your Weights and Biases account, then to the "prod"-tagged run and then click in the "Graph view" tab, you should see a graph similar to this:

That's the end of the pipeline and means everything went well.

In case of errors

When you make an error writing your conda.yml file, you might end up with an environment for the pipeline or one of the components that is corrupted. Most of the time mlflow realizes that and creates a new one every time you try to fix the problem. However, sometimes this does not happen, especially if the problem was in the pip dependencies. In that case, you might want to clean up all conda environments created by mlflow and try again. In order to do so, you can get a list of the environments you are about to remove by executing:

conda info --envs | grep mlflow | cut -f1 -d" "

If you are ok with that list, execute this command to clean them up:

NOTE: this will remove ALL the environments with a name starting with mlflow. Use at your own risk

for e in $(conda info --envs | grep mlflow | cut -f1 -d" "); do conda uninstall --name $e --all -y;done

This will iterate over all the environments created by mlflow and remove them.

Next Steps

Some next steps to improve the pipeline are:

  • [] Add more feature engineering to the process in order to improve the model scoring metrics.
  • [] Make use of optuna to improve hyperparameter optimization. A guide on how to implement optuna in the mlflow/hydra setup can be found in this medium article
  • [] Improve testing routines and their documentation with great expectations.
  • [] Extend the pipeline routines to incorporate more models, like xgboost and svm.
  • [] shoot a video explaining how to tag an artifact as production.
  • [] add model deployment instructions (online,offline, docker image)
  • [] add hooks (black formating)
  • [] add descriptions to MLproject of basic_cleaning folder

Important urls: