Spaces:
Sleeping
Sleeping
| title: IRIS Classification Lambda | |
| emoji: 🏢 | |
| colorFrom: indigo | |
| colorTo: blue | |
| sdk: gradio | |
| sdk_version: 5.5.0 | |
| app_file: app.py | |
| pinned: false | |
| short_description: IRIS Classification Lambda | |
| # IRIS classification task with AWS Lambda | |
| [](https://www.python.org/downloads/) | |
| [](https://hub.docker.com/repository/docker/cvachet/iris-classification-lambda) | |
| [](https://github.com/psf/black) | |
|  | |
|  | |
|  | |
| **Aims:** Categorization of different species of iris flowers (Setosa, Versicolor, and Virginica) | |
| based on measurements of physical characteristics (sepals and petals). | |
| **Method:** Use of Decision Tree Classifier | |
| **Architecture:** | |
| - Front-end: user interface via Gradio library | |
| - Back-end: use of AWS Lambda function to run deployed ML model | |
| You can try out our deployed [Hugging Face Space](https://huggingface.co/spaces/cvachet/iris_classification_lambda | |
| )! | |
| ---- | |
| **Table of contents:** | |
| - [Local development](#1-local-development) | |
| - [AWS deployment](#2-deployment-to-aws) | |
| - [Hugging Face deployment](#3-deployment-to-hugging-face) | |
| - [Docker Hub deployment](#4-deployment-to-docker-hub) | |
| ---- | |
| ## 1. Local development | |
| ### 1.1 Training the ML model | |
| bash | |
| > python train.py | |
| ### 1.2. Docker container | |
| - Building the docker image | |
| bash | |
| > docker build -t iris-classification-lambda . | |
| - Running the docker container | |
| bash | |
| > docker run --name iris-classification-lambda-cont -p 8080:8080 iris-classification-lambda | |
| ### 1.3. Execution via command line | |
| Example of a prediction request | |
| bash | |
| > curl -X POST "http://localhost:8080/2015-03-31/functions/function/invocations" -H "Content-Type: application/json" -d '{"features": [[6.5, 3.0, 5.8, 2.2], [6.1, 2.8, 4.7, 1.2]]}' | |
| python | |
| > python3 inference_api.py --url http://localhost:8080/2015-03-31/functions/function/invocations -d '{"features": [[6.5, 3.0, 5.8, 2.2], [6.1, 2.8, 4.7, 1.2]]}' | |
| ### 1.4. Execution via user interface | |
| Use of Gradio library for web interface | |
| **Note:** The environment variable ```AWS_API``` should point to the local container | |
| > export AWS_API=http://localhost:8080 | |
| Command line for execution: | |
| > python3 app.py | |
| The Gradio web application should now be accessible at http://localhost:7860 | |
| ## 2. Deployment to AWS | |
| ### 2.1. Pushing the docker container to AWS ECR | |
| <details> | |
| Steps: | |
| - Create new ECR Repository via aws console | |
| Example: ```iris-classification-lambda``` | |
| - Optional for aws cli configuration (to run above commands): | |
| > aws configure | |
| - Authenticate Docker client to the Amazon ECR registry | |
| > aws ecr get-login-password --region <aws_region> | docker login --username AWS --password-stdin <aws_account_id>.dkr.ecr.<aws_region>.amazonaws.com | |
| - Tag local docker image with the Amazon ECR registry and repository | |
| > docker tag iris-classification-lambda:latest <aws_account_id>.dkr.ecr.<aws_region>.amazonaws.com/iris-classification-lambda:latest | |
| - Push docker image to ECR | |
| > docker push <aws_account_id>.dkr.ecr.<aws_region>.amazonaws.com/iris-classification-lambda:latest | |
| </details> | |
| [Link to AWS ECR Documention](https://docs.aws.amazon.com/AmazonECR/latest/userguide/docker-push-ecr-image.html) | |
| ### 2.2. Creating and testing a Lambda function | |
| <details> | |
| **Steps**: | |
| - Create function from container image | |
| Example name: ```iris-classification``` | |
| - Notes: the API endpoint will use the ```lambda_function.py``` file and ```lambda_hander``` function | |
| - Test the lambda via the AWS console | |
| Example JSON object: | |
| ``` | |
| { | |
| "features": [[6.5, 3.0, 5.8, 2.2], [6.1, 2.8, 4.7, 1.2]] | |
| } | |
| ``` | |
| Advanced notes: | |
| - Steps to update the Lambda function with latest container via aws cli: | |
| > aws lambda update-function-code --function-name iris-classification --image-uri <aws_account_id>.dkr.ecr.<aws_region>.amazonaws.com/iris-classification-lambda:latest | |
| </details> | |
| ### 2.3. Creating an API via API Gateway | |
| <details> | |
| **Steps**: | |
| - Create a new ```Rest API``` (e.g. ```iris-classification-api```) | |
| - Add a new resource to the API (e.g. ```/classify```) | |
| - Add a ```POST``` method to the resource | |
| - Integrate the Lambda function to the API | |
| - Notes: using proxy integration option unchecked | |
| - Deploy API with a specific stage (e.g. ```test``` stage) | |
| </details> | |
| Example AWS API Endpoint: | |
| ```https://<api_id>.execute-api.<aws_region>.amazonaws.com/test/classify``` | |
| ### 2.4. Execution for deployed model | |
| Example of a prediction request | |
| bash | |
| > curl -X POST "https://<api_id>.execute-api.<aws_region>.amazonaws.com/test/classify" -H "Content-Type: application/json" -d '{"features": [[6.5, 3.0, 5.8, 2.2], [6.1, 2.8, 4.7, 1.2]]}' | |
| python | |
| > python3 inference_api.py --url https://<api_id>.execute-api.<aws_region>.amazonaws.com/test/classify -d '{"features": [[6.5, 3.0, 5.8, 2.2], [6.1, 2.8, 4.7, 1.2]]}' | |
| ## 3. Deployment to Hugging Face | |
| This web application is available on Hugging Face | |
| Hugging Face space URL: | |
| https://huggingface.co/spaces/cvachet/iris_classification_lambda | |
| Note: This space uses the ML model deployed on AWS Lambda | |
| ## 4. Deployment to Docker Hub | |
| This web application is available on Docker Hub as a docker image | |
| URL: | |
| https://hub.docker.com/r/cvachet/iris-classification-lambda | |