mirror of
https://github.com/Azure/MachineLearningNotebooks.git
synced 2025-12-20 09:37:04 -05:00
431 lines
14 KiB
Plaintext
431 lines
14 KiB
Plaintext
{
|
|
"cells": [
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"Copyright (c) Microsoft Corporation. All rights reserved. \n",
|
|
"\n",
|
|
"Licensed under the MIT License."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"# YOLO Real-time Object Detection using ONNX on AzureML\n",
|
|
"\n",
|
|
"This example shows how to convert the TinyYOLO model from CoreML to ONNX and operationalize it as a web service using Azure Machine Learning services and the ONNX Runtime.\n",
|
|
"\n",
|
|
"## What is ONNX\n",
|
|
"ONNX is an open format for representing machine learning and deep learning models. ONNX enables open and interoperable AI by enabling data scientists and developers to use the tools of their choice without worrying about lock-in and flexibility to deploy to a variety of platforms. ONNX is developed and supported by a community of partners including Microsoft, Facebook, and Amazon. For more information, explore the [ONNX website](http://onnx.ai).\n",
|
|
"\n",
|
|
"## YOLO Details\n",
|
|
"You Only Look Once (YOLO) is a state-of-the-art, real-time object detection system. For more information about YOLO, please visit the [YOLO website](https://pjreddie.com/darknet/yolo/)."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"## Prerequisites\n",
|
|
"\n",
|
|
"To make the best use of your time, make sure you have done the following:\n",
|
|
"\n",
|
|
"* Understand the [architecture and terms](https://docs.microsoft.com/azure/machine-learning/service/concept-azure-machine-learning-architecture) introduced by Azure Machine Learning\n",
|
|
"* Go through the [00.configuration.ipynb](../00.configuration.ipynb) notebook to:\n",
|
|
" * install the AML SDK\n",
|
|
" * create a workspace and its configuration file (config.json)"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"# Check core SDK version number\n",
|
|
"import azureml.core\n",
|
|
"\n",
|
|
"print(\"SDK version:\", azureml.core.VERSION)"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"#### Install necessary packages\n",
|
|
"\n",
|
|
"You'll need to run the following commands to use this tutorial:\n",
|
|
"\n",
|
|
"```sh\n",
|
|
"pip install coremltools\n",
|
|
"pip install onnxmltools\n",
|
|
"```"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"## Convert model to ONNX\n",
|
|
"\n",
|
|
"First we download the CoreML model. We use the CoreML model listed at https://coreml.store/tinyyolo. This may take a few minutes."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"!wget https://s3-us-west-2.amazonaws.com/coreml-models/TinyYOLO.mlmodel"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"Then we use ONNXMLTools to convert the model."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"import onnxmltools\n",
|
|
"import coremltools\n",
|
|
"\n",
|
|
"# Load a CoreML model\n",
|
|
"coreml_model = coremltools.utils.load_spec('TinyYOLO.mlmodel')\n",
|
|
"\n",
|
|
"# Convert from CoreML into ONNX\n",
|
|
"onnx_model = onnxmltools.convert_coreml(coreml_model, 'TinyYOLOv2')\n",
|
|
"\n",
|
|
"# Save ONNX model\n",
|
|
"onnxmltools.utils.save_model(onnx_model, 'tinyyolov2.onnx')\n",
|
|
"\n",
|
|
"import os\n",
|
|
"print(os.path.getsize('tinyyolov2.onnx'))"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"## Deploying as a web service with Azure ML\n",
|
|
"\n",
|
|
"### Load Azure ML workspace\n",
|
|
"\n",
|
|
"We begin by instantiating a workspace object from the existing workspace created earlier in the configuration notebook."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"from azureml.core import Workspace\n",
|
|
"\n",
|
|
"ws = Workspace.from_config()\n",
|
|
"print(ws.name, ws.location, ws.resource_group, sep = '\\n')"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"### Registering your model with Azure ML\n",
|
|
"\n",
|
|
"Now we upload the model and register it in the workspace."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"from azureml.core.model import Model\n",
|
|
"\n",
|
|
"model = Model.register(model_path = \"tinyyolov2.onnx\",\n",
|
|
" model_name = \"tinyyolov2\",\n",
|
|
" tags = {\"onnx\": \"demo\"},\n",
|
|
" description = \"TinyYOLO\",\n",
|
|
" workspace = ws)"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"#### Displaying your registered models\n",
|
|
"\n",
|
|
"You can optionally list out all the models that you have registered in this workspace."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"models = ws.models()\n",
|
|
"for m in models:\n",
|
|
" print(\"Name:\", m.name,\"\\tVersion:\", m.version, \"\\tDescription:\", m.description, m.tags)"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"### Write scoring file\n",
|
|
"\n",
|
|
"We are now going to deploy our ONNX model on Azure ML using the ONNX Runtime. We begin by writing a score.py file that will be invoked by the web service call. The `init()` function is called once when the container is started so we load the model using the ONNX Runtime into a global session object."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"%%writefile score.py\n",
|
|
"import json\n",
|
|
"import time\n",
|
|
"import sys\n",
|
|
"import os\n",
|
|
"from azureml.core.model import Model\n",
|
|
"import numpy as np # we're going to use numpy to process input and output data\n",
|
|
"import onnxruntime # to inference ONNX models, we use the ONNX Runtime\n",
|
|
"\n",
|
|
"def init():\n",
|
|
" global session\n",
|
|
" model = Model.get_model_path(model_name = 'tinyyolov2')\n",
|
|
" session = onnxruntime.InferenceSession(model)\n",
|
|
"\n",
|
|
"def preprocess(input_data_json):\n",
|
|
" # convert the JSON data into the tensor input\n",
|
|
" return np.array(json.loads(input_data_json)['data']).astype('float32')\n",
|
|
"\n",
|
|
"def postprocess(result):\n",
|
|
" return np.array(result).tolist()\n",
|
|
"\n",
|
|
"def run(input_data_json):\n",
|
|
" try:\n",
|
|
" start = time.time() # start timer\n",
|
|
" input_data = preprocess(input_data_json)\n",
|
|
" input_name = session.get_inputs()[0].name # get the id of the first input of the model \n",
|
|
" result = session.run([], {input_name: input_data})\n",
|
|
" end = time.time() # stop timer\n",
|
|
" return {\"result\": postprocess(result),\n",
|
|
" \"time\": end - start}\n",
|
|
" except Exception as e:\n",
|
|
" result = str(e)\n",
|
|
" return {\"error\": result}"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"### Create container image\n",
|
|
"First we create a YAML file that specifies which dependencies we would like to see in our container."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"from azureml.core.conda_dependencies import CondaDependencies \n",
|
|
"\n",
|
|
"myenv = CondaDependencies.create(pip_packages=[\"numpy\",\"onnxruntime\"])\n",
|
|
"\n",
|
|
"with open(\"myenv.yml\",\"w\") as f:\n",
|
|
" f.write(myenv.serialize_to_string())"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"Then we have Azure ML create the container. This step will likely take a few minutes."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"from azureml.core.image import ContainerImage\n",
|
|
"\n",
|
|
"image_config = ContainerImage.image_configuration(execution_script = \"score.py\",\n",
|
|
" runtime = \"python\",\n",
|
|
" conda_file = \"myenv.yml\",\n",
|
|
" description = \"TinyYOLO ONNX Demo\",\n",
|
|
" tags = {\"demo\": \"onnx\"}\n",
|
|
" )\n",
|
|
"\n",
|
|
"\n",
|
|
"image = ContainerImage.create(name = \"onnxyolo\",\n",
|
|
" models = [model],\n",
|
|
" image_config = image_config,\n",
|
|
" workspace = ws)\n",
|
|
"\n",
|
|
"image.wait_for_creation(show_output = True)"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"In case you need to debug your code, the next line of code accesses the log file."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"print(image.image_build_log_uri)"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"We're all set! Let's get our model chugging.\n",
|
|
"\n",
|
|
"### Deploy the container image"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"from azureml.core.webservice import AciWebservice\n",
|
|
"\n",
|
|
"aciconfig = AciWebservice.deploy_configuration(cpu_cores = 1, \n",
|
|
" memory_gb = 1, \n",
|
|
" tags = {'demo': 'onnx'}, \n",
|
|
" description = 'web service for TinyYOLO ONNX model')"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"The following cell will likely take a few minutes to run as well."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"from azureml.core.webservice import Webservice\n",
|
|
"from random import randint\n",
|
|
"\n",
|
|
"aci_service_name = 'onnx-tinyyolo'+str(randint(0,100))\n",
|
|
"print(\"Service\", aci_service_name)\n",
|
|
"\n",
|
|
"aci_service = Webservice.deploy_from_image(deployment_config = aciconfig,\n",
|
|
" image = image,\n",
|
|
" name = aci_service_name,\n",
|
|
" workspace = ws)\n",
|
|
"\n",
|
|
"aci_service.wait_for_deployment(True)\n",
|
|
"print(aci_service.state)"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"In case the deployment fails, you can check the logs. Make sure to delete your aci_service before trying again."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"if aci_service.state != 'Healthy':\n",
|
|
" # run this command for debugging.\n",
|
|
" print(aci_service.get_logs())\n",
|
|
" aci_service.delete()"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"## Success!\n",
|
|
"\n",
|
|
"If you've made it this far, you've deployed a working web service that does object detection using an ONNX model. You can get the URL for the webservice with the code below."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"print(aci_service.scoring_uri)"
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "markdown",
|
|
"metadata": {},
|
|
"source": [
|
|
"When you are eventually done using the web service, remember to delete it."
|
|
]
|
|
},
|
|
{
|
|
"cell_type": "code",
|
|
"execution_count": null,
|
|
"metadata": {},
|
|
"outputs": [],
|
|
"source": [
|
|
"#aci_service.delete()"
|
|
]
|
|
}
|
|
],
|
|
"metadata": {
|
|
"authors": [
|
|
{
|
|
"name": "onnx"
|
|
}
|
|
],
|
|
"kernelspec": {
|
|
"display_name": "Python 3.6",
|
|
"language": "python",
|
|
"name": "python36"
|
|
},
|
|
"language_info": {
|
|
"codemirror_mode": {
|
|
"name": "ipython",
|
|
"version": 3
|
|
},
|
|
"file_extension": ".py",
|
|
"mimetype": "text/x-python",
|
|
"name": "python",
|
|
"nbconvert_exporter": "python",
|
|
"pygments_lexer": "ipython3",
|
|
"version": "3.5.6"
|
|
}
|
|
},
|
|
"nbformat": 4,
|
|
"nbformat_minor": 2
|
|
} |