Skip to content

This is a repository for a semantic segmentation inference API of a video using the BMW-IntelOpenVINO-Segmentation-Inference-API.

Notifications You must be signed in to change notification settings

obeidahmad/Video-Inference-API

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Video-Inference-API

This is a repository for a semantic segmentation inference API of a video using the BMW-IntelOpenVINO-Segmentation-Inference-API.

Prerequisites

  • OS:
    • Ubuntu 20.04
    • Windows 10 pro/enterprise
  • Docker

Check for prerequisites

To check if you have docker-ce installed:

docker --version

Install prerequisites

Ubuntu

Use the following command to install docker on Ubuntu:

chmod +x install_prerequisites.sh && source install_prerequisites.sh

Windows 10

To install Docker on Windows, please follow the link.

Build The Docker Image

In order to build the project run the following command from the project's root directory:

docker build -t video_inference_api -f docker/Dockerfile .

Run The Docker Container

If you wish to deploy this API using docker, please issue the following run command.

To run the API, go the to the API's directory and run the following:

Using Linux based docker:

docker run -itv $(pwd)/data:/data -p <port_of_your_choice>:8080 video_inference_api

Using Windows based docker:

Using PowerShell:
docker run -itv ${PWD}/data:/data -p <port_of_your_choice>:8080 video_inference_api
Using CMD:
docker run -itv %cd%/data:/data -p <port_of_your_choice>:8080 video_inference_api

Connect with other container:

Make sure all containers are in the same network. Add the following option to the docker command:

--net <network_name>

Refer to Use bridge networks | Docker Documentation for more information about bridge networks


The <docker_host_port> can be any unique port of your choice.
The <network_name> is a bridged network name of your choice.

The API file will run automatically, and the service will listen to http requests on the chosen port.

API Endpoints

To see all available endpoints, open your favorite browser and navigate to:

http://<machine_IP>:<docker_host_port>/docs

Endpoints summary

/inference (POST)

Perform inference on a video using the specified model and returns the resulting video as a response.

More Information

Refer to BMW-IntelOpenVINO-Segmentation-Inference-API for more information on how to add models

About

This is a repository for a semantic segmentation inference API of a video using the BMW-IntelOpenVINO-Segmentation-Inference-API.

Topics

Resources

Stars

Watchers

Forks