Product architecture

The analytic framework of IBM® Video Analytics is composed of the Metadata Ingestion Lookup and Signaling (MILS), the Semantic Streams Engine (SSE), and the Deep Learning Engine (DLE).

The following diagram shows a high-level view of the product architecture and the flow of information between the various components of Video Analytics.

A diagram of the solution architecture

Video data and metadata for alerts and recorded video flow through the system in the following sequence.

  1. Video from cameras is uploaded by file or streamed from a video management system (VMS).
  2. The SSE requests and receives video.
  3. The SSE uses the DLE for object detection and attribute analysis.
  4. The SSE generates metadata, and sends it to the server.
  5. The MILS indexes and stores the metadata.
  6. The operator client either searches the metadata or receives alerts from MILS.
  7. Live or recorded video can be viewed from the video player in the operator client.

Through the extensible capability of Video Analytics, each industry can build a surveillance solution. A solution can be built on top of the framework by building a set of plug-ins:

The following diagram shows the basic system architecture of Video Analytics.

Communication between MILS, SSE, and DLE servers over network

The SSE and DLE are distributed across 1 or more servers. On each server is an engine service, and within each engine service are 1 or more engines. Communication between the SSE servers and the MILS, and between the DLE and SSE servers, takes place over an Internet Protocol network.

Parent topic:
Solution overview
Related concepts:
Key concepts of Video Analytics