Integration with other media services

Updated: May 06, 2022

The mm-stream service uses other platform services to read audio and video content.

It can access the content using:

Whether the content must be stored on the same node or it can be read from another node depends on the particular service used. For example, for media files visible through NFS, mm-stream can access these remote files through POSIX calls. Likewise, the Camera Framework allows clients to read camera feeds originating from another node as long as it's mounted in the local filesystem.

The following diagram shows the flow of media content from these various services through mm-stream to remote clients:

Diagram showing media flow from platform services through mm-stream to remote clients over the network
Figure 1. Integration of mm-stream with other services

Communication between the network media server and the services that provide media input is done through shared libraries. These are loaded by mm-stream and they read the inputs based on configuration parameters such as audio sampling rate, video dimensions, and frame rate. Remote clients access the content from mm-stream through RTP and RTSP. Pure RTP is used for multicasting, when multiple media players (clients) receive a media stream with minimal bandwidth cost. RTSP is used when a media player wants to control streaming through playback commands (e.g., play, pause).

The mm-stream service uses the OpenMAX AL API to access the input media, encode it, and output it. This API is implemented by a media engine included in the Multimedia for QNX SDP; for details, see the QNX Additions to OpenMAX AL guide. Remote clients can use this same media engine or the mm-renderer service (see the Multimedia Renderer Developer's Guide for details) to play the content streamed over the network.