Multimedia Applications: Streaming, VoIP, and Characteristics

Choose a study mode

Play Quiz
Study Flashcards
Spaced Repetition
Chat to Lesson

Podcast

Play an AI-generated podcast conversation about this lesson

Questions and Answers

What is the primary reason for using buffering techniques in multimedia applications like video streaming?

  • To allow users to download the entire video file before playback begins.
  • To reduce the overall bandwidth consumption.
  • To ensure that the video plays back at a consistent rate without freezing or speeding up. (correct)
  • To enable users to skip advertisements within the video content.

Why are glitches in audio generally more noticeable than glitches in video during a video conference?

  • The human ear is more sensitive to disruptions in audio than the human eye is to disruptions in video. (correct)
  • Audio data is less compressed than video data in video conferences.
  • Audio signals have a higher bit rate compared to video.
  • Video conferences use more advanced error correction techniques for audio.

Which of the following is a characteristic of streaming stored video?

  • It must be broadcast to many simultaneous users in different geographic locations.
  • It is highly sensitive to delays, with even short delays being unacceptable.
  • It is less interactive compared to other forms of streaming.
  • It can be implemented using a peer-to-peer model instead of a client-server model. (correct)

What is the significance of 'quantization' in the context of encoding analog audio into a digital format?

<p>It involves taking samples per second, then 'rounding' each sample's value to a discrete number. (B)</p> Signup and view all the answers

Which of the following functions does a signaling protocol perform in traditional telephony?

<p>Handling user location, session establishment, and call participation management. (B)</p> Signup and view all the answers

Why is end-to-end delay a critical QoS metric for VoIP applications?

<p>It affects how naturally people can converse, as excessive delay leads to people talking over each other. (D)</p> Signup and view all the answers

What is the main trade-off when using a longer jitter buffer in VoIP applications?

<p>It reduces packet loss but adds to the end-to-end delay. (A)</p> Signup and view all the answers

Why do VoIP protocols typically use UDP instead of TCP, despite TCP providing reliability?

<p>TCP's retransmission of lost packets can cause unacceptable delays for real-time communication. (C)</p> Signup and view all the answers

How does Forward Error Correction (FEC) address packet loss in VoIP?

<p>By transmitting redundant data alongside the main transmission, allowing the receiver to reconstruct lost data. (C)</p> Signup and view all the answers

What is the primary goal of 'interleaving' as a method for dealing with packet loss in VoIP?

<p>To ensure that lost audio chunks are not consecutive, making the gaps less noticeable. (D)</p> Signup and view all the answers

What is the advantage of using HTTP for video delivery over specialized video servers?

<p>HTTP enables the use of existing CDN infrastructure and simplifies bypassing middleboxes and firewalls. (D)</p> Signup and view all the answers

What is the main difference between progressive download and streaming?

<p>Progressive download downloads the entire video file at once, while streaming downloads the video in segments as it is played. (D)</p> Signup and view all the answers

Why is single-bitrate encoding not optimal for video streaming?

<p>It cannot adapt to the diversity of network conditions and user devices. (C)</p> Signup and view all the answers

What is the purpose of a 'manifest file' in video streaming?

<p>To provide metadata about the video content, including available bitrates and segment URLs. (A)</p> Signup and view all the answers

In Dynamic Adaptive Streaming over HTTP (DASH), what does 'dynamic streaming' refer to?

<p>The automated adjustment of video bitrate based on network conditions and device capabilities. (D)</p> Signup and view all the answers

Which of the following factors typically characterizes in a good Quality of Experience (QoE) for video streaming?

<p>High video quality and minimal re-buffering. (D)</p> Signup and view all the answers

In rate-based adaptation, what signal is primarily used to facilitate the selection of bitrate?

<p>The network throughput. (A)</p> Signup and view all the answers

If the available bandwidth is 10 Mbps and the bitrate of the video chunk is 2 Mbps, what is the buffer-filling rate?

<p>5 (C)</p> Signup and view all the answers

What steps can be taken to make the estimation of continuous throughput to discrete a better?

<p>Select the maximum bitrate that is less that the estimate of the thoroughput (C)</p> Signup and view all the answers

How does congestion affect the throughput of a video?

<p>It can have an impact on video quality as the TCP flow has another competing flow. (B)</p> Signup and view all the answers

In what situation would rate-based adaptation lead to an overly high estimate of future bandwidth?

<p>When the bandwidth abruptly decreases after a period of high throughput. (B)</p> Signup and view all the answers

Which of the following is a potential issue with using buffer-based adaptation for video streaming algorithms?

<p>It may lead to initial high-quality downloads that are unnecessary. (B)</p> Signup and view all the answers

In constant bitrate (CBR) encoding, what remains fixed over time?

<p>The output size of the video. (A)</p> Signup and view all the answers

What happens during the Filling State of progressive download?

<p>The client tries to fill a video as fast as possible until the buffer fills up. (C)</p> Signup and view all the answers

In what situation is variable bitrate more computationally expensive than constant bitrate?

<p>With the same bit rate. (B)</p> Signup and view all the answers

What can happen if an I-frame is partially lost?

<p>RGB matrices can not be obtained correctly. (C)</p> Signup and view all the answers

Other than lossless and lossy, what are two good ways to compress that are key components?

<p>Spatial and temporal redundancy. (A)</p> Signup and view all the answers

How does higher weighting impact high spectral frequencies?

<p>It is used to achieve desired image quality. (B)</p> Signup and view all the answers

In what situation would using the difference between frames to encode be wasteful (bad)?

<p>When a scene changes. (D)</p> Signup and view all the answers

Which example sequence of frames is an appropriate sequence?

<p>IBBBPIΒΡΡΡΙΒΒΒΒΙ (B)</p> Signup and view all the answers

Regarding Digital Rights Management, what has it encouraged?

<p>Content providers to put their content on the Internet. (D)</p> Signup and view all the answers

Generally from a user's perspective, what is a higher bit rate associated with?

<p>Better video quality. (C)</p> Signup and view all the answers

Flashcards

Multimedia Application

Network application that employs audio or video.

High Bit Rate (Video)

The rate at which video data is transmitted, typically between 100 kbps to over 3Mbps.

Compressing Video

Reducing the size of video data to save bandwidth, which can be lossy or lossless.

Multimedia Application Categories

Streaming audio and video, conversational voice and video, and streaming live audio and video.

Signup and view all the flashcards

Audio Encoding

The process of converting analog audio into a digital format by sampling and quantization.

Signup and view all the flashcards

Quantization

Quantization is rounding each audio sample's value to a discrete number.

Signup and view all the flashcards

Encoding Scheme Categories

Narrowband, broadband, and multimode.

Signup and view all the flashcards

Signaling protocols

Techniques such as user location, session establishment, session negotiation, and call participation management.

Signup and view all the flashcards

VoIP QoS Metrics

End-to-end delay, jitter, and packet loss.

Signup and view all the flashcards

End-to-End Delay

The total delay from sender to receiver, including encoding, packetization, network delays, and decoding.

Signup and view all the flashcards

Jitter

Variation in delay between received packets, which can disrupt the reconstruction of the audio stream

Signup and view all the flashcards

Packet Loss

The ratio of lost packets, which can negatively impact speech quality in VoIP.

Signup and view all the flashcards

VoIP Packet Loss Solutions

FEC transmits redundant data, interleaving mixes audio chunks, and error concealment guesses lost packets.

Signup and view all the flashcards

Forward Error Correction (FEC)

Transmitting redundant data alongside the main transmission to allow the receiver to replace lost data.

Signup and view all the flashcards

Interleaving

Mixing chunks of audio together so that lost chunks aren't consecutive.

Signup and view all the flashcards

Error Concealment

Estimating or 'guessing' what the lost audio packet might be based on surrounding snippets.

Signup and view all the flashcards

Streaming Stored Video

Video starts playing quickly, is interactive, and provides continuous playout.

Signup and view all the flashcards

Streaming Live Video

Video content delivered simultaneously to many clients.

Signup and view all the flashcards

Conversational Video over IP

Real-time, interactive communication often with multiple participants.

Signup and view all the flashcards

Reliable Transport Protocol

Ensuring data is delivered reliably over the internet by using TCP

Signup and view all the flashcards

Video Streaming Steps

Video is first created, then compressed via an encoding algorithm, DRM secured, delivered from a server, and decoded at the user end.

Signup and view all the flashcards

Video Compression Redundancy

Spatial redundancy (within images) and temporal redundancy (across images).

Signup and view all the flashcards

JPEG Steps

Transform to YCbCr, divide into 8x8 blocks, Discrete Cosine Transform, Quantization, and encodes the data

Signup and view all the flashcards

Frames encoding

encode the difference between consecutive frames (P or B frames) rather than the entire frame.

Signup and view all the flashcards

Constant bitrate encoding

The output size of the video is constant over time

Signup and view all the flashcards

Variable bitrate encoding

The output size of the video varies based on the underlying scene complexity

Signup and view all the flashcards

Byte-Range Requests

Client pacing of video by sending byte-range requests.

Signup and view all the flashcards

Playout Buffer

Video is pre-fetched and stored in a buffer to account for Internet variations.

Signup and view all the flashcards

Multiple Bitrates Encoding

Segment is encoded at multiple bitrates and client requests with a specified quality.

Signup and view all the flashcards

What is DASH

Video is chunked, encoded at multiple bitrates, where algorithm choose chunk based on network estimates.

Signup and view all the flashcards

Bitrate Adaptation

The client adapts the video bitrate or the quality of the video chunks to download based on network conditions.

Signup and view all the flashcards

QoE Characteristics

Low re-buffering, high video quality, and low video quality variations.

Signup and view all the flashcards

Bitrate Adaptation Signals

The network throughput and video buffer.

Signup and view all the flashcards

Rate-Based Adaptation Steps

Estimate the future bandwidth and select max bitrate below estimate.

Signup and view all the flashcards

Study Notes

  • Multimedia applications involve audio and video, presenting distinct challenges and requiring specific techniques.

Video and Audio Characteristics

  • Video has a high bit rate, ranging from 100 kbps to over 3 Mbps, varying with video quality.
  • Audio has a lower bit rate than video.
  • Audio glitches are more noticeable than video glitches.
  • Both video and audio can be compressed at varying quality levels.

Types of Multimedia Applications

  • Streaming stored audio and video,Udacity.
  • Conversational voice and video over IP (VoIP), Skype.
  • Streaming live audio and video, GATech graduation ceremony.

Streaming Stored Video

  • Begins playing within seconds of data receipt.
  • Interactive, allowing pause, fast forward, skip, and rewind.
  • Continuous playout without freezing.
  • Stored on a CDN or implemented with a peer-to-peer model.

Streaming Live Audio and Video

  • Similar to streaming stored content but with differences.
  • Involves many simultaneous users across different locations.
  • Sensitive to delay, but a ten-second delay is generally acceptable.

Conversational Voice and Video over IP (VoIP)

  • VoIP is phone service over the Internet.
  • Often involves multiple participants.
  • Highly delay-sensitive, where delays over 400ms can be frustrating.
  • Loss-tolerant, with techniques to conceal glitches and the ability to ask for repeats.

How VoIP Works

  • VoIP is transmitted over the Internet, which operates on a "best effort" basis without guaranteed delivery or timing.

VoIP Topics

  • Encoding.
  • Signaling.
  • QoS (Quality of Service) metrics.

Audio Encoding

  • Analog audio, represented as a continuous wave, is encoded into digital format using discrete data.
  • Audio is encoded by sampling thousands of times per second and rounding each sample's value, a process called quantization.

Encoding Technique

  • PCM (Pulse Code Modulation) takes 8000 samples per second with 8-bit values for speech and 44,100 samples per second with 16-bit values for audio CDs.
  • Encoding schemes include narrowband, broadband, and multimode, each with different characteristics and tradeoffs.
  • Audio, like video, can be compressed, but there are tradeoffs.

Signaling Protocols

  • Used to set up and tear down calls.
  • Responsible for user location, session establishment, session negotiation, and call participation management.
  • VoIP uses signaling protocols like SIP (Session Initiation Protocol).
  • VoIP faces signaling challenges due to interfacing with the PSTN.

QoS metrics for VoIP

  • End-to-end delay.
  • Jitter.
  • Packet loss.

End-to-End Delay

  • Includes encoding, packetizing, network, playback buffer, and decoding delays.
  • Below 150ms is unnoticeable, 150-400ms is noticeable but acceptable, and over 400ms is unacceptable.
  • VoIP applications often discard packets delayed beyond a threshold.

Jitter

  • Jitter interferes with reconstructing the analog voice stream, leading to gaps in audio.
  • Audio gaps should be kept below 30ms to maintain intelligibility.
  • The jitter buffer smooths out delay variations by buffering and playing packets at a steady rate.
  • Tradeoffs exist between longer buffers that reduce packet discard but add end-to-end delay, and shorter buffers that reduce delay but increase packet loss.

Packet Loss

  • Packet loss is inevitable.
  • TCP is not suitable for VoIP due to retransmission delays and congestion control impacts.
  • VoIP protocols use UDP.
  • Packet loss occurs if packets never arrive or arrive after their scheduled playout. Tolerance for packet loss can range from 1 to 20 percent, depending on the voice codec and other factors.

Methods of Dealing with Packet Loss:

  • FEC (Forward Error Correction) transmits redundant data to replace lost data.
  • Interleaving mixes audio chunks to minimize consecutive losses.
  • Error concealment estimates lost audio packets based on neighboring snippets. FEC consumes more bandwidth and increases playout delay. Interleaving increases latency and is less useful for VoIP. Error concealment is computationally efficient but may not always be accurate.

Live/On-Demand Streaming Introduction

  • Streaming media content accounts for 60-70% of Internet traffic.
  • Enabled by increased bandwidth, efficient video compression, and Digital Rights Management.
  • Types of streamed content: live (sports events, concerts) and on-demand (Netflix, YouTube).
  • The constraint for live streaming is that there is not a lot of room for pre-fetching content.

Video Streaming

  • The video content is first created, compressed, secured with DRM, and hosted on a server.
  • Content providers use their own data centers or third-party CDNs for scalability.
  • End-users download, decode, and render the video.
  • Goal: reduce size by removing similar information without significantly compromising quality.
  • Lossy compression offers higher bandwidth savings but cannot recover original quality.

Image Compression Steps (JPEG)

  • Transform RGB image into YCbCr components.
  • Divide the image into 8x8 blocks and apply Discrete Cosine Transformation (DCT).
  • Compress the matrix of coefficients using a Quantization table (lossy step).
  • Perform lossless encoding to store the coefficients.

Video Compression and Temporal Redundancy

  • Encode the first frame (I-frame) as a JPEG and then encode the difference between frames (P-frame).
  • Insert an I-frame periodically (e.g., every 15 frames) to avoid using the difference all the way, and is known as a Group of Pictures (GoP).
  • Encode a frame as a function of the past and future I- or P-frames (B-frame).

VBR vs CBR

  • VBR (variable bitrate) offers better image quality than CBR (constant bitrate) but is more computationally expensive.
  • Video compression requires heavy computation, with specialized hardware encoders for real-time live streaming.

Transport Protocols

  • Videos need to be decoded the client side, and can fail partially if data is lost
  • TCP provides reliability and congestion control, making it suitable for video delivery
  • The intial vision was to use specialized video servers that know the state of the client and can pause.

HTTP protocol.

  • Use existing HTTP protocol, stateless
  • Clients are intelligent, download
  • Can use CDN's
  • Easier to bypass middleboxes

Progressive Download vs Streaming

  • Send an HTTP GET request for video.
  • waste of network
  • Video buffer at the client to store this content in memory
  • client tries to pace it
  • byte-range requests
  • streaming without stalls.
  • Accounts for variations by using playout buffer:
  • video buffer is 5 seconds.
  • full, the client will wait for it to get depleted.
Streaming is in two states:
  • Filling state
  • Steady state

How to Handle Network and User Device Diversity

  • Multiple bitrates encoding because:
  • device over
  • network environments over

Bitrate Adaptation in DASH

  • Client dynamically adjusts the video bitrate based on the network conditions and device type, Dynamic Streaming over HTTP or DASH where dynamic streaming signifies the dynamic bitrate adaptation.

Goal of Bitrate Adaptation:

  • A good quality of experience (QoE) is usually characterized by the following:
  • Low or zero re-buffering:
  • High video quality
  • low video quality variations
  • low startup latency

Bitrate Adaptation Algorithms

  • Signals to select the bitrate depends on,
  • network Throughput
  • video Buffer

Rate-Based Adaptation Mechanisms

  • involves estimating the future bandwidth.
  • continuous throughput is mapped to discrete bitrate. Reasons for an added Factor:
  • Conservative in our estimate
  • VBR-encoded
  • transport-layer overheads.
Issues with Bitrate Adaptation
  1. bandwidth is 5 Mbps for the first 20 seconds
  2. Available

UNDER-Estimation with Rate-Based Adaption

DASH clients have an ON-OFF pattern in the steady state. This happens when the video client has the buffer filled up. And it is waiting for to deplete:

  • the TCP connection reset the congestion window.
  • two competing TCP-flows they would have gotten their fair share.

Buffer-Based adaptation

  • The bitrate of is a function of the buffer occupancy is for a higher bandwidth the bitrate
  • Using buffer-based adaptation can overcome the errors in bandwidth estimation that we saw in rate-based adaptation.
  • But it has its own issues
  • Low or zero re-buffering
  • Large buffer to implement Bitrate Adaptation Algorithms

Studying That Suits You

Use AI to generate personalized quizzes and flashcards to suit your learning preferences.

Quiz Team

Related Documents

More Like This

Use Quizgecko on...
Browser
Browser