Post on 07-Feb-2018
Cairo University
Faculty of Engineering Electronics and Electrical Communications Department
Professional Masters Program – Major Telecommunications
Part 1: Voice Over IP (VoIP)
Dr. Wagdy Anis Aziz, Adjunct Doctor Senior Manager, Core Network Support, Mobinil
wanis@mobinil.com +201222201073
EPS – 3GPP Architecture Domains
• From an operator’s perspective, in order to provide a data-only mobile broadband service, the infrastructure must be upgraded to EPS. • EPS provides components that allow existing 2G/ G access networks to utilize EPC components.
• For those incumbent 2G/ G operators the existing CS network can provide access to voice calls in the short term, but the deployment of
IMS in conjunction with EPS would provide an All-IP network with access to speech services.
• Our focus here is how EPS is supporting voice services. • There are two fundamentally different ways that voice services can be realized for LTE users; using circuit-switched or IP Multimedia
Subsystem (IMS) technologies.
• Voice services based on circuit-switched technology Circuit-switched fallback CSFB
• Voice services with IMS technology Single-Radio Voice Call Continuity SRVCC
Why IP Networks?
1. Cost reduction - There can be a real savings in long distance
telephone costs which is extremely important to most companies, particularly those with international markets.
2. Simplification - An integrated voice/data network allows more standardization and reduces total equipment needs.
3. Consolidation -The ability to eliminate points of failure,
consolidate accounting systems and combine operations is obviously more efficient.
4. Advanced Applications -The long run benefits of IP include support for multimedia and multileveled applications, something which today's telephone system can't compete with.
Migration of all 2G/3G network interfaces to IP
Interface Vendor Status Main Benefits
3G
IuCs (RNC>>CS core)
Huawei Validated * important for the implementation of Iu-Flex feature which is needed for the completion of the MSC pool
project NSN Validated in lab, live trial in May
ALU trial planned in June
IuPS (RNC>>PS core)
Huawei Validated and deployed *Saving on transport side due to better utilization of pooled BW.
*Saving on PaCo side due to better SGSN resources utilization.
NSN Validated and deployed
ALU Validated and deployed
Iur (RNC>>RNC)
Huawei Validated and deployed * Saving on TX side due to better utilization of pooled
BW. NSN Validated and deployed
ALU No Need
Iub Dual Stack (nodeB>>RNC)
Huawei Validated and deployed * Saving on TX side due to higher transmission efficiency
* Allows reaching higher throughput per site by using the IP BW offered by MBH.
NSN Validated and deployed
ALU Validated
Iub Full IP (nodeB>>RNC)
Huawei trial planned in May * More savings on TX side
* Voice & Data sharing the same transmission resources pool allows for higher throughput during low voice
traffic period.
NSN trial planned in June
ALU trial planned in June
2G
A (BSC>>CS core)
Huawei live trial ongoing * Saving (CAPEX&OPEX) from BSS side due to removal of TC
*Better voice quality after removing transcoding and enabling TrFO calls.
*Important for the implementation of A-Flex feature which is needed for the completion of the MSC pool
project.
NSN trials planned in May
ALU trial planned after B12 upgrade
Gb (BSC>>PS core)
Huawei live trial ongoing * Offer higher data throughput by better utilization of the pooled resources of packet core and transport
networks NSN trial planned in May
ALU Validated
Abis (BTS>>BSC)
Huawei No plans yet * Saving on TX side due to statistical multiplexing (but needs introducing of new Synchronization methods) NSN No plans yet
ALU No plans yet
• What is VoIP?
• Why is VoIP Attractive?
• How Does it Work?
• Voice Coding
• VoIP Signaling Standards
• QoS Impairment Factors
• QoS Measurement Methods
• Design of VoIP Network Using SIP
Topics
VoIP Main Factors
VoIP Codecs – G Series , AMR. vocoder (VOice enCODER),)
VoIP Signaling Protocols – H.323 , SIP , H.248 , Megaco , SIGTRAN,
BICC .
VoIP QoS Impairment Parameters – Delay , Packet Loss , Jitter.
VoIP QoS Measurement Methods – Subjective Methods (MOS) and
Objective Methods ( PESQ , E-Model)
• “VoIP” stands for Voice over Internet Protocol.
• VoIP is a technique that encodes voice to the low rates and route the relatively low bandwidth signals as packetized "data", over dedicated transmission facilities or the "Internet" using the Internet Protocol.
• VoIP is the ability to make telephone calls and send faxes over IP-based data networks with a suitable quality of service (QoS) and superior cost/benefit.
What is VoIP?
1. Cost reduction - As described, there can be a real savings in long distance telephone costs which is extremely important to most companies, particularly those with international markets.
2. Simplification - An integrated voice/data network allows more standardization and reduces total equipment needs.
3. Consolidation -The ability to eliminate points of failure, consolidate accounting systems and combine operations is obviously more efficient.
4. Advanced Applications -The long run benefits of VoIP include support for multimedia and multileveled applications, something which today's telephone system can't compete with.
Why is VoIP Attractive?
Introduction to VoIP Codecs
• In VoIP applications, voice call is the mandatory service even when a video session is
enabled. A VoIP tool (e.g., Skype,….) normally provides many voice codecs which can be
selected or updated manually or automatically.
• Typical voice codecs used in VoIP include ITU-T standards such as 64 kb/s G.711 PCM, 8
kb/s G.729 and 5.3/6.3 kb/s G.723.1; 3GPP standards such as AMR; opensource codecs
such as iLBC and proprietary codecs such as Skype’s SILK codec which has variable bit
rates in the range of 6 to 40 kb/s.
• Some codecs can only operate at a fixed bit rate, whereas many advanced codecs can have
variable bit rates which may be used for adaptive VoIP applications to improve voice quality.
• Voice codecs or speech codecs are based on different speech compression techniques
which aim to remove redundancy from the speech signal to achieve compression and to
reduce transmission and storage costs.
• In practice, speech compression codecs are normally compared with the 64 kb/s PCM
codec which is regarded as the reference for all speech codecs. Speech codecs with the
lowest data rates (e.g., 2.4 or 1.2 kb/s Vocoder) are used mainly in secure communications.
• In general, the higher the speech bit rate, the higher the speech quality and the greater the
bandwidth and storage requirements.
• In practice, it is always a trade-off between bandwidth utilization and speech quality.
Speech Compression and Coding Techniques
• Speech compression aims to remove redundancy in speech
representation to reduce transmission bandwidth and
storage space (and further to reduce cost).
There are in general three basic speech compression
techniques, which are:
1- Waveform-Based Coding
2- Parametric-Based Coding
3- Hybrid Coding Techniques.
Speech Compression and Coding Techniques (1)
1- Waveform-Based Coding
• As the name implied, waveform based speech compression is mainly to
remove redundancy in the speech waveform and to reconstruct the
speech waveform at the decoder side as closely as possible to the
original speech waveform.
• Waveform-based speech compression techniques are simple and
normally low in implementation complexity, whereas their compression
ratios are also low. The typical bit rate range for waveform-based
speech compression coding is from 64 kb/s to 16 kb/s. At bit rate lower
than 16 kb/s, the quantization error for waveform-based speech
compression coding is too high, and this results in lower speech quality.
Typical waveform-based speech compression codecs are PCM and
ADPCM (Adaptive Differential PCM)
Speech Compression and Coding Techniques (1)
1- Waveform-Based Coding
• Typical ones are Pulse Code Modulation (PCM) and Adaptive Differential PCM (ADPCM)
• For PCM, it uses non-uniform quantization to have more fine quantization steps for small speech
signal and coarse quantization steps for large speech signal (logarithmic compression). Statistics have shown that small speech signal has higher percentage in overall speech representations. Smaller quantization steps will have lower quantization error, thus better Signal-to-Noise Ratio (SNR) for PCM coding.
• There are two PCM codecs, namely PCM μ-law which is standardized for use in North America and Japan, and PCM A-law for use in Europe and the rest of the world. ITU-T G.711 was standardized by ITU-T for PCM codecs in 1988.
• For both PCM A-law and μ-law, each sample is coded using 8 bits (compressed from 16-bit linear PCM data per sample), this yields the PCM transmission rate of 64 kb/s when 8 kHz sample rate is applied (8000 samples/s × 8 bits/sample = 64 kb/s). 64 kb/s PCM is normally used as a reference point for all other speech compression codecs.
• ADPCM, proposed by Jayant in 1974 at Bell Labs, was developed to further compress PCM codec
based on correlation between adjacent speech samples. Consisting of adaptive quantiser and adaptive predictor, a block diagram for ADPCM encoder and decoder (codec).
Speech Compression and Coding Techniques (1)
1- Waveform-Based Coding
• If an ADPCM sample is coded into 4 bits, the produced ADPCM bit rate is 4 × 8
= 32 kb/s. This means that one PCM channel (at 64 kb/s) can transmit two
ADPCM channels at 32 kb/s each. If an ADPCM sample is coded into 2 bits,
then ADPCM bit rate is 2 × 8 = 16 kb/s. One PCM channel can transmit four
ADPCM at 16 kb/s each. ITU-T G.726 defines ADPCM bit rate at 40, 32, 24 and
16 kb/s which corresponds to 5, 4, 3, 2 bits of coding for each ADPCM sample.
• The higher the ADPCM bit rate, the higher the numbers of the quantization
levels, the lower the quantization error, and thus the better the voice quality.
This is why the quality for 40 kb/s ADPCM is better than that of 32 kb/s. The
quality of 24 kb/s ADPCM is also better than that of 16 kb/s.
Speech Compression and Coding Techniques (2)
2- Parametric-Based Coding
• Parametric-based speech coding is based on the principles of how speech is produced.
• Parametric compression only sends relevant parameters related with speech production to the
receiver side and reconstructs the speech from the speech production model, Thus, high
compression ratio can be achieved. The most typical example of parametric compression is
Linear Prediction Coding (LPC), proposed by Atal in 1971 at Bell Labs. It was designed to
emulate the human speech production mechanisms and the compression can reach the bit rate
as lower as 800 bit/s (Compression Ratio reaches 80 when compared to 64 kb/s PCM). It
normally operates at bit rates from 4.8 to 1.2 kb/s. The LPC based speech codecs can achieve
high compression rate, however, the voice quality is also low.
• Compared to waveform-based codecs, parametric-based codecs are higher in implementation complexity, but can achieve better compression ratio.
• A typical parametric codec is Linear Prediction Coding (LPC) vocoder which has a bit rate
from 1.2 to 4.8 kb/s and is normally used in secure wireless communications systems when
transmission bandwidth is very limited.
Speech Compression and Coding Techniques (3)
3- Hybrid Coding Techniques
• Hybrid Coding Techniques were proposed to combine the features of both waveform-
based and parametric-based coding (and hence the name of hybrid coding). It keeps the
nature of parametric coding which includes vocal tract filter and pitch period analysis, and
voiced/unvoiced decision.
• Instead of using an impulse period train to represent the excitation signal for voiced
speech segment, it uses waveform-like excitation signal for voiced, unvoiced or transition
(containing both voiced or unvoiced) speech segments.
• Many different techniques are explored to represent waveform-based excitation signals
such as multi-pulse excitation, codebook excitation and vector quantization.
• The most well known one, so called “Codebook Excitation Linear Prediction (CELP)”
has created a huge success for hybrid speech codecs in the range of 4.8 kb/s to 16 kb/s
for mobile/wireless/satellite communications achieving toll quality (MOS over 4.0) or
communications quality (MOS over 3.5).
• Almost all modern speech codecs (such as G.729, G.723.1, AMR, iLBC and SILK
codecs) belong to the hybrid compression coding with majority of them based on
CELP techniques.
Speech Compression and Coding Techniques (3)
The typical hybrid compression codecs include the following from several Standardization bodies,
such as the International Telecommunication Union, Telecommunication Standardisation Sector
(ITU-T), Européen Télécommunication Standards Institute (ETSI):
• LD-CELP: Low Delay CELP, used in ITU-T G.728 at 16 kb/s .
• CS-ACELP: Conjugate-Structure Algebraic-Code-Excited Linear Prediction, used in ITU-T G.729
at 8 kb/s.
• RPE/LTP: Regular Pulse Excitation/Long Term Prediction, used in ETSI GSM Full-Rate (FR) at 13
kb/s.
• VSELP: Vector Sum Excited Linear Prediction: ETSI GSM Half-Rate (HR) at 5.6 kb/s.
• ACELP: Algebraic CELP, used in ETSI GSM Enhanced Full-Rate (EFR) at 12.2 kb/s and ETSI
AMR from 4.75 to 12.2 kb/s.
• ACELP/MP-MLQ: Algebraic CELP/Multi Pulse—Maximum Likelihood Quantization, used in ITU-T
G.723.1 at 5.3/6.3 kb/s.
Speech Analysis & Speech Synthesis
Two disciplines play a role:
• Speech analysis: is that portion of voice processing that converts speech to digital forms suitable for storage on computer systems and transmission on digital (data or telecommunications) networks.
Speech analysis processes are also called digital speech encoding (or coding).
• Speech synthesis: is that portion of voice processing that reconverts speech data from a digital form to a form suitable for human usage. These functions are essentially the inverse of speech analysis.
speech synthesis is also called speech decoding.
Major Techniques Used in Speech Coding
Waveform Coding
(Analysis/Synthesis) in the Time Domain
G.711 PCM and G.726,G.727 ADPCM
Vocoding
(Analysis/Synthesis) in the Frequency Domain
Hybrid Coders
That combine the characteristics of the two main
types G.728 LD-CELP / G.729 A CS-CELP / G.723.1
ACELP
Major Parameters of Standard Codecs
Origin Standard Type
Codec
Bit rate
Kb/S
Voice
Frame (ms)
Look
ahead
(ms)
Algor.
delay (ms) le
ITU-T
G.711 PCM 64
0.125 0 0.125
0
G.726
G.727 ADPCM
16 50
24 25
32 7
40 0.125 0 0.125 2
G.728 LD-CELP 12.8
0.625 0 0.625 20
16 7
G.729(a) CS-ACEP 8 20 5 15 11
G.723.1 ACELP 5.3
30 7.5 37.5 19
MP-MLQ 6.3 15
ETSI GSM-FR RPE-LTP 13 20 0 20 20
GSM-HR VSEPL 6.5 20 0 20 23
Major Parameters of Standard Codecs (AMR)
AMR Rates from 3GPP
4.75 KBIT/S
• 5.15 KBIT/S
• 5.90 KBIT/S
• 6.70 KBIT/S
• 7.40 KBIT/S
• 7.95 KBIT/S
• 10.2 KBIT/S
• 12.2 KBIT/S • Sampling frequency 8 kHz/13-bit (160 samples for 20 ms frames). • AMR utilizes Voice Activity Detection (VAD) and Comfort Noise Generation (CNG)
to reduce bandwidth usage during silence periods. • Algorithmic delay is 20 ms per frame. • The complexity of the algorithm is rated at 5, using a relative scale where G.711 is
1 and G.729a is 15. • PSQM testing under ideal conditions yields Mean Opinion Scores of 4.14 for AMR
(12.2 kbit/s), compared to 4.45 for G.711 (µ-law) • PSQM testing under network stress yields Mean Opinion Scores of 3.79 for AMR
(12.2 kbit/s), compared to 4.13 for G.711 (µ-law)
Narrowband to Fullband Speech Audio Compression
• In the last sections, we mainly discussed Narrowband (NB) speech compression, aimed at
speech spectrum from 0 to 4 kHz. Not only used in VoIP systems, this 0 to 4 kHz
narrowband speech, expanded from speech frequency range of 300 Hz to 3400 Hz, has also
been used in traditional digital telephony in the Public Switched Telephone Networks (PSTN)
• In VoIP and mobile applications, there is a trend in recent years to use Wideband (WB)
speech to provide high fidelity speech transmission quality. For WB speech, the
speech spectrum is expanded to 0–7 kHz, with sampling rate at 16 kHz.
• Compared to 0–4 kHz narrowband speech, wideband speech will have more higher
frequency components and have high speech quality.
• There are currently three wideband speech compression methods which have been used
in different wideband speech codecs standardized by ITU-T or ETSI, They are:
1. Waveform compression based on sub-band (SB) ADPCM: such as ITU-T G.722.
2. Hybrid compression based on CELP: such as AMR-WB or ITU-T G.722.2.
3. Transform compression coding: such as ITU-T G.722.1.
Narrowband to Fullband Speech Audio Compression
• The above table summarizes the Narrowband, Wideband, Super-wideband and
Fullband speech/audio compression coding basic information, including signal
bandwidth, sampling rate, typical bit rate range and standards examples.
• Example:
WB-AMR is now used for 3G Mobile Networks to enhance Voice Quality.
Narrowband to Fullband Speech Audio Compression - Summary
• we have discussed key narrowband to full band speech compression
codecs standardized by ITU-T, ETSI and IETF. We now summarize them in
the following table which includes each codec’s basic information such as
which standardization body was involved, which year was standardized,
codec type, Narrowband (NB), Wideband (WB), Super-wideband (SWB) or
Fullband (FB), bit rate (kb/s), length of speech frame (ms), bits per
ample/frame (coded bits per sample or per frame), look-ahead time (ms),
and coding’s algorithmic delay (ms).
• From this table, you should be able to see the historic development of
speech compression coding standards (from 64 kb/s, 32 kb/s, 16 kb/s, 8
kb/s to 6.4/5.3 kb/s) for achieving high compression efficiency, the mobile
codecs development from GSM to AMR for 2G and 3G applications, the
development from single rate codec, dual-rate codec, 8- mode codec to
variable rate codec for achieving high application flexibility, and the trend
from narrowband (NB) codecs to wideband codecs (WB) for achieving high
speech quality (even for High Definition voice).
Narrowband to Fullband Speech Audio Compression - Summary
• This development has made speech compression codecs more efficient and
more flexible for many different applications including VoIP.
• In the table, the columns on coded bits per sample/frame and speech frame
for each codec will help you to understand payload size and to calculate
VoIP bandwidth which will be covered in RTP transport protocol.
• The columns on look-ahead time and codec’s algorithmic delay will help to
understand codec delay and VoIP end-to-end delay, a key QoS metric,
which will be discussed in detail next part.
• It has to be mentioned that many VoIP phones (hardphones or softphones)
have incorporated many different NB and even WB codecs. How to
negotiate which codec to be used at each VoIP terminal and how to change
the codec/mode/bit rate during a VoIP session on the fly will be discussed
on SIP/SDP signaling.
How does it work ?
• VoIP digitizes voice and shoots it across the Internet.
• A conversation is digitized and encapsulated into packets.
• Packets are sent across the Internet - and re-assembled at the destination and reconverted back into audio.
• Across the Internet packets can use many different routes (Packet Switching), whereas in a traditional telephone call a single dedicated circuit is required for each call (Circuit Switching).
• The routers that route traffic on the Internet are a fraction of cost of switches on traditional long distance phone networks. All this means cheaper phone calls.
Voice To/From IP
Analog
Digital
CODEC: Analog to Digital
Voice
Compress
Create Voice Datagram
Add Header (RTP, UDP, IP, etc.)
IP-Network
Process Header
Voice
Re-sequence
Decompress
CODEC: Digital to Analog
IP-Network
Analog
Digital
Speech Codecs (Vocoders)
A VoIP telephone call
The most common, standardized encoding algorithms and their coding rate and speech quality.
Vocoder Attributes
• Vocoder speech quality is a function of bit rate, complexity, and processing delay.
There usually is a strong interdependence between all these attributes and they
may have to be traded off against each other.
• For example, low-bit-rate vocoders tend to have more delay than higher bit rate
vocoders. Low-bit-rate vocoders also require higher VLSI complexity to
implement. As might be expected, low-bit-rate vocoders often have lower speech
quality than the higher bit rate vocoders.
Vocoder Attributes
1. Bit Rate: • Most vocoders operate at a fixed bit rate regardless of the input signal
characteristics; however, the goal is to make the vocoder variable-rate. For
simultaneous voice and data applications, a compromise is to create a silence
compression algorithm as shown in the below table .
2. Delay: The delay in a speech coding system usually consists of two major components:
• Frame delay
• Speech processing delay
Vocoder Attributes
3. Vocoders's Complexity : • Vocoders are often implemented on Digital Signal Processor (DSP) hardware.
Complexity can be measured in terms of computing speed in Million Instructions Per
Second (MIPS), of Random Access Memory (RAM), and of Read Only Memory
(ROM). Complexity determines cost.
4. Quality : • The measure used in comparisons is how good the speech sounds under ideal
conditions-namely, clean speech, no transmission errors, and only one encoding
(note, however, that in the real world these ideal conditions are often not met
because there can be large amounts of such background noise as street noise,
office noise, air conditioning noise, etc.). Quality is measured by different methods
but it finally mapped to Mean Opinion Score (MOS) value which is ranked from 1
to 5 to represent the degree of voice quality.
The Used Codec Parameters
Codec
Coding
Rate
(Kbps)
Speech
quality
)MOS(
Complexity
)MIPS(
Look Ahead
)ms(
Frame
Size
)ms(
Algorithmic
Delay
)ms(
G.723.1 5.3 or 6.3 3.8 Highest (20) 7.5 30 37.5
G.729 8 4 High (14-20) 5 10 15
G.711 64 4.4 Lowest (<15) 0 0.125 0
VoIP Per Call Bandwidth
• The header consists of 40 bytes: For IP (20 bytes) + User Datagram Protocol (UDP) (8 bytes) +Real-Time Transport Protocol (RTP) (12 bytes) headers.
• 6 bytes for Multilink Point-to-Point Protocol (MP) or Frame Relay Forum (FRF) (L2) header.
• 1 byte for the end-of-frame flag on MP and Frame Relay frames.
• 18 bytes for Ethernet L2 headers, including 4 bytes of Frame Check Sequence (FCS) or Cyclic Redundancy Check (CRC).
Bandwidth Calculation Formulas
The following calculations are used:
• Total packet size = (L2 header: MP or FR or Ethernet) + (IP/UDP/RTP header) + (voice payload size)
• PPS = (codec bit rate) / (voice payload size)
• Packet Per Second
• Bandwidth = total packet size x PPS
Sample Calculation (1)
For example, the required bandwidth for a G.729 call (8 Kbps codec bit rate) with RTP, MP and the default 20 bytes of voice payload is:
• Total packet size (bytes) = (MP header of 7 bytes) + (IP/UDP/RTP header of 40 bytes) + (voice payload of 20 bytes) = 67 bytes
• Total packet size (bits) = (67 bytes) x 8 bits per byte = 536 bits
• PPS = (8 Kbps codec bit rate) / (160 bits) = 50 pps
Note: 160 bits = 20 bytes (default voice payload) x 8 bits per byte
• Bandwidth per call = voice packet size (536 bits) x 50 pps = 26.8 Kbps
Sample Calculation (2)
• Another example, the required bandwidth for a G.729 call (8
Kbps codec bit rate) with Compressed RTP (cRTP) , MP and
the default 20 bytes of voice payload is:
• Total packet size (bytes) = (MP header of 7 bytes) +
(compressed IP/UDP/RTP header of 2 bytes) + (voice payload
of 20 bytes) = 29 bytes
• Total packet size (bits) = (29 bytes) x 8 bits per byte = 232 bits
• PPS = (8 Kbps codec bit rate) / (160 bits) = 50 pps
• Note: 160 bits = 20 bytes (default voice payload) x 8 bits per
byte
• Bandwidth per call = voice packet size (232 bits) x 50 pps =
11.6 Kbps
RTP header Compression
• All VoIP packets are made up of two components:
• voice samples and IP/UDP/RTP headers. Although the voice samples are
compressed by the Digital Signal Processor (DSP) and may vary in size based
on the codec used.
• These headers are a constant 40 bytes in length. When compared to the 20
bytes of voice samples in a default G.729 call, these headers make up a
considerable amount of overhead. Using cRTP, these headers can be
compressed to two or four bytes. This compression offers significant VoIP
bandwidth savings. For example, a default G.729 VoIP call consumes 24 Kb
without cRTP, but only 12 Kb with cRTP enabled.
Voice Activity Detection (VAD)
• With circuit-switched voice networks, all voice calls use 64 Kbps fixed-bandwidth
links regardless of how much of the conversation is speech and how much is
silence. With VoIP networks, all conversation and silence is packetized. Using
Voice Activity Detection (VAD), packets of silence can be suppressed.
• Over time and as an average on a volume of more than 24 calls, VAD may
provide up to a 35 percent bandwidth savings. The savings are not realized on
every individual voice call, or on any specific point measurement.
• For the purposes of network design and bandwidth engineering, VAD should not
be taken into account, especially on links that carry fewer than 24 voice calls
simultaneously.
• Various features such as music on hold and fax render VAD ineffective. When
the network is engineered for the full voice call bandwidth, all savings provided
by VAD are available to data applications.
• VAD also provides Comfort Noise Generation (CNG). Because you can mistake
silence for a disconnected call, CNG provides locally generated white noise so
the call appears normally connected to both parties.
Assignment
- Write a report about your previous experiences related to
Voice over Internet Protocol (VoIP) in your company, and
your company's approach to deal with the explosion
happening for VoIP applications. How it will handle the high
QoS requirements for VoIP and how it will deal with the
VoIP QoS Impairment Parameters (Delay, Jitter, Packet
Loss … etc)?
- Deadline: Monday March 16th.