Chapter Five
Fundamental Concepts of Video
Complied by: Nigusu Y. (nigusu02@gmail.com )
1
Introduction
o Video is an electronic medium for the recording, copying, playing,
broadcasting and displaying of motion visual images.
o Video is a series of images which are displayed on screen at fast speed ( e.g
30 images per second)
o It projects single images at a fast rate producing the illusion of continuous
motion.
o The rate at which the frames are projected is generally between 24 and 30
frames per second (fps).
o The rate at which these images are presented is referred to as the Frame
Rate .
2
o A single image is called frame and video is a series of frames.
o An image just like conventional images is modeled as a matrix of pixels.
o To model smooth motion psychophysical studies have shown that a rate of
30 frames a second is good enough to simulate smooth motion.
o Video content is any content format that features or includes videos like
live videos, recorded presentations, movies, webinars or animated GIF.
3
Types of colour video Signals
o Component video
Each primary is sent as a separate video signal.
Best color reproduction.
Requires more bandwidth and good synchronization of the three
components
Component video takes the different components of the video and
breaks them into separate signals.
Improvements to component video have led to many video formats,
including S-Video, RGB etc.
Component cables are designed for HDTV.
4
o Composite Video
Color (chrominance) and luminance signals are mixed into a single carrier
wave.
Some interference between the two signals is unavoidable.
Composite analog video has all its components (brightness, color,
synchronization information, etc.) combined into one signal.
Due to the compositing (or combining) of the video components, the
quality of composite video is marginal at best.
The results are color bleeding, low clarity and high generational loss.
Composite cables are designed for older, standard definition (SD)
systems.
The yellow plug carries the video, while the red and white cables carry the
audio 5
o Separated video
– Compromise between component analog video and the composite video.
– It uses two lines, one for luminance and another for composite
chrominance signal
– The National Television Standards Committee ( NTSC) standard used in
North America and Japanese Television uses a 768 by 484 display.
– The Phase Alternative system (PAL) standard for European television is
slightly larger at 768 by 576.
– Since the resolution between analogue video and computers is different,
conversion of analogue video to digital video at times can be down-sizing
of the video and the loss of some resolution.
6
Types of Videos
o Analog Video
– Analog video is a video signal represented by one or more analog signals
– Analog technology requires information representing images and sound
to be in a real-time continuous scale electric signal between sources and
receivers.
– It is used throughout the television industry.
– For television, images and sound are converted into electric signals by
transducers.
– Distortion of images (loosing the original shape) and noise (colour
fluctuation and boosting images) are common problems for analog video.
7
Con’t
– In an analogue video signal, each frame is represented by a fluctuating
voltage signal known as analogue waveform.
– One of the earliest formats for analog video was composite video.
– Analog formats are vulnerable to loss due to transmission noise effects.
– Quality loss is also possible from one generation to another.
– This type of loss is like photocopying, in which a copy of a copy is
never as good as the original.
8
o Digital Video
– Digital video is an electronic representation of moving visual
images (video) in the form of encoded digital data.
– It comprises a series of digital images displayed in rapid succession.
– Computer based digital video is defined as a series of individual
images and associated audio.
– These elements are stored in a format in which both elements (pixel
and sound sample) are represented as a series of binary digits (bits).
9
Con’t
There are two significant advantages for using computers for digital
video :
o The ability to randomly access the storage of video
o Compress the video stored.
o Advantages:
– Direct random access –> good for nonlinear video editing
– No problem for repeated recording
– No need for blanking and sync pulse
o Almost all digital video uses component video
10
Factors of Digital video
o Frame Rate
The standard for displaying any type of non-film video is 30 frames
per second (film is 24 frames per second).
This means that the video is made up of 30 (or 24) pictures or frames
for every second of video.
Additionally these frames are split in half (odd lines and even lines),
to form what are called fields.
11
o Colour Resolution
Color resolution refers to the number of colors displayed on
the screen at one time.
Computers deal with color in an RGB (red-green-blue)
format, while video uses a variety of formats.
One of the most common video formats is called YUV.
Although there is no direct correlation between RGB and
YUV, they are similar in that they both have varying levels of
color depth (maximum number of colour).
12
o Spatial Resolution
– The third factor is spatial resolution or in other words, "How big is the
picture?".
– Since PC and Macintosh computers generally have resolutions in
excess of 640 by 480, most people assume that this resolution is the
video standard.
o Image Quality
– The last, and most important factor is video quality that is acceptable
for your application.
– For some this may be 1/4 screen, 15 frames per second (fps), at 8 bits
per pixel.
– Other require a full screen (768 by 484), full frame rate video, at 24
bits per pixel (16.7 million colours). 13
Displaying Video
There are two ways of scanning / displaying video on screen:
Interlaced scan
Progressive scan
o Interlaced Scanning
Interlaced video displays even and odd scan lines as separate fields.
Interlaced scanning writes every second line of the picture during a
scan, and writes the other half during the next sweep.
Doing that we only need 25/30 pictures per second.
This idea of splitting up the image into two parts became known as
interlacing and the splitted up pictures as fields.
14
o During the first scan the upper field is written on screen.
o The first, 3rd, 5th, etc. line is written and after writing each line the
electron beam moves to the left again before writing the next line.
o Once all the odd lines have been written the electron beam travels back to
the upper left of the screen and starts writing the even lines.
15
Originally used in traditional analog SD (Standard Definition)
broadcasts since it was more efficient in transmitting video.
Examples interlaced video displaying
o Television
Scans 625, 525 horizontal lines top to bottom
Scan line using interlacing system
Scan 25-30HZ for fulltime
Uses limited colour palette and restricted luminance (lightness or
darkness)
16
o Progressive Scanning
Progressive scan video content displays both the even and odd
scan lines (the entire video frame) on the TV at the same time.
PC CRT displays are fundamentally different from TV screens.
Monitor writes a whole picture per scan.
Progressive scan updates all the lines on the screen at the same
time, 60 times every second.
This is known as progressive scanning and today all PC screens
write a picture like this.
17
Examples progressive video displaying
o Computer
– Scans 480 horizontal lines from
– Scan each line progressively
– Scan full frame at a rate of typically 66.67 HZ or higher
– Use RGB color model
18
Recording Video
o CCDs (Charge Coupled Devices) a chip containing a series of tiny, light-
sensitive photo-sites that forms the heart of all electronic and digital
cameras.
o Digital camera uses lens which focuses the image onto a CCD, which
then converts the image into electrical pulses saved into memory.
o In short, just as the film in a conventional camera records an image
when light hits it, the CCD records the image electronically.
o The photo-sites convert light into electrons.
19
Video Broadcasting Standards/ TV standards
The video broadcasting standards are: PAL, NTSC, SECAM and HDTV
o PAL (Phase Alternate Line)
PAL uses 625 horizontal lines at a field rate of 50 fields per second (or 25
frames per second).
Only 576 of these lines are used for picture information with the remaining
49 lines used for sync or holding additional information such as closed
captioning.
It is used in Australia, New Zealand, United Kingdom, and Europe.
Scans 625 lines per frame, 25 frames per second (40 msec/frame)
Interlaced, each frame is divided into 2 fields, 312.5 lines/field
For color representation, PAL uses YUV color model
o In PAL,
– 5.5 MHz is allocated to Y,
– 1.8 MHz each to U and V 20
o SECAM (Sequential Color with Memory)
SECAM uses the same bandwidth as PAL but transmits the color
information sequentially.
It is used in France, East Europe, etc
SECAM (System Electronic Pour Couleur Avec Memoire) is very
similar to PAL.
It specifies the same number of scan lines and frames per second.
It is the broadcast standard for France, Russia, and parts of Africa and
Eastern Europe.
21
o NTSC (National Television Standards Committee)
NTSC is a black-and-white and color compatible 525-line system that
scans a nominal 30 interlaced television picture frames per second.
Used in USA, Canada, and Japan.
525 scan lines per frame, 30 frames per second
Interlaced, each frame is divided into 2 fields i.e. 262.5 lines/field
20 lines reserved for control information at the beginning of each field
So a maximum of 485 lines of visible data
Each line takes 63.5 microseconds to scan.
For color representation, NTSC uses YIQ color model.
Basic Compression Idea
22
o HDTV (High Definition Television)
High-Definition television (HDTV) means broadcast of television signals
with a higher resolution than traditional formats (NTSC, SECAM, PAL)
allow.
Except for early analog formats in Europe and Japan, HDTV is broadcasted
digitally and its the introduction of digital television (DTV).
Modern plasma television uses this broadcast
It consists of 720-1080 lines and higher number of pixels (as many as 1920
pixels).
Having a choice in between progressive and interlaced is one advantage of
HDTV.
HDTV signal is digital resulting in crystal clear, noise-free pictures and CD
quality sound. 23
Color Models in Video
YUV and YIQ are the two commonly used color models in video.
o YUV Color Model
Initially, for Phase Alternative system (PAL) analog video, it is now also
used in CCIR 601 standard for digital video
Chrominance is defined as the difference between a color and a reference
white at the same luminance.
The YUV color model is the basic color model used in analogue color TV
broadcasting.
Initially YUV is the re-coding of RGB for transmission efficiency
(minimizing bandwidth) and for downward compatibility with black-and
white television.
The YUV color space is “derived” from the RGB space.
It comprises the luminance (Y) and two color difference (U, V) components
24
It can be represented by U and V -- the color differences.
U=B-Y
V=R-Y
– If b/w image, then U = V = 0. --> No chrominance!
The luminance can be computed as a weighted sum of red, green and blue
components;
The color difference(chrominance) components are formed by
subtracting luminance from blue and from red.
The principal advantage of the YUV model in image processing is
decoupling of luminance and color information.
25
o The importance of this decoupling is that the luminance component of an image
can be processed without affecting its color component.
o Notice: In the middle it is completely black, which is where U and V are zero,
and Y is as well and if there is no chrominance the picture can be B and W.
o U is the axis from blue to yellow and V is the axis from magenta to cyan.
o Y ranges from 0 to 1 (or 0 to 255 in digital formats), while U and V range from -
0.5 to 0.5 (or -128 to 127 in signed digital form, or 0 to 255 in unsigned form).
26
YIQ Color Model
o The YIQ colour space model is use in U.S. commercial colour television
broadcasting (NTSC).
o It is a rotation of the RGB colour space such that the Y axis contains the
luminance information, allowing backwards-compatibility with black-and-
white colour TV's, which display only this axis of the colour space.
o The chrominance information is contained in the I (orange-blue) and Q (purple-
green) axes, which are roughly orthogonal.
o The human visual system is much more sensitive to changes in the I axis than in
the Q axis, allowing the Q axis to be transmitted with less fidelity, conserving
bandwidth.
27
Thanks!
28