images and videos formats
images and videos formats
IMAGES-VIDEOS
● An image consists of a rectangular array of dots called pixels. The size of the image is
specified in terms of width X height, in numbers of the pixels.
● The physical size of the image, in inches or centimeters, depends on the resolution of the
device on which the image is displayed. The resolution is usually measured in DPI (Dots
Per Inch).
● An image will appear smaller on a device with a higher resolution than on one with a
lower resolution. For color images, one needs enough bits per pixel to represent all the
colors in the image. The number of the bits per pixel is called the depth of the image.
1-bit images-
➢ An image is a set of pixels. Note that a pixel is a picture element in digital image.
In 1-bit images, each pixel is stored as a single bit (0 or 1). A bit has only two
states either on or off, white or black, true or false.
➢ Therefore, such an image is also referred to as a binary image, since only two
states are available.
➢ 1-bit image is also known as 1-bit monochrome images because it contains one
color that is black for off state and white for on state.
➢ A 1-bit image with resolution 640*480 needs a storage space of 640*480 bits.
640 x 480 bits. = (640 x 480) / 8 bytes = (640 x 480) / (8 x 1024) KB= 37.5KB.
The clarity or quality of 1-bit image is very low.
➢ Each pixel of 8-bit gray level image is represented by a single byte (8 bits).
Therefore each pixel of such image can hold 28=256 values between 0 and 255.
➢ Therefore each pixel has a brightness value on a scale from black (0 for no
brightness or intensity) to white (255 for full brightness or intensity).
➢ For example, a dark pixel might have a value of 15 and a bright one might be 240.
➢ A grayscale digital image is an image in which the value of each pixel is a single
sample, which carries intensity information.
➢ Images are composed exclusively of gray shades, which vary from black being at
the weakest intensity to white being at the strongest.
➢ Grayscale images carry many shades of gray from black to white. Grayscale
images are also called monochromatic, denoting the presence of only one (mono)
color (chrome).
➢ An image is represented by bitmap. A bitmap is a simple matrix of the tiny dots
(pixels) that form an image and are displayed on a computer screen or printed.
➢ A 8-bit image with resolution 640 x 480 needs a storage space of 640 x 480
bytes=(640 x 480)/1024 KB= 300KB. Therefore an 8-bit image needs 8 times
more storage space than 1-bit image.
● The number of entries in the palette determines the maximum number of colors
which can appear on screen simultaneously.
● The width of each entry in the palette determines the number of colors which the
wider full palette can represent.
A common example would be a palette of 256 colors that is the number of entries is 256 and thus
each entry is addressed by an 8-bit pixel value. Each color can be chosen from a full palette, with
a total of 16.7 million colors that is the each entry is of 24 bits and 8 bits per channel which sets
the total combinations of 256 levels for each of the red, green and blue components 256 x 256 x
256 =16,777,216 colors.
This is the native Photoshop file format created by Adobe. You can import this format
directly into most desktop publishing applications.
➔ The RGB color model is an additive color model in which red, green and blue
light are added together in various ways to reproduce a broad array of colors.
➔ The name of the model comes from the initials of the three additive primary
colors red, green, and blue.
➔ The main purpose of the RGB color model is for the sensing, representation, and
display of images in electronic systems, such as televisions and computers,
though it has also been used in conventional photography
➔ . Before the electronic age, the RGB color model already had a solid theory
behind it, based in human perception of colors.
➔ For Example- Computer Monitor, Television etc.
➔ In the CMYK model, white is the natural color of the background, while black
results from a full combination of colored inks.
➔ To save cost on ink, and to produce deeper black tones, unsaturated and dark
colors are produced by using black ink instead of the combination of cyan,
magenta, and yellow.
➔ The CMYK model uses printing inks.
➔ For Example- Paint, Pigments, and color filter etc.
Advantages:
1. Easy to Implement.
2. It uses color space for applications.
3. No transformation for data display.
Disadvantages:
1. We cannot transfer the color values from one to another device.
2. Complex to determine the particular color.
Consider you have an color image , means you have three different arrays of RED, GREEN
and BLUE. Now if you want to convert it into CMY, here’s what you have to do. You have
to subtract it by the maximum number of levels – 1. Each matrix is subtracted and its
respective CMY matrix is filled with result.
Color Look-Up Table
● The color Look Up table is a technique or process to convert a range of input
colors into another range of colors. It is also called as “CLUT.”
● The color Lookup table has existed in the graphics card. Graphics Card is also
called “Display Adapter.”
● The Color Look-Up table provides us various colors that are used to modify the
color of the objects.
● Either we can use the colors available in the palette, or we can create the colors of
our choice in the color window.
● In image processing, the lookup table is used to change the input into the more
required output format. The Color Look-Up table is used to store and index the
color value of an image.
Look Up File: The Look-Up file is a two-dimensional table that is used to contain the
data. The Look Up data is stored in a disk file.
Color Palettes: The color palettes are defined as mathematical tables used to determine
the color of the pixel that is displayed on the screen.
In the Macintosh Operating system, it is known as “Color Look-Up table.”
In Windows operating system, it is known as the “Color palette.”
It is used to stores a set of bytes instead of the color of the image.
Advantages:
1. Easy to modify.
2. Space Efficient.
Disadvantages:
1. It does not maintain any changes in the history.
2. There is a need to determine and maintain the reference.
Direct Coding
● “Direct Coding is a technique or process which is used to provide a certain amount
of memory storage space for a pixel to encode the pixel with the color.”
For Example-
● If we assign one bit for each primary color to 3-bit for each pixel.
● This 3-bit representation allows the two intensity levels of each primary to
separate: 0(Off) or 1(On),then each pixel can occupy one color out of eight colors
that are similar to the corner of the RGB color cube.
0 0 0 Black
0 0 1 Blue
0 1 0 Green
0 1 1 Cyan
1 0 0 Red
1 0 1 Magenta
1 1 0 Yellow
1 1 1 White
● Mostly 3 bytes, or 24 bits per pixel are used in industries and companies, with 1
byte for each primary color. We can allot each primary color to have 265 different
intensity levels, similar to binary values from 00000000 to 11111111.
● The color of the pixel can be 265 x 265 x 265 or 16.7 million. The representation
of black and white, the grayscale image, needs just 1 bit per pixel.
● The bit value 0 represents black, and 1 represents white. The direct coding
technique is simple, and it also supports various applications.
file formats:- JPEG, PNG, GIF etc. file formats:- PDF, EPS etc
Basically it is used for online logos, online ads, Basically it is used for business
digital graphics, photographs for website, social cards, stationary, stickers, posters,
media, or apps etc. brochures etc.
Video signal
A video signal is a signal produced by the video adapter that allows a
display device, such as a computer monitor, to display a picture.
S-Video
● S-video was one of a number of enhancements in bringing the signal from the video
cassette player to TVs, and separated video signal (luminance) and color signal
(chrominance).
● S-video is a technology for transmitting video signals over a cable by dividing the video
information into two separate signals: one for color (chrominance), and the other for
brightness (luminance).
● When sent to a television, this produces sharper images than Composite Video, where the
video information is transmitted as a single signal over one wire. This is because
televisions are designed to display separate Luminance (Y) and Chrominance (C) signals.
● Computer monitors are designed for RGB (short for Red, Green, Blue monitor) signals.
Most digital video devices such as digital cameras and game players produce video in
RGB format. The images are clearer when displayed on a computer monitor. When
displayed on a standard television, however, they look better in S-Video format than in
Composite Video format.
● To use S-Video, the device sending the signals must support S-Video output and the
device receiving the signals must have an S-Video input jack. Then you need an S-Video
cable to connect the two devices.
● S-Video cable doesn't always come standard with a TV, and usually must be purchased
separately.
● S-Video cables carry four or more wires wrapped together in an insulated sleeve, with
S-Video connectors at either end. It is only for video and requires separate audio cables,
but it provides a slightly better picture than a composite video cable.
● Like composite video, S-video connectors are widely used on VCRs, DVD players and
receivers. The audio for both composite video and S-video uses common left/right stereo
connections.
● As a result, there is less crosstalk between the color information and the crucial
gray-scale information.
● S-Video cables are used for computer-to-TV output for business or home use.
Component Video
● Component video is a video signal that has been split into two or more components. In
popular use, it refers to a type of analog video information that is transmitted or stored as
three separate signals.
● Component analog video signals do not use R, G, and B components but rather a
colorless component, termed luma, which provides brightness information (as in
black-and-white video).
● This combines with one or more color-carrying components, termed chroma, that give
only color information.
● In component video, the luminance (Y) and two color difference signals (U and V or I
and Q) are separated into three separate analog signals that can be transmitted over three
separate wires.
● Component video is used in professional video production and provides the best quality
and the most accurate reproduction of colors.
● Component Video gives the best color reproduction since there is no crosstalk between
the three channels. Component video requires more bandwidth and good synchronization
of the three components.
● Component video cables come in three-wire sets shown below.
Analog video:
one channel, it is called composite video as is the case, among others with NTSC,
● Analog video may be carried in separate channels, as in two channel S - Video (YC) and
● Analog video is used in both consumer and professional television production applications.
● However, digital video signal formats with higher quality have been adopted, including serial
digital interface (SDI), Firewire (IEEE 1394), Digital Visual Interface (DVI) and High -
● Most TV is still sent and received as an analog signal. Once the electrical signal is received,
we may assume that brightness is at least a monotonic function of voltage, if not necessarily
● An analog signal f(t) samples a time - varying image. So - called progressive scanning traces
through a complete picture (a frame) row - wise for each time interval. A high - resolution
is used. Here, the odd - numbered lines are traced first, then the even - numbered lines. This
results in "odd" and "even" fields — two fields make up one frame.
● In fact, the odd lines (starting from 1) end up at the middle of a line at the end of the odd
field, and the even scan starts at a half - way point. The following figure shows the scheme
used. First the solid (odd) lines are traced— P to Q, then R to S, and so on, ending at T —
then the even field starts at U and ends at V. The scan fines are not horizontal because a small
● Interlacing was invented because, when standards were being defined, it was difficult to
transmit the amount of information in a full frame quickly enough to avoid flicker. The
● Because of interlacing, the odd and even lines are displaced in time from each other. This is
generally not noticeable except when fast action is taking place onscreen, when blurring may
occur. For example, in the video in the following figure, the moving helicopter is blurred
● Since it is sometimes necessary to change the frame rate, resize, or even produce stills from
an interlaced source video, various schemes are used to de - interlace it. The simplest de -
interlacing method consists of discarding one field and duplicating the scan lines of the other
field, which results in the information in one field being lost completely. Other, more
● CRT displays are built like fluorescent lights and must flash 50 to 70 times per second to
appear smooth. In Europe, this fact is conveniently tied to their 50 Hz electrical system, and
they use video digitized at 25 frames per second (fps); in North America, the 60 Hz electric
● The jump from Q to R and so on is called the horizontal retrace, during which the electronic
beam in the CRT is blanked. The jump from T to U or V to P is called the vertical retrace.
NTSC Video
● NTSC, named for the National Television System Committee, is the analog television system
that is used in most of North America, parts of South America (except Brazil, Argentina,
Uruguay, and French Guiana), Myanmar, South Korea, Taiwan, Japan, the Philippines, and
● The first NTSC standard was developed in 1941 and had no provision for color television. In
1953 a second modified version of the NTSC standard was adopted, which allowed color
television broadcasting compatible with the existing stock of black - and - white receivers.
● NTSC was the first widely adopted broadcast color system and remained dominant where it
had been adopted until the first decade of the 21st century, when it was replaced with digital
ATSC.
● Digital broadcasting permits higher - resolution television, but digital standard definition
television in these countries continues to use the frame rate and number of lines of resolution
● systems using the NTSC frame rate and resolution (such as DVDs) are still referred to
informally as "NTSC". NTSC baseband video signals are also still often used in video
playback (typically of recordings from existing libraries using existing equipment) and in
● Different video formats provide different numbers of samples per line, as listed in the above
table. Laser disks have about the same resolution as Hi - 8. (In comparison, mini DV 1/4 -
inch tapes for digital video are 480 lines by 720 samples per line.)
PAL Video
uses 625 scan lines per frame, at 25 frames per second (or 40 msec / frame), with a 4 : 3
● Its broadcast TV signals are also used in composite video. This important standard is widely
used in Western Europe, China, India and many other parts of the world.
● PAL uses the YUV color model with an 8 MHz channel, allocating a bandwidth of 5.5 MHz
● The color subcarrier frequency is fsc ≈ 4.43 MHz. To improve picture quality, chroma signals
have alternate signs (e.g., +U and — U) in successive scan lines; hence the name "Phase
Alternating Line.
SECAM Video
● SECAM, which was invented by the French, is the third major broadcast TV standard.
● SECAM also uses 625 scan lines per frame, at 25 frames per second, with a 4:3 aspect ratio
● The original design called for a higher number of scan lines (over 800), but the final version
● SECAM and PAL are similar, differeing slightly in their color coding scheme. In SECAM, U
and V signals are modulated using separate color subcarriers at 4.25 MHz and 4.41 MHz,
respectively.
● They are sent in alternate lines - that is, only one of the U or V signals will be sent on each
scan line.
Digital video
● Digital video comprises a series of orthogonal bitmap digital images displayed in rapid
succession at a constant rate. In the context of video these images are called frames. We
measure the rate at which frames are displayed in frames per second (FPS).
● Since every frame is an orthogonal bitmap digital image it comprises a raster of pixels. If it
has a width of W pixels and a height of Hpixels we say that the frame size is WxH.
● Pixels have only one property, their color. The color of a pixel is represented by a fixed
number of bits. The more bits the more subtle variations of colors can be reproduced. This is
An example video can have a duration (T) of 1 hour (3600sec), a frame size of 640 x 480 (W x H) at
a color depth of 24bits and a frame rate of 25fps. This example video has the following properties:
Chroma Subsampling
● Since humans see color with much less spatial resolution than black and white, it
● The first number (in this case 4), refers to the size of the sample. The two following
numbers both refer to chroma. They are both relative to the first number and define
the horizontal and vertical sampling respectively.
● A signal with chroma 4:4:4 has no compression (so it is not subsampled) and
transports both luminance and color data entirely.
● In a four by two array of pixels, 4:2:2 has half the chroma of 4:4:4, and 4:2:0 has a
quarter of the color information available.
● The 4:2:2 signal will have half the sampling rate horizontally, but will maintain full
sampling vertically.
● 4:2:0, on the other hand, will only sample colors out of half the pixels on the first
row and ignores the second row of the sample completely.
Scheme 4:2:0, along with others, is commonly used in JPEG and MPEG.
for resolutions of 1,280 by 720 in which all the rows are refreshed
introduced).
in each cycle.
● While 720p gives slightly better images than 1080i for scenes with a
great deal of motion, 1080i gives slightly greater details, which result in
simultaneously.