Uploaded by Achenyu Bk

Multimedia Data Processing Notes 2024-2025

advertisement
SWE 116 : Multimedia data processing / COMPUTER GRAPHICS
➢ Computer graphics: 3 credits (45 hours); L, T, SPW
1. Digital images ▪ Bitmap images ▪ Vector images ▪ Characteristics of bitmap images ▪ Image
compression ▪ Final improvements of images ▪ Practicals on the creation buttons and images for
the Web
2. Sound ▪ Definition of sound ▪ Characteristics of sound ▪ Digitalization of sound ▪ Mono and
stereo sound ▪ Size of a sound file ▪ Sound compression
3. Video ▪ Definition of video ▪ Analogue video ▪ Digital video ▪ Compression of digital video
(notion on Codec)
Introduction
Multimedia in principle means data of more than one medium. It usually refers to data representing
multiple types of medium to capture information and experiences related to objects and events.
Commonly used forms of data are numbers, alphanumeric, text, images, audio, and video. In
common usage, people refer a data set as multimedia only when time-dependent data such as audio
and video are involved.
Computer Graphics
Computer graphics is the field of computer science that deals with generating and manipulating visual content using
computers.
This encompasses a wide range of applications, including: 2D & 3D Graphics, Rendering, Animation, Visualization,
User Interfaces etc.
Advantages of Computer Graphics
Enhanced Visualization: They provide clear and detailed visual representations of complex data, making it easier
to understand and interpret information.
Realism and Engagement: High-quality graphics create immersive experiences in entertainment, education, and
training, capturing audience attention and enhancing engagement.
Speed and Efficiency: Graphics software allows for rapid design iterations and modifications, significantly
speeding up the creative process compared to traditional methods.
Cost-Effective Prototyping: Creating digital prototypes reduces material costs and time, allowing for better testing
and refinement before physical production.
Flexibility: Computer graphics can easily be modified and adapted for various applications, making them versatile
for different industries and purposes.
Accessibility: Digital graphics can be shared easily across platforms, reaching wider audiences and facilitating
collaboration regardless of geographical locations.
Improved Communication: Visual aids can transcend language barriers, conveying ideas and concepts more
effectively than text alone.
Interactivity: Many computer graphics applications allow for user interaction, enhancing user experiences in fields
like gaming and web design.
Opportunities computer graphics offers
Computer graphics offers a diverse range of opportunities across various sectors. Here are some key areas:
1.
2.
3.
4.
5.
6.
Video Game Development: Roles in game design, 3D modeling, animation, and visual effects for interactive
entertainment.
Film and Animation: Opportunities in animation studios for creating visual effects, character design, and
3D animation in movies and television.
Graphic Design: Careers in branding, advertising, and marketing, focusing on creating visuals for print and
digital media.
Web Design: Designing user interfaces (UI) and user experiences (UX) for websites and applications,
integrating graphics to enhance usability.
Architectural Visualization: Creating 3D renderings and walkthroughs for architects and real estate
developers to visualize projects before construction.
Medical Visualization: Working with medical imaging technologies to create visual representations of
complex biological data for diagnosis and research.
7.
Data Visualization: Designing visual representations of data for businesses and organizations, making
complex information more understandable.
8. Virtual Reality (VR) and Augmented Reality (AR): Developing immersive experiences and applications
that blend digital graphics with the real world.
9. Education and Training: Creating educational content and simulations that use graphics to enhance learning
experiences in various fields.
10. Research and Development: Engaging in innovative projects in computer graphics technology, exploring
new rendering techniques, graphics hardware, and software applications.
Definition of some Terms
1) Multimedia
Multimedia refers to the integration of various content forms, including text, audio, images, animations, video, and
interactivity, to create engaging and dynamic presentations. It combines different media types to enhance
communication and information delivery.
Types of Media:
•
•
•
•
•
Text: Written content that provides information or context.
Audio: Sound elements, including music, voiceovers, and sound effects.
Images: Static visuals like photographs, illustrations, and infographics.
Animation: Moving images that can illustrate concepts or tell stories.
Video: Recorded moving images that convey information or entertainment
2) Multimedia Application
A multimedia application is a software program that combines multiple forms of media—such as text, audio, images,
video, and animation—to create interactive and engaging content. These applications can be used for various purposes
across different fields. E.g Video Editing Software, Social Media Platforms: Apps like Instagram and TikTok etc
3) Multimedia System
A multimedia system is a computing system that integrates multiple forms of media—such as text, audio, images,
video, and animations—to create, store, process, and present information. These systems are designed to manage and
deliver rich content in an interactive manner.
Key components of a multimedia system
Hardware:
•
•
•
Computers or Servers: The central processing units that run multimedia applications.
Input Devices: Such as microphones, cameras, and scanners for capturing audio, video, and images.
Output Devices: Monitors, speakers, and printers for displaying and playing back multimedia content.
Software:
•
•
•
Multimedia Authoring Tools: Applications for creating multimedia content (e.g., Adobe Creative Suite).
Media Players: Software for playing back audio and video files (e.g., VLC Media Player).
Editing Software: Programs for editing text, images, audio, and video.
Content:
The actual multimedia elements, such as videos, audio files, graphics, and animations, that are used in the system.
Storage:
•
Systems for storing large multimedia files, including hard drives, cloud storage, and databases.
Networking:
•
Infrastructure for transmitting multimedia content over the internet or local networks, enabling streaming and
sharing.
Features of Multimedia System
Integration of Multiple Media Types: Multimedia systems combine text, audio, images, video, and animations
to provide a richer and more engaging experience.
Interactivity: Users can interact with the content, making choices that affect the outcome or the information
presented, enhancing user engagement.
Real-time Processing: Many multimedia applications require real-time processing to deliver smooth playback and
interaction, especially in gaming and video conferencing.
High Bandwidth Requirements: Multimedia content, particularly video and audio, often requires significant
bandwidth for transmission and storage, necessitating efficient network capabilities.
Synchronization: Effective multimedia systems synchronize various media elements (like audio and video) to
ensure they work together seamlessly.
Non-linear Navigation: Unlike traditional linear presentations (like a book), multimedia systems often allow users
to navigate content in a non-linear fashion, enabling personalized experiences.
Rich User Interfaces: They typically feature visually appealing and user-friendly interfaces that enhance the
overall experience and accessibility.
Storage and Compression: Multimedia systems utilize various storage techniques and compression algorithms to
manage large media files efficiently, ensuring quick access and minimal load times.
Multi-user Capability: Many multimedia systems support multiple users, allowing for collaborative projects or
shared experiences, especially in educational and gaming contexts.
Adaptability: They can be tailored for different platforms and devices, such as smartphones, tablets, and
computers, ensuring accessibility across various environments.
1) Digital Images
A digital image is a representation of a real image as a set of numbers that can be
stored and handled by a digital computer.
A digital image consists of a large number of small spots of color. These spots are
called pixels, a contraction of “picture elements.” When displayed on a monitor or printed
on paper, pixels are so small and so closely packed together that the collective effect on
the human eye is a continuous pattern of colors. However, if you look closely at magnified
computer displays, for example as produced by projectors, you can sometimes see
individual pixels as small rectangles.
* Bitmap or Raster Images: Raster images are the types of images you get from
a digital camera or the camera on your phone; the same goes for images pulled from
documents via scanner. The common image file types on the internet – jpgs, gifs,
and the like – are generally raster images. Any image that’s made up of pixels – the
tiny, individually-colored units that combine to make the whole image – is raster.
This is why these images blur when enlarged: A close view of the image reveals
the individual pixels that make the image and breaks the appearance of a smooth
transition across these pixels. This blurring is called pixelating.
Raster images are still the best form of image for photographs, especially regarding
editing functions from software like Photoshop, designed to work well with the
pixel-based format. It’s possible, of course, to print raster images at different sizes,
but the pixel count in the image needs to meet a minimum that correlates with the
size of the piece being printed.
* Vector Images: Vector images are different from raster images. First, they aren’t
made up of pixels. Vector images, which are generally made or translated through
illustration software, are actually made up of individual lines and shapes (called
objects) that combine to make a whole image, through mathematical calculations
that define the shapes and directions of the lines.
Vector art is not created through a camera; instead, it’s created through illustration
software like Adobe Illustrator and Corel Draw and commonly saved as .ai, .pdf,
.eps and .svg files.
Although vector images lack the complex details that you can create with photographs (raster
images), vector art shines in it’s own way: Recall the earlier reference to pixelation. You won’t
find any in a vector image, no matter what size the image is stretched or shrunk to.
This is why vector images are an excellent format for the creation of printed products such
as postcards, brochures, banners, and signs. Logos or illustrations as a vector image can be
stretched to any size for printing without losing image quality or ever seeing any blur or pixelation,
which means your graphics will look great regardless of whether they’re printed on small business
cards or jumbo-sized banners.
The Bottom Line
In short, raster images are best for detailed images like photographs and can include more varied
shades of colors and gradients as they utilize pixels. Raster images need to be printed in the
appropriate size and medium for the image.
Vector images don’t utilize pixels, work well for things like logos and illustrations, and can be
printed at any size. Vector images scale infinitely, which makes them the perfect format for
company logos that will be applied on various mediums.
Knowing the difference can help you understand the best choice of image for any printing project
your business takes on. Your U.S. Press account manager can help you with any further questions
or details regarding raster images, vector images, and how to use both to the best benefit of your
company.
Assignment
Bring out 5 Differences between Vector and Bitmap Images.
* Characteristics of Bitmap Images
•
•
•
•
Bitmap graphics are resolution dependent - they have a fixed resolution based on the
number of pixels they consist of.
Bitmap graphics handle complex, multi-coloured images well - This is why they're
typically used for digital photography and other detailed graphics.
Bitmap graphics can become pixelated when enlarged - Increasing the size of a bitmap
graphic can often result in a loss of quality.
Bitmap graphics have larger file sizes - Compared to vector graphics, bitmap graphics
tend to have larger file sizes as they store individual colours for each pixel.
* Image Compression
Image compression is a process applied to a graphics file to minimize its size in bytes without
degrading image quality below an acceptable threshold. By reducing the file size, more images
can be stored in a given amount of disk or memory space. The image also requires
less bandwidth when being transmitted over the internet or downloaded from a webpage,
reducing network congestion and speeding up content delivery.
Types of image compression
The methods used to compress image files typically fall into one of two categories: lossy and
lossless.
1) Lossy compression reduces an image file size by permanently removing less critical
information, particularly redundant data. Lossy compression can significantly reduce file size,
but it can also reduce image quality to the point of distortion, especially if the image is overly
compressed. However, quality can be maintained when compression is carefully applied.
One of the challenges with lossy compression is that it's irreversible. Once it has been applied to
an image, that image can never be restored to its original state. If lossy compression is applied
repeatedly to the same image, it gets increasingly distorted. That said, lossy compression has
proved to be a valuable strategy for the web, where a moderate amount of image degradation can
often be tolerated
Note: The most common example of lossy compression is JPEG, an image compression format
used extensively on the web and in digital photography. This widely recognized format is
supported by numerous tools and applications. Additionally, compression can be applied in
degrees, making it possible to use JPEG compression that best strikes a balance between file size
and quality.
2) Lossless Compression method applies compression without removing critical data or
reducing image quality and results in a compressed image that can be restored to its
original state with no degradation or distortion. However, lossless compression doesn't
reduce the file size nearly as much as lossy compression, offering little advantage in
terms of storage space, network bandwidth or download speeds. Lossless compression is
generally used in situations where image quality is more important than disk space or
network performance, such as for product images or to showcase artwork.
Note:
- One of the most common lossless formats is PNG, a widely used format that reduces file size
by identifying patterns and compressing those patterns together. Although PNG files are
generally larger than JPEG files, websites use them extensively when more image detail is
needed, such as for logos, icons, screenshots or images with text. Another familiar lossless
format is BMP, a proprietary approach to image compression introduced by Microsoft and used
primarily for Microsoft products, particularly Windows computers.
- GIF is compression format that falls into the lossless category, although there is some
confusion as to whether it is lossy or lossless. GIF images are limited to 256 colors, so
converting an image with more colors to GIF results in a loss of quality, which is sometimes
attributed to lossy compression. But the compression algorithms used by GIF are lossless. If
quality is lost, it's due to issues related to converting the file. Currently, the GIF format is used
primarily for simple videos and animations.
- A compression format that's making headway is Google's WebP, an image format developed
exclusively for the web. Unlike most compression techniques, WebP supports both lossless and
lossy compression, making it highly versatile. WebP images generally use less disk space than
other formats but offer comparable quality. Most major browsers support WebP images.
Compression can also be used for non-image file types such as text or program files, but its use
tends to be limited to lossless compression. In text and program files, it is crucial that
compression be lossless because a single error can damage the meaning of a text file or cause a
program not to run. The zip file format is an example of lossless compression commonly used
for text files or even entire directories of files.
In image compression, a small loss in quality is usually not noticeable. In situations where there
is some tolerance for loss, greater compression can be applied to files than when there is no
tolerance for loss. For this reason, graphic images can usually be compressed to a much greater
degree than text or program files.
Final Improvements of Images
There's a reason it's said that a picture is worth a thousand words. And when it comes to building
your brand, images are a great way to communicate clearly within your audience. Hence the
reason, it is good to have a good Image Resolution.
How to Improve Images Resolution
1) Cropping: Enhance focus and composition
Cropping can remove unwanted or distracting areas, improve the composition, or help emphasize
a focal point.
N.B: If you plan on doing significant cropping, try to start with a large image as possible. Because
the more you crop, the more you reduce your image quality and resolution.
When cropping, it is adviced to follow what’s known as the rule of thirds. This technique imagines
that your photo is divided into thirds, both vertically and horizontally, with four lines (two vertical,
two horizontal).
The four points where those lines would intersect form guidelines to place your focal point, or the
most important area of your image.
2) Blurring: Enhance background images
It is used for smoothing or reducing noise in images and uncluttered background.
While it is intentionally applied in some cases, it can also occur as a result of image compression or
other processing techniques, potentially affecting the clarity and accuracy of computer vision tasks.
Note: With some light to moderate blurring, you can retain recognizable shapes or scenes in your
background photos.
3) Saturation: Enhance or reduce color intensity
Saturation has to do with color intensity, so more saturated colors are bolder and brighter (closer
to their purest form) while less saturated colors are more faded (closer to gray). Complete
desaturation leaves you with a black-and-white photo
Note: While any photo editing technique can be overdone, too much saturation can look
particularly strange and unnatural. So, unless you’re after a particular effect, be careful to go easy
on the saturation, or else you’ll end up with an image that has an almost glowing, neon look.
4) Contrast: Enhance highlights and shadows
Increasing contrast is a good way to make your image pop (wow) and add a little drama.
Upping the contrast produces lighter lights, darker darks, and a wider range of tones in between—
making your image look just a little better than what you’d see in real life.
On the other hand, reducing contrast can give an image a more flat, even tone.
As with saturation, too much contrast is usually not a good thing. Highlights can get blown out
(too bright) and shadows can get too dark, which means you lose detail and dimension in those
areas.
5) Brightness: Enhance overall lighting
The brightness will make everything brighter (or darker). This should be done incrementally and
with attention to not creating areas that are overly bright, and you can often improve and balance
your results by adjusting the contrast at the same time.
6) Filters: Enhance and correct photos
Thanks to Instagram and other apps, filters have become a popular photo editing option. They can
be used to add special or artistic effects to your images, but they can also serve as a shortcut to
correct issues in your image.
However, most images shouldn’t need filters. Try to only apply one if you’re trying to achieve a
specific effect.
7) Arrangement: Enhance your layout with grids
A clean and organized layout is an asset to any design project. If you’re not sure where to start,
aligning all your design elements to a grid is always a safe bet. Grids are also a nice way to
showcase multiple images
8) Frames: Enhance image shape and style
Like actual picture frames, you hang on a wall, framing images in design is traditionally used to
draw attention to the image. Frames can be simple or decorative, a single line or an illustrated
design, colored or not, depending on the style and mood of your project.
9) Layering: Enhance images with screens or overlays
Screens (also called overlays) are a common approach. They are transparent blocks of color that
sit on top of your image. Adjusting the opacity of the screen determines how well your image will
show through.
10) Text: Enhance images with typography on top
Images and text are two foundations of graphic design (opens in a new tab or window). Combine
them, and they’re a powerhouse of visual communication. So, try adding relevant information to
an image by placing text on top (or picking an image that complements the subject of your text).
11) Auto Enhance: Automatically correct and enhance photos
Save time and effort with the Auto Enhance effect. With a single tap, easily adjust the
enhancement level. From correcting a dark image and making a picture clearer to fixing a photo’s
brightness and adjusting to your desired color aesthetic, the Auto Enhance effect lets you make
these changes quickly.
12) Photo Effects: Magically enhance photos
2) SOUND
Sound is a pressure wave which is created by a vibrating object.
Sound is created when something vibrates and sends waves of energy (vibration) into our ears.
Audible noise (sound) is generated by physically vibrating material. The vibration produces
pressure wave in the air, the pressure wave travel through the air, and ultimately cause our
eardrums to vibrate. The vibration of our ear drums is converted into electrical pulses sent to the
brain, which perceives the sound.
The five main characteristics of sound waves include wavelength, amplitude, frequency, time
period and velocity.
• Wavelength: The most important characteristic of sound waves may be the wavelength.
Sound consists of a longitudinal wave that includes compressions and rarefactions as they
travel through a given medium. The distance that one wave travels before it repeats itself
is the wavelength. It is the combined length of a compression and the adjacent rarefaction,
or the distance between the centers of two consecutive rarefactions or compressions.
•
Amplitude: The amplitude is the size of a given wave. Think of it as sort of like the wave’s
height as opposed to its length. The amplitude is more accurately defined as the maximum
displacement of the particles the sound wave disturbs as it passes through a medium.
•
Frequency: The frequency of sound refers to the number of sound waves a sound produces
per second. A low-frequency sound has fewer waves, while a high-frequency sound has
more. Sound frequency is measured in hertz (HZ) and is not dependent upon the medium
the sound is passing through.
•
Time Period – The time period is almost the opposite of the frequency. It is the time
required to produce a single complete wave, or cycle. Each vibration of the vibrating body
producing the sound is equal to a wave.
•
Velocity – finally, the velocity of the wave, sometimes referred to as the speed, is the
amount of distance in meters per second that a wave travels in one second.
DIGITALISATION OF SOUND
Note: If we want to use a digital version of sound wave, we must form digitized representations
of audio information.
Digitization is the process of representing various types of information in a form that can be stored
and processed by a digital device. It is the combined operations of sampling and quantization, also
called analog-todigital (A/D) conversion.
Unlike analog storage media such as magnetic tape or vinyl records, computers store audio
information digitally as a series of zeroes and ones. In digital storage, the original waveform is
broken up into individual snapshots called samples. This process is typically known
as digitizing or sampling the audio, but it is sometimes called analog-to-digital conversion.
When you record from a microphone into a computer, for example, analog-to-digital converters
transform the analog signal into digital samples that computers can store and process.
Steps Of Digitization Of Sound.
•
Sampling - Audio sampling is the process of transforming a musical source into a digital
file. Digital audio recording does this by taking samples of the audio source along the
soundwaves at regular intervals. The more samples you take - known as the ‘sample rate’
- the more closely the final digital file will resemble the original. A higher sample rate
tends to deliver a better-quality audio reproduction.
•
Quantization - quantization is the studio-software process of transforming
performed musical notes, which may have some imprecision due to expressive
performance, to an underlying musical representation that eliminates the imprecision. The
process results in notes being set on beats and on exact fractions of beats
The purpose of quantization in music processing is to provide a more beat-accurate timing
of sounds
•
Encoding - Encoding is the process of changing digital audio from one format to another.
There are two broad types of formats - uncompressed and compressed. Uncompressed
audio is mainly found in the PCM format of audio CDs. Generally, audio encoding means
going from uncompressed PCM, to some kind of compressed audio format.
Monophonic and Stereophonic Sound
Mono (monophonic) sound is single-channel audio where all the instruments are mixed into one
signal, intended to be heard as if emanating from one position.
Stereo (stereophonic) sound is achieved by using two audio channels feeding 2 separate speakers.
This creates a more three-dimensional sound, and better resembles how we hear things in the world
at large.
What is Stereo sound?
Stereo is sound recorded with two microphones and or played back through two channels (a left
and a right that are outputted to two (usually separate) speakers, and could be anything from laptop
speaker etc)
All our listening systems are configured in stereo; our phones, laptop speakers, headphones,
the sound in films, the sound in video games, the PA systems in clubs and festivals, and so on.
There two types of stereo Sound.
There is True Stereo and Pseudo Stereo.
True stereo is when a sound source like an instrument or dialogue is recorded with two
microphones.
Pseudo stereo or, simulated stereo is anything that plays through two channels but isn’t recorded
in true stereo. For example, if you record a guitar with one microphone, the guitar will be in mono
but played back in stereo, because all audio playback systems play sound through two channels.
Stereo mixing allows for precise placement and separation of different elements in the mix. By
panning (separating) instruments across the stereo field, you make it easier for listeners to
distinguish individual elements within the music.
Mixing in stereo opens up a world of creative possibilities. Moving sounds across the stereo field,
or sculpting stereo sounds can create unique sonic landscapes making your music more engaging
and dramatic.
What is Mono Sound?
Mono sound is any sound – in most cases, music, that is recorded and or played back
using one audio channel (all the instruments are mixed into one signal, intended to be heard as if
emanating from one position).
For example, one microphone recording a guitar is a mono recording, because you’re
using one channel (with one mic) to pick up the sound of the guitar.
Even if you record multiple instruments, using one microphone, it is still a mono recording because
everything is being captured and played back through one channel.
It doesn’t matter how many sounds or instruments there are in the recording. If it is all sent
through one channel, it is mono.
Mono sound remains most commonly used for voice recordings because of its advantages in
reducing background noise and capturing a voice in the clearest, most natural way possible.
ADVANTAGES AND DISADVANTAGES OF MONO VS STEREO
SOUND
PROS AND CONS OF MONO SOUND
Pros:
Clarity
And
Focus:
Mono sound offers a clear and focused listening experience, making it ideal for speech and singleinstrument recordings.
Compatibility:
Works well on all speakers and systems, ensuring consistent playback.
Simplicity
In
Mixing:
Mixing in mono is more straightforward, as it doesn’t involve balancing multiple channels in a
stereo field.
Cons:
Lack
Of
Depth:
Mono sound does not provide stereo sound’s spatial depth and immersive experience.
Limited
Creative
Options:
Offers fewer opportunities for creative sound placement and effects.
PROS AND CONS OF STEREO SOUND
Pros:
Immersive
Experience:
Stereo provides a more realistic and engaging listening experience, with sounds appearing to come
from different directions.
Creative
Flexibility:
Allows for innovative use of the sound space, enhancing the artistic expression in music.
Depth
And
Realism:
Adds depth and realism to recordings, making them feel more lifelike.
Cons:
Complexity:
Requires more careful mixing and mastering to maintain balance and prevent phase issues.
Inconsistency
Across
Playback
Systems:
Stereo sound can vary greatly depending on the listening environment and equipment.
Differences between Mono vs. Stereo sound
1) Cost
Mono sound is easy and inexpensive, whereas stereo sound is more expensive for recording and
reproduction.
2) Recording
Mono sound is easy to record, requiring only basic equipment while working on production,
whereas stereo sound requires technical knowledge and audio skills to record, apart from recording
equipment. The recording artist needs to know the relative position of the objects and events.
3) Usage
Mono sound is generally used for alert systems, radio talk shows, hearing aids, public address
systems, telephone, mobile communication, etc. In addition, movies, television, music players, FM
radio stations, podcasts, and other apps like Spotify, Gaana use stereo sound.
4) Playback
Mono has only one channel of playback while stereo has two channels of playback.
Note.
Even though there are many benefits and user experience advantages of stereo sound over the
mono sound. The importance of mono sound is not reduced. Some of the important vocals and
sounds are still recorded in mono sound.
Mono VS Stereo Sound: Which One is Better
There is no definitive answer about which one is better. It all depends on our own preferences and
situations. For example, if you have a home theater system in your house, then you may need to
use stereo sound to get a better experience with even more channels. The stereo sound is ideal for
watching a movie and if there are lots of music and environmental sounds, as it allows you to
immerse yourself in the story.
If you only wear one earbud at a time, you would better choose mono sound. Then you will hear
the entire song and all its parts.
When converting a signal to a sound, you only need to use only one channel. When you use
multiple channels to convert multiple signals to sounds, then it will produce stereo sound. Your
preference for mono or stereo sound is entirely subjective, as everyone, like sound, is unique!
SIZE OF A SOUND FILE
Music files for standard, everyday listeners are typically between 96 kbps and 160 kbps on
streaming platforms like Spotify. But the highest quality MP3 can be 256 kbps to 320 kbps, and
CDs are even better, at 1411 kbps. These sizes are for final products; in studios and editing rooms,
raw file sizes are much larger.
The five major that impact both audio quality and audio file size include:
1. Sample Rate: Measured in Hertz, this is a frequency that represents the number of
“snapshots” per second taken from a sound. For example, when you listen to a song, you’re
not hearing exactly what was recorded in a studio. Instead, the song is made up of snapshots
2.
3.
4.
5.
of the sound (the frame rate for video files). Faster frequencies are more detailed, but also
result in larger audio or music files.
Bitrate: How much sound data is recorded in each sample, and then transferred into audio.
Again, a higher bitrate means better audio quality – and a larger file size.
Compression: Lossless and lossy compression are the two types of audio compression,
and these impact audio file sizes the most. “Generally, you should go with lossless
compression when you want to store a nearly perfect copy of the source material and lossy
compression when the imperfect copy is good enough for day-to-day usage.” The less
compression used, the better playback you’ll get.
File Format: Audio files come in a variety of types, each of which provides different
performance specs that influence the file size. File formats include:
o AIFF/AIF – Audio Interchange File Format, used for storing sound data on Apple
PCs and devices
o RF64 – For multichannel sound in broadcasting and audio archiving
o W64 – Wave64, supported on Sony editing programs
o WAV – Waveform Audio File Format, for storing audio bitstream on PCs,
particularly Windows
o WMA – Windows Media Audio
Channels: Audio files are played through channels. The most common are mono (one) and
stereo (two), but the number of channels can grow to include all recorded waveforms in an
audio file. For instance, MP3 files are typically stereo, with a left and right channel. Each
channel adds more data to a file size, with mono typically being the smallest, stereo being
roughly twice as large, and so on.
3) VIDEOS
Definition:
Videos are an electronic form of capturing moving media for storage and retrieval purposes.
Similar to images, we obtain videos using image-acquisition devices such as video cameras,
smartphones, and camcorders. So, we can divide videos into two categories: analog and digital.
1) Analog Video
Analog video signals are in the form of waves (like a radio wave).
Initially, we videos are aquired as analog signals and stored in mediums such as magnetic
tapes. A sensitive plate captures a scene at an instance, and electrodes read line by line
from left to right. So, we call “frame” a single reading from top to bottom of a
photosensitive plate by an electrode. Consequently, a complete video consists of several
frames displayed sequentially at a standard speed.
2) Digital Video
Digital video signals are in the form of 1's and 0's (it's either on or it's off).
Digital video is capable of higher quality and, eventually, a much lower cost than earlier
analog technology.
The digitization of videos is basically the digitization of frames. After digitization, we
can visualize each frame as a separate image. Besides, the digitalization process converts
videos to a form directly readable by computers. Digital videos allow the manipulation of
several video parameters, such as the frame rate, depth, size, and resolution.
Initially, the semiconductor-based sensors record the frames that make up digital movies.
A frame structure is a matrix of elements holding pixel values. The number of rows and
columns indicates the frame size.
Characteristics of Videos
1) Frame Rate
The frame rate specifies the speed of the video. For example, the rate of 20
frames/second indicates reading and displaying 20 frames each second.
2) Color depth
Color depth defines how visually appealing the video looks. Bits per pixel indicate
the number of colors possibly a pixel can display. Note that the color depth can change
according to the devices used for capturing and displaying videos.
3) Compression Method
The compression method used and the number of pixels utilized to represent the frames
define the quality of the videos.
4) Aspect Ratio
Aspect ratio describes the proportional relationship between the width and height of
video screens and video picture elements. All popular video formats are rectangular,
and this can be described by a ratio between width and height.
5) Color Model
The color model uses the video color representation and maps encoded color values to
visible colors reproduced by the system.
6) Video Quality
Video quality can be measured with formal metrics like peak signal-to-noise
ratio (PSNR) or through subjective video quality assessment using expert observation.
Storage Meduims
Digitalizing videos enabled several mediums to store them.
Among the most commonly used storage mediums are tapes, floppy disks, compact
disks (CDs), digital versatile disks (DVDs), and universal serial bus (USB) drives.
Compairing Analog and Digital Videos
Analog
Digital
Signal
Analog signal is a continuous signal
which represents physical
measurements.
Digital signals are discrete time signals
generated by digital modulation.
Waves
Denoted by sine waves
Denoted by square waves
Representation
Uses continuous range of values to
represent information
Uses discrete or discontinuous values to
represent information
Example
Human voice in
electronic devices.
analog
Computers, CDs, DVDs, and other digital
electronic devices.
Technology
Analog
technology
waveforms as they are.
records
Samples analog waveforms into a limited set of
numbers and records them.
Data
transmissions
Subjected to deterioration by noise
during transmission and write/read
cycle.
Can be noise-immune without deterioration
during transmission and write/read cycle.
Response
Noise
to More likely to get affected reducing
accuracy
Less affected since noise response are analog in
nature
Flexibility
Analog hardware is not flexible.
Digital hardware is flexible in implementation.
Uses
Can be used in analog devices only.
Best suited for audio and video
transmission.
Best suited
electronics.
Applications
Thermometer
PCs, PDAs
Bandwidth
Analog signal processing can be
done in real time and consumes less
bandwidth.
There is no guarantee that digital signal
processing can be done in real time and
consumes more bandwidth to carry out the
same information.
Memory
Stored in the form of wave signal
Stored in the form of binary bit
Power
Analog
power
Digital instrument drawS only negligible power
Cost
Low cost and portable
instrument
air,
draws
large
for
Computing
and
Cost is high and not easily portable
digital
Analog
Digital
Impedance
Low
High order of 100 megaohm
Errors
Analog instruments usually have a
scale which is cramped at lower end
and give considerable observational
errors.
Digital instruments are free from observational
errors like parallax and approximation errors.
Video Formats
As a video typically consists of multiple images, it demands higher storage space. Due to the size
of a video, it’s difficult for us to retrieve it from the storage medium and share it. However, we
require space-efficient resources for real-time applications. For these reasons, we can use
many compression formats.
A video file occupies large memory spaces based on many factors, such as frame size and video
length. We can store videos in many file formats. Each of these formats employs a particular
compression method.
Some videos contain similar-looking frames. For example, videos captured with slow-moving
objects or cameras at a fixed position may have many identical adjacent frames. Thus each video
format or compression standards use efficient techniques to handle these identical frames.
•
•
•
•
•
•
Examples of Video formats are: MOV
MP4
AVI
WMF
FLV
WebM
Pros and cons of Video
Pros
1) Video content is calls attention and has high engagement
2) Video content increases conversion rates
3) Video is the powerhouse for multimedia
4) Video content is perfectly suitable for mobile devices
5) Video is the engine of social media
Cons
1) Video content can be expensive
2) Producing a video takes a lot of time
3) Video content might not suit all types of audience
4) It’s hard to update videos or keep them unfading
Video Compression
A video codec is an integral part of video processing. It encompasses an algorithm or software that
executes two primary tasks – compression and decompression of video files.
The compression phase occurs before a video file is transmitted or stored. During compression,
the codec minimally reduces the size of the video file, preserved in a digital format. This
compression is vital to make the data easier to store or transmit over the internet. The smaller the
file size, the less bandwidth is needed for data transfer.
Decompression, on the other hand, happens at the receiving end. The compressed video data is
expanded back into its original form, enabling regular playback. The decoding happens on your
device when you play a video on a media player or streaming service. The codec translates the
compressed data into an uncompressed video, allowing you to view it.
Download