[Verano 2016 - Ive - Coralys]
Arrays help us store and work with groups of data of the same type. The data is stored in consecutive memory spaces, which can be accessed by using the name of the array with indexes or subscripts that indicate the position where the data is stored. Repetition structures provide us a simple way of accessing the data within an array. In this laboratory experience, you will be exposed to simple sound processing algorithms in order to practice the use of loops to manipulate arrays.
This laboratory experience is an adaptation of the nifty assignment presented by Daniel Zingaro in [1].
Practice the use of loops to manipulate arrays.
Learn simple algorithms to process sound.
Practice modular programming.
Before coming to the laboratory you should have:
Reviewed the basic concepts related to arrays and loops.
Studied the left
and right
attributes of the QAudioBuffer::S16S
class in the Qt
multimedia library.
Studied the concepts and instructions for the laboratory session.
Taken the Pre-Lab quiz, available in Moodle.
Sounds are vibrations that propagate through elastic media such as air, water, and solids. The sound waves are generated by a sound source, such as the vibrating diaphragm of a stereo speaker [2]. Sound waves consist of areas of high and low pressure called compressions and rarefractions, respectively.
Microphones turn sound waves into electrical signals. These electrical signals can be digitized, i.e. converted to a stream of numbers, where each number is the intensity of the electrical signal at an instant in time. The sample rate is the number of samples of a sound signal taken per second. For example, a sample rate of 44,100 samples per second is used in CD-quality recordings. This means that every second, for each of the channels (left and right), 44,100 samples of the audio signal are taken and converted to numbers.
Figure 1: Illustration of the steps involved in sound digitalization. The pressure wave is converted to a voltage signal by the microphone. The voltage signal is sampled and digitized by the analog to digital converter to obtain a number value for each sample. The stream of numbers constitutes the digitized sound. Taken from [3].
Question:
How many bytes would it take to store a song that is exactly 180 seconds long and is recorded in stereo CD-quality? Assume that each sample is converted to a number of 2 bytes.
Answer:
180 seconds * 44,100 samples/second * 2 bytes/sample * 2 channels = 31,752,000 bytes = 31.75 MBytes.
Fortunately, there is sound data compression techniques such as MP3 and Ogg, that reduce the amount of memory required to store CD-quality music.
Digital sound processing techniques can be used to enhance sound quality by removing noise and echo, to perform data compression, and to improve transmission. Digital sound processing also plays an important role in voice recognition applications and in scientific research such as in biodiversity recognition using sound sensors [4]. Digital sound can also be easily manipulated to produce special effects.
Since digital sound recordings are in essence, a collection of numeric values that represent a sound wave, digital sound processing can be as simple as applying arithmetic operations over those values. For example, let’s say that you are given a digital sound recording. The louder the recording, the higher the absolute values of the numbers that it contains. To decrease the volume of the whole recording we could multiply each value by a positive number smaller than 1.
Figure 2. One of the simplest sound processing tasks: changing the volume of a sound wave by multiplying each point by a positive number smaller than 1 (in this case 0.5).
For this laboratory experience you will use the multimedia libraries of Qt
. To complete the exercises, you will need to understand the left
and right
members of the QAudioBuffer::S16S
class. For the purpose of this laboratory experience, we use the name AudioBuffer
to refer to QAudioBuffer::S16S
.
Each object of the class AudioBuffer
will have the variable members left
and right
that contain the left and right values of the stereo sound sample. These variables are public and you can access their content by writing the name of the object, followed by a period and the name of the variable. To represent a sound signal, we use an array of AudioBuffer
objects. Each element in the array is an object that contains the left and right values of the signal at an instant in time (remember that each second contains 44,100 samples). For instance, if we have an array of AudioBuffer
objects, called frames
, then frames[i].left
refers to the left channel value of the sound at sample i
.
Figure 3. In the figure, frame
is an array of AudioBuffer
objects. During this laboratory experience, sound signals will be represented by an array of AudioBuffer
objects. An object with index i
stores the values of the left and right channels of sample i
.
The HalfVolume
function in the following example illustrates how to read and modify an array of AudioBuffer
objects:
void HalfVolume(AudioBuffer frames[], int N){
// for each sample in the signal, reduce its value to half
for (int i=0; i < N; i++) {
frames[i].left = frames[i].left / 2;
frames[i].right = frames[i].right / 2;
}
}
!INCLUDE “../../eip-diagnostic/sound-processing/en/diag-sound-processing-01.html”
!INCLUDE “../../eip-diagnostic/sound-processing/en/diag-sound-processing-02.html”
!INCLUDE “../../eip-diagnostic/sound-processing/en/diag-sound-processing-03.html”
The SoundProcessing
project contains the skeleton of an application to process stereo sound. The application you will complete will allow the user to apply four different algorithms to process sound. The sub-directory called WaveSamples
contains sound files for you to test your implementations.
A cheap (but many times ineffective) way to remove the vocals from a recording is by taking advantage of the fact that voice is commonly recorded in both left and right channels, while the rest of the instruments may not. If this is the case, then we can remove vocals from a recording by subtracting the left and right channels.
Load the project SoundProcessing
into QtCreator
. There are two ways to do this:
SoundProcessing.pro
located in the folder /home/eip/labs/arrays-soundprocessing
of your virtual machine.Bitbucket
: Use a terminal and write the command git clone http:/bitbucket.org/eip-uprrp/arrays-soundprocessing
to download the folder arrays-soundprocessing
from Bitbucket
. Double click the file SoundProcessing.pro
located in the folder that you downloaded to your computer.Compile and run the program. You will see a graphical interface to process sound and recordings.
Load any of the wave files love.wav
, cartoon.wav
, or grace.wav
by clicking the Search
button in the right side of the Audio In
label, and play it clicking the Play Audio In
button.
In this exercise, your task is to complete the function RemoveVocals
in file audiomanip.cpp
so it can remove voices from a recording. The function receives an array of objects of the class AudioBuffer
, and the size of the array.
Algorithm:
For each sample in the array, compute the difference of the sample’s left channel minus its right channel, divide it by 2, and use this value as the new value for both the left and right channels of the corresponding sample.
Play the output sound file with the application by clicking the Play Audio Out
button.
A common sound effect is the gradual intensification of the recording’s volume, or fade in. This is the result of constantly increasing the value of consecutive samples in the array of sound samples.
Load and play any of the wave files rain.wav
, or water.wav
just as in Exercise 1.
Your task is to complete the function AudioFadeIn
in file audiomanip.cpp
so it gradually intensifies the volume of a recording up to a certain moment. The function receives an array of objects of the class AudioBuffer
, the size of the array, and a fade in length that will be applied to the AudioBuffer
. For example, if fade_length
is 88200
, the fade in should not affect any sample in position 88200
or higher.
Reproduce the following recordings from the WaveSamples
folder:
rain-fi.wav
water-fi.wav
The recordings were created using the fade in filter with fade_length
set to 88200
. You should be able to listen how the water and the rain linearly fades in over the first two seconds, and then remains at the same volume throughout the recording. Notice that, since we are using sounds recorded at 44100
samples per second, 88200
corresponds to two seconds of the recording.
Algorithm:
To apply a fade in to a sound, we multiply successive samples by constantly increasing fractional numbers between 0
and 1
. Multiplying samples by 0
silences them, and multiplying by 1
keeps them the same; multiplying by a factor between 0
and 1
scales their volume by that factor. Both channels of the samples should be multiplied by the same factor.
For instance, if fade_length
is 4, the filter will be applied to the first 4 samples:
Sample Number | Multiply by factor |
---|---|
0 | 0 |
1 | 0.25 |
2 | 0.5 |
3 | 0.75 |
>= 4 | 1 (Do not modify the sample) |
Notice that we have 4 samples and the factor used to multiply the sample in each channel stars at 0
and increases 0.25
each time until reaching 1
.
Another common sound effect is the gradual decrease of volume in a recording. This is the result of constantly decreasing the value of consecutive samples in the array of sound samples.
Load and play any of the wave files rain.wav
, or water.wav
just like in the previous exercises.
Your task in this exercise is to complete the function AudioFadeOut
in the file audiomanip.cpp
so it will fade out the volume starting from a sample up to the end of the recording. The function receives an array of objects of the class AudioBuffer
, the size of the array, and a fade out length that will be applied to the AudioBuffer
. For example, if fade_length
is 88200
, the fade-out should not affect any sample numbered 88200
or lower.
Reproduce the following recordings from the WaveSamples
folder:
rain.fo.wav
water.fo.wav
The recordings were created using the fade out filter with fade_length
set to 88200
. You should be able to listen how the water and the rain is played at maximum volume and then in the last two seconds the sound starts to linearly fade out.
Algorithm:
The multiplicative factors for fade_out
are the same as for fade_in
, but are applied in the reverse order. For example, if fade_length
were 4
, the samples in the fourth-before-last positions would be multiplied by 0.75
(in both channels), the samples in the third-before-last positions would be multiplied by 0.5
, the samples in the penultimate positions would be multiplied by 0.25
, the samples in the final positions would be multiplied by 0.0
.
The sound effect we want to produce in this exercise is to start hearing sound from the left channel, then fading from that channel, intensifying in the right channel, and ending up completely on the right channel.
Load and play the airplane.wav
just like in the previous exercises.
Your task is to complete the function LeftToRight
in file audiomanip.cpp
so the sound “moves” from the left channel to the right channel. The function receives an array of objects of class AudioBuffer
, the size of the array, and a pan length that will be applied to the AudioBuffer
. For example, if pan_length
is 88200
, the pan should not affect any sample in position 88200
or higher.
Play the airplane.out.wav
recording. You should be able to listen how the airplane sound starts completely at the left, then slowly moves to the right, reaching the extreme right by the final sample.
Algorithm:
Getting a sound to move from left to right like this requires a fade-out on the left channel and a fade-in on the right channel. For instance, if pan_length
is 4
, the filter will be applied to the first 4 samples:
Sample Number | Multiply left channel by factor | Multiply right channel by factor |
---|---|---|
0 | 0.75 | 0 |
1 | 0.5 | 0.25 |
2 | 0.25 | 0.5 |
3 | 0 | 0.75 |
>= 4 | (Do not modify the sample) | (Do not modify the sample) |
Use “Deliverable” in Moodle to upload the audiomanip.cpp
file. Remember to use good programming techniques, include the names of the programmers involved, and document your program.
[1] Daniel Zingaro, http://nifty.stanford.edu/2012/zingaro-stereo-sound-processing/
[2] http://en.wikipedia.org/wiki/Sound
[3] http://homepages.udayton.edu/~hardierc/ece203/sound_files/image001.jpg.
[4] Arbimon, A web based network for storing, sharing, and analyzing acoustic information. http://arbimon.com/
[5] https://somnathbanik.wordpress.com/2012/10/22/digital-signal-processing-featured-project/
[7] http://diveintodotnet.com/2014/12/02/programming-basics-what-are-strings/