Assignment 15.1.1

Non-Linear Editing

 Non-linear editing is editing film and video using software programs such Final Cut Pro, Adobe Premier to arrange static and time-based media assets into a presentation that can tell a story or communicate a message.  The media assets are created at different times and locations. The editor uses these pieces, which are stored in the computer in a project folder, to build a narrative that has a beginning, middle, and end.

The editing programs today use many of the tools of the early editing techniques in a digital format such as the razor blade tool. The program’s razor blade tool digitally cuts the digital image in a manner similar to the actual razor blade tool would slice film during the process of machine-to-machine editing.

Assignment 15.1.2

The four parts of an edited sequence are:

  1. Scripted action and dialog when the video has a narrative progress.
  2. Unscripted action and dialog when the video is a piece of journalism or a documentary.
  3. Titles and graphics.
  4. Music and sound effects.

Scripted action and dialog involves a performer reciting a written dialog off-book or committed to memory. Performing does not go smoothly, so multiple attempts or takes are made. These takes are assembled in the editing room before the performance is presented to the audience. Sometimes promptors are used, such as during an awards show such as the Oscars or Grammys. A prompter is similar to a television except
a devised script would scroll down the screen.The performer would then rehearse using the device reciting the dialog as it moves down the screen. When the show is filmed the performer recites the dialog. This prevents the performer from forgetting or omitting critical parts of the dialog.

Unscripted action and dialog are created during a documentary or journalistic interview. Involving a non-performer, information is recorded and a storyline created around the interview. Another type of unscripted dialog is the voiceover (VO), in which, an individual would comment on a series of images without being seen. Reporters and narrators use this
in order to be unobserved during a critical point of the video. B-Roll are additional shots provided when actual images cannot be used, such when the crime occurs or the building where a private celebrity party is held.

Titles and graphics are stationary fonts and type that add to the storyline. Sometimes introductory when used at the beginning of a video such as title credits or at the end of the film in ending credits, they add additional information to the images and story being presented. They can be informative such as a graph or the name of the individual being interviewed. Graphics could be additional written information regarding a product for commercials or dates of historical events.

Music and sound effects are often prepared off-camera and recorded. These need to be added at precise moments on the video. To maintain quality of the sound, WAV and AIFF compression formats are used without regard to file size. MP3s , the consumer compression format, lose the fidelity of sound because of loss of data during the compression process and the data cannot be retrieved. The loss data creates a distortion that is noticeable.

Discussion 14.1.1

Redundancy is the amount of space wasted in the storage media recording pictorial information digitally. Redundancy occurs in two ways: temporal redundancy and spatial redundancy. When a video image is taken, the image is shot in frames per second. In this series of images, a pixel can remain in a fixed location for a period of time. This is temporal redundancy.

Spatial redundancy occurs within one frame. Pixels can be spread or compressed in an image so the space the image consumes in storage can be large if the pixels are spread out. Shrinking the pixels reduces the file size.

Assignment 14.1.2

The format most commonly used for uploading video to an online hosting site is MPEG-4-AVCHD. AVCHD works well with the flash memory card formats, Secured Digital (SD) and Compact Flash (CF) which are inexpensive types of memory. AVCHD has editing programs that do not need additional step of transcoding the formats to begin the editing process. The AVCHD standard makes it easier for video files to be recognized by editing software through the use of the .MTS (MPEG Transport Stream). The files names can be altered but the videos will remain intact. This allows consumer electronic devices such as prosumer cameras, gaming consoles, Blu-ray players, and other devices to read the coding and render the video and audio with fewer flaws that a “rookie” operator could make through mis-identifying the file formats.

The AVCHD format allows the operator to define the rates the video signal is recorded in the flash medium. The cameraperson can set image quality ranges from low to high bit rates. The higher the bit rate, the better quality of image and the lowest amount of compression. But the reduced compression means larger file sizes and creates a need for a large SD card size. Yet, large size flash cards are becoming less expensive allowing longer videos to be shot by a prosumer camera.

Assigment 14.1.1

The two methods of video compression are intraframe (I-frame) compression and interframe compression. Compression itself is a mathematically-based algorithm that allows the reduction of the file size of images while maintaining visual quality. The way compression works is similar to compression in photographic files and audio files. Video compression takes advantage of the physical characteristics of filming which is a vast number of individual shots called frames. Frames have similar characteristics to movie film which is actually individual photos taken by a camera at a high rate of speed except that the construction of frames are with computer pixels rather than a chemical emulsion exposed to light.

In compression, the software program relies on the principle of redundant information in multiple images. Redundancy occurs in two types: Spatial redundancy and temporal redundancy. If a pixel occurs in the same location for a number of images, the program notes it in the first image and removes it in subsequent images until the location ceases. The program is using temporal redundancy to compress the image. The second type of redundancy, spatial, the pixels are in the process of stretching, so the software prevents the pixels from expanding.

Video software uses either intraframe compression or interframe compression for reducing the file size. Intraframe works with the issue of spatial redundancy by reducing the expansion of the I-frame 10:1, maintaining the integrity of the data within the image. But the file is still large because the program doesn’t make use of temporal redundancy. Interframe compression addresses both issues of spatial and temporal redundancy.

A software program using interframe compression starts with creating I-frames like the intraframe program would create but it goes further by analyzing the frames to create the I-frame at fixed intervals of 15 frames. These frames are organized into a GOP or group of pictures. In the interframe method, the I-frame is renamed a keyframe, which acts as the reference frame, holding the color value of the pixel in place, similar to a sample in PCM.

 

Assignment 13.1.1

Polar Patterns

Omnidirectional

The omnidirectional microphone has a polar pattern that picks up sound from every direction. This polar pattern requires a quiet room with little noise. The omnidirectional is used for orchestra, theatre, and choir performances. The microphones would be hung from the rafters of the performance hall, in strategic locations, in order to capture the range of all the instruments and/or voices. The audience members are generally quiet during the performance, although if an audience member coughs or sneezes during a quiet passage, that will be picked up as well.

Bidirectional

The bidirectional microphone has a polar pattern that picks up sound from two directions, the front and the back of the microphone element. This pattern blocks sound waves entering from the sides of the microphone. The bidirectional microphone would be used for interviews so any extraneous noise from the audience is blocked and the microphone can be shared. The interviewer and interviewee would be able to hold their discussion without an additional set up.

Cardioid

The cardioid microphone has a “heart-shaped” polar pattern, in which only the sound waves coming from one direction is picked up by the microphone element. This type of microphone is used by speakers during a presentation and vocalists needing to isolate the sound of their voices.

Discussion 13.1.1

Dynamic Range

Dynamic range is the differences in loudness within the performance. A music composition has different loudness for artistic emphasis of emotion. If reading a music sheet, there would be written commentary, such as piano or “soft” or forte “loud”, by the composer telling the singer(s) or instrumentalists how to “sound” the piece. An audio engineer needs to take into consideration how to maintain the fidelity of music and the dynamic range, keeping both balanced for the aural effect the composer or presenter is creating.

Assignment 12.1.1

amplitude1

a) Amplitude

Amplitude is the total height of a sound wave. A sound wave is a pressure wave. A physical characteristic of amplitude is loudness, a logarithmic measurement call a decibel unit was assigned to amplitude to measure the amount of loudness or sound pressure. The lower the units of decibels (dBs), the quieter the sound. The louder the sound the more damage is done to human hearing. Human hearing becomes damaged at 140 decibels and hearing is permanently damaged at 150 decibels.

wavelength

b) Wave Length

Wave length is the distance of one sound vibration measured from crest to crest for one second.

Frequency2

c) Frequency

Frequency is the number of complete wave cycles that occur over a set period of time. Measured as Hertz per second, frequency has the physical characteristic as the pitch of sound which can range from low to high. To calibrate the audio in a production, the common frequency tone is 1 kHz or 1,000 Hertz. The movement of the the cycle has three phases: Compression, Rest, and Rarefaction.

Compression occurs as the wave of molecules moves upward, peaking at the crest and has the greatest amount of pressure. At the highest point, when the energy of compression can go no further, rest occurs. Rest only lasts for a micro-second because the energy of the pressure wave is beginning to shift directions due to elasticity of the pressure’s molecules. The molecules then begin to lose their energy and start pulling apart. This loss of energy and separation is the rarefaction phase and the soundwave descends to the lowest point, the trough.

Frequency plays an important part in audio production as this creates the aural impression. The perception of sound by humans is the human frequency spectrum. The normal range for hearing in this spectrum is between 20 Hz to 20,000 Hz. This spectrum’s range is broken into three bands, bass, midrange, and treble. These bands are used determine the audio output in sound systems using a 3-band equalizer.

Assignment 12.1.3

Coder is the short form of coder-decoder, a computer program/algorithm that is designed to encode and decode audio and video signals into sequences of binary data or raw digital bitstream. Nyquist’s Pulse or PCM is a type of codec which reduces a large file to a manageable file size for storage. MP3 is a well known codec used for digital audio storage and playback.

A container format is a file format the bundles a codec in order to keep the separate audio and video data together during processing for an end-user. The most common containers used are .wav for the Windows operating system and .aiff formats for Apple operating system. Both container systems are not structurally different because the coder or PCM audio stream doesn’t change once the container opens in a CD player. The container “talks” to the operating system to let it know what kind of data is inside and how to uncompress it.

Assignment 12.1.2

Sampling Rate

In the early days of audio recording, sound was recorded as an analog signal. An analog signal is a continuous electric impulse of varying amplitude. Stored as vibrations on wax or plastic materials, the signal would be picked up by an electrically charged phonograph needle as the hard surface spun. The audio signal had a one to one correlation that allowed a precise reproduction. When technology shifted from analog signaling to digital signals, the format for recording signal changed from a linear model to a mathematically discrete pulse code moderation or PCM.

PCM breaks down the continuous linear signal into “points” called samples. If I were to compare a sound sample, it would be similar to a geological core sample for oil. Sound samples are mathematically calculated at certain points along the analog signal. The sample takes a measurement of the sound and converts the sound into a binary string. This binary string is inputted into the computer digital system as a raw data file.

The PCM’s audio stream maintains its fidelity by using two properties, sampling rate and bit depth. Sampling rate is the number of samples taken each second. The bit depth is the number of bits in a waveform. The size of a sampling rate is twice the frequency which for human hearing is between 20 Hz and 20,000 Hz or 20 kHz. So the sampling rate at it’s highest for human hearing is 40 kilohertz. Any higher and the sound pressure will damage hearing.