MarketVideo
Company Profile

Video

Video is an electronic medium used for the recording, copying, playback, transmission, and display of moving visual images, with or without accompanying audio. Video technology was initially developed for live transmission and later expanded to include recording and storage through analog formats such as magnetic tape. Since the late 20th century, digital video has become the dominant form, enabling efficient compression, storage, editing, and distribution across broadcast television, physical media, and internet-based platforms. Advances in digital imaging, compression standards, and network infrastructure have significantly influenced media production, communication, entertainment, education, and information dissemination worldwide.

Etymology
The word video comes from the Latin video, "I see," the first-person singular present indicative of videre, "to see". ==History==
History
Analog video composite video signal (analog) Video developed from facsimile systems developed in the mid-19th century. Mechanical video scanners, such as the Nipkow disk, were patented as early as 1884, but it took several decades before practical video systems could be developed. Whereas the medium of film records using a sequence of miniature photographic images visible to the naked eye, video encodes images electronically, turning them into analog or digital electronic signals for transmission and recording. Video was originally exclusively live technology, and was first developed for mechanical television systems. These were quickly replaced by cathode-ray tube (CRT) television systems. Live video cameras used an electron beam, which would scan a photoconductive plate with the desired image and produce a voltage signal proportional to the brightness in each part of the image. The signal could then be sent to televisions, where another beam would receive and display the image. Charles Ginsburg led an Ampex research team to develop one of the first practical video tape recorders (VTR). In 1951, the first of these captured live images from television cameras by writing the camera's electrical signal onto magnetic videotape. VTRs sold for around US$50,000 in 1956, and videotapes cost US$300 per one-hour reel. However, prices gradually dropped over the years, and in 1971, Sony began selling videocassette recorder (VCR) decks and tapes into the consumer market. Digital video Digital video is capable of higher quality and, eventually, a much lower cost than its analog predecessor. After the commercial introduction of the DVD, in 1997, and later the Blu-ray Disc, in 2006, sales of videotape and recording equipment fell. Advances in computer technology allow even inexpensive personal computers and smartphones to capture, store, edit, and transmit digital video, further reducing the cost of video production and allowing programmers and broadcasters to move to tapeless production. The advent of digital broadcasting and the subsequent digital television transition are in the process of relegating analog video to the status of a legacy technology in most parts of the world. The development of high-resolution video cameras with improved dynamic range and broader color gamuts, along with the introduction of high-dynamic-range digital intermediate data formats with improved color depth, has caused digital video technology to converge with film technology. Since 2013, the use of digital cameras in Hollywood has surpassed the use of film cameras. ==Characteristics==
Characteristics
Frame rate Frame rate—the number of still pictures per unit of time—ranges from six or eight frames per second (frame/s or fps) for older mechanical cameras to 120 or more for new professional cameras. The PAL and SECAM standards specify 25 fps, while NTSC specifies 29.97 fps. Film is shot at a slower frame rate of 24 frames per second, which slightly complicates the process of transferring film to video. The minimum frame rate to achieve persistence of vision (the illusion of a moving image) is about 16 frames per second. Interlacing vs. progressive-scan systems Video can be interlaced or progressive. In progressive scan systems, each refresh period updates all scan lines in each frame, in sequence. When displaying a natively progressive broadcast or recorded signal, the result is the optimum spatial resolution of both the stationary and moving parts of the image. Interlacing was invented as a way to reduce flicker in early mechanical and CRT video displays, without increasing the number of complete frames per second. Interlacing retains detail while requiring lower bandwidth compared to progressive scanning. In interlaced video, the horizontal scan lines of each complete frame are treated as if numbered consecutively and captured as two fields: an odd field (upper field) consisting of the odd-numbered lines and an even field (lower field) consisting of the even-numbered lines. Analog display devices reproduce each frame, effectively doubling the frame rate as far as perceptible overall flicker is concerned. When the image capture device acquires the fields one at a time, rather than dividing up a complete frame after it is captured, the frame rate for motion is effectively doubled as well, resulting in smoother, more lifelike reproduction of rapidly moving parts of the image when viewed on an interlaced CRT display. When displaying a natively interlaced signal on a progressive scan device, the overall spatial resolution is degraded by simple line doubling—artifacts, such as flickering or comb effects in moving parts of the image, appear unless special signal processing eliminates them. A procedure known as deinterlacing can optimize the display of an interlaced video signal from an analog, DVD, or satellite source on a progressive scan device such as an LCD television, digital video projector, or plasma panel. Deinterlacing cannot, however, produce video quality that is equivalent to true progressive scan source material. Pixels on computer monitors are usually square, but pixels used in digital video often have non-square aspect ratios, such as those used in the PAL and NTSC variants of the CCIR 601 digital video standard and the corresponding anamorphic widescreen formats. The 720 by 480 pixel raster uses thin pixels on a 4:3 aspect ratio display and fat pixels on a 16:9 display. Color model and depth The color model uses the video color representation and maps encoded color values to visible colors reproduced by the system. There are several such representations in common use: typically, YIQ is used in NTSC television, YUV is used in PAL television, YDbDr is used by SECAM television, and YCbCr is used for digital video. The number of distinct colors a pixel can represent depends on the color depth expressed in the number of bits per pixel. A common way to reduce the amount of data required in digital video is by chroma subsampling (e.g., 4:4:4, 4:2:2, etc.). Because the human eye is less sensitive to details in color than brightness, the luminance data for all pixels is maintained, while the chrominance data is averaged for a number of pixels in a block, and the same value is used for all of them. For example, this results in a 50% reduction in chrominance data using 2-pixel blocks (4:2:2) or 75% using 4-pixel blocks (4:2:0). This process does not reduce the number of possible color values that can be displayed, but it reduces the number of distinct points at which the color changes. Stereoscopy Stereoscopic video for 3D film and other applications can be displayed using several different methods: • Two channels: a right channel for the right eye and a left channel for the left eye. Both channels may be viewed simultaneously by using light-polarizing filters 90 degrees off-axis from each other on two video projectors. These separately polarized channels are viewed wearing eyeglasses with matching polarization filters. • Anaglyph 3D, where one channel is overlaid with two color-coded layers. This left and right layer technique is occasionally used for network broadcasts or recent anaglyph releases of 3D movies on DVD. Simple red/cyan plastic glasses provide the means to view the images discretely to form a stereoscopic view of the content. • One channel with alternating left and right frames for the corresponding eye, using LCD shutter glasses that synchronize to the video to alternately block the image for each eye, so the appropriate eye sees the correct frame. This method is most common in computer virtual reality applications, such as in a Cave Automatic Virtual Environment, but reduces effective video framerate by a factor of two. ==Formats==
Formats{{anchor|Video formats}}
Different layers of video transmission and storage each provide their own set of formats to choose from. For transmission, there is a physical connector and signal protocol (see List of video connectors). A given physical link can carry certain display standards that specify a particular refresh rate, display resolution, and color space. Many analog and digital recording formats are in use, and digital video clips can also be stored on a computer file system as files, which have their own formats. In addition to the physical format used by the data storage device or transmission medium, the stream of ones and zeros that is sent must be in a particular digital video coding format, for which a number is available. Analog video Analog video is a video signal represented by one or more analog signals. Analog color video signals include luminance (Y) and chrominance (C). When combined into one channel, as is the case among others with NTSC, PAL, and SECAM, it is called composite video. Analog video may be carried in separate channels, as in two-channel S-Video (YC) and multi-channel component video formats. Analog video is used in both consumer and professional television production applications. Digital video Digital video signal formats have been adopted, including serial digital interface (SDI), Digital Visual Interface (DVI), High-Definition Multimedia Interface (HDMI) and DisplayPort Interface. BNC_connector_%28male%29.jpg|Serial digital interface (SDI) Dvi-cable.jpg|Digital Visual Interface (DVI) HDMI-Connector.jpg|HDMI Displayport-cable.jpg|DisplayPort ==Transport medium==
Transport medium
Video can be transmitted or transported in a variety of ways including wireless terrestrial television as an analog or digital signal, coaxial cable in a closed-circuit system as an analog signal. Broadcast or studio cameras use a single or dual coaxial cable system using serial digital interface (SDI). See List of video connectors for information about physical connectors and related signal standards. Video may be transported over networks and other shared digital communications links using, for instance, MPEG transport stream, SMPTE 2022 and SMPTE 2110. ==Display standards==
Display standards
Digital television Digital television broadcasts use the MPEG-2 and other video coding formats and include: • ATSC – United States, Canada, Mexico, KoreaDigital Video Broadcasting (DVB) – EuropeISDBJapanISDB-Tb – uses the MPEG-4 video coding format – Brazil, ArgentinaDigital multimedia broadcasting (DMB) – Korea Analog television Analog television broadcast standards include: • Field-sequential color system (FCS) – US, Russia; obsolete • Multiplexed Analogue Components (MAC) – Europe; obsolete • Multiple sub-Nyquist sampling encoding (MUSE) – Japan • NTSCUnited States, Canada, JapanEDTV-II Clear-Vision - NTSC extension, Japan • PALEurope, Asia, OceaniaPAL-M – PAL variation, BrazilPAL-N – PAL variation, Argentina, Paraguay and UruguayPALplus – PAL extension, EuropeRS-343 (military) • SECAMFrance, former Soviet Union, Central AfricaCCIR System ACCIR System BCCIR System GCCIR System HCCIR System ICCIR System M An analog video format consists of more information than the visible content of the frame. Preceding and following the image are lines and pixels containing metadata and synchronization information. This surrounding margin is known as a blanking interval or blanking region; the horizontal and vertical front porch and back porch are the building blocks of the blanking interval. Computer displays Computer display standards specify a combination of aspect ratio, display size, display resolution, color depth, and refresh rate. A list of common resolutions is available. ==Recording==
Recording
video cassette tape Early television was almost exclusively a live medium, with some programs recorded to film for historical purposes using Kinescope. The analog video tape recorder was commercially introduced in 1951. The following list is in rough chronological order. All formats listed were sold to and used by broadcasters, video producers, or consumers; or were important historically. • VERA (BBC experimental format ca. 1952) • 2" Quadruplex videotape (Ampex 1956) • 1" Type A videotape (Ampex) • 1/2" EIAJ (1969) • U-matic 3/4" (Sony) • 1/2" Cartrivision (Avco) • VCR, VCR-LP, SVR1" Type B videotape (Robert Bosch GmbH) • 1" Type C videotape (Ampex, Marconi and Sony) • 2" Helical Scan Videotape (IVC) (1975) • Betamax (Sony) (1975) • VHS (JVC) (1976) • Video 2000 (Philips) (1979) • 1/4" CVC (Funai) (1980) • Betacam (Sony) (1982) • VHS-C (JVC) (1982) • HDVS (Sony) (1984) • Video8 (Sony) (1986) • Betacam SP (Sony) (1987) • S-VHS (JVC) (1987) • Pixelvision (Fisher-Price) (1987) • UniHi 1/2" HD (1988) • Betacam IMX (Sony) • D-VHS (JVC) • D-TheaterD1 (Sony) • D2 (Sony) • D3D5 HDD6 (Philips) • Digital-S D9 (JVC) • Digital Betacam (Sony) • Digital8 (Sony) • DV (including DVC-Pro) • HDCAM (Sony) • HDVProHD (JVC) • MicroMVMiniDV Optical storage mediums offered an alternative, especially in consumer applications, to bulky tape formats. • Blu-ray Disc (Sony) • China Blue High-definition Disc (CBHD) • DVD (was Super Density Disc, DVD Forum) • Professional DiscUniversal Media Disc (UMD) (Sony) • Enhanced Versatile Disc (EVD, Chinese government-sponsored) • HD DVD (NEC and Toshiba) • HD-VMDCapacitance Electronic DiscLaserdisc (MCA and Philips) • Television Electronic Disc (Teldec and Telefunken) • VHD (JVC) • Video CD ==Digital encoding formats==
Digital encoding formats
A video codec is software or hardware that compresses and decompresses digital video. In the context of video compression, codec is a portmanteau of encoder and decoder, while a device that only compresses is typically called an encoder, and one that only decompresses is a decoder. The compressed data format usually conforms to a standard video coding format. The compression is typically lossy, meaning that the compressed video lacks some information present in the original video. A consequence of this is that decompressed video has lower quality than the original, uncompressed video because there is insufficient information to accurately reconstruct the original video. • CCIR 601 (ITU-T) • H.261 (ITU-T) • H.263 (ITU-T) • H.264/MPEG-4 AVC (ITU-T + ISO) • H.265M-JPEG (ISO) • MPEG-1 (ISO) • MPEG-2 (ITU-T + ISO) • MPEG-4 (ISO) • Ogg-TheoraVP8-WebMVC-1 (SMPTE) == See also ==
tickerdossier.comtickerdossier.substack.com