Interlaced scan refers to one of two common methods for "painting" a video image on an electronic display screen (the second is progressive scan) by scanning or displaying each line or row of pixels. This technique uses two fields to create a frame. One field contains all the odd lines in the image, the other contains all the even lines of the image. A PAL based television display, for example, scans 50 fields every second (25 odd and 25 even). The two sets of 25 fields work together to create a full frame every 1/25th of a second, resulting in a display of 25 frames per second.
The interlaced scan pattern in a CRT (cathode ray tube) display completes such a scan too, but only for every second line. This is carried out from the top left corner to the bottom right corner of a CRT display. This process is repeated again, only this time starting at the second row, in order to fill in those particular gaps left behind while performing the first progressive scan on alternate rows only.
Such scan of every second line is called interlacing. A field is an image that contains only half of the lines needed to make a complete picture. The afterglow of the phosphor of CRTs, in combination with the persistence of vision results in two fields being perceived as a continuous image which allows the viewing of full horizontal detail with half the bandwidth that would be required for a full progressive scan while maintaining the necessary CRT refresh rate to prevent flicker.
Only CRTs can display interlaced video directly – other display technologies require some form of deinterlacing.
But this solution could not be used for television — storing a full video frame and scanning it twice would require a frame buffer, a method that did not become feasible until the late 1980s. In addition, avoiding on-screen interference patterns caused by studio lighting and the limits of vacuum tube technology required that CRTs for TV be scanned at AC line frequency. (This was 60 Hz in the US, 50 Hz Europe.) In 1936 when the analog standards were being set in the UK, CRTs could only scan at around 200 lines in 1/50th of a second. By using interlace, a pair of 202.5-line fields could be superimposed to become a sharper 405 line frame. The vertical scan frequency remained 50 Hz, so flicker was not a problem, but visible detail was noticeably improved. As a result, this system was able to supplant John Logie Baird's 240 line mechanical progressive scan system that was also being used at the time.
From the 1940s onward, improvements in technology allowed the US and the rest of Europe to adopt systems using progressively more bandwidth to scan higher line counts, and achieve better pictures. However the fundamentals of interlaced scanning were at the heart of all of these systems. The US adopted the 525 line system known as NTSC, Europe adopted the 625 line system, and the UK switched from its 405 line system to 625 in order to avoid having to develop a unique method of color TV. France switched from its unique 819 line system to the more European standard of 625. It should be noted that although the term PAL is often used to describe the line and frame standard of the TV system, this is in fact incorrect and refers only to the method of superimposing the colour information on the standard 625 line broadcast. The French adopted their own SECAM system which was also adopted by some other countries, notably Russia and its satellites. PAL has been used on some otherwise NTSC broadcasts notably in Brazil.
Interlaced video reduces the signal bandwidth by a factor of two, for a given line count and refresh rate.
Alternatively, a given bandwidth can be used to provide an interlaced video signal with twice the display refresh rate for a given line count (versus progressive scan video). A higher refresh rate reduces flicker on CRT monitors. The higher refresh rate improves the portrayal of motion, because objects in motion are captured and their position is updated on the display more often. The human visual system averages the rapidly displayed still pictures into a moving picture image, and so interlace artifacts aren't usually objectionable when viewed at the intended field rate, on an interlaced video display.
For a given bandwidth and refresh rate, interlaced video can be used to provide a higher spatial resolution than progressive scan. For instance, 1920x1080 pixel resolution interlaced HDTV with a 60 Hz field rate (known as 1080i60) has a similar bandwidth to 1280x720 pixel progressive scan HDTV with a 60 Hz frame rate (720p60), but approximately 50% more spatial resolution.
Note that this is assuming an analog or uncompressed digital video signal. With digital video compression, as used in all current digital TV standards, interlacing introduces some additional inefficiencies over fully progressive video, and so the bandwidth savings are significantly less than half.
Because modern computer video displays are progressive scan systems, interlaced video will have visible artifacts when it is displayed on computer systems. Computer systems are frequently used to edit video and this disparity between computer video display systems and television signal formats means that the video content being edited cannot be viewed properly unless separate video display hardware is utilized.
To minimize the artifacts caused by interlaced video display on a progressive scan monitor, a process called deinterlacing can be utilized. This process is not perfect, and it generally results in a lower resolution, particularly in areas with objects in motion. Deinterlacing systems are integrated into progressive scan television displays in order to provide the best possible picture quality for interlaced video signals.
Interlace introduces a potential problem called interline twitter. This aliasing effect only shows up under certain circumstances, when the subject being shot contains vertical detail that approaches the horizontal resolution of the video format. For instance, a person on television wearing a shirt with fine dark and light stripes may appear on a video monitor as if the stripes on the shirt are "twittering". Television professionals are trained to avoid wearing clothing with fine striped patterns to avoid this problem. High-end video cameras or Computer Generated Imagery systems apply a low-pass filter to the vertical resolution of the signal in order to prevent possible problems with interline twitter.
This animation demonstrates the interline twitter effect. The two interlaced images use half the bandwidth of the progressive one. The interlaced scan (second from left) precisely duplicates the pixels of the progressive image (far left), but interlace causes details to twitter. Real interlaced video blurs such details to prevent twitter, as seen in the third image from the left, but such softening (or anti-aliasing) comes at the cost of resolution. A line doubler could never restore the third image to the full resolution of the progressive image. Note – Because the frame rate has been slowed down, you will notice additional flicker in simulated interlaced portions of this image.
Interline twitter is the primary reason that interlacing is unacceptable for a computer display. Each scanline on a high-resolution computer monitor is typically used to display discrete pixels that do not span the scanlines above or below. When the overall interlaced framerate is 30 frames per second, a pixel that spans only one scanline is visible for 1/30th of a second followed by 1/30th of a second of darkness, reducing the per-line/per-pixel framerate to 15 frames per second.
To avoid this problem, sharp detail is typically never displayed on standard interlaced television. When computer graphics are shown on a standard television, the screen is treated as if it were half the resolution of what it actually is or even lower. If text is displayed, it will be large enough so that horizontal lines are never just one scanline wide. Most fonts used in television programming have wide, fat strokes, and do not include fine-detail serifs that would make the twittering more visible.
Despite arguments against it and the calls by many prominent technological companies such as Microsoft to leave interlacing to history, interlacing continues to be supported by the television standard setting organizations and is still being included in new digital video transmission formats such as DV, DVB (including its HD modifications), and ATSC.
By the mid-1980s computers had outgrown these video systems and needed better displays. The Apple IIgs suffered from the use of the old scanning method, with the highest display resolution being 640x200, resulting in a severely distorted tall narrow pixel shape, making the display of realistic proportioned images difficult. Solutions from various companies varied widely. Because PC monitor signals did not need to be broadcast, they could consume far more than the 6, 7 and 8 MHz of bandwidth that NTSC and PAL signals were confined to. Apple Inc. built a custom 342p display into the Macintosh, and EGA for IBM compatible PCs was 350p. The Commodore Amiga created a true interlaced NTSC signal (as well as RGB variations). This ability resulted in the Amiga dominating the video production field until the mid 1990s, but the interlaced display mode caused flicker problems for more traditional PC applications where single-pixel detail is required. 1987 saw the introduction of VGA, which PCs soon standardized on, Apple only followed suit some years later with the Mac when the VGA standard was improved to match Apple's proprietary 24 bit colour video standard also introduced in 1987.
In the late 1980s and early 1990s, monitor and graphics card manufacturers introduced newer high resolution standards that once again included interlace. These monitors ran at very high refresh rates, intending that this would alleviate flicker problems. Such monitors proved very unpopular. While flicker was not obvious on them at first, eyestrain and lack of focus nevertheless became a serious problem. The industry quickly abandoned this practice, and for the rest of the decade all monitors included the assurance that their stated resolutions were "non-interlace". This experience is why the PC industry today remains against interlace in HDTV, and lobbied for the 720p standard. Also the industry is lobbing beyond 720p, actually 1080/60p for NTSC legacy countries, and 1080/50p for PAL legacy countries.
"Image Reconstruction System, Apparatus, and Method Employing Non-Sequential Scanning Scheme Using Real-Time Feedback" in Patent Application Approval Process
Aug 22, 2013; By a News Reporter-Staff News Editor at Politics & Government Week -- A patent application by the inventors LEE, Jae Hak...
Wipo Publishes Patent of Mss, Doak Arthur G. for "Sequential Scanning of Multiple Wavelengths" (American Inventor)
Sep 03, 2012; GENEVA, Sept. 3 -- Publication No. WO/2012/115817 was published on Aug. 30. Title of the invention: "SEQUENTIAL SCANNING OF...
US Patent Issued to Kabushiki Kaisha Toshiba on Feb. 28 for "Sequential Scanning Conversion Device and Method" (Japanese Inventor)
Mar 01, 2012; ALEXANDRIA, Va., March 1 -- United States Patent no. 8,125,566, issued on Feb. 28, was assigned to Kabushiki Kaisha Toshiba...