Video Conferencing

A Video conference  (also known as a videoteleconference) is a set of interactive telecommunication technologies which allow two or more locations to interact via two-way video and audio transmissions simultaneously. It has also been called 'visual collaboration' and is a type of groupware.
Videoconferencing differs from videophone calls in that it's designed to serve a conference rather than individuals. It is an intermediate form of videotelephony, first deployed commercially by AT&T during the early 1970s using their Picturephone technology.

History

Videoconferencing uses telecommunications of audio and video to bring people at different sites together for a meeting. This can be as simple as a conversation between two people in private offices (point-to-point) or involve several sites (multi-point) with more than one person in large rooms at different sites. Besides the audio and visual transmission of meeting activities, videoconferencing can be used to share documents, computer-displayed information, and whiteboards.

Simple analog videoconferences could be established as early as the invention of the television. Such videoconferencing systems usually consisted of two closed-circuit television systems connected via coax cable or radio. An example of that was the German Reich Postzentralamt (Post Office) network set up in Berlin and several other cities from 1936 to 1940.
During the first manned space flights, NASA used two radiofrequency (UHF or VHF) links, one in each direction. TV channels routinely use this kind of videoconferencing when reporting from distant locations, for instance. Then mobile links to satellites using specially equipped trucks became rather common.
This technique was very expensive, though, and could not be used for applications such as telemedicine, distance education, and business meetings. Attempts at using normal telephony networks to transmit slow-scan video, such as the first systems developed by AT&T, failed mostly due to the poor picture quality and the lack of efficient video compression techniques. The greater 1 MHz bandwidth and 6 Mbit/s bit rate of Picturephone in the 1970s also did not cause the service to prosper.
It was only in the 1980s that digital telephony transmission networks became possible, such as ISDN, assuring a minimum bit rate (usually 128 kilobits/s) for compressed video and audio transmission. During his time, there was also research into other forms of digital video and audio communication. Many of these technologies, such as the Media space, are not as widely used today as videoconferencing but were still an important area of research. The first dedicated systems started to appear in the market as ISDN networks were expanding throughout the world. One of the first commercial Videoconferencing systems sold to companies came from PictureTel Corp. who had an Initial Public Offering in November, 1984. Videoconferencing systems throughout the 1990s rapidly evolved from very expensive proprietary equipment, software and network requirements to standards based technology that is readily available to the general public at a reasonable cost.
Finally, in the 1990s, IP (Internet Protocol) based videoconferencing became possible, and more efficient video compression technologies were developed, permitting desktop, or personal computer (PC)-based videoconferencing. In 1992 CU-SeeMe was developed at Cornell by Tim Dorcey et al. In 1995 the First public videoconference and peacecast between the continents of North America and Africa took place, linking a technofair in San Francisco with a techno-rave and cyberdeli in Cape Town. At the Winter Olympics opening ceremony in Nagano, Japan, Seiji Ozawa conducted the Ode to Joy from Beethoven's Ninth Symphony simultaneously across five continents in near-real time.
In the 2000s, videotelephony was popularized via free Internet services such as Skype and iChat, web plugins and on-line telecommunication programs which promoted low cost, albeit low-quality, videoconferencing to virtually every location with an Internet connection.
In May 2005, the first high definition video conferencing systems, produced by LifeSize Communications, were displayed at the Interop trade show in Las Vegas, Nevada, able to provide 30 frames per second at a 1280 by 720 display resolution. Polycom introduced its first high definition video conferencing system to the market in 2006. Currently, high definition resolution has now become a standard feature, with most major suppliers in the videoconferencing market offering it.

Technology
The core technology used in a videoconferencing system is digital compression of audio and video streams in real time. The hardware or software that performs compression is called a codec (coder/decoder). Compression rates of up to 1:500 can be achieved. The resulting digital stream of 1s and 0s is subdivided into labeled packets, which are then transmitted through a digital network of some kind (usually ISDN or IP). The use of audio modems in the transmission line allow for the use of POTS, or the Plain Old Telephone System, in some low-speed applications, such as videotelephony, because they convert the digital pulses to/from analog waves in the audio spectrum range.
The other components required for a videoconferencing system include :
  •  Audio input : Microphones, CD/DVD player, cassette player, or any other source of PreAmp   audio outlet.
  •   Audio output : Usually loudspeakers associated with the display device or telephone
  •   Data transfer : Analog or digital telephone network, LAN or Internet
  •  Computer  : A data processing unit that ties together the other components, does the  compressing and decompressing, and initiates and maintains the data linkage via the network.
There are basically two kinds of videoconferencing systems:
  1. Dedicated systems have all required components packaged into a single piece of equipment, usually a console with a high quality remote controlled video camera. These cameras can be controlled at a distance to pan left and right, tilt up and down, and zoom. They became known as PTZ cameras. The console contains all electrical interfaces, the control computer, and the software or hardware-based codec. Omnidirectional microphones are connected to the console, as well as a TV monitor with loudspeakers and/or a video projector. There are several types of dedicated videoconferencing devices:
    • Large group videoconferencing are non-portable, large, more expensive devices used for large rooms and auditoriums.
    • Small group videoconferencing are non-portable or portable, smaller, less expensive devices used for small meeting rooms.
    • Individual videoconferencing are usually portable devices, meant for single users, have fixed cameras, microphones and loudspeakers integrated into the console.
  2. Desktop systems are add-ons (hardware boards, usually) to normal PCs, transforming them into videoconferencing devices. A range of different cameras and microphones can be used with the board, which contains the necessary codec and transmission interfaces. Most of the desktops systems work with the H.323 standard. Videoconferences carried out via dispersed PCs are also known as e-meetings.

Conferencing layers

The components within a Conferencing System can be divided up into several different layers: User Interface, Conference Control, Control or Signal Plane and Media Plane.
Video Conferencing User Interfaces could either be graphical or voice responsive. Many of us have encountered both types of interfaces, normally we encounter graphical interfaces on the computer or television, and Voice Responsive we normally get on the phone, where we are told to select a number of choices by either saying it or pressing a number. User interfaces for conferencing have a number of different uses; it could be used for scheduling, setup, and making the call. Through the User Interface the administrator is able to control the other three layers of the system.
Conference Control performs resource allocation, management and routing. This layer along with the User Interface creates meetings (scheduled or unscheduled) or adds and removes participants from a conference.
Control (Signaling) Plane contains the stacks that signal different endpoints to create a call and/or a conference. Signals can be, but aren’t limited to, H.323 and Session Initiation Protocol (SIP) Protocols. These signals control incoming and outgoing connections as well as session parameters.
The Media Plane controls the audio and video mixing and streaming. This layer manages Real-Time Transport Protocols, User Datagram Packets (UDP) and Real-Time Transport Control Protocols (RTCP). The RTP and UDP normally carry information such the payload type which is the type of codec, frame rate, video size and many others. RTCP on the other hand acts as a quality control Protocol for detecting errors during streaming.

Multipoint videoconferencing

Simultaneous videoconferencing among three or more remote points is possible by means of a Multipoint Control Unit (MCU). This is a bridge that interconnects calls from several sources (in a similar way to the audio conference call). All parties call the MCU unit, or the MCU unit can also call the parties which are going to participate, in sequence. There are MCU bridges for IP and ISDN-based videoconferencing. There are MCUs which are pure software, and others which are a combination of hardware and software. An MCU is characterised according to the number of simultaneous calls it can handle, its ability to conduct transposing of data rates and protocols, and features such as Continuous Presence, in which multiple parties can be seen on-screen at once. MCUs can be stand-alone hardware devices, or they can be embedded into dedicated videoconferencing units.
The MCU consists of two logical components:
  1. A single multipoint controller (MC), and
  2. Multipoint Processors (MP), sometimes referred to as the mixer.
The MC controls the conferencing while it is active on the signaling plane, which is simply where the system manages conferencing creation, endpoint signaling and in-conferencing controls. This component negotiates parameters with every endpoint in the network and controls conferencing resources While the MC controls resources and signaling negotiations, the MP operates on the media plane and receives media from each endpoint. The MP generates output streams from each endpoint and redirects the information to other endpoints in the conference.
Some systems are capable of multipoint conferencing with no MCU, stand-alone, embedded or otherwise. These use a standards-based H.323 technique known as "decentralized multipoint", where each station in a multipoint call exchanges video and audio directly with the other stations with no central "manager" or other bottleneck. The advantages of this technique are that the video and audio will generally be of higher quality because they don't have to be relayed through a central point. Also, users can make ad-hoc multipoint calls without any concern for the availability or control of an MCU. This added convenience and quality comes at the expense of some increased network bandwidth, because every station must transmit to every other station directly.