ITU-T Study Group 16
   HOME
*





ITU-T Study Group 16
The ITU-T Study Group 16 (SG16) is a statutory group of the ITU Telecommunication Standardization Sector (ITU-T) concerned with multimedia coding, systems and applications, such as video coding standards. It is responsible for standardization of the "H.26x" line of video coding standards, the "T.8xx" line of image coding standards, and related technologies, as well as various collaborations with the World Health Organization, including on safe listing (H.870) accessibility of e-health (F.780.2), it is also the parent body of VCEG and various Focus Groups, such as the ITU-WHO Focus Group on Artificial Intelligence for Health. Administratively, SG16 is a statutory meeting of the World Telecommunication Standardization Assembly (WTSA), which creates the ITU-T Study Groups and appoints their management teams. The secretariat is provided by the Telecommunication Standardization Bureau (under Director Chaesub Lee). WTSA instructed ITU to hold the Global Standards Symposium as a part of ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Standards Organization
A standards organization, standards body, standards developing organization (SDO), or standards setting organization (SSO) is an organization whose primary function is developing, coordinating, promulgating, revising, amending, reissuing, interpreting, or otherwise contributing to the usefulness of technical standards to those who employ them. Such an organization works to create uniformity across producers, consumers, government agencies, and other relevant parties regarding terminology, product specifications (e.g. size, including units of measure), protocols, and more. Its goals could include ensuring that Company A's external hard drive works on Company B's computer, an individual's blood pressure measures the same with Company C's sphygmomanometer as it does with Company D's, or that all shirts that should not be ironed have the same icon (a clothes iron crossed out with an X) on the label. Most standards are voluntary in the sense that they are offered for adoption by people ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Multimedia
Multimedia is a form of communication that uses a combination of different content forms such as text, audio, images, animations, or video into a single interactive presentation, in contrast to traditional mass media, such as printed material or audio recordings, which features little to no interaction between users. Popular examples of multimedia include video podcasts, audio slideshows and animated videos. Multimedia also contains the principles and application of effective interactive communication such as the building blocks of software, hardware, and other technologies. Multimedia can be recorded for playback on computers, laptops, smartphones, and other electronic devices, either on demand or in real time (streaming). In the early years of multimedia, the term "rich media" was synonymous with interactive multimedia. Over time, hypermedia extensions brought multimedia to the World Wide Web. Terminology The term ''multimedia'' was ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Video Quality
Video quality is a characteristic of a video passed through a video transmission or processing system that describes perceived video degradation (typically, compared to the original video). Video processing systems may introduce some amount of distortion or artifacts in the video signal that negatively impacts the user's perception of a system. For many stakeholders in video production and distribution, assurance of video quality is an important task. Video quality evaluation is performed to describe the quality of a set of video sequences under study. Video quality can be evaluated objectively (by mathematical models) or subjectively (by asking users for their rating). Also, the quality of a system can be determined offline (i.e., in a laboratory setting for developing new codecs or services), or in-service (to monitor and ensure a certain level of quality). From analog to digital video Since the world's first video sequence was recorded and transmitted, many video processing sys ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Video Compression
In information theory, data compression, source coding, or bit-rate reduction is the process of encoding information using fewer bits than the original representation. Any particular compression is either lossy or lossless. Lossless compression reduces bits by identifying and eliminating statistical redundancy. No information is lost in lossless compression. Lossy compression reduces bits by removing unnecessary or less important information. Typically, a device that performs data compression is referred to as an encoder, and one that performs the reversal of the process (decompression) as a decoder. The process of reducing the size of a data file is often referred to as data compression. In the context of data transmission, it is called source coding; encoding done at the source of the data before it is stored or transmitted. Source coding should not be confused with channel coding, for error detection and correction or line coding, the means for mapping data onto a signal. C ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Video Codec
A video codec is software or hardware that compresses and decompresses digital video. In the context of video compression, ''codec'' is a portmanteau of ''encoder'' and ''decoder'', while a device that only compresses is typically called an '' encoder'', and one that only decompresses is a ''decoder''. The compressed data format usually conforms to a standard video coding format A video coding format (or sometimes video compression format) is a content representation format for storage or transmission of digital video content (such as in a data file or bitstream). It typically uses a standardized video compression algori .... The compression is typically lossy compression, lossy, meaning that the compressed video lacks some information present in the original video. A consequence of this is that decompressed video has lower quality than the original, uncompressed video because there is insufficient information to accurately reconstruct the original video. There are complex ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Moving Picture Experts Group
The Moving Picture Experts Group (MPEG) is an alliance of working groups established jointly by ISO and IEC that sets standards for media coding, including compression coding of audio, video, graphics, and genomic data; and transmission and file formats for various applications.John Watkinson, ''The MPEG Handbook'', p. 1 Together with JPEG, MPEG is organized under ISO/IEC JTC 1/SC 29 – ''Coding of audio, picture, multimedia and hypermedia information'' (ISO/IEC Joint Technical Committee 1, Subcommittee 29). MPEG formats are used in various multimedia systems. The most well known older MPEG media formats typically use MPEG-1, MPEG-2, and MPEG-4 AVC media coding and MPEG-2 systems transport streams and program streams. Newer systems typically use the MPEG base media file format and dynamic streaming (a.k.a. MPEG-DASH). History MPEG was established in 1988 by the initiative of Dr. Hiroshi Yasuda ( NTT) and Dr. Leonardo Chiariglione (CSELT). Chiariglione was the group' ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Video Coding Experts Group
The Video Coding Experts Group or Visual Coding Experts Group (VCEG, also known as Question 6) is a working group of the ITU Telecommunication Standardization Sector (ITU-T) concerned with standards for compression coding of video, images, audio, and other signals. It is responsible for standardization of the "H.26x" line of video coding standards, the "T.8xx" line of image coding standards, and related technologies. Administratively, VCEG is the informal name of Question 6 (Visual, audio and signal coding) of Working Party 3 (Audiovisual technologies and intelligent immersive applications) of ITU-T Study Group 16 (Multimedia and related digital technologies). Its abbreviated title is ITU-T Q.6/SG16, or more simply, ITU-T Q6/16. The goal of VCEG is to produce ITU-T Recommendations (international standards) for video coding and image coding methods appropriate for conversational (e.g. videoconferencing and video telephony) and non-conversational (e.g., streaming, broadcast, file d ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Joint Photographic Experts Group
The Joint Photographic Experts Group (JPEG) is the joint committee between ISO/IEC JTC 1/SC 29 and ITU-T Study Group 16 that created and maintains the JPEG, JPEG 2000, JPEG XR, JPEG XT, JPEG XS, JPEG XL, and related digital image standards. It also has the responsibility for maintenance of the JBIG and JBIG2 standards that were developed by the former Joint Bi-level Image Experts Group. Within ISO/IEC JTC 1, JPEG is Working Group 1 (WG 1) of Subcommittee 29 (SC 29) and has the formal title ''JPEG Coding of digital representations of images'', where it is one of eight working groups in SC 29. In the ITU-T (formerly called the CCITT), its work falls in the domain of the ITU-T Visual Coding Experts Group (VCEG), which is Question 6 of Study Group 16. JPEG has typically held meetings three or four times annually in North America, Asia and Europe (with meetings held virtually during the COVID-19 pandemic). The chairman of JPEG (termed its ''Convenor'' in ISO/IEC terminology) is ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Federated Learning
Federated learning (also known as collaborative learning) is a machine learning technique that trains an algorithm across multiple decentralized edge devices or servers holding local data samples, without exchanging them. This approach stands in contrast to traditional centralized machine learning techniques where all the local datasets are uploaded to one server, as well as to more classical decentralized approaches which often assume that local data samples are identically distributed. Federated learning enables multiple actors to build a common, robust machine learning model without sharing data, thus allowing to address critical issues such as data privacy, data security, data access rights and access to heterogeneous data. Its applications are spread over a number of industries including defense, telecommunications, IoT, and pharmaceutics. A major open question at the moment is how inferior models learned through federated data are relative to ones where the data are pooled. ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Multi-party Computation
Secure multi-party computation (also known as secure computation, multi-party computation (MPC) or privacy-preserving computation) is a subfield of cryptography with the goal of creating methods for parties to jointly compute a function over their inputs while keeping those inputs private. Unlike traditional cryptographic tasks, where cryptography assures security and integrity of communication or storage and the adversary is outside the system of participants (an eavesdropper on the sender and receiver), the cryptography in this model protects participants' privacy from each other. The foundation for secure multi-party computation started in the late 1970s with the work on mental poker, cryptographic work that simulates game playing/computational tasks over distances without requiring a trusted third party. Note that traditionally, cryptography was about concealing content, while this new type of computation and protocol is about concealing partial information about data while comp ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  




Homomorphic Encryption
Homomorphic encryption is a form of encryption that permits users to perform computations on its encrypted data without first decrypting it. These resulting computations are left in an encrypted form which, when decrypted, result in an identical output to that produced had the operations been performed on the unencrypted data. Homomorphic encryption can be used for privacy-preserving outsourced storage and computation. This allows data to be encrypted and out-sourced to commercial cloud environments for processing, all while encrypted. For sensitive data, such as health care information, homomorphic encryption can be used to enable new services by removing privacy barriers inhibiting data sharing or increase security to existing services. For example, predictive analytics in health care can be hard to apply via a third party service provider due to medical data privacy concerns, but if the predictive analytics service provider can operate on encrypted data instead, these priva ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Trustworthy AI
Trustworthy AI is a programme of work of the ITU (United Nations Specialized Agency for ICT) under its AI for Good programme. The programme advances the standardization of a number of Privacy-enhancing technologies (PETs), including homomorphic encryption, federated learning, secure multi-party computation, differential privacy, zero-knowledge proof. Privacy-Enhancing Technologies are complex and the working can be considered counterintuitive. For instance, homomorphic encryption enciphers data, yet if this data is numerical, it is still possible to compute on this data (the outcomes still being encrypted but decryptable by the data owner). Other PETs exhibit similar characteristics. Often these technologies are developed with the explicit goal of enabling their analysis in juristictions different from the data creation (under e.g. GDPR). As such, the programme being led by two international organizations develops international standards to work in this context. The PET ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]