MPEG-4: Audio/video and synthetic graphics/audio for mixed media

Research output: Contribution to journalArticleResearchpeer review

Authors

  • Peter K. Doenges
  • Tolga K. Capin
  • Fabio Lavagetto
  • Joern Ostermann
  • Igor S. Pandzic
  • Eric D. Petajan

External Research Organisations

  • Evans & Sutherland
  • École polytechnique fédérale de Lausanne (EPFL)
  • University of Genova
  • Lucent
  • University of Geneva
  • Bell Laboratories
  • AT&T Labs
View graph of relations

Details

Original languageEnglish
Pages (from-to)433-463
Number of pages31
JournalSignal Processing: Image Communication
Volume9
Issue number4
Publication statusPublished - May 1997
Externally publishedYes

Abstract

MPEG-4 addresses coding of digital hybrids of natural and synthetic, aural and visual (A/V) information. The objective of this synthetic/natural hybrid coding (SNHC) is to facilitate content-based manipulation, interoperability, and wider user access in the delivery of animated mixed media. SNHC will support non-real-time and passive media delivery, as well as more interactive, real-time applications. Integrated spatial-temporal coding is sought for audio, video, and 2D/3D computer graphics as standardized A/V objects. Targets of standardization include mesh-segmented video coding, compression of geometry, synchronization between A/V objects, multiplexing of streamed A/V objects, and spatial-temporal integration of mixed media types. Composition, interactivity, and scripting of A/V objects can thus be supported in client terminals, as well as in content production for servers, also more effectively enabling terminals as servers. Such A/V objects can exhibit high efficiency in transmission and storage, plus content-based interactivity, spatial-temporal scalability, and combinations of transient dynamic data and persistent downloaded data. This approach can lower bandwidth of mixed media, offer tradeoffs in quality versus update for specific terminals, and foster varied distribution methods for content that exploit spatial and temporal coherence over buses and networks. MPEG-4 responds to trends at home and work to move beyond the paradigm of audio/video as a passive experience to more flexible A/V objects which combine audio/video with synthetic 2D/3D graphics and audio.

Keywords

    2D/3D graphics, Facial/avatar animation, Geometry/texture compression, Natural/synthetic hybrid coding, Real-time communications, Scalable multimedia, Structured synthetic audio, Text-to-speech synthesis, Text/graphics integration, Virtual environments

ASJC Scopus subject areas

Cite this

MPEG-4: Audio/video and synthetic graphics/audio for mixed media. / Doenges, Peter K.; Capin, Tolga K.; Lavagetto, Fabio et al.
In: Signal Processing: Image Communication, Vol. 9, No. 4, 05.1997, p. 433-463.

Research output: Contribution to journalArticleResearchpeer review

Doenges PK, Capin TK, Lavagetto F, Ostermann J, Pandzic IS, Petajan ED. MPEG-4: Audio/video and synthetic graphics/audio for mixed media. Signal Processing: Image Communication. 1997 May;9(4):433-463. doi: 10.1016/S0923-5965(97)00007-6
Doenges, Peter K. ; Capin, Tolga K. ; Lavagetto, Fabio et al. / MPEG-4 : Audio/video and synthetic graphics/audio for mixed media. In: Signal Processing: Image Communication. 1997 ; Vol. 9, No. 4. pp. 433-463.
Download
@article{a6b5d27242af4b51ac362787b9067d08,
title = "MPEG-4: Audio/video and synthetic graphics/audio for mixed media",
abstract = "MPEG-4 addresses coding of digital hybrids of natural and synthetic, aural and visual (A/V) information. The objective of this synthetic/natural hybrid coding (SNHC) is to facilitate content-based manipulation, interoperability, and wider user access in the delivery of animated mixed media. SNHC will support non-real-time and passive media delivery, as well as more interactive, real-time applications. Integrated spatial-temporal coding is sought for audio, video, and 2D/3D computer graphics as standardized A/V objects. Targets of standardization include mesh-segmented video coding, compression of geometry, synchronization between A/V objects, multiplexing of streamed A/V objects, and spatial-temporal integration of mixed media types. Composition, interactivity, and scripting of A/V objects can thus be supported in client terminals, as well as in content production for servers, also more effectively enabling terminals as servers. Such A/V objects can exhibit high efficiency in transmission and storage, plus content-based interactivity, spatial-temporal scalability, and combinations of transient dynamic data and persistent downloaded data. This approach can lower bandwidth of mixed media, offer tradeoffs in quality versus update for specific terminals, and foster varied distribution methods for content that exploit spatial and temporal coherence over buses and networks. MPEG-4 responds to trends at home and work to move beyond the paradigm of audio/video as a passive experience to more flexible A/V objects which combine audio/video with synthetic 2D/3D graphics and audio.",
keywords = "2D/3D graphics, Facial/avatar animation, Geometry/texture compression, Natural/synthetic hybrid coding, Real-time communications, Scalable multimedia, Structured synthetic audio, Text-to-speech synthesis, Text/graphics integration, Virtual environments",
author = "Doenges, {Peter K.} and Capin, {Tolga K.} and Fabio Lavagetto and Joern Ostermann and Pandzic, {Igor S.} and Petajan, {Eric D.}",
year = "1997",
month = may,
doi = "10.1016/S0923-5965(97)00007-6",
language = "English",
volume = "9",
pages = "433--463",
journal = "Signal Processing: Image Communication",
issn = "0923-5965",
publisher = "Elsevier",
number = "4",

}

Download

TY - JOUR

T1 - MPEG-4

T2 - Audio/video and synthetic graphics/audio for mixed media

AU - Doenges, Peter K.

AU - Capin, Tolga K.

AU - Lavagetto, Fabio

AU - Ostermann, Joern

AU - Pandzic, Igor S.

AU - Petajan, Eric D.

PY - 1997/5

Y1 - 1997/5

N2 - MPEG-4 addresses coding of digital hybrids of natural and synthetic, aural and visual (A/V) information. The objective of this synthetic/natural hybrid coding (SNHC) is to facilitate content-based manipulation, interoperability, and wider user access in the delivery of animated mixed media. SNHC will support non-real-time and passive media delivery, as well as more interactive, real-time applications. Integrated spatial-temporal coding is sought for audio, video, and 2D/3D computer graphics as standardized A/V objects. Targets of standardization include mesh-segmented video coding, compression of geometry, synchronization between A/V objects, multiplexing of streamed A/V objects, and spatial-temporal integration of mixed media types. Composition, interactivity, and scripting of A/V objects can thus be supported in client terminals, as well as in content production for servers, also more effectively enabling terminals as servers. Such A/V objects can exhibit high efficiency in transmission and storage, plus content-based interactivity, spatial-temporal scalability, and combinations of transient dynamic data and persistent downloaded data. This approach can lower bandwidth of mixed media, offer tradeoffs in quality versus update for specific terminals, and foster varied distribution methods for content that exploit spatial and temporal coherence over buses and networks. MPEG-4 responds to trends at home and work to move beyond the paradigm of audio/video as a passive experience to more flexible A/V objects which combine audio/video with synthetic 2D/3D graphics and audio.

AB - MPEG-4 addresses coding of digital hybrids of natural and synthetic, aural and visual (A/V) information. The objective of this synthetic/natural hybrid coding (SNHC) is to facilitate content-based manipulation, interoperability, and wider user access in the delivery of animated mixed media. SNHC will support non-real-time and passive media delivery, as well as more interactive, real-time applications. Integrated spatial-temporal coding is sought for audio, video, and 2D/3D computer graphics as standardized A/V objects. Targets of standardization include mesh-segmented video coding, compression of geometry, synchronization between A/V objects, multiplexing of streamed A/V objects, and spatial-temporal integration of mixed media types. Composition, interactivity, and scripting of A/V objects can thus be supported in client terminals, as well as in content production for servers, also more effectively enabling terminals as servers. Such A/V objects can exhibit high efficiency in transmission and storage, plus content-based interactivity, spatial-temporal scalability, and combinations of transient dynamic data and persistent downloaded data. This approach can lower bandwidth of mixed media, offer tradeoffs in quality versus update for specific terminals, and foster varied distribution methods for content that exploit spatial and temporal coherence over buses and networks. MPEG-4 responds to trends at home and work to move beyond the paradigm of audio/video as a passive experience to more flexible A/V objects which combine audio/video with synthetic 2D/3D graphics and audio.

KW - 2D/3D graphics

KW - Facial/avatar animation

KW - Geometry/texture compression

KW - Natural/synthetic hybrid coding

KW - Real-time communications

KW - Scalable multimedia

KW - Structured synthetic audio

KW - Text-to-speech synthesis

KW - Text/graphics integration

KW - Virtual environments

UR - http://www.scopus.com/inward/record.url?scp=0031140089&partnerID=8YFLogxK

U2 - 10.1016/S0923-5965(97)00007-6

DO - 10.1016/S0923-5965(97)00007-6

M3 - Article

AN - SCOPUS:0031140089

VL - 9

SP - 433

EP - 463

JO - Signal Processing: Image Communication

JF - Signal Processing: Image Communication

SN - 0923-5965

IS - 4

ER -

By the same author(s)