Agency: European Commission | Branch: FP7 | Program: CP | Phase: ICT-2013.1.6 | Award Amount: 5.23M | Year: 2013
ICoSOLE aims at developing a platform that enables users to experience live events which are spatially spread out, such as festivals (e.g. Gentse feesten in Belgium, Glastonbury in the UK), parades, marathons or bike races, in an immersive way by combining high-quality spatial video and audio and user generated content. The project will develop a platform for a context-adapted hybrid broadcast-Internet service, providing efficient tools for capture, production and distribution of audiovisual content captured by a heterogeneous set of devices spread over the event site.\nThe approach uses a variety of sensors, ranging from mobile consumer devices over professional broadcast capture equipment to panoramic and/or free-viewpoint video and spatial audio. Methods for streaming live high-quality audiovisual content from mobile capture devices to content acquisition, processing and editing services will be developed.\nIn order to combine the heterogeneous capture sources, ICoSOLE will research and develop approaches for integration of content from professional and consumer capture devices, including mobile (and moving) sensors, based on metadata and content analysis. Methods for fusing visual and audio information into a format agnostic data representation will be developed, which enable rendering video and audio for virtual viewer/listener positions.\nICoSOLE will develop efficient tools for media production professionals to select, configure and review the content sources being used. These tools capture, extract and annotate metadata during the production process and integrate this metadata throughout the entire production chain to the end user.\nContent will be provided via broadcast, enhanced by additional content transported via broadband and novel interaction possibilities for second screen and web consumption. The content will also be provided in an adapted form to mobile devices, with specific location-based functionalities for users at or near the place of the event.
Agency: European Commission | Branch: FP7 | Program: CP | Phase: ICT-2011.1.5 | Award Amount: 6.46M | Year: 2011
SCENE will develop novel scene representations for digital media that go beyond the ability of either sample based (video) or model-based (CGI) methods to create and deliver richer media experiences. The SCENE representation and its associated tools will make it possible to capture 3D video, combine video seamlessly with CGI, manipulate and deliver it to either 2D or 3D platforms in either linear or interactive form. The outcomes will be:\n Improved ways of capturing and fusing scene data from multiple cameras and other kinds of sensors, so that structural, depth and surface appearance can be contained in a single representation\n A novel scene camera, which can capture appearance and depth to create the necessary input for a scene representation from a single viewpoint\n Improved methods for video scene analysis of stereo or multi-view sequences from which to extract spatio-temporally consistent structural information\n A video scene representation that goes beyond current image based and computer graphic approaches to obtain an inherently 3D, spatio-temporally consistent worldview\n Algorithms and data structures for transforming scene representations into forms that are adapted for efficient storage, distribution and rendering in the context of various application scenarios\n Methods of rendering the scene representation at optimal quality for different purposes and receiving platforms, such as postproduction, virtual studios and network delivery to set-top boxes and mobile devices\n Methods for merging, editing and manipulating 3D scene information in a re-usable and adaptable manner, supporting interaction, scene modification and object replacement\n Media samples that show the use of the technology and representation for digital cinema, virtual studio and network delivery to the home and mobile platforms\n An industry-based non-profit SCENE Foundation, to promote the representation and co-ordinate standardisation activities after the end of the project.
Agency: European Commission | Branch: FP7 | Program: CP | Phase: ICT-2011.1.5 | Award Amount: 5.36M | Year: 2011
TOSCA-MP aims to develop user-centric content annotation and search tools for professionals in networked media production and archiving (television, radio, online), addressing their specific use cases and workflow requirements. This will be achieved by scalable and distributed content processing methods performing advanced multimodal information extraction and semantic enrichment. Other key technology areas will include search methods across heterogeneous networked content repositories and novel user interfaces. An open standards based service oriented framework integrates the components of the system.\nTOSCA-MP will enable professionals in media production and archiving to seamlessly access content and indexes from distributed heterogeneous repositories in the network. This will be achieved by providing technologies that allow instant access to a large network of distributed multimedia databases, including beyond state-of-the-art metadata linking and alignment. The distributed repositories can be accessed through a single user interface that provides novel methods for result presentation, semi-automatic annotation and means of providing implicit user feedback.\nTOSCA-MP enables content holders to leverage scalable distributed processing in the network. The networked approach enables the use of in-house or external service models. Thus, content holders can use these services without giving content out of their hands. Advanced genre-adaptive indexing methods provide multilingual speech processing, object detection, quality analysis, semantic enrichment based on visual and speech analysis, enrichment from external sources and multimodal content clustering.\nThe project will develop models of key user tasks in the audiovisual media production workflow. These models are used to adapt the components of the system to the specific and dynamic requirements of real user tasks in the media production domain, and to evaluate the tools in a cost-effective way.
Agency: European Commission | Branch: FP7 | Program: CP | Phase: ICT-2009.1.5 | Award Amount: 14.15M | Year: 2010
FascinatE will create an innovative end-to-end system and associated standards for future immersive and interactive TV services. It will allow users to navigate around an ultra-high resolution video panorama, showing a live or recorded event, with matching accompanying audio. The output will be adapted to their particular device, covering anything from a mobile handset to an immersive panoramic display with surround sound, delivering a true personalized multi-screen experience.\nAt the production side, this requires new scene capturing systems, using multiple microphones and cameras with different fields-of-view and frame rates. These various video signals, together with metadata describing their relative alignment, constitute a novel layered scene representation. From this, any particular portion can be rendered at any desired resolution. This represents a paradigm shift in production technology, from todays format-specific view of an area selected by a cameraman to a format-agnostic representation of the whole scene. This approach is considered to be a more intelligent and future-proof alternative to other approaches, which just increase the resolution of the pictures (e.g. to 8k).\nScript metadata will describe shot framing as suggested by the supervising director. Rule-based systems will frame these regions in a subjectively-appealing manner, taking into account knowledge of how to adapt them to different display sizes, as well as the personal preferences and interactions of the user.\nIntelligent network components will tailor the transmitted data to suit the screen size and selected view for each terminal. For low-power devices, the component itself will render the desired view, whereas for powerful devices, better performance will be achieved through selectively transmitting portions of the relevant scene layers.\nAt the user terminal, novel interaction methods will allow viewers to choose either a script-driven view or to freely explore the scene themselves
Agency: European Commission | Branch: FP7 | Program: CP | Phase: ICT-2007.1.5 | Award Amount: 15.21M | Year: 2008
Film or cinema has been the driving force for the entertainment industry, setting the standards of quality, providing the most compelling experience, and feeding the distribution chains of other media (broadcast TV, cable and satellite channels, DVD, video, games et cetera). The creation of a complete `3-D capable chain is expected to follow a similar path. The media industry knows that astonishing the public is still a route to large audiences and financial success. 2020 3D Media proposes to research, develop, and demonstrate novel forms of compelling entertainment experience based on technologies for the capture, production, networked distribution and display of sounds and images in three-dimensions. 2020 3D Media will add extra dimensions to Digital Cinema and create new forms of stereoscopic and immersive networked media for the home and public spaces. The goal is to research and develop technologies to support the acquisition, coding, editing, networked distribution, and display of stereoscopic and immersive audiovisual content to provide novel forms of compelling entertainment experience in the home or public spaces. The users of the resulting technologies will be media industry professionals across the current film, TV and `new media sectors to make programme material addressing the general public. The key will be the creation of technologies for creating and presenting surround video as a viable system, based on recognised standards. This will require innovations and knew knowledge in: - Technologies and formats for 3D sound and image capture and coding, including novel high-resolution cameras - Technologies and methods for 3-D postproduction of sound and images - Technologies for the distribution and display of spatial media - The creative application of spatial media technologies
Agency: European Commission | Branch: FP7 | Program: CP | Phase: ICT-2007.1.2 | Award Amount: 12.76M | Year: 2008
IRMOS will design, develop, integrate and validate a Service Oriented Infrastructure that enables a broad range of interactive real-time applications. It will support the development and deployment of real time applications in a distributed, managed, secure and cost effective way. The infrastructure will be demonstrated in sectors with major economic and social importance by focusing on film production, virtual and augmented reality, and interactive collaborative learning. The infrastructure will be accompanied by specification languages, toolkits and standards compliant interfaces to ensure the widest possible take-up in applications involving complex value chains and real time needs, for example in security, safety and emergency response scenarios.Todays Service Oriented Infrastructures (consisting of networks, computer systems, applications and their combination) there is no provision for Quality of Service assurances and so they are insufficient for the rapid construction and provision of many interactive real-time applications, they also lack dependability and resilience. This prevents business at all levels, especially SMEs, from capitilising on the business models and services that have so much potential to emerge in this area. Whilst European business and individuals are becoming increasingly agile, todays Service Oriented Infrastructures are simply failing to keep pace.The IRMOS vision is motivated by combining the benefits of SOAs with the ability to support real-time applications with stringent performance, reliability and temporal constraints. We will allow real time applications to be exploited far more effectively and with dramatically lower costs. Instead of costly investment in dedicated infrastructure, it will be possible for European businesses to make full use of commodity resources by using a flexible service oriented approach that achieves guaranteed Quality of Service.
Deutsche Thomson Ohg and Furtwangen University of Applied Sciences | Date: 2011-07-20
The present invention relates to a method (500) for processing image data. The method includes dividing (506) each of a first plurality of images corresponding to a plurality of overlapping color value ranges into a first plurality of components of contiguous pixels. Each of a second plurality of images corresponding to the plurality of overlapping color value ranges are divided (522) into a second plurality of components of contiguous pixels. A plurality of motion vectors are identified (526) that generate a smallest distance between the first plurality of components and the second plurality of components, wherein the distance comprises a difference in color values.
Deutsche Thomson Ohg | Date: 2010-01-20
Multimedia application data formats often use data packets and may contain packetized hierarchical layers with a base layer (BL) and one or more enhancement layers (EL). Packets that comprise variable length coded data typically provide padding bits, which can be used for adding checksum information. The checksum information refers to both, the decoded base layer and decoded enhancement layer packet data, and is used for detecting synchronization loss between the layers. Though the packets provide different individual amounts of padding bits, this variable amount of bits per packet is sufficient for adding checksum information for synchronization related purposes. This allows e.g. out-of-sync detection even if the enhancement layer encoding provides no overhead information for synchronization, in particular no packet headers and no packet synchronization words.
Deutsche Thomson Ohg | Date: 2010-01-20
The invention relates to a new type of quality of service architecture for a network assembly and a corresponding computer network system. The Internet technology as approved by the IETF organization has defined different services. One is the integrated services (IntServ) and the other the differentiated services (DiffServ) architecture. Another is the RSVP Protocol according to which Quality of Service can be implemented with accuracy and richer functionality. The invention shows a way how less sophisticated QoS managers inside the network with only DiffServ or Intserv capability, can be used for enhancing the QoS functionality based on so-called RSVP shadowing messages, and easy to implement RSVP shadowing stacks.
Deutsche Thomson Ohg | Date: 2010-01-06
The invention relates to a digital signal processing technique that changes the length of an audio signal and, thus, effectively its play-out speed. This is used in the professional market for frame rate conversion in the film industry or sound effects in music production. Furthermore, consumer electronics devices, like e.g. mp3-players, voice recorders or answering machines, make use of time scaling for fast forward or slow-motion audio play-out. A method for time scaling a sequence of input signal values which makes use of a modified waveform similarity overlap add approach is proposed. According to said method, the waveform similarity overlap add approach is modified such that a similarity measure between two signal sub-sequences is weighted in dependence on a temporal distance between said two signal sub-sequences. Taking the temporal distance into account enables to bias the WSOLA approach towards preferred temporal distances.