Cesson-Sévigné, France
Cesson-Sévigné, France

Time filter

Source Type

Hajj-Ahmad A.,University of Maryland University College | Baudry S.,Technicolor RandD France | Chupeau B.,Technicolor RandD France | Doerr G.,Technicolor RandD France
IH and MMSec 2015 - Proceedings of the 2015 ACM Workshop on Information Hiding and Multimedia Security | Year: 2015

Cryptography-based content protection is an efficient means to protect multimedia content during transport. Nevertheless, content is eventually decrypted at rendering time, leaving it vulnerable to piracy e.g. using a camcorder to record movies displayed on an LCD screen. Such type of piracy naturally imprints a visible flicker signal in the pirate video due to the interplay between the rendering and acquisition devices. The parameters of such flicker are inherently tied to the characteristics of the pirate devices such as the backlight of the LCD screen and the read-out time of the camcorder. In this article, we introduce a forensic methodology to estimate such parameters by analyzing the flicker signal present in pirate recordings. Experimental results clearly showcase that the accuracy of these estimation techniques offers efficient means to tell-tale which devices have been used for piracy thanks to the variety of factory settings used by consumer electronics manufacturers. © 2015 ACM.


Penet C.,Technicolor RandD France | Demarty C.-H.,Technicolor RandD France | Gravier G.,French National Center for Scientific Research | Gros P.,French Institute for Research in Computer Science and Automation
Proceedings - International Workshop on Content-Based Multimedia Indexing | Year: 2013

This article investigates a novel use of the well-known audio words representations to detect specific audio events, namely gunshots and explosions, in order to get more robustness towards soundtrack variability in Hollywood movies. An audio stream is processed as a sequence of stationary segments. Each segment is described by one or several audio words obtained by applying product quantization to standard features. Such a representation using multiple audio words constructed via product quantisation is one of the novelties described in this work. Based on this representation, Bayesian networks are used to exploit the contextual information in order to detect audio events. Experiments are performed on a comprehensive set of 15 movies, made publicly available. Results are comparable to the state of the art results obtained on the same dataset but show increased robustness to decision thresholds, however limiting the range of possible operating points in some conditions. Late fusion provides a solution to this issue. © 2013 IEEE.


Pinson M.H.,National Telecommunications and Information Administration NTIA | Schmidmer C.,OPTICOM | Janowski L.,AGH University of Science and Technology | Pepion R.,University of Nantes | And 6 more authors.
2013 5th International Workshop on Quality of Multimedia Experience, QoMEX 2013 - Proceedings | Year: 2013

In 2011, the Video Quality Experts Group (VQEG) ran subjects through the same audiovisual subjective test at six different international laboratories. That small dataset is now publically available for research and development purposes. © 2013 IEEE.


Bartczak B.,University of Kiel | Vandewalle P.,Philips | Grau O.,Center House | Briand G.,Technicolor RandD France | And 7 more authors.
IEEE Transactions on Broadcasting | Year: 2011

This paper discusses an approach to 3D-Television that is based on the Layered Depth Video (LDV) format. The LDV format contains explicit depth and occlusion information, allowing for the generation of novel viewpoints for stereoscopic and auto-stereoscopic multi-view displays. Thus, the format is effectively invariant to the display type and also allows the depth impression to be easily changed to best meet viewers' preferences for visual comfort. The major aspects of a content delivery chain based on the LDV format are discussed in this paper. The requirements placed on data acquisition are introduced, and a multi-camera system, which is well suited for LDV compliant data capture, is presented. Also discussed is the conversion of different input data streams, like standard stereo videos, multi-view data supplemented by depth data, and videos from wide baseline setups, to the LDV format. Moreover, the advantages of the LDV format in editing and mixing are examined. The paper also presents a transmission system based on currently available coding and transmission standards. Optimization of the bandwidth via different approaches to the compression of the LDV signal is analyzed, and the results of conducted experiments in this field are discussed. Finally, the aspects of perceptual human factors for the proper evaluation of 3D-TV services and the implemented LDV system are examined. This contribution reflects the efforts of the EU-funded project 3D4YOU to unify all aspects of 3D-TV production. © 2011 IEEE.


Yasakethu L.,Technicolor RandD France | Blonde L.,Technicolor RandD France | Doyen D.,Technicolor RandD France | Huynh-Thu Q.,Technicolor RandD France
2011 International Conference on 3D Imaging, IC3D 2011 | Year: 2011

3D cinema and 3DTV are at two different levels in the screen size spectrum. When the same stereoscopic-3D content is viewed on a cinema screen and 3DTV screen, it will produce a different 3D impression. As a result, it is difficult to fulfill the requirements of 3DTV with content captured for 3D cinema. Thus, it is important to properly address the issue of 3DTV content creation to avoid possible delays in the deployment of 3DTV. In this paper, we first explore the effects of using the same content for 3D cinema and 3DTV and then analyze the performance of several disparity based transformations for 3D cinema to 3DTV content conversion, by subjective testing. Effectiveness of the transformations is analyzed in terms of both depth quality and visual comfort of 3D experience. We show that by using a simple shift-based disparity transformation technique, it is possible to enhance the 3DTV experience from a common input signal which is originally captured for cinema viewing. © 2011 IEEE.


Rolland-Neviere X.,Technicolor RandD France | Doerr G.,Technicolor RandD France | Alliez P.,French Institute for Research in Computer Science and Automation
ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings | Year: 2015

Radial-based 3D watermarking alters the distances between the center of mass of the 3D mesh and its vertices. These watermarking systems are inherently sensitive to cropping. To address this limitation, this paper introduces a complementary blind resynchronization module to transmit critical synchronization information to the watermark decoder. Spherical patterns formed by several secret landmark vertices are embedded alongside the payload and blindly retrieved by the decoder, thereby conveying the synchronization information needed. Experimental results showcase significant improvement against cropping, while preserving performances against valumetric attacks thanks to a control parameter that automatically switches between alternate resynchronization modes. © 2015 IEEE.


Turkan M.,Technicolor RandD France | Thoreau D.,Technicolor RandD France | Guillotel P.,Technicolor RandD France
2013 IEEE International Conference on Image Processing, ICIP 2013 - Proceedings | Year: 2013

We describe a self-content single-image super-resolution algorithm based on multi-scale neighbor embeddings of small image patches. Given an input low-resolution patch, we gradually expand its size by relying on local geometric similarities of low- and high-resolution patch spaces under small scaling factors. We characterize the local geometry with K-similar patches taken from an exemplar set and we collect exemplar patch pairs from the input image and its appropriately rescaled versions. While ensuring local images compatibility with an optimization on K, we satisfy image smoothness by patch overlapping. We further enforce global consistency through an adaptive back-projection. Our experimental results show better performance on synthesizing natural looking textures and sharp edges with less artifacts when compared to other methods. © 2013 IEEE.


Eluard M.,Technicolor RandD France | Maetz Y.,Technicolor RandD France | Doerr G.,Technicolor RandD France
2014 IEEE International Conference on Image Processing, ICIP 2014 | Year: 2014

Collaborative creation workflows are calling for softer protection mechanisms than conventional encryption. Geometry-preserving encryption relates to protection techniques for 3D objects that alter the visual representation of the object after rendering but preserves some geometrical primitives, e.g. The bounding box. The advantage of this approach is the ability of legacy renderers to readily display composite scenes that contain both protected and unprotected objects. In this paper, we introduce a collection of geometry-preserving encryption algorithms, whose visual impact can be adjusted with a tuning parameter, and then report on the influence of the deformation on the actual rendering time of the object. Experimental results clearly highlight that special care needs to be taken in order to preserve interactivity, that is a de facto requirement for collaborative creation. © 2014 IEEE.


Rolland-Neviere X.,Technicolor RandD France | Doerr G.,Technicolor RandD France | Alliez P.,French Institute for Research in Computer Science and Automation
2014 IEEE International Conference on Image Processing, ICIP 2014 | Year: 2014

Modulating the distances between the vertices and the center of mass of a triangular mesh is a popular approach to watermark 3D objects. Prior work has formulated this approach as a quadratic programming problem which minimizes the geometric distortion while embedding the watermark payload in the histogram of distances. To enhance this framework, we introduce two watermarking components, namely the spread transform and perceptual shaping based on roughness information. Benchmarking results showcase the benefits of these addons with respect to the fidelity-robustness trade-off. © 2014 IEEE.

Loading Technicolor RandD France collaborators
Loading Technicolor RandD France collaborators