Dissertations / Theses on the topic 'Analyse de contenu vidéo'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Analyse de contenu vidéo.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Cámara, Chávez Guillermo. "Analyse du contenu vidéo par apprentissage actif." Cergy-Pontoise, 2007. http://www.theses.fr/2007CERG0380.
Full textThis thesis presents work towards a unified framework for semi-automated video indexing and interactive retrieval. To create an efficient index, a set of representative key frames are selected from the entire video content. We developed an automatic shot boundary detection algorithm to get rid of parameters and thresholds. We adopted a SVM classifier due to its ability to use very high dimensional feature spaces while at the same time keeping strong generalization guarantees from few training examples. We deeply evaluated the combination of features and kernels and present interesting results obtained, for shot extraction TRECVID 2006 Task. We then propose an interactive video retrieval system: RETINVID, to significantly reduce the number of key frames annotated by the user. The key frames are selected based on their ability to increase the knowledge of the data. We perform an experiment against the 2005 TRECVID benchmark for high-level task
Decombas, Marc. "Compression vidéo très bas débit par analyse du contenu." Thesis, Paris, ENST, 2013. http://www.theses.fr/2013ENST0067/document.
Full textThe objective of this thesis is to find new methods for semantic video compatible with a traditional encoder like H.264/AVC. The main objective is to maintain the semantic and not the global quality. A target bitrate of 300 Kb/s has been fixed for defense and security applications. To do that, a complete chain of compression has been proposed. A study and new contributions on a spatio-temporal saliency model have been done to extract the important information in the scene. To reduce the bitrate, a resizing method named seam carving has been combined with the H.264/AVC encoder. Also, a metric combining SIFT points and SSIM has been created to measure the quality of objects without being disturbed by less important areas containing mostly artifacts. A database that can be used for testing the saliency model but also for video compression has been proposed, containing sequences with their manually extracted binary masks. All the different approaches have been thoroughly validated by different tests. An extension of this work on video summary application has also been proposed
Moinard, Matthieu. "Codage vidéo hybride basé contenu par analyse/synthèse de données." Phd thesis, Telecom ParisTech, 2011. http://tel.archives-ouvertes.fr/tel-00830924.
Full textBlanc, Katy. "Description de contenu vidéo : mouvements et élasticité temporelle." Thesis, Université Côte d'Azur (ComUE), 2018. http://www.theses.fr/2018AZUR4212/document.
Full textVideo recognition gain in performance during the last years, especially due to the improvement in the deep learning performances on images. However the jump in recognition rate on images does not directly impact the recognition rate on videos. This limitation is certainly due to this added dimension, the time, on which a robust description is still hard to extract. The recurrent neural networks introduce temporality but they have a limited memory. State of the art methods for video description usually handle time as a spatial dimension and the combination of video description methods reach the current best accuracies. However the temporal dimension has its own elasticity, different from the spatial dimensions. Indeed, the temporal dimension of a video can be locally deformed: a partial dilatation produces a visual slow down during the video, without changing the understanding, in contrast with a spatial dilatation on an image which will modify the proportions of the shown objects. We can thus expect to improve the video content classification by creating an invariant description to these speed changes. This thesis focus on the question of a robust video description considering the elasticity of the temporal dimension under three different angles. First, we have locally and explicitly described the motion content. Singularities are detected in the optical flow, then tracked along the time axis and organized in chain to describe video part. We have used this description on sport content. Then we have extracted global and implicit description thanks to tensor decompositions. Tensor enables to consider a video as a multi-dimensional data table. The extracted description are evaluated in a classification task. Finally, we have studied speed normalization method thanks to Dynamical Time Warping methods on series. We have showed that this normalization improve the classification rates
Dhollande, Nicolas. "Optimisation du codage HEVC par des moyens de pré-analyse et/ou pré-codage du contenu." Thesis, Rennes 1, 2016. http://www.theses.fr/2016REN1S113.
Full textThe High Efficiency Video Coding (HEVC) standard was released in 2013 which reduced network bandwidth by a factor of 2 compared to the prior standard H.264/AVC. These gains are achieved by a very significant increase in the encoding complexity. Especially with the industrial demand to shift in format from High Definition (HD) to Ultra High Definition (UHD), one can understand the relevance of complexity reduction techniques to develop cost-effective encoders. In our first contribution, we attempted new strategies to reduce the encoding complexity of Intra-pictures. We proposed a method with inference rules on the coding modes from the modes obtained with pre-encoding of the UHD video down-sampled in HD. We, then, proposed a fast partitioning method based on a pre-analysis of the content. The first method reduced the complexity by a factor of 3x and the second one, by a factor of 6, with a loss of compression efficiency of 5%. As a second contribution, we adressed the Inter-pictures. By implementing inference rules in the UHD encoder, from a HD pre-encoding pass, the encoding complexity is reduced by a factor of 3x when both HD and UHD encodings are considered, and by 9.2x on just the UHD encoding, with a loss of compression efficiency of 3%. Combined with an encoding configuration imitating a real system, our approach reduces the complexity by a factor of close to 2x with 4% of loss. These strategies built during this thesis offer encouraging prospects for implementation of low complexity HEVC UHD encoders. They are fully adapted to the WebTV/OTT segment that is playing a growing part in the video delivery, in which the video signal is encoded with different resolution to reach heterogeneous devices and network capacities
Wehbe, Hassan. "Synchronisation automatique d'un contenu audiovisuel avec un texte qui le décrit." Thesis, Toulouse 3, 2016. http://www.theses.fr/2016TOU30104/document.
Full textWe address the problem of automatic synchronization of an audiovisual content with a procedural text that describes it. The strategy consists in extracting pieces of information about the structure from both contents, and in matching them depending on their types. We propose two video analysis tools that respectively extract: * Limits of events of interest using an approach inspired by dictionary quantization. * Segments that enclose a repeated action based on the YIN frequency analysis method. We then propose a synchronization system that merges results coming from these tools in order to establish links between textual instructions and the corresponding video segments. To do so, a "Confidence Matrix" is built and recursively processed in order to identify these links in respect with their reliability
Souvannavong, Fabrice. "Indexation et recherche de plans vidéo par le contenu sémantique." Phd thesis, Télécom ParisTech, 2005. http://pastel.archives-ouvertes.fr/pastel-00001298.
Full textDemarty, Claire-Hélène. "Segmentation et structuration d'un document vidéo pour la caractérisation et l'indexation de son contenu sémantique." Phd thesis, École Nationale Supérieure des Mines de Paris, 2000. http://pastel.archives-ouvertes.fr/pastel-00003303.
Full textDon, Anthony. "Indexation et navigation dans les contenus visuels : approches basées sur les graphes." Bordeaux 1, 2006. http://www.theses.fr/2006BOR13258.
Full textLe, Guen Benjamin. "Adaptation du contenu spatio-temporel des images pour un codage par ondelettes." Phd thesis, Université Rennes 1, 2008. http://tel.archives-ouvertes.fr/tel-00355207.
Full textDans cette thèse, nous proposons d'aborder le problème d'adaptativité sous un angle différent. L'idée est de déformer le contenu d'une image pour l'adapter au noyau d'ondelette séparable standard. La déformation est modélisée par un maillage déformable et le critère d'adaptation utilisé est le coût de description de l'image déformée. Une minimisation énergétique similaire à une estimation de mouvement est mise en place pour calculer les paramètres du maillage. A l'issue de cette phase d'analyse, l'image est représentée par une image déformée de moindre coût de codage et par les paramètres de déformation. Après codage, transmission et décodage de ces inforrnations, l'image d'origine peut être synthétisée en inversant la déformation. Les performances en compression de ce schéma par analyse-synthèse spatiales sont étudiées et comparées à celles de JPEG2000. Visuellement, on observe une meilleure reconstruction des contours des images avec une atténuation significative de l'effet rebond.
Conservant l'idée d'adapter le contenu des images à un noyau de décomposition fixe, nous proposons ensuite un schéma de codage par analyse-synthèse spatio-temporelles dédié à la vidéo. L'analyse prend en entrée un groupe d'images (GOF) et génère en sortie un groupe d'images déformées dont le contenu est adapté à une décomposition 3D horizontale-verticale-temporelle fixe. Le schéma est conçu de sorte qu'une seule géométrie soit estimée et transmise pour l'ensemble du GOF. Des résultats de compression sont présentés en utilisant le maillage déformable pour modéliser la géométrie et le mouvement. Bien qu'une seule géométrie soit encodée, nous montrons que son coût est trop important pour permettre une amélioration significative de la qualité visuelle par rapport à un schéma par analyse-synthèse exploitant uniquement le mouvement.
Guironnet, Mickaël. "Méthodes de résumé de vidéo à partir d'informations bas niveau, du mouvement de caméra ou de l'attention visuelle." Université Joseph Fourier (Grenoble), 2006. http://www.theses.fr/2006GRE10155.
Full textThe growing volume of video leads to the need of new tools for indexing. One of the possible tools is video summary which provides a fast overview to the user. The objective of this thesis is to extract from visual information, a summary containing the “message” of video. We chose to study three new methods of video summary using different types of visual features. The first method of summary rests on low level features (color, orientation and motion). The combination of these features which is based on a fuzzy inference system allows a hierarchical summary to be built. We show the interest of such a summary in an application of query by example. The second method of summary is built from camera motion. This higher level feature is thought by the filmmaker and so induces information on the content. A method of camera motion classification based on Transferable Belief Model is achieved. The method of summary is elaborated according to rules about the magnitude and the chain of the identified motions. The third method of summary is developed from visual attention. To know the places where the glance is directed during the video playback is higher level information and relevant to create the summary. A spatio-temporal attention model is proposed, and then used to detect the change of content in time in order to build the summary
Derbas, Nadia. "Contributions à la détection de concepts et d'événements dans les documents vidéos." Thesis, Grenoble, 2014. http://www.theses.fr/2014GRENM035/document.
Full textA consequence of the rise of digital technology is that the quantity of available collections of multimedia documents is permanently and strongly increasing. The indexing of these documents became both very costly and impossible to do manually. In order to be able to analyze, classify and search multimedia documents, indexing systems have been defined. However, most of these systems suffer quality or practicability issues. Their performance is limited and depends on the data volume and data variability. Indexing systems analyze multimedia documents, looking for static concepts (bicycle, chair,...), or events (wedding, protest,...). Therefore, the variability in shapes, positions, lighting or orientation of objects hinders the process. Another aspect is that systems must be scalable. They should be able to handle big data while using reasonable amount of computing time and memory.The aim of this thesis is to improve the general performance of content-based multimedia indexing systems. Four main contributions are brought in this thesis for improving different stages of the indexing process. The first one is an "early-early fusion method" that merges different information sources in order to extract their deep correlations. This method is used for violent scenes detection in movies. The second contribution is a weakly supervised method for basic concept (objects) localization in images. This can be used afterwards as a new descriptor to help detecting complex concepts (events). The third contribution tackles the noise reduction problem on ambiguously annotated data. Two methods are proposed: a shot annotation generator, and a shot weighing method. The last contribution is a generic descriptor optimization method, based on PCA and non-linear transforms.These four contributions are tested and evaluated using reference data collections, including TRECVid and MediaEval. These contributions helped our submissions achieving very good rankings in those evaluation campaigns
Niaz, Usman. "Amélioration de la détection des concepts dans les vidéos en coupant de plus grandes tranches du monde visuel." Thesis, Paris, ENST, 2014. http://www.theses.fr/2014ENST0040/document.
Full textVisual material comprising images and videos is growing ever so rapidly over the internet and in our personal collections. This necessitates automatic understanding of the visual content which calls for the conception of intelligent methods to correctly index, search and retrieve images and videos. This thesis aims at improving the automatic detection of concepts in the internet videos by exploring all the available information and putting the most beneficial out of it to good use. Our contributions address various levels of the concept detection framework and can be divided into three main parts. The first part improves the Bag of Words (BOW) video representation model by proposing a novel BOW construction mechanism using concept labels and by including a refinement to the BOW signature based on the distribution of its elements. We then devise methods to incorporate knowledge from similar and dissimilar entities to build improved recognition models in the second part. Here we look at the potential information that the concepts share and build models for meta-concepts from which concept specific results are derived. This improves recognition for concepts lacking labeled examples. Lastly we contrive certain semi-supervised learning methods to get the best of the substantial amount of unlabeled data. We propose techniques to improve the semi-supervised cotraining algorithm with optimal view selection
Bendraou, Youssef. "Détection des changements de plans et extraction d'images représentatives dans une séquence vidéo." Thesis, Littoral, 2017. http://www.theses.fr/2017DUNK0458/document.
Full textWith the recent advancement in multimedia technologies, in conjunction with the rapid increase of the volume of digital video data and the growth of internet ; it has becom mandatory to have the hability browse and search through information stored in large multimedia databases. For this purpose, content based video retrieval (CBVR) has become an active area of research durinf the last decade. The objective of this thesis is to present applications for temporal video segmentation and video retrieval based on different mathematical models. A shot is considered as the elementary unit of a video, and is defined as a continuous sequence of frames taken from a single camera, representing an action during time. The different types of transitions that may occur in a video sequence are categorized into : abrupt and gradual transition. In this work, through statistical analysis, we segment a video into its constituent units. This is achieved by identifying transitions between adjacent shots. The first proposed algorithm aims to detect abrupt shot transitions only by measuring the similarity between consecutive frames. Given the size of the vector containing distances, it can be modeled by a log normal distribution since all the values are positive. Gradual shot transition identification is a more difficult task when compared to cut detection. Generally, a gradual transition may share similar characteristics as a dynamic segment with camera or object motion. In this work, singular value decomposition (SVD) is performed to project features from the spatial domain to the singular space. Resulting features are reduced and more refined, which makes the remaining tasks easier. The proposed system, designed for detecting both abrupt and gradual transitions, has lead to reliable performances achieving high detection rates. In addition, the acceptable computational time allows to process in real time. Once a video is partitioned into its elementary units, high-level applications can be processed, such as the key-frame extraction. Selecting representative frames from each shot to form a storyboard is considered as a static and local video summarization. In our research, we opted for a global method based on local extraction. Using refined centrist features from the singular space, we select representative frames using modified k-means clustering based on important scenes. This leads to catch pertinent frames without redoudancy in the final storyboard
Silverston, Thomas. "Peer-to-Peer video live streaming : measurement experiments and traffic analysis." Paris 6, 2009. http://www.theses.fr/2009PA066303.
Full textCaron, André. "Recherche par le contenu adaptée à la surveillance vidéo." Mémoire, Université de Sherbrooke, 2011. http://savoirs.usherbrooke.ca/handle/11143/4911.
Full textMarechal, Nicolas. "Génération de contenu graphique." Phd thesis, Université Claude Bernard - Lyon I, 2010. http://tel.archives-ouvertes.fr/tel-00843000.
Full textGarboan, Adriana. "Traçage de contenu vidéo : une méthode robuste à l'enregistrement en salle de cinéma." Phd thesis, Ecole Nationale Supérieure des Mines de Paris, 2012. http://pastel.archives-ouvertes.fr/pastel-00871762.
Full textGarboan, Adriana. "Traçage de contenu vidéo : une méthode robuste à l’enregistrement en salle de cinéma." Thesis, Paris, ENMP, 2012. http://www.theses.fr/2012ENMP0097/document.
Full textSine qua non component of multimedia content distribution on the Internet, video fingerprinting techniques allow the identification of content based on digital signatures(fingerprints) computed from the content itself. The signatures have to be invariant to content transformations like filtering, compression, geometric modifications, and spatial-temporal sub-sampling/cropping. In practice, all these transformations are non-linearly combined by the live camcorder recording use case.The state-of-the-art limitations for video fingerprinting can be identified at three levels: (1) the uniqueness of the fingerprint is solely dealt with by heuristic procedures; (2) the fingerprinting matching is not constructed on a mathematical ground, thus resulting in lack of robustness to live camcorder recording distortions; (3) very few, if any, full scalable mono-modal methods exist.The main contribution of the present thesis is to specify, design, implement and validate a new video fingerprinting method, TrackART, able to overcome these limitations. In order to ensure a unique and mathematical representation of the video content, the fingerprint is represented by a set of wavelet coefficients. In order to grant the fingerprints robustness to the mundane or malicious distortions which appear practical use-cases, the fingerprint matching is based on a repeated Rho test on correlation. In order to make the method efficient in the case of large scale databases, a localization algorithm based on a bag of visual words representation (Sivic and Zisserman, 2003) is employed. An additional synchronization mechanism able to address the time-variants distortions induced by live camcorder recording was also designed.The TrackART method was validated in industrial partnership with professional players in cinematography special effects (Mikros Image) and with the French Cinematography Authority (CST - Commision Supérieure Technique de l'Image et du Son). The reference video database consists of 14 hours of video content. The query dataset consists in 25 hours of replica content obtained by applying nine types of distortions on a third of the reference video content. The performances of the TrackART method have been objectively assessed in the context of live camcorder recording: the probability of false alarm lower than 16 10-6, the probability of missed detection lower than 0.041, precision and recall equal to 0.93. These results represent an advancement compared to the state of the art which does not exhibit any video fingerprinting method robust to live camcorder recording and validate a first proof of concept for the developed statistical methodology
Cámara, Chávez Guillermo Philipp-Foliguet Sylvie. "Analyse du contenu vidéo par apprentissage actif." [s.l.] : [s.n.], 2009. http://biblioweb.u-cergy.fr/theses/07CERG0380.pdf.
Full textThèse soutenue en co-tutelle. Titre provenant de l'écran titre. Bibliogr. p. 157-174.
Truong, Arthur. "Analyse du contenu expressif des gestes corporels." Thesis, Evry, Institut national des télécommunications, 2016. http://www.theses.fr/2016TELE0015/document.
Full textNowadays, researches dealing with gesture analysis suffer from a lack of unified mathematical models. On the one hand, gesture formalizations by human sciences remain purely theoretical and are not inclined to any quantification. On the other hand, the commonly used motion descriptors are generally purely intuitive, and limited to the visual aspects of the gesture. In the present work, we retain Laban Movement Analysis (LMA – originally designed for the study of dance movements) as a framework for building our own gesture descriptors, based on expressivity. Two datasets are introduced: the first one is called ORCHESTRE-3D, and is composed of pre-segmented orchestra conductors’ gestures, which have been annotated with the help of lexicon of musical emotions. The second one, HTI 2014-2015, comprises sequences of multiple daily actions. In a first experiment, we define a global feature vector based upon the expressive indices of our model and dedicated to the characterization of the whole gesture. This descriptor is used for action recognition purpose and to discriminate the different emotions of our orchestra conductors’ dataset. In a second approach, the different elements of our expressive model are used as a frame descriptor (e.g., describing the gesture at a given time). The feature space provided by such local characteristics is used to extract key poses of the motion. With the help of such poses, we obtain a per-frame sub-representation of body motions which is available for real-time action recognition purpose
Zhao, Shuji. "Catégorisation par le contenu sémantique d'objets vidéo : recherche et reconnaissance d'acteurs dans les films." Thesis, Cergy-Pontoise, 2011. http://www.theses.fr/2011CERG0511/document.
Full textIn this thesis, we propose a new video object retrieval and recognition system based on visual content.From video sequences, we detect, then extract video objects such as face and car, and define the continuous content made of regions containing this object in successive frames. From this volume, called Track, we extract spatio-temporally consistent visual features to define the video object representation: Spatio-Temporal Tube.To evaluate the similarity between complex tube objects, we design a Spatio-Temporal Tube Kernel (STTK) function. Based on this kernel similarity we present both supervised and active learning strategies embedded in Support Vector Machine framework. Additionally, we propose a multi-class classification framework dealing with highly unbalanced datasets.Our approach is successfully evaluated on real movie databases. Our machine learning approach outperforms the state of the art methods for multi-class actor recognition. Our method is also evaluated for actor retrieval task and on a car database showing hence promising results for car identification task and the potential of extension to any category of video objects
Ben, Abdelali Abdessalem. "Etude de la conception d’architectures matérielles dédiées pour les traitements multimédia : indexation de la vidéo par le contenu." Dijon, 2007. http://www.theses.fr/2007DIJOS075.
Full textThis thesis constitutes a contribution to the study of content based automatic video indexing aiming at designing hardware architectures dedicated to this type of multimedia application. The content based video indexing represents an important domain that is in constant development for different types of applications such as the Internet, the interactive TV, the personal video recorders (PVR) and the security applications. The proposed study is done through concrete AV analysis techniques for video indexing and it is carried out according to different aspects related to application, technology and methodology. It is included in the context of dedicated hardware architectures design and exploitation of the new embedded systems technologies for the recent multimedia applications. Much more interest is given to the reconfigurable technology and to the new possibilities and means of the FPGA devices utilization. The first stage of this thesis is devoted to the study of the automatic content based video indexing domain. It is about the study of features and the new needs of indexing systems through the approaches and techniques currently used as well as the application fields of the new generations of these systems. This is in order to show the interest of using new architectures and technological solutions permitting to support the new requirements of this domain. The second stage is dedicated to the validation and the optimization of some visual descriptors of the MPEG-7 standard for the video temporal segmentation. This constitutes a case study through an important example of AV content analysis techniques. The proposed study constitutes also a stage of preparation for the hardware implementation of these techniques in the context of hardware accelerators design for real time automatic video indexing. Different Algorithm Architecture Adequacy aspects have been studied through the proposition of various algorithmic transformations that can be applied for the considered algorithms. The third stage of this thesis is devoted to study the design of dedicated hardware operators for video content analysis techniques as well as the exploitation of the new reconfigurable systems technologies for designing SORC dedicated to the automatic video indexing. Several hardware architectures have been proposed for the MPEG-7 descriptors and different concepts related to the exploitation of reconfigurable technology and SORC have been studied as well (methodologies and tools for designing such systems on chip, technology and methods for the dynamic and partial reconfiguration, FPGA based hardware platforms, SORC structure for video indexing, etc. )
Fagette, Antoine. "Détection de foule et analyse de comportement par analyse vidéo." Thesis, Paris 6, 2014. http://www.theses.fr/2014PA066709.
Full textThis thesis focuses on the similarity between a fluid and a crowd and on the adaptation of the particle video algorithm for crowd tracking and analysis. This interrogation ended up with the design of a complete system for crowd analysis out of which, this thesis has addressed three main problems: the detection of the crowd, the estimation of its density and the tracking of the flow in order to derive some behavior features.The contribution to crowd detection introduces a totally unsupervised method for the detection and location of dense crowds in images without context-awareness. After retrieving multi-scale texture-related feature vectors from the image, a binary classification is conducted to identify the crowd and the background.The density estimation algorithm is tackling the problem of learning regression models when it comes to large dense crowds. In such cases, the learning is impossible on real data as the ground truth is not available. Our method relies on the use of synthetic data for the learning phase and proves that the regression model obtained is valid for a use on real data.Our adaptation of the particle video algorithm leads us to consider the cloud of particles as statistically representative of the crowd. Therefore, each particle has physical properties that enable us to assess the validity of its behavior according to the one expected from a pedestrian, and to optimize its motion guided by the optical flow. This leads us to three applications: the detection of the entry and exit areas of the crowd in the image, the detection of dynamic occlusions and the possibility to link entry areas with exit ones, according to the flow of the pedestrians
Sarda, Elisa. "Les effets des jeux vidéo à contenu sexiste sur l'objectivation de la femme et sur les stéréotypes de genre." Thesis, Université Grenoble Alpes (ComUE), 2017. http://www.theses.fr/2017GREAH026/document.
Full textIn our society, women are sexually objectified and are the target of sexist behaviors. Media and video games are a main channel of sexism and objectification of women. Several studies show that video game can impact players’ behavior. However the influence of sexist video games on negative perception of women is rather indefinite. In this thesis, we study the possible relationship between video games and the negative perception of women, to focus on its psychological mechanism. We argue that video games can work as a prime, activating in players’ memory some association between women and object, or some association between self-concept and masculinity, which in turn can increase negative perception of women. We also hypothesize that men, or players who highly identified with sexist game characters, are most impacted by sexist content of video games.In three studies we showed that there is an association between playing sexist video games and sexist attitudes, or women objectification. However we do not find that men are most impacted than women. Thus in other studies we test the moderating role of identification with sexist game character. In one study we show that players who highly identified with sexist game character associated more their self-concept with masculinity and gave more importance to women appearance rather than to her competence. In two other studies, we show that playing with sexist video game increases implicit associations of women with objects (especially for participants who highly identified with sexist game characters); however in the last study we do not replicate this result. Taken together, these studies show that regular video game playing is related to negative perception of women. Our results also suggest that identification plays a role in the impact of sexist video games and they encourage us to consider sexist content of video games as a prime that can make accessible some mental representations about one self or about women
Ruiloba, Rosario I. "Analyse et description du montage vidéo numérique." Paris 6, 2001. http://www.theses.fr/2001PA066528.
Full textVeneau, Emmanuel. "Macro-segmentation multi-critère et classification de séquences par le contenu dynamique pour l'indexation vidéo." Rennes 1, 2002. http://www.theses.fr/2002REN10013.
Full textPéretié, Guilhem. "Segmentation spatio-temporelle temps-réel de flux vidéo pour un encodage dépendant de son contenu." Bordeaux 1, 2007. http://www.theses.fr/2007BOR13355.
Full textThe work presented in this document was accomplished within the framework of a purse CIFRE, i. E. A partnership between the university, a PhD Student and a company. They correspond to two projects being integrated in a common environment: The extraction of the content natural or encoded videos for their characterization. Developed in an enterprise environment, they are meant to offer concrete and innovating solutions vis-a-vis the technological and economical challenges with which the enterprise is confronted. They are presented in two parts. The first one will be dealing with the pixel-based characterization of the images according to the visual attention of an observer, for a selective reduction of information in preparation to the encoding phase. The second part will describe a method of adaptation of a video flow (transrating) in real time, by putting a modification of the flow «done on the flight » of pre-encoded videos
Grandin, Pascal. "Analyse du contenu informationnel d'un consensus de marché." Paris 9, 1992. https://portail.bu.dauphine.fr/fileviewer/index.php?doc=1992PA090021.
Full textThe market consensus is an average of forecastsof earnings per share and their growth rate in mid-term by several independant sources. First, the way financial analysts revise their forecasts is described. Then it is shown that aggregating several forecasts yields a forecast which is better on average. This comes from the diversification of idiosyncratic risks in an aggregate of individual forecasts. It is also shown that consensus forecasts are useful for portfolio management, even though some of the information conveyed by these forecasts is already incorporated in stodk prices. Last, the dispersion in individual forecasts is a good measure of risk in the stock market
Steinmetz, Nadine. "Context-aware semantic analysis of video metadata." Phd thesis, Universität Potsdam, 2013. http://opus.kobv.de/ubp/volltexte/2014/7055/.
Full textThe Semantic Web provides information contained in the World Wide Web as machine-readable facts. In comparison to a keyword-based inquiry, semantic search enables a more sophisticated exploration of web documents. By clarifying the meaning behind entities, search results are more precise and the semantics simultaneously enable an exploration of semantic relationships. However, unlike keyword searches, a semantic entity-focused search requires that web documents are annotated with semantic representations of common words and named entities. Manual semantic annotation of (web) documents is time-consuming; in response, automatic annotation services have emerged in recent years. These annotation services take continuous text as input, detect important key terms and named entities and annotate them with semantic entities contained in widely used semantic knowledge bases, such as Freebase or DBpedia. Metadata of video documents require special attention. Semantic analysis approaches for continuous text cannot be applied, because information of a context in video documents originates from multiple sources possessing different reliabilities and characteristics. This thesis presents a semantic analysis approach consisting of a context model and a disambiguation algorithm for video metadata. The context model takes into account the characteristics of video metadata and derives a confidence value for each metadata item. The confidence value represents the level of correctness and ambiguity of the textual information of the metadata item. The lower the ambiguity and the higher the prospective correctness, the higher the confidence value. The metadata items derived from the video metadata are analyzed in a specific order from high to low confidence level. Previously analyzed metadata are used as reference points in the context for subsequent disambiguation. The contextually most relevant entity is identified by means of descriptive texts and semantic relationships to the context. The context is created dynamically for each metadata item, taking into account the confidence value and other characteristics. The proposed semantic analysis follows two hypotheses: metadata items of a context should be processed in descendent order of their confidence value, and the metadata that pertains to a context should be limited by content-based segmentation boundaries. The evaluation results support the proposed hypotheses and show increased recall and precision for annotated entities, especially for metadata that originates from sources with low reliability. The algorithms have been evaluated against several state-of-the-art annotation approaches. The presented semantic analysis process is integrated into a video analysis framework and has been successfully applied in several projects for the purpose of semantic video exploration of videos.
Spengler, Alexander A. "Analyse probabiliste du contenu de pages web : représentation des sémantiques de contenu dans le paradigme bayésien." Paris 6, 2011. http://www.theses.fr/2011PA066590.
Full textCiancaglini, Vincenzo. "Du routage par clé au routage par contenu : interconnexion des systèmes et applications de diffusion vidéo." Phd thesis, Université Nice Sophia Antipolis, 2013. http://tel.archives-ouvertes.fr/tel-00875653.
Full textde, Cuetos Philippe. "Streaming de Vidéos Encodées en Couches sur Internet avec Adaptation au Réseau et au Contenu." Phd thesis, Télécom ParisTech, 2003. http://pastel.archives-ouvertes.fr/pastel-00000489.
Full textAllaert, Benjamin. "Analyse des expressions faciales dans un flux vidéo." Thesis, Lille 1, 2018. http://www.theses.fr/2018LIL1I021/document.
Full textFacial expression recognition has attracted great interest over the past decade in wide application areas, such as human behavior analysis, e-health and marketing. In this thesis we explore a new approach to step forward towards in-the-wild expression recognition. Special attention has been paid to encode respectively small/large facial expression amplitudes, and to analyze facial expressions in presence of varying head pose. The first challenge addressed concerns varying facial expression amplitudes. We propose an innovative motion descriptor called LMP. This descriptor takes into account mechanical facial skin deformation properties. When extracting motion information from the face, the unified approach deals with inconsistencies and noise, caused by face characteristics. The main originality of our approach is a unified approach for both micro and macro expression recognition, with the same facial recognition framework. The second challenge addressed concerns important head pose variations. In facial expression analysis, the face registration step must ensure that minimal deformation appears. Registration techniques must be used with care in presence of unconstrained head pose as facial texture transformations apply. Hence, it is valuable to estimate the impact of alignment-related induced noise on the global recognition performance. For this, we propose a new database, called SNaP-2DFe, allowing to study the impact of head pose and intra-facial occlusions on expression recognition approaches. We prove that the usage of face registration approach does not seem adequate for preserving the features encoding facial expression deformations
Etievent, Emmanuel. "Assistance à l'indexation vidéo par analyse du mouvement." Lyon, INSA, 2002. http://theses.insa-lyon.fr/publication/2002ISAL0015/these.pdf.
Full textSalman, Zahida. "La Presse féminine irakienne analyse de contenu et enquête /." Lille 3 : ANRT, 1987. http://catalogue.bnf.fr/ark:/12148/cb37601027n.
Full textSalman, Zahida. "La presse féminine irakienne : analyse de contenu et enquête." Paris 4, 1986. http://www.theses.fr/1986PA040015.
Full textWhen it first appeared, the Iraqi feminine press had a strong link with the Arabic feminist movements, especially with the Iraqi and the Egyptian ones. The various feminine magazines that appears one after the other in Iraq during the first half of this century never lasted long and were not very largely spread. "al mar'a" (the woman), the only feminine magazine now published in Iraq is edited twice a month with a circulation of about 30,000 copies. It is directed by the general union of Iraqi women. It treats of about 20 different items. Looking closer to the content of this magazine, one can notice that there is much more texts than pictures and nearly advertisings. The more important headings deal about cultural matters and the feminine condition. A poll made on 414 readers of both sexes show that they are generally young (16 to 30 years old), they have varied social ori0 gins, and the magazine has a great importance in the family life. Through the readers'letters, the concern of men towards this magazine is even more stressed, as well as the importance of the items about marriage, problems between husband and wife or with the family. The readers' mail shows several purposes: psychotherapical, social and cultural, entertaining and educational. The magazine emphasizes on the fact that the woman's promotion goes together with the true participation in the areas of education, economics and politics, without forgetting her classic family duty
Reguig, Miloud. "Analyse des "Lâfitât" de Aḥmad Maṭar (contenu et formé)." Aix-Marseille 1, 1998. http://www.theses.fr/1998AIX10062.
Full textPoullot, Sébastien. "Scalable Content-Based Video Copy Detection for Stream Monitoring and Video Mining." Paris, CNAM, 2009. http://www.theses.fr/2009CNAM0627.
Full textLe paysage vidéo a récemment été profondément bouleversé par de nombreuses innovations technologiques. Les méthodes et acteurs de la distribution et de la production vidéo ont notamment fortement évolués. Le nombre de canaux de diffusion télévisuels augmente continuellement et parallèlement Internet supporte de nombreux sites communautaires et blogs comportant de la vidéo. Les utilisateurs finaux sont devenus eux-mêmes auteurs et créateurs, le volume qu'ils génèrent concurrence largement celui produit par les professionnels. On peut aussi noter que les logiciels d'édition vidéo sont aujourd'hui grand public et que la personnalisation de contenus est simple et très en vogue. Les professionnels aussi réutilisent largement de vieux contenus pour faire du neuf. Une conséquence directe est l'augmentation croissante du nombre de copies diffusées et hébergées sur les réseaux. L'existence de ces copies soulèvent le problème de la protection des droits. Un ayant droit peut exprimer légitimement le besoin d'être rémunéré si un oeuvre lui appartenant est diffusé sur une chaîne. L'INA est chargé d'accomplir cette tâche en France et donc de surveiller les différents canaux pour noter ces rediffusions. Le challenge tient aux volumes à protéger et à surveiller. Le nombre d'heures numérisées est de l'ordre du demi million et le nombre de canaux de la centaine. Les documentalistes ne peuvent gérer une telle connaissance ni un tel afflux. Un pré travail automatique par ordinateur est obligatoire: un système de surveillance vidéo par le contenu. Celui-ci est chargé de lire les flux vidéos diffusés et de décider si dans ces flux apparaissent des vidéos issues de la base référence à protéger. La détection par le contenu signifie l'utilisation du signal vidéo pour faire cette reconnaissance. Les vidéos représentent de gros volumes de données, et l'exploitation du signal complet n'est pas envisageable. Par conséquent on résume les vidéos par des descripteurs, sorte de quantificateurs du signal. Le problème de la surveillance repose alors sur la recherche de descripteurs dans une base de descripteurs de référence. Ces bases contiennent des milliards de descripteurs qui sont des vecteurs de moyenne ou grande dimension (20 à quelques centaines). Un tel système pour être viable demande alors un système d'indexation des descripteurs pour effectuer des recherches rapides. Après cette recherche un processus prend la décision en utilisant les descripteurs issus de la recherche. Dans cette thèse nous présentons un nouveau schéma d'indexation, appelé Zgrid, pour faire la recherche rapide. Ce schéma permet de faire une recherche approximative. Nous l'avons amélioré par des analyses de distribution des données dans l'espace de description. Par ailleurs nous proposons un nouveau modèle des distortions subies par les descripteurs lors des processus de copies et un modèle de densité locale pour corriger la recherche, celle-ci est alors plus sélective et moins consommatrice de temps. L'utilisation croisée de ces différentes propositions permet de suivre en temps réel différé un flux vidéo et de le comparer à une base de référence de 280,000 heures de vidéo avec un simple PC. L'existence de nombreuses copies peut aussi présenter des avantages. La détection des différentes occurrences d'un même contenu peut permettre par exemple de mutualiser des annotations ou d'aider à la navigation dans les bases vidéos. Le problème prend alors une autre dimension avec une complexité quadratique: on doit rechercher l'ensemble des descripteurs d'une base sur cette même base, ce qu'on appelle communément une auto jointure par similarité. Pour réduire la complexité de cette tâche nous proposons ici un nouveau descripteur dit Glocal qui utilise des descripteurs locaux pour construire un descripteur global au niveau de l'image. Ce changement de niveau permet par ailleurs de réduire aussi la complexité du processus de décision finale. Nous proposons aussi un nouveau système d'indexation adapté à l'auto jointure par similarité et à ce descripteur. La réduction globale des temps de calculs permet de trouver les occurrences dans une base de 10,000 heures avec un simple PC mais aussi de trouver ces occurrences dans une petite base (moins de 100 heures) en 30 secondes. On peut ainsi envisager des applications « off-line » pour les administrateurs de site vidéos et « online » pour les utilisateurs
Benchaita, Walid. "Stabilisation et optimisation des réseaux de diffusion de contenu." Thesis, Paris 6, 2016. http://www.theses.fr/2016PA066593.
Full textToday, many devices are capable to capture full HD videos and use their network connections to access the Internet. The popularization of these devices and continuous efforts to increase network quality has brought a proper environment for the rise of live streaming. Associated with the large scale of Users Generated Content (UGC), live streaming presents new challenges. Content Delivery Networks (CDN)are considered as the potential solution to deliver this rising content volume. Although CDN solutions are progressively integrated with the network infrastructure, they still show technological limitations in dealing with the increasing amount of bandwidth-hungry and demanding applications. In this thesis, the main target of our contributions is request routing, which is a content delivery mechanism that has a key impact on scale and performance of the CDN, as well as on the perceived Quality of Experience (QoE). First, we present a flexible scheme and an optimization algorithm, based on Lyapunov theory, for request routing in CDNs. Our online approach provides a stable quality of service to clients, while improving content delivery delays. It also reduces data transport costs for operators and outperforms existing techniques in terms of peak traffic management.Second, to overcome the limitations of the redirection mechanism used in current request routing solutions, we introduce a new approach to content delivery incorporating Information-Centric Networking (ICN) principles without requiring any change in the underlying network. This solution improves content delivery performance and enables the implementation of cost efficient request routing strategies
Käs, Christian. "Analyse et indexation des flux vidéos dans le domaine compressé H. 264." Thesis, Bordeaux 1, 2010. http://www.theses.fr/2010BOR14012/document.
Full textThe amount of generated audiovisual content continues to increase. In this work, we concentrate on the challenge of automatic video analysis and indexing, which builds the basis for efficient information access and retrieval. Taking advantage of the fact that video in most cases is stored and transmitted in encoded form, we pursue the approach of compressed domain processing, which represents a reasonable trade-off between computational complexity and quality of analysis results. The most valuable information encoded in H.264 streams is the motion vectors (MVs) that are associated with macroblocks in temporally predicted frames of type B and P. We present a number of completely unsupervised techniques to determine the camera motion and to detect and track moving objects from the extracted MV fields. We furthermore present the potential of compressed domain processing through several example applications, including traffic surveillance and video copy detection
Aguirre, Francisco. "Trajectographie sous-marine par analyse de séquences d'images vidéo." Brest, 1990. http://www.theses.fr/1990BRES2009.
Full textHurtut, Thomas. "Analyse et recherche d'oeuvres d'art 2D selon le contenu pictural." Phd thesis, Télécom ParisTech, 2008. http://tel.archives-ouvertes.fr/tel-00426645.
Full textAubin, Isabelle. "Pornographisation de la culture populaire : analyse de contenu des vidéoclips." Thèse, Université d'Ottawa / University of Ottawa, 2011. http://hdl.handle.net/10393/19976.
Full textHarb, Hadi. "Classification du signal sonore en vue d'une indexation par le contenu des documents multimédia." Ecully, Ecole centrale de Lyon, 2003. http://bibli.ec-lyon.fr/exl-doc/hharb.pdf.
Full textHumans have a remarkable ability to categorise audio signals into classes, such as speech, music, explosion, etc. . . The thesis studies the capacity of developing audio classification algorithms inspired by the human perception of the audio semantic classes in the multimedia context. A model of short therm auditory memory is proposed in order to explain some psychoacoustic effects. The memory model is then simplified to constitute the basis of the Piecewise Gaussian Modelling (PGM) features. The PGM features are coupled to a mixture of neural networks to form a general audio signal classifier. The classifier was successfully applied to speech/music classification, gender identification, action detection and musical genre recognition. A synthesis of the classification effort was used in order to structure a video into "audio scenes" and "audio chapters". This work has permitted the development of an autoamtic audio indexer prototype, CYNDI
Law-To, Julien. "From genericity to distinctiveness of video content description : application to video copy detection." Versailles-St Quentin en Yvelines, 2007. http://www.theses.fr/2007VERS0025.
Full textMes travaux de thèse portent sur l’indexation et la recherche dans de grandes bases de vidéos. Partant d’une description visuelle de l’image basée sur plusieurs natures de points d’intérêt, notre approche aboutit à une représentation de plus haut niveau, associant descripteurs visuels locaux, leurs trajectoires ainsi qu’une interprétation en termes de comportement de ces descripteurs locaux au sein de la vidéo. Cette méthode permet une description fine de la vidéo tout en réduisant la redondance temporelle qui lui est intrinsèquement liée. Une application cruciale dans la gestion de patrimoines numériques est la traçabilité du catalogue vidéo. Dans ce contexte, nous proposons ViCopT, un système de détection de copie par le contenu. Une validationde sa robustesse et de sa discriminance a été réalisée sur une base de 1000h et a montrée la pertinence de nos choix. Les hautes performances de ViCopT ont été mesurées dans des évaluations comparatives tant au niveau européen qu'international
Hervieu, Alexandre. "Analyse de trajectoires vidéos à l'aide de modélisations markoviennes pour l'interprétation de contenus." Rennes 1, 2009. ftp://ftp.irisa.fr/techreports/theses/2009/hervieu.pdf.
Full textThis thesis deals with the use of trajectories extracted from videos. The approach is invariant to translation, to rotation and to scaling and takes into account both shape and dynamics-related information on the trajectories. A hidden Markov model (HMM) is proposed to handle lack of observations and parameters are properly estimated. A similarity measure between HMM is used to tackle three dynamic video content understanding tasks: recognition, clustering and detection of unexpected events. Hierarchical semi-Markov chains are developed to process interacting trajectories. The interactions between trajectories are taken into used to recognize activity phases. Our method has been evaluated on sets of trajectories extracted from squash and handball video. Applications of such interaction-based models have also been extended to 3D gesture and action recognition and clustering. The results show that taking into account the interactions may be of great interest for such applications
Tano, Richard. "Determining multimediastreaming content." Thesis, Umeå universitet, Institutionen för fysik, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-50376.
Full textDetta examensarbete skrevs av Richard Tano student på Umeå universitet åt Ericsson Luleå. Övervakning av nätets prestanda är av yttersta vikt för nätverksleverantörer. Detta görs med modeller för att utvärdera QoS (Quality of Service) som överensstämmer med ITU-T rekommendationer. Vid bestämning av kvaliten på videoströmmar är det mer meningsfullt att utvärdera QoE (Quality of Experience) för att få insikt i hur användaren uppfattar kvaliten. Detta graderas i värden av MOS (Mean opinion score). En viktig aspekt för att bestämma QoE är typen av videoinnehåll, vilket är korrelerat till videons kodningskomplexitet och MOS värden. I detta arbete undersöktes möjligheterna att förbättra kvalitetsuppskattningsmodellerna under uppfyllande av ITU-T studygroup 12 (q.14). Metoder undersöktes och en algoritm utvecklades som använder tidsserieanalys av paketstatistik för uppskattning av videoströmmars MOS-värden. Metoder som ingår i algoritmen är en nyutvecklad frekventa mönster metod tillsammans med regressions analys. En modell som använder algoritmen från låg till hög bithastighet definierades. Den nya modellen gav omkring 20% förbättrad precision i uppskattning av MOS-värden jämfört med existerande referensmodell. Även en algoritm som enbart använder regressionsstatistik och modellerande av statistiska parametrar utvecklades. Denna algoritm levererade jämförbara resultat med föregående algoritm men gav även kraftigt förbättrad effektivitet.
Horn, Johanna, and Daniel Severus. "Exploring the Trust Generating Factors of Video Tutorials." Thesis, Högskolan i Gävle, Företagsekonomi, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:hig:diva-23651.
Full textCammas, Nathalie. "Codage vidéo scalable par maillages et ondelettes t+2D." Rennes 1, 2004. https://hal.archives-ouvertes.fr/tel-01131881.
Full text