Дисертації з теми "Analyse de données vidéo"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся з топ-50 дисертацій для дослідження на тему "Analyse de données vidéo".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Переглядайте дисертації для різних дисциплін та оформлюйте правильно вашу бібліографію.
Moinard, Matthieu. "Codage vidéo hybride basé contenu par analyse/synthèse de données." Phd thesis, Telecom ParisTech, 2011. http://tel.archives-ouvertes.fr/tel-00830924.
Повний текст джерелаMongy, Sylvain. "Modélisation et analyse du comportement des utilisateurs exploitant des données vidéo." Phd thesis, Université des Sciences et Technologie de Lille - Lille I, 2008. http://tel.archives-ouvertes.fr/tel-00842718.
Повний текст джерелаRacine, Victor. "Quantification des dynamiques cellulaires par analyse de données de vidéo-microscopie 3D+t." Paris 6, 2006. http://www.theses.fr/2006PA066480.
Повний текст джерелаThis thesis presents several approaches aiming to analyze fluorescence microscopy images in the field of cell biology. It is essentially focused on techniques for localization and tracking of multimolecular complexes in multidimensional data (2D, 2D+t, 3D and 3D+t). The first part of this work is dedicated to the extraction and characterization of fluorescent biological structures using wavelet based segmentation. Various biological studies performed in collaboration with various research groups are detailed, such as a study about morphometric analysis of organelles of cell constrained by micro patterns or the localization of the mid1p protein in yeasts. In a second part, a tracking algorithm of labeled molecular structures is presented. It is based on the linking over time of segmented objects by minimizing the association costs by a simulated annealing technique. This method is particularly well adapted in a lot of biological situations, because it allows modeling of many events like birth, death, fusion and fission. A single molecule (mono-disperse DNA 166kbp) dynamics analysis has been made using this tracking technique in order to extract the variations of the molecular diffusion coefficient. The third part describes a set of analyzes with the goal to study the spatial and temporal localization of transport intermediates involved in the membrane trafficking tagged by chimerical GFP-Rab6A and GFP-Rab6A’ proteins. Several complementary approaches are used to extract quantitative information and to describe the underneath biological processes
Allart, Thibault. "Apprentissage statistique sur données longitudinales de grande taille et applications au design des jeux vidéo." Thesis, Paris, CNAM, 2017. http://www.theses.fr/2017CNAM1136/document.
Повний текст джерелаThis thesis focuses on longitudinal time to event data possibly large along the following tree axes : number of individuals, observation frequency and number of covariates. We introduce a penalised estimator based on Cox complete likelihood with data driven weights. We introduce proximal optimization algorithms to efficiently fit models coefficients. We have implemented thoses methods in C++ and in the R package coxtv to allow everyone to analyse data sets bigger than RAM; using data streaming and online learning algorithms such that proximal stochastic gradient descent with adaptive learning rates. We illustrate performances on simulations and benchmark with existing models. Finally, we investigate the issue of video game design. We show that using our model on large datasets available in video game industry allows us to bring to light ways of improving the design of studied games. First we have a look at low level covariates, such as equipment choices through time and show that this model allows us to quantify the effect of each game elements, giving to designers ways to improve the game design. Finally, we show that the model can be used to extract more general design recommendations such as dificulty influence on player motivations
Gadchaux, Christian. "Inclusion des étudiants en situation de handicap à l’université : approches croisées anthropologique et didactique, analyse de données d’entretiens cliniques et de données vidéo." Rouen, 2016. http://www.theses.fr/2016ROUEL027.
Повний текст джерелаThe university growth in the number of disabled students (2005-103 law), and is facing the complex challenge of inclusion (2013-595 law). The law on equal opportunities and rights made it compulsory admission, and the law school and refounding the Republic requires the establishment of inclusive university. We question the issue of inclusion of students with disabilities in both the socialization aspect of the anthropological approach in the aspect of the production-transmission of academic knowledge and their accessibilisation by the didactic approach. The inventory of the issue is in a critical historical analysis of texts and institutions, in order to realize the conceptual evolution of the concepts according to their origin. We analyze and critique of official documents and institutional positioning. Building on scientific research we return to our account some conceptual elements (liminality, stigmata, rites of interaction, rites of institution) to continue the anthropological approach to disability. Didactics of joint action that we present seems quickly respond to relational needs inclusion in the academic higher education. Our thesis adopts a clinical and collaborative approach, giving pride of place to actors, students with disabilities or not, and teachers. We listen to three students with disabilities, three regular students (all Bachelor, two of Sociology and Anthropology) and two teachers (one experienced, the other rookie). These clinical interviews collected at Rogerian method of active listening give rise to a set of corpus analyzed in the thesis and delivered in full in appendices to serve as empirical contributions to scientific research. We urge cross approach, anthropological and didactic, description and video analysis filmed TD sociology and cross analysis of experienced teachers. What about the relationship and the production-transmission of knowledge in tutorials, with and for students with disabilities at the University today? If the clinical lighting does not allow us to report comprehensively on developments of the current reality, at least it allows us better to ask some questions in the light of societal and anthropological issues of inclusion of cognitive socialization university
Garcia, garcia Miguel. "Analyse de l'hypovigilance au volant par fusion d'informations environnementales et d'indices vidéo." Thesis, Université Grenoble Alpes (ComUE), 2018. http://www.theses.fr/2018GREAT120.
Повний текст джерелаDriver hypovigilance (whether caused by distraction or drowsiness) is one of the major threats to road safety. This thesis is part of the Toucango project, hold by the start-up Innov+, which aims to build a real-time hypovigilance detector based on the fusion of near infra-red video evidence and environmental information. The objective of this thesis is therefore to propose techniques for extracting relevant indices as well as multimodal fusion algorithms that can be embedded in the system for real-time operation. In order to work near ground truth conditions, a naturalistic driving database has been created with the collaboration of several transport companies. We first present a scientific state of the art and a study of the solutions available on the market for hypovigilance detection. Then, we propose several methods based on image (for the detection of relevant indices on the head, eyes, mouth and face) and data processing (for environmental indices based on geolocation). We carry out a study on the environmental factors related to hypovigilance and develop a contextual risk estimation system. Finally, we propose multimodal fusion techniques of these indices with the objective of detecting several hypovigilance behaviors: visual or cognitive distraction, engagement in a secondary task, sleep deprivation, microsleep and drowsiness
Buisson, Alexandre. "Implémentation efficace d'un codeur vidéo hiérarchique granulaire sur une architecture à processeurs multimedia." Rennes 1, 2002. http://www.theses.fr/2002REN10083.
Повний текст джерелаYe, Zakaria. "Analyse de Performance des Services de Vidéo Streaming Adaptatif dans les Réseaux Mobiles." Thesis, Avignon, 2017. http://www.theses.fr/2017AVIG0219/document.
Повний текст джерелаDue to the growth of video traffic over the Internet in recent years, HTTP AdaptiveStreaming (HAS) solution becomes the most popular streaming technology because ithas been succesfully adopted by the different actors in Internet video ecosystem. Itallows the service providers to use traditional stateless web servers and mobile edgecaches for streaming videos. Further, it allows users to access media content frombehind Firewalls and NATs.In this thesis we focus on the design of a novel video streaming delivery solutioncalled Backward-Shifted Coding (BSC), a complementary solution to Dynamic AdaptiveStreaming over HTTP (DASH), the standard version of HAS. We first describe theBackward-Shifted Coding scheme architecture based on the multi-layer Scalable VideoCoding (SVC). We also discuss the implementation of BSC protocol in DASH environment.Then, we perform the analytical evaluation of the Backward-Sihifted Codingusing results from queueing theory. The analytical results show that BSC considerablydecreases the video playback interruption which is the worst event that users can experienceduring the video session. Therefore, we design bitrate adaptation algorithms inorder to enhance the Quality of Experience (QoE) of the users in DASH/BSC system.The results of the proposed adaptation algorithms show that the flexibility of BSC allowsus to improve both the video quality and the variations of the quality during thestreaming session.Finally, we propose new caching policies to be used with video contents encodedusing SVC. Indeed, in DASH/BSC system, cache servers are deployed to make contentsclosed to the users in order to reduce network latency and improve user-perceived experience.We use Linear Programming to obtain optimal static cache composition tocompare with the results of our proposed algorithms. We show that these algorithmsincrease the system overall hit ratio and offload the backhaul links by decreasing thefetched content from the origin web servers
Leny, Marc. "Analyse et enrichissement de flux compressés : application à la vidéo surveillance." Thesis, Evry, Institut national des télécommunications, 2010. http://www.theses.fr/2010TELE0031/document.
Повний текст джерелаThe increasing deployment of civil and military videosurveillance networks brings both scientific and technological challenges regarding analysis and content recognition over compressed streams. In this context, the contributions of this thesis focus on: - an autonomous method to segment in the compressed domain mobile objects (pedestrians, vehicles, animals …), - the coverage of the various compression standards commonly used in surveillance (MPEG-2, MPEG-4 Part 2, MPEG-4 Part 10 / H.264 AVC), - an optimised multi-stream processing chain from the objects segmentation up to their tracking and description. The developed demonstrator made it possible to bench the performances of the methodological approaches chosen for a tool dedicated to help investigations. It identifies vehicles from a witness description in databases of tens of hours of video. Moreover, while dealing with corpus covering the different kind of content expected from surveillance (subway stations, crossroads, areas in countryside or border surveillance …), the system provided the following results: - simultaneous real time analysis of up to 14 MPEG-2 streams, 8 MPEG-4 Part 2 streams or 3 AVC streams on a single core (2.66 GHz; 720x576 video, 25 fps), - 100% vehicles detected over the length of traffic surveillance footages, with a image per image detection near 95%, - a segmentation spreading over 80 to 150% of the object area (under or over-segmentation linked with the compressed domain). These researches led to 9 patents linked with new services and applications that were made possible thanks to the suggested approaches. Among these lie tools for Unequal Error Protection, Visual Cryptography, Watermarking or Steganography
Goeller, Adrien. "Contribution à la perception augmentée de scènes dynamiques : schémas temps réels d’assimilation de données pour la mécanique du solide et des structures." Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLC013/document.
Повний текст джерелаThe development of sensors has always followed the ambition of industrial and scientific people to observe the unobservable. High speed cameras are part of this adventure, revealing invisible dynamics such as cracks formation or subtle mosquito flight. Industrial high speed vision is a very competitive domain in which cameras stand out through their acquisition speed. This thesis aims to broaden their capacity by augmenting the initial acquisition with dynamic models. This work proposes to develop methods linking in real time a model with a real system. Aimed benefits are interpolation, prediction and identification. Three parts are developed. The first one is based on video processing and submits to use kinematic elementary and generic models. An algorithm of motion estimation for large movements is proposed but the generic nature does not allow a sufficient knowledge to be conclusive. The second part proposes using sequential data assimilation methods known as Kalman filters. A scheme to assimilate video data with a mechanical model is successfully implemented. An application of data assimilation in modal analysis is developed. Two multi sensors real time assimilation schemes for nonlinear modal identification are proposed. These schemes are integrated in two applications on 3D reconstruction and motion magnification
Selmi, Mouna. "Reconnaissance d’activités humaines à partir de séquences vidéo." Thesis, Evry, Institut national des télécommunications, 2014. http://www.theses.fr/2014TELE0029/document.
Повний текст джерелаHuman activity recognition (HAR) from video sequences is one of the major active research areas of computer vision. There are numerous application HAR systems, including video-surveillance, search and automatic indexing of videos, and the assistance of frail elderly. This task remains a challenge because of the huge variations in the way of performing activities, in the appearance of the person and in the variation of the acquisition conditions. The main objective of this thesis is to develop an efficient HAR method that is robust to different sources of variability. Approaches based on interest points have shown excellent state-of-the-art performance over the past years. They are generally related to global classification methods as these primitives are temporally and spatially disordered. More recent studies have achieved a high performance by modeling the spatial and temporal context of interest points by encoding, for instance, the neighborhood of the interest points over several scales. In this thesis, we propose a method of activity recognition based on a hybrid model Support Vector Machine - Hidden Conditional Random Field (SVM-HCRF) that models the sequential aspect of activities while exploiting the robustness of interest points in real conditions. We first extract the interest points and show their robustness with respect to the person's identity by a multilinear tensor analysis. These primitives are then represented as a sequence of local "Bags of Words" (BOW): The video is temporally fragmented using the sliding window technique and each of the segments thus obtained is represented by the BOW of interest points belonging to it. The first layer of our hybrid sequential classification system is a Support Vector Machine that converts each local BOW extracted from the video sequence into a vector of activity classes’ probabilities. The sequence of probability vectors thus obtained is used as input of the HCRF. The latter permits a discriminative classification of time series while modeling their internal structures via the hidden states. We have evaluated our approach on various human activity datasets. The results achieved are competitive with those of the current state of art. We have demonstrated, in fact, that the use of a low-level classifier (SVM) improves the performance of the recognition system since the sequential classifier HCRF directly exploits the semantic information from local BOWs, namely the probability of each activity relatively to the current local segment, rather than mere raw information from interest points. Furthermore, the probability vectors have a low-dimension which prevents significantly the risk of overfitting that can occur if the feature vector dimension is relatively high with respect to the training data size; this is precisely the case when using BOWs that generally have a very high dimension. The estimation of the HCRF parameters in a low dimension allows also to significantly reduce the duration of the HCRF training phase
Zipstein, Marc. "Les Méthodes de compression de textes : algorithmes et performances." Paris 7, 1990. http://www.theses.fr/1990PA077107.
Повний текст джерелаBeaudoin, Jean-Michel. "Étude comportementale de l'intérêt des porcs à l'engraissement envers des objets d'enrichissement et évaluation de l'utilisation d'une analyse automatisée basée sur des données d'accélérométrie comparativement à l'analyse vidéo." Master's thesis, Université Laval, 2017. http://hdl.handle.net/20.500.11794/27943.
Повний текст джерелаHassan, Hassan. "Modélisation et analyse de performances du trafic multimédia dans les réseaux hétérogènes." Phd thesis, Université Paul Sabatier - Toulouse III, 2006. http://tel.archives-ouvertes.fr/tel-00130060.
Повний текст джерелаMathonat, Romain. "Rule discovery in labeled sequential data : Application to game analytics." Thesis, Lyon, 2020. http://www.theses.fr/2020LYSEI080.
Повний текст джерелаIt is extremely useful to exploit labeled datasets not only to learn models and perform predictive analytics but also to improve our understanding of a domain and its available targeted classes. The subgroup discovery task has been considered for more than two decades. It concerns the discovery of rules covering sets of objects having interesting properties, e.g., they characterize a given target class. Though many subgroup discovery algorithms have been proposed for both transactional and numerical data, discovering rules within labeled sequential data has been much less studied. In that context, exhaustive exploration strategies can not be used for real-life applications and we have to look for heuristic approaches. In this thesis, we propose to apply bandit models and Monte Carlo Tree Search to explore the search space of possible rules using an exploration-exploitation trade-off, on different data types such as sequences of itemset or time series. For a given budget, they find a collection of top-k best rules in the search space w.r.t chosen quality measure. They require a light configuration and are independent from the quality measure used for pattern scoring. To the best of our knowledge, this is the first time that the Monte Carlo Tree Search framework has been exploited in a sequential data mining setting. We have conducted thorough and comprehensive evaluations of our algorithms on several datasets to illustrate their added-value, and we discuss their qualitative and quantitative results. To assess the added-value of one or our algorithms, we propose a use case of game analytics, more precisely Rocket League match analysis. Discovering interesting rules in sequences of actions performed by players and using them in a supervised classification model shows the efficiency and the relevance of our approach in the difficult and realistic context of high dimensional data. It supports the automatic discovery of skills and it can be used to create new game modes, to improve the ranking system, to help e-sport commentators, or to better analyse opponent teams, for example
Eseholi, Tarek Saad Omar. "Optimisation des techniques de compression d'images fixes et de vidéo en vue de la caractérisation des matériaux : applications à la mécanique." Thesis, Valenciennes, 2018. http://www.theses.fr/2018VALE0047.
Повний текст джерелаThis PhD. thesis focuses on the optimization of fixed image and video compression techniques for the characterization of materials in mechanical science applications, and it constitutes a part of MEgABIt (MEchAnic Big Images Technology) research project supported by the Polytechnic University Hauts-de-France (UPHF). The scientific objective of the MEgABIt project is to investigate the ability to compress large volumes of data flows from mechanical instrumentation of deformations with large volumes both in the spatial and frequency domain. We propose to design original processing algorithms for data processing in the compressed domain in order to make possible at the computational level the evaluation of the mechanical parameters, while preserving the maximum of information provided by the acquisitions systems (high-speed imaging, tomography 3D). In order to be relevant image compression should allow the optimal computation of morpho-mechanical parameters without causing the loss of the essential characteristics of the contents of the mechanical surface images, which could lead to wrong analysis or classification. In this thesis, we use the state-of-the-art HEVC standard prior to image analysis, classification or storage processing in order to make the evaluation of the mechanical parameters possible at the computational level. We first quantify the impact of compression of video sequences from a high-speed camera. The experimental results obtained show that compression ratios up to 100: 1 could be applied without significant degradation of the mechanical surface response of the material measured by the VIC-2D analysis tool. Then, we develop an original classification method in the compressed domain of a surface topography database. The topographical image descriptor is obtained from the prediction modes calculated by intra-image prediction applied during the lossless HEVC compression of the images. The Support vector machine (SVM) is also introduced for strengthening the performance of the proposed system. Experimental results show that the compressed-domain topographies classifier is robust for classifying the six different mechanical topographies either based on single or multi-scale analyzing methodologies. The achieved lossless compression ratios up to 6:1 depend on image complexity. We evaluate the effects of surface filtering types (high-pass, low-pass, and band-pass filter) and the scale of analysis on the efficiency of the proposed compressed-domain classifier. We verify that the high analysis scale of high-frequency components of the surface profile is more appropriate for classifying our surface topographies with accuracy of 96 %
Picot, Antoine. "Détection d'hypovigilance chez le conducteur par fusion d'informations physiologiques et vidéo." Phd thesis, Grenoble INPG, 2009. http://tel.archives-ouvertes.fr/tel-00440959.
Повний текст джерелаDevanne, Maxime. "3D human behavior understanding by shape analysis of human motion and pose." Thesis, Lille 1, 2015. http://www.theses.fr/2015LIL10138/document.
Повний текст джерелаThe emergence of RGB-D sensors providing the 3D structure of both the scene and the human body offers new opportunities for studying human motion and understanding human behaviors. However, the design and development of models for behavior recognition that are both accurate and efficient is a challenging task due to the variability of the human pose, the complexity of human motion and possible interactions with the environment. In this thesis, we first focus on the action recognition problem by representing human action as the trajectory of 3D coordinates of human body joints over the time, thus capturing simultaneously the body shape and the dynamics of the motion. The action recognition problem is then formulated as the problem of computing the similarity between shape of trajectories in a Riemannian framework. Experiments carried out on four representative benchmarks demonstrate the potential of the proposed solution in terms of accuracy/latency for a low-latency action recognition. Second, we extend the study to more complex behaviors by analyzing the evolution of the human pose shape to decompose the motion stream into short motion units. Each motion unit is then characterized by the motion trajectory and depth appearance around hand joints, so as to describe the human motion and interaction with objects. Finally, the sequence of temporal segments is modeled through a Dynamic Naive Bayesian Classifier. Experiments on four representative datasets evaluate the potential of the proposed approach in different contexts, including recognition and online detection of behaviors
Codreanu, Tatiana. "Analyse des comportements et expérience utilisateurs d'une plateforme de visioconférence pour l'enseignement des langues : Le cas de la phase 1.0 de VISU." Thesis, Lyon 2, 2014. http://www.theses.fr/2014LYO20048.
Повний текст джерелаThis research studies the behavior of users of a desktop videoconferencing platform VISU (designed to deliver online courses), through the multimodal pedagogical communication of two groups of teachers: trainee tutors and experienced teachers based in different locations who taught French as a Foreign Language to a group of students from UC Berkeley in 2010. A team of researchers and developers worked together with the tutors in this experiment to improve the usability of the platform, which was under construction during the course of the study. In understanding the experience of users while using the tool, due to the performance limitations of the tool under construction, a new method of analysing data was devised in which user behavior was studied through discourse analysis, mimogestuality and the usage of tools including chat and webcam. Specifically, we analysed the usage of different tools (text based tools) as well as the webcam’s effects on user behavior. Then, through a micro analysis of the pedagogical instruction, we identified the role of these different communication tools used by the tutors in conveying the meaning of the task to be carried out. Interviews, questionnaires and perceptions of students and tutors were gathered to inform the analysis of their experiences and their feelings. This research therefore attempts, through the description of the synchronous teaching communication, to better understand the different practices, mainly discursive and mimo-gestural, of tutors and students engaged in the multimodal learning. In addition, a key significance of this study is that it demonstrates the value of considering user experience (UX) in studies involving language learning through technology. At the same time, it also indicates the value of including discourse analysis and mimogestuality in user experience research involving interactive pedagogical tools
Schmoll, Laurence. "Concevoir un scénario de jeu vidéo sérieux pour l'enseignement-apprentissage des langues ou comment dominer un oxymore." Thesis, Strasbourg, 2016. http://www.theses.fr/2016STRAC014.
Повний текст джерелаThis research examines the use of video games in foreign language classrooms. It investigates the integration of this digital medium, originating as an informal leisure activity, in a formal language learning classroom setting. The ‘learning video game’, which refers to both the act of learning and the act of play, can be seen as an oxymoron encompassing seriousness as well as fun. This work focuses on the fun aspect. This work is motivated by the following research question: how can the use of a learning video game foster a playful attitude, hypothetically conducive to a greater involvement of the learner, while also potentially leading to increaseand complexification in student foreign language production and interaction? As a part of the Eveil 3D-project, this study analyses reception from a sample of learners to two learning games which have the same scenario but were completed according to two different game design models. The findings provide both a game design model and a scenario template in order to support foreign language learning through the use of the video game
Lozano, Espinosa Rafael. "Intégration de données vidéo dans un SGBD à objets." Université Joseph Fourier (Grenoble), 2000. http://www.theses.fr/2000GRE10050.
Повний текст джерелаAlibay, Manu. "Fusion de données capteurs étendue pour applications vidéo embarquées." Thesis, Paris, ENMP, 2015. http://www.theses.fr/2015ENMP0032/document.
Повний текст джерелаThis thesis deals with sensor fusion between camera and inertial sensors measurements in order to provide a robust motion estimation algorithm for embedded video applications. The targeted platforms are mainly smartphones and tablets. We present a real-time, 2D online camera motion estimation algorithm combining inertial and visual measurements. The proposed algorithm extends the preemptive RANSAC motion estimation procedure with inertial sensors data, introducing a dynamic lagrangian hybrid scoring of the motion models, to make the approach adaptive to various image and motion contents. All these improvements are made with little computational cost, keeping the complexity of the algorithm low enough for embedded platforms. The approach is compared with pure inertial and pure visual procedures. A novel approach to real-time hybrid monocular visual-inertial odometry for embedded platforms is introduced. The interaction between vision and inertial sensors is maximized by performing fusion at multiple levels of the algorithm. Through tests conducted on sequences with ground-truth data specifically acquired, we show that our method outperforms classical hybrid techniques in ego-motion estimation
Diot, Fabien. "Graph mining for object tracking in videos." Thesis, Saint-Etienne, 2014. http://www.theses.fr/2014STET4009/document.
Повний текст джерелаDetecting and following the main objects of a video is necessary to describe its content in order to, for example, allow for a relevant indexation of the multimedia content by the search engines. Current object tracking approaches either require the user to select the targets to follow, or rely on pre-trained classifiers to detect particular classes of objects such as pedestrians or car for example. Since those methods rely on user intervention or prior knowledge of the content to process, they cannot be applied automatically on amateur videos such as the ones found on YouTube. To solve this problem, we build upon the hypothesis that, in videos with a moving background, the main objects should appear more frequently than the background. Moreover, in a video, the topology of the visual elements composing an object is supposed consistent from one frame to another. We represent each image of the videos with plane graphs modeling their topology. Then, we search for substructures appearing frequently in the database of plane graphs thus created to represent each video. Our contributions cover both fields of graph mining and object tracking. In the first field, our first contribution is to present an efficient plane graph mining algorithm, named PLAGRAM. This algorithm exploits the planarity of the graphs and a new strategy to extend the patterns. The next contributions consist in the introduction of spatio-temporal constraints into the mining process to exploit the fact that, in a video, the motion of objects is small from on frame to another. Thus, we constrain the occurrences of a same pattern to be close in space and time by limiting the number of frames and the spatial distance separating them. We present two new algorithms, DYPLAGRAM which makes use of the temporal constraint to limit the number of extracted patterns, and DYPLAGRAM_ST which efficiently mines frequent spatio-temporal patterns from the datasets representing the videos. In the field of object tracking, our contributions consist in two approaches using the spatio-temporal patterns to track the main objects in videos. The first one is based on a search of the shortest path in a graph connecting the spatio-temporal patterns, while the second one uses a clustering approach to regroup them in order to follow the objects for a longer period of time. We also present two industrial applications of our method
Fagette, Antoine. "Détection de foule et analyse de comportement par analyse vidéo." Thesis, Paris 6, 2014. http://www.theses.fr/2014PA066709.
Повний текст джерелаThis thesis focuses on the similarity between a fluid and a crowd and on the adaptation of the particle video algorithm for crowd tracking and analysis. This interrogation ended up with the design of a complete system for crowd analysis out of which, this thesis has addressed three main problems: the detection of the crowd, the estimation of its density and the tracking of the flow in order to derive some behavior features.The contribution to crowd detection introduces a totally unsupervised method for the detection and location of dense crowds in images without context-awareness. After retrieving multi-scale texture-related feature vectors from the image, a binary classification is conducted to identify the crowd and the background.The density estimation algorithm is tackling the problem of learning regression models when it comes to large dense crowds. In such cases, the learning is impossible on real data as the ground truth is not available. Our method relies on the use of synthetic data for the learning phase and proves that the regression model obtained is valid for a use on real data.Our adaptation of the particle video algorithm leads us to consider the cloud of particles as statistically representative of the crowd. Therefore, each particle has physical properties that enable us to assess the validity of its behavior according to the one expected from a pedestrian, and to optimize its motion guided by the optical flow. This leads us to three applications: the detection of the entry and exit areas of the crowd in the image, the detection of dynamic occlusions and the possibility to link entry areas with exit ones, according to the flow of the pedestrians
Ruiloba, Rosario I. "Analyse et description du montage vidéo numérique." Paris 6, 2001. http://www.theses.fr/2001PA066528.
Повний текст джерелаCámara, Chávez Guillermo. "Analyse du contenu vidéo par apprentissage actif." Cergy-Pontoise, 2007. http://www.theses.fr/2007CERG0380.
Повний текст джерелаThis thesis presents work towards a unified framework for semi-automated video indexing and interactive retrieval. To create an efficient index, a set of representative key frames are selected from the entire video content. We developed an automatic shot boundary detection algorithm to get rid of parameters and thresholds. We adopted a SVM classifier due to its ability to use very high dimensional feature spaces while at the same time keeping strong generalization guarantees from few training examples. We deeply evaluated the combination of features and kernels and present interesting results obtained, for shot extraction TRECVID 2006 Task. We then propose an interactive video retrieval system: RETINVID, to significantly reduce the number of key frames annotated by the user. The key frames are selected based on their ability to increase the knowledge of the data. We perform an experiment against the 2005 TRECVID benchmark for high-level task
Tapu, Ruxandra Georgina. "Segmentation and structuring of video documents for indexing applications." Phd thesis, Institut National des Télécommunications, 2012. http://tel.archives-ouvertes.fr/tel-00843596.
Повний текст джерелаJerbi, Khaled. "Synthese matérielle haut niveau des programmes flot de données." Rennes, INSA, 2012. https://tel.archives-ouvertes.fr/tel-00827163.
Повний текст джерелаThe evolution of video processing algorithms involved the advent of several standards. These standards present many common algorithms but designers are not able to reuse them because of their monolithic description. To solve this problem, iso/iec mpeg committee created the reconfigurable video coding (rvc) standard based on the idea that processing algorithms can be defined as a library of components that can be updated separately. Thus, these components of the modular library are standardized instead of the whole decoder. Mpeg rvc framework aims at providing a unified high-level specification of current mpeg coding technologies using a dataflow language called cal actor language (cal). Rvc presents also a compilation framework of the cal for hardware and software targets, but hardware compilers cannot compile high-level features which are omnipresent in most advanced designs. In this thesis, the cal language is used to develop a baseline of the lar still image coder. The problem of hardware generation is later resolved using automatic transformations of the high-level features into their equivalent low-level ones. These transformations are validated using different designs
Cagnazzo, Marco. "CODAGE DES DONNÉES VISUELLES : EFFICACITÉ, ROBUSTESSE, TRANSMISSION." Habilitation à diriger des recherches, Université Pierre et Marie Curie - Paris VI, 2013. http://tel.archives-ouvertes.fr/tel-00859677.
Повний текст джерелаAllaert, Benjamin. "Analyse des expressions faciales dans un flux vidéo." Thesis, Lille 1, 2018. http://www.theses.fr/2018LIL1I021/document.
Повний текст джерелаFacial expression recognition has attracted great interest over the past decade in wide application areas, such as human behavior analysis, e-health and marketing. In this thesis we explore a new approach to step forward towards in-the-wild expression recognition. Special attention has been paid to encode respectively small/large facial expression amplitudes, and to analyze facial expressions in presence of varying head pose. The first challenge addressed concerns varying facial expression amplitudes. We propose an innovative motion descriptor called LMP. This descriptor takes into account mechanical facial skin deformation properties. When extracting motion information from the face, the unified approach deals with inconsistencies and noise, caused by face characteristics. The main originality of our approach is a unified approach for both micro and macro expression recognition, with the same facial recognition framework. The second challenge addressed concerns important head pose variations. In facial expression analysis, the face registration step must ensure that minimal deformation appears. Registration techniques must be used with care in presence of unconstrained head pose as facial texture transformations apply. Hence, it is valuable to estimate the impact of alignment-related induced noise on the global recognition performance. For this, we propose a new database, called SNaP-2DFe, allowing to study the impact of head pose and intra-facial occlusions on expression recognition approaches. We prove that the usage of face registration approach does not seem adequate for preserving the features encoding facial expression deformations
Etievent, Emmanuel. "Assistance à l'indexation vidéo par analyse du mouvement." Lyon, INSA, 2002. http://theses.insa-lyon.fr/publication/2002ISAL0015/these.pdf.
Повний текст джерелаMarine, Cadoret. "Analyse factorielle de données de catégorisation. : Application aux données sensorielles." Rennes, Agrocampus Ouest, 2010. http://www.theses.fr/2010NSARG006.
Повний текст джерелаIn sensory analysis, holistic approaches in which objects are considered as a whole are increasingly used to collect data. Their interest comes on a one hand from their ability to acquire other types of information as the one obtained by traditional profiling methods and on the other hand from the fact they require no special skills, which makes them feasible by all subjects. Categorization (or free sorting), in which subjects are asked to provide a partition of objects, belongs to these approaches. The first part of this work focuses on categorization data. After seeing that this method of data collection is relevant, we focus on the statistical analysis of these data through the research of Euclidean representations. The proposed methodology which consists in using factorial methods such as Multiple Correspondence Analysis (MCA) or Multiple Factor Analysis (MFA) is also enriched with elements of validity. This methodology is then illustrated by the analysis of two data sets obtained from beers on a one hand and perfumes on the other hand. The second part is devoted to the study of two data collection methods related to categorization: sorted Napping® and hierarchical sorting. For both data collections, we are also interested in statistical analysis by adopting an approach similar to the one used for categorization data. The last part is devoted to the implementation in the R software of functions to analyze the three kinds of data that are categorization data, hierarchical sorting data and sorted Napping® data
Derathé, Arthur. "Modélisation de la qualité de gestes chirurgicaux laparoscopiques." Thesis, Université Grenoble Alpes, 2020. https://thares.univ-grenoble-alpes.fr/2020GRALS021.pdf.
Повний текст джерелаSous cœlioscopie, le traitement chirurgical permet une meilleure prise en charge du patient, et sa pratique est de plus en plus fréquente en routine clinique. Cette pratique présente néanmoins ses difficultés propres pour le chirurgien, et nécessite une formation prolongée pendant l’internat et en post-internat. Pour faciliter cette formation, il est notamment possible de développer des outils d’évaluation et d’analyse de la pratique chirurgicale.Dans cette optique, l’objectif de ce travail de thèse est d’étudier la faisabilité d’une méthodologie proposant, à partir d’un traitement algorithmique, des analyses à portée clinique pertinente pour le chirurgien. J’ai donc traité les problèmes suivants : Il m’a fallu recueillir et annoter un jeu de données, implémenter un environnement d’apprentissage dédié à la prédiction d’un aspect spécifique de la pratique chirurgicale, et proposer une approche permettant de traduire mes résultats algorithmiques sous une forme pertinente pour le chirurgien. Dès que cela était possible, nous avons cherché à valider ces différentes étapes de la méthodologie
Käs, Christian. "Analyse et indexation des flux vidéos dans le domaine compressé H. 264." Thesis, Bordeaux 1, 2010. http://www.theses.fr/2010BOR14012/document.
Повний текст джерелаThe amount of generated audiovisual content continues to increase. In this work, we concentrate on the challenge of automatic video analysis and indexing, which builds the basis for efficient information access and retrieval. Taking advantage of the fact that video in most cases is stored and transmitted in encoded form, we pursue the approach of compressed domain processing, which represents a reasonable trade-off between computational complexity and quality of analysis results. The most valuable information encoded in H.264 streams is the motion vectors (MVs) that are associated with macroblocks in temporally predicted frames of type B and P. We present a number of completely unsupervised techniques to determine the camera motion and to detect and track moving objects from the extracted MV fields. We furthermore present the potential of compressed domain processing through several example applications, including traffic surveillance and video copy detection
Gomes, Da Silva Alzennyr. "Analyse des données évolutives : application aux données d'usage du Web." Phd thesis, Université Paris Dauphine - Paris IX, 2009. http://tel.archives-ouvertes.fr/tel-00445501.
Повний текст джерелаGomes, da Silva Alzennyr. "Analyse des données évolutives : Application aux données d'usage du Web." Paris 9, 2009. https://portail.bu.dauphine.fr/fileviewer/index.php?doc=2009PA090047.
Повний текст джерелаNowadays, more and more organizations are becoming reliant on the Internet. The Web has become one of the most widespread platforms for information change and retrieval. The growing number of traces left behind user transactions (e. G. : customer purchases, user sessions, etc. ) automatically increases the importance of usage data analysis. Indeed, the way in which a web site is visited can change over time. These changes can be related to some temporal factors (day of the week, seasonality, periods of special offer, etc. ). By consequence, the usage models must be continuously updated in order to reflect the current behaviour of the visitors. Such a task remains difficult when the temporal dimension is ignored or simply introduced into the data description as a numeric attribute. It is precisely on this challenge that the present thesis is focused. In order to deal with the problem of acquisition of real usage data, we propose a methodology for the automatic generation of artificial usage data over which one can control the occurrence of changes and thus, analyse the efficiency of a change detection system. Guided by tracks born of some exploratory analyzes, we propose a tilted window approach for detecting and following-up changes on evolving usage data. In order measure the level of changes, this approach applies two external evaluation indices based on the clustering extension. The proposed approach also characterizes the changes undergone by the usage groups (e. G. Appearance, disappearance, fusion and split) at each timestamp. Moreover, the refereed approach is totally independent of the clustering method used and is able to manage different kinds of data other than usage data. The effectiveness of this approach is evaluated on artificial data sets of different degrees of complexity and also on real data sets from different domains (academic, tourism, e-business and marketing)
Aguirre, Francisco. "Trajectographie sous-marine par analyse de séquences d'images vidéo." Brest, 1990. http://www.theses.fr/1990BRES2009.
Повний текст джерелаDecombas, Marc. "Compression vidéo très bas débit par analyse du contenu." Thesis, Paris, ENST, 2013. http://www.theses.fr/2013ENST0067/document.
Повний текст джерелаThe objective of this thesis is to find new methods for semantic video compatible with a traditional encoder like H.264/AVC. The main objective is to maintain the semantic and not the global quality. A target bitrate of 300 Kb/s has been fixed for defense and security applications. To do that, a complete chain of compression has been proposed. A study and new contributions on a spatio-temporal saliency model have been done to extract the important information in the scene. To reduce the bitrate, a resizing method named seam carving has been combined with the H.264/AVC encoder. Also, a metric combining SIFT points and SSIM has been created to measure the quality of objects without being disturbed by less important areas containing mostly artifacts. A database that can be used for testing the saliency model but also for video compression has been proposed, containing sequences with their manually extracted binary masks. All the different approaches have been thoroughly validated by different tests. An extension of this work on video summary application has also been proposed
Liu, Ming. "Analyse et optimisation du système asiatique de diffusion terrestre et mobile de la télévision numérique." Phd thesis, INSA de Rennes, 2011. http://tel.archives-ouvertes.fr/tel-00662247.
Повний текст джерелаWeber, Jonathan. "Segmentation morphologique interactive pour la fouille de séquences vidéo." Phd thesis, Université de Strasbourg, 2011. http://tel.archives-ouvertes.fr/tel-00643585.
Повний текст джерелаPérinel, Emmanuel. "Segmentation en analyse de données symboliques : le cas de données probabilistes." Paris 9, 1996. https://portail.bu.dauphine.fr/fileviewer/index.php?doc=1996PA090079.
Повний текст джерелаSibony, Eric. "Analyse mustirésolution de données de classements." Thesis, Paris, ENST, 2016. http://www.theses.fr/2016ENST0036/document.
Повний текст джерелаThis thesis introduces a multiresolution analysis framework for ranking data. Initiated in the 18th century in the context of elections, the analysis of ranking data has attracted a major interest in many fields of the scientific literature : psychometry, statistics, economics, operations research, machine learning or computational social choice among others. It has been even more revitalized by modern applications such as recommender systems, where the goal is to infer users preferences in order to make them the best personalized suggestions. In these settings, users express their preferences only on small and varying subsets of a large catalog of items. The analysis of such incomplete rankings poses however both a great statistical and computational challenge, leading industrial actors to use methods that only exploit a fraction of available information. This thesis introduces a new representation for the data, which by construction overcomes the two aforementioned challenges. Though it relies on results from combinatorics and algebraic topology, it shares several analogies with multiresolution analysis, offering a natural and efficient framework for the analysis of incomplete rankings. As it does not involve any assumption on the data, it already leads to overperforming estimators in small-scale settings and can be combined with many regularization procedures for large-scale settings. For all those reasons, we believe that this multiresolution representation paves the way for a wide range of future developments and applications
Bouabdallah, Amine. "Contributions à la fiabilisation du transport de la vidéo." Phd thesis, Institut National Polytechnique de Toulouse - INPT, 2010. http://tel.archives-ouvertes.fr/tel-01068061.
Повний текст джерелаCammas, Nathalie. "Codage vidéo scalable par maillages et ondelettes t+2D." Rennes 1, 2004. https://hal.archives-ouvertes.fr/tel-01131881.
Повний текст джерелаRamadan, Wassim. "Amélioration de la transmission de contenus vidéo et de données dans les réseaux sans-fil." Phd thesis, Université de Franche-Comté, 2011. http://tel.archives-ouvertes.fr/tel-00802909.
Повний текст джерелаGoavec-Merou, Gwenhael. "Générateur de coprocesseur pour le traitement de données en flux (vidéo ou similaire) sur FPGA." Thesis, Besançon, 2014. http://www.theses.fr/2014BESA2056/document.
Повний текст джерелаUsing Field Programmable Gate Arrays (FPGA) is one of the very few solution for real time processingdata flows of several hundreds of Msamples/second. However, using such componentsis technically challenging beyond the need to become familiar with a new kind of dedicateddescription language and ways of describing algorithms, understanding the hardware behaviouris mandatory for implementing efficient processing solutions. In order to circumvent these difficulties,past researches have focused on providing solutions which, starting from a description ofan algorithm in a high-abstraction level language, generetes a description appropriate for FPGAconfiguration. Our contribution, following the strategy of block assembly based on the skeletonmethod, aimed at providing a software environment called CoGen for assembling various implementationsof readily available and validated processing blocks. The resulting processing chainis optimized by including FPGA hardware characteristics, and input and output bandwidths ofeach block in order to provide solution fitting best the requirements and constraints. Each processingblock implementation is either generated automatically or manually, but must complywith some constraints in order to be usable by our tool. In addition, each block developer mustprovide a standardized description of the block including required resources and data processingbandwidth limitations. CoGen then provides to the less experienced user the means to assemblethese blocks ensuring synchronism and consistency of data flow as well as the ability to synthesizethe processing chain in the available hardware resources. This working method has beenapplied to video data flow processing (threshold, contour detection and tuning fork eigenmodesanalysis) and on radiofrequency data flow (wireless interrogation of sensors through a RADARsystem, software processing of a frequency modulated stream, software defined radio)
Aaron, Catherine. "Connexité et analyse des données non linéaires." Phd thesis, Université Panthéon-Sorbonne - Paris I, 2005. http://tel.archives-ouvertes.fr/tel-00308495.
Повний текст джерелаDarlay, Julien. "Analyse combinatoire de données : structures et optimisation." Phd thesis, Université de Grenoble, 2011. http://tel.archives-ouvertes.fr/tel-00683651.
Повний текст джерелаOperto, Grégory. "Analyse structurelle surfacique de données fonctionnelles cétrébrales." Aix-Marseille 3, 2009. http://www.theses.fr/2009AIX30060.
Повний текст джерелаFunctional data acquired by magnetic resonance contain a measure of the activity in every location of the brain. If many methods exist, the automatic analysis of these data remains an open problem. In particular, the huge majority of these methods consider these data in a volume-based fashion, in the 3D acquisition space. However, most of the activity is generated within the cortex, which can be considered as a surface. Considering the data on the cortical surface has many advantages : on one hand, its geometry can be taken into account in every processing step, on the other hand considering the whole volume reduces the detection power of usually employed statistical tests. This thesis hence proposes an extension of the application field of volume-based methods to the surface-based domain by adressing problems such as projecting data onto the surface, performing surface-based multi-subjects analysis, and estimating results validity
Le, Béchec Antony. "Gestion, analyse et intégration des données transcriptomiques." Rennes 1, 2007. http://www.theses.fr/2007REN1S051.
Повний текст джерелаAiming at a better understanding of diseases, transcriptomic approaches allow the analysis of several thousands of genes in a single experiment. To date, international standard initiatives have allowed the utilization of large quantity of data generated using transcriptomic approaches by the whole scientific community, and a large number of algorithms are available to process and analyze the data sets. However, the major challenge remaining to tackle is now to provide biological interpretations to these large sets of data. In particular, their integration with additional biological knowledge would certainly lead to an improved understanding of complex biological mechanisms. In my thesis work, I have developed a novel and evolutive environment for the management and analysis of transcriptomic data. Micro@rray Integrated Application (M@IA) allows for management, processing and analysis of large scale expression data sets. In addition, I elaborated a computational method to combine multiple data sources and represent differentially expressed gene networks as interaction graphs. Finally, I used a meta-analysis of gene expression data extracted from the literature to select and combine similar studies associated with the progression of liver cancer. In conclusion, this work provides a novel tool and original analytical methodologies thus contributing to the emerging field of integrative biology and indispensable for a better understanding of complex pathophysiological processes