Dissertations / Theses on the topic 'Camera motion estimation'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Camera motion estimation.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Kim, Jae-Hak, and Jae-Hak Kim@anu edu au. "Camera Motion Estimation for Multi-Camera Systems." The Australian National University. Research School of Information Sciences and Engineering, 2008. http://thesis.anu.edu.au./public/adt-ANU20081211.011120.
Full textKim, Jae-Hak. "Camera motion estimation for multi-camera systems /." View thesis entry in Australian Digital Theses Program, 2008. http://thesis.anu.edu.au/public/adt-ANU20081211.011120/index.html.
Full textSrestasathiern, Panu. "Line Based Estimation of Object Space Geometry and Camera Motion." The Ohio State University, 2012. http://rave.ohiolink.edu/etdc/view?acc_num=osu1345401748.
Full textHannuksela, J. (Jari). "Camera based motion estimation and recognition for human-computer interaction." Doctoral thesis, University of Oulu, 2008. http://urn.fi/urn:isbn:9789514289781.
Full textKurz, Christian [Verfasser], and Hans-Peter [Akademischer Betreuer] Seidel. "Constrained camera motion estimation and 3D reconstruction / Christian Kurz. Betreuer: Hans-Peter Seidel." Saarbrücken : Saarländische Universitäts- und Landesbibliothek, 2014. http://d-nb.info/1063330734/34.
Full textHughes, Lloyd Haydn. "Enhancing mobile camera pose estimation through the inclusion of sensors." Thesis, Stellenbosch : Stellenbosch University, 2014. http://hdl.handle.net/10019.1/95917.
Full textENGLISH ABSTRACT: Monocular structure from motion (SfM) is a widely researched problem, however many of the existing approaches prove to be too computationally expensive for use on mobile devices. In this thesis we investigate how inertial sensors can be used to increase the performance of SfM algorithms on mobile devices. Making use of the low cost inertial sensors found on most mobile devices we design and implement an extended Kalman filter (EKF) to exploit their complementary nature, in order to produce an accurate estimate of the attitude of the device. We make use of a quaternion based system model in order to linearise the measurement stage of the EKF, thus reducing its computational complexity. We use this attitude estimate to enhance the feature tracking and camera localisation stages in our SfM pipeline. In order to perform feature tracking we implement a hybrid tracking algorithm which makes use of Harris corners and an approximate nearest neighbour search to reduce the search space for possible correspondences. We increase the robustness of this approach by using inertial information to compensate for inter-frame camera rotation. We further develop an efficient bundle adjustment algorithm which only optimises the pose of the previous three key frames and the 3D map points common between at least two of these frames. We implement an optimisation based localisation algorithm which makes use of our EKF attitude estimate and the tracked features, in order to estimate the pose of the device relative to the 3D map points. This optimisation is performed in two steps, the first of which optimises only the translation and the second optimises the full pose. We integrate the aforementioned three sub-systems into an inertial assisted pose estimation pipeline. We evaluate our algorithms with the use of datasets captured on the iPhone 5 in the presence of a Vicon motion capture system for ground truth data. We find that our EKF can estimate the device’s attitude with an average dynamic accuracy of ±5°. Furthermore, we find that the inclusion of sensors into the visual pose estimation pipeline can lead to improvements in terms of robustness and computational efficiency of the algorithms and are unlikely to negatively affect the accuracy of such a system. Even though we managed to reduce execution time dramatically, compared to typical existing techniques, our full system is found to still be too computationally expensive for real-time performance and currently runs at 3 frames per second, however the ever improving computational power of mobile devices and our described future work will lead to improved performance. From this study we conclude that inertial sensors make a valuable addition into a visual pose estimation pipeline implemented on a mobile device.
AFRIKAANSE OPSOMMING: Enkel-kamera struktuur-vanaf-beweging (structure from motion, SfM) is ’n bekende navorsingsprobleem, maar baie van die bestaande benaderings is te berekeningsintensief vir gebruik op mobiele toestelle. In hierdie tesis ondersoek ons hoe traagheidsensors gebruik kan word om die prestasie van SfM algoritmes op mobiele toestelle te verbeter. Om van die lae-koste traagheidsensors wat op meeste mobiele toestelle gevind word gebruik te maak, ontwerp en implementeer ons ’n uitgebreide Kalman filter (extended Kalman filter, EKF) om hul komplementêre geaardhede te ontgin, en sodoende ’n akkurate skatting van die toestel se postuur te verkry. Ons maak van ’n kwaternioon-gebaseerde stelselmodel gebruik om die meetstadium van die EKF te lineariseer, en so die berekeningskompleksiteit te verminder. Hierdie afskatting van die toestel se postuur word gebruik om die fases van kenmerkvolging en kameralokalisering in ons SfM proses te verbeter. Vir kenmerkvolging implementeer ons ’n hibriede volgingsalgoritme wat gebruik maak van Harris-hoekpunte en ’n benaderde naaste-buurpunt-soektog om die soekruimte vir moontlike ooreenstemmings te verklein. Ons verhoog die robuustheid van hierdie benadering, deur traagheidsinligting te gebruik om vir kamerarotasies tussen raampies te kompenseer. Verder ontwikkel ons ’n doeltreffende bondelaanpassingsalgoritme wat slegs optimeer oor die vorige drie sleutelraampies, en die 3D punte gemeenskaplik tussen minstens twee van hierdie raampies. Ons implementeer ’n optimeringsgebaseerde lokaliseringsalgoritme, wat gebruik maak van ons EKF se postuurafskatting en die gevolgde kenmerke, om die posisie en oriëntasie van die toestel relatief tot die 3D punte in die kaart af te skat. Die optimering word in twee stappe uitgevoer: eerstens net oor die kamera se translasie, en tweedens oor beide die translasie en rotasie. Ons integreer die bogenoemde drie sub-stelsels in ’n pyplyn vir postuurafskatting met behulp van traagheidsensors. Ons evalueer ons algoritmes met die gebruik van datastelle wat met ’n iPhone 5 opgeneem is, terwyl dit in die teenwoordigheid van ’n Vicon bewegingsvasleggingstelsel was (vir die gelyktydige opneming van korrekte postuurdata). Ons vind dat die EKF die toestel se postuur kan afskat met ’n gemiddelde dinamiese akkuraatheid van ±5°. Verder vind ons dat die insluiting van sensors in die visuele postuurafskattingspyplyn kan lei tot verbeterings in terme van die robuustheid en berekeningsdoeltreffendheid van die algoritmes, en dat dit waarskynlik nie die akkuraatheid van so ’n stelsel negatief beïnvloed nie. Al het ons die uitvoertyd drasties verminder (in vergelyking met tipiese bestaande tegnieke) is ons volledige stelsel steeds te berekeningsintensief vir intydse verwerking op ’n mobiele toestel en hardloop tans teen 3 raampies per sekonde. Die voortdurende verbetering van mobiele toestelle se berekeningskrag en die toekomstige werk wat ons beskryf sal egter lei tot ’n verbetering in prestasie. Uit hierdie studie kan ons aflei dat traagheidsensors ’n waardevolle toevoeging tot ’n visuele postuurafskattingspyplyn kan maak.
Fathollahi, Ghezelghieh Mona. "Estimation of Human Poses Categories and Physical Object Properties from Motion Trajectories." Scholar Commons, 2017. http://scholarcommons.usf.edu/etd/6835.
Full textAlmatrafi, Mohammed Mutlaq. "Optical Flow for Event Detection Camera." University of Dayton / OhioLINK, 2019. http://rave.ohiolink.edu/etdc/view?acc_num=dayton1576188397203882.
Full textLee, Hong Yun. "Deep Learning for Visual-Inertial Odometry: Estimation of Monocular Camera Ego-Motion and its Uncertainty." The Ohio State University, 2019. http://rave.ohiolink.edu/etdc/view?acc_num=osu156331321922759.
Full textEkström, Marcus. "Road Surface Preview Estimation Using a Monocular Camera." Thesis, Linköpings universitet, Datorseende, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-151873.
Full textJosefsson, Mattias. "3D camera with built-in velocity measurement." Thesis, Linköpings universitet, Datorseende, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-68632.
Full textI dagens industri används ofta 3D-kameror för att inspektera produkter. Kameran producerar en 3D-modell samt en intensitetsbild genom att sätta ihop en serie av profilbilder av objektet som erhålls genom lasertriangulering. I många av dessa uppställningar används en fysisk encoder som återspeglar hastigheten på till exempel transportbandet som produkten ligger på. Utan den här encodern kan bilden som kameran fångar bli förvrängd på grund av hastighetsvariationer. I det här examensarbetet presenteras en metod för att integrera funktionaliteten av encodern in i kamerans mjukvara. För att göra detta krävs att ett mönster placeras längs med objektet som ska bli skannat. Mönstret återfinns i bilden fångad av kameran och med hjälp av detta mönster kan hastigheten bestämmas och objektets korrekta proportioner återställas.
Lam, Hong Kwai. "Automatic white balancing for digital camera and fast motion vector re-estimation for arbitrary downscaling of compressed video /." View abstract or full-text, 2004. http://library.ust.hk/cgi/db/thesis.pl?ELEC%202004%20LAMH.
Full textIncludes bibliographical references (leaves 55-58). Also available in electronic version. Access restricted to campus users.
Fallqvist, Marcus. "Automatic Volume Estimation Using Structure-from-Motion Fused with a Cellphone's Inertial Sensors." Thesis, Linköpings universitet, Datorseende, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-144194.
Full textI rapporten framgår hur volymen av storskaliga objekt, nämligen grus-och stenhögar,kan bestämmas i utomhusmiljö med hjälp av en mobiltelefons kamerasamt interna sensorer som gyroskop och accelerometer. Projektet är beställt avEscenda Engineering med motivering att ersätta mer komplexa och resurskrävandesystem med ett enkelt handhållet instrument. Implementationen använderbland annat de vanligt förekommande datorseendemetoderna Kanade-Lucas-Tommasi-punktspårning, Struktur-från-rörelse och 3D-karvning tillsammans medenklare sensorfusion. I rapporten framgår att volymestimering är möjligt mennoggrannheten begränsas av sensorkvalitet och en bias.
Deigmoeller, Joerg. "Intelligent image cropping and scaling." Thesis, Brunel University, 2011. http://bura.brunel.ac.uk/handle/2438/4745.
Full textEfstratiou, Panagiotis. "Skeleton Tracking for Sports Using LiDAR Depth Camera." Thesis, KTH, Medicinteknik och hälsosystem, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-297536.
Full textSkelettspårning kan åstadkommas med hjälp av metoder för uppskattning av mänsklig pose. Djupinlärningsmetoder har visat sig vara det främsta tillvägagångssättet och om man använder en djupkamera med ljusdetektering och varierande omfång verkar det vara möjligt att utveckla ett markörlöst system för rörelseanalysmjukvara. I detta projekt används ett tränat neuralt nätverk för att spåra människor under sportaktiviteter och för att ge feedback efter biomekanisk analys. Implementeringar av fyra olika filtreringsmetoder för mänskliga rörelser presenteras, kalman filter, utjämnare med fast intervall, butterworth och glidande medelvärde. Mjukvaran verkar vara användbar vid fälttester för att utvärdera videor vid 30Hz. Detta visas genom analys av inomhuscykling och släggkastning. En ickestatisk kamera fungerar ganska bra vid mätningar av en stilla och upprättstående person. Det genomsnittliga absoluta felet är 8.32% respektive 6.46% då vänster samt höger knävinkel användes som referens. Ett felfritt system skulle gynna såväl idrottssom hälsoindustrin.
Szolgay, Daniel. "Video event detection and visual data pro cessing for multimedia applications." Thesis, Bordeaux 1, 2011. http://www.theses.fr/2011BOR14313/document.
Full textThis dissertation (i) describes an automatic procedure for estimating the stopping condition of non-regularized iterative deconvolution methods based on an orthogonality criterion of the estimated signal and its gradient at a given iteration; (ii) presents a decomposition method that splits the image into geometric (or cartoon) and texture parts using anisotropic diffusion with orthogonality based parameter estimation and stopping condition, utilizing the theory that the cartoon and the texture components of an image should be independent of each other; (iii) describes a method for moving foreground object extraction in sequences taken by wearable camera, with strong motion, where the camera motion compensated frame differencing is enhanced with a novel kernel-based estimation of the probability density function of the background pixels. The presented methods have been thoroughly tested and compared to other similar algorithms from the state-of-the-art
Zins, Matthieu. "Color Fusion and Super-Resolution for Time-of-Flight Cameras." Thesis, Linköpings universitet, Datorseende, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-141956.
Full textSchmidt, Eric. "Measuring the Speed of a Floorball Shot Using Trajectory Detection and Distance Estimation With a Smartphone Camera : Using OpenCV and Computer Vision on an iPhone to Detect the Speed of a Floorball Shot." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-190823.
Full textDušek, Stanislav. "Určení parametrů pohybu ze snímků kamery." Master's thesis, Vysoké učení technické v Brně. Fakulta elektrotechniky a komunikačních technologií, 2009. http://www.nusl.cz/ntk/nusl-217782.
Full textVerplaetse, Christopher James. "Inertial-optical motion-estimating camera for electronic cinematrography." Thesis, Massachusetts Institute of Technology, 1997. http://hdl.handle.net/1721.1/29128.
Full textZhang, Chunxiao. "Estimation of 3D human motion kinematics model from multiple cameras." Thesis, University of Central Lancashire, 2009. http://clok.uclan.ac.uk/19932/.
Full textOreifej, Omar. "Robust Subspace Estimation Using Low-Rank Optimization. Theory and Applications in Scene Reconstruction, Video Denoising, and Activity Recognition." Doctoral diss., University of Central Florida, 2013. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/5684.
Full textPh.D.
Doctorate
Electrical Engineering and Computing
Engineering and Computer Science
Computer Engineering
Lao, Yizhen. "3D Vision Geometry for Rolling Shutter Cameras." Thesis, Université Clermont Auvergne (2017-2020), 2019. http://www.theses.fr/2019CLFAC009/document.
Full textMany modern CMOS cameras are equipped with Rolling Shutter (RS) sensors which are considered as low cost, low consumption and fast cameras. In this acquisition mode, the pixel rows are exposed sequentially from the top to the bottom of the image. Therefore, images captured by moving RS cameras produce distortions (e.g. wobble and skew) which make the classic algorithms at best less precise, at worst unusable due to singularities or degeneracies. The goal of this thesis is to propose a general framework for modelling and solving structure from motion (SfM) with RS cameras. Our approach consists in addressing each sub-task of the SfM pipe-line (namely image correction, absolute and relative pose estimation and bundle adjustment) and proposing improvements.The first part of this manuscript presents a novel RS correction method which uses line features. Unlike existing methods, which uses iterative solutions and make Manhattan World (MW) assumption, our method R4C computes linearly the camera instantaneous-motion using few image features. Besides, the method was integrated into a RANSAC-like framework which enables us to detect curves that correspond to actual 3D straight lines and reject outlier curves making image correction more robust and fully automated.The second part revisits Bundle Adjustment (BA) for RS images. It deals with a limitation of existing RS bundle adjustment methods in case of close read-out directions among RS views which is a common configuration in many real-life applications. In contrast, we propose a novel camera-based RS projection algorithm and incorporate it into RSBA to calculate reprojection errors. We found out that this new algorithm makes SfM survive the degenerate configuration mentioned above.The third part proposes a new RS Homography matrix based on point correspondences from an RS pair. Linear solvers for the computation of this matrix are also presented. Specifically, a practical solver with 13 point correspondences is proposed. In addition, we present two essential applications in computer vision that use RS homography: plane-based RS relative pose estimation and RS image stitching. The last part of this thesis studies absolute camera pose problem (PnP) and SfM which handle RS effects by drawing analogies with non-rigid vision, namely Shape-from-Template (SfT) and Non-rigid SfM (NRSfM) respectively. Unlike all existing methods which perform 3D-2D registration after augmenting the Global Shutter (GS) projection model with the velocity parameters under various kinematic models, we propose to use local differential constraints. The proposed methods outperform stat-of-the-art and handles configurations that are critical for existing methods
Šimetka, Vojtěch. "3D Rekonstrukce historických míst z obrázků na Flickru." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2015. http://www.nusl.cz/ntk/nusl-234976.
Full textSvoboda, Ondřej. "Analýza vlastností stereokamery ZED ve venkovním prostředí." Master's thesis, Vysoké učení technické v Brně. Fakulta strojního inženýrství, 2019. http://www.nusl.cz/ntk/nusl-399416.
Full textMagerand, Ludovic. "Calcul de pose dynamique avec les caméras CMOS utilisant une acquisition séquentielle." Thesis, Clermont-Ferrand 2, 2014. http://www.theses.fr/2014CLF22534/document.
Full textComputer Vision, a field of Computer Science, is about extracting information from cameras. Their sensors can be produced using the CMOS technology which is widely used on mobile devices due to its low cost and volume. This technology allows a fast acquisition of an image by sequentially exposin the scan-line. However this method produces some deformation in the image if there is a motion between the camera and the filmed scene. This effect is known as Rolling Shutter and various methods have tried to remove these artifacts. Instead of correcting it, previous works have shown methods to extract information on the motion from this effect. These methods rely on a extension of the usual geometrical model of cameras by taking into account the sequential acquisition and the motion, supposed uniform, between the sensor and the scene. From this model, it’s possible to extend the usual pose estimation (estimation of position and orientation of the camera in the scene) to also estimate the motion parameters. Following on from this approach, we will present an extension to non-uniform motions based on a smoothing of the derivatives of the motion parameters. Afterwards, we will present a polynomial model of the Rolling Shutter and a global optimisation method to estimate the motion parameters. Well implemented, this enables to establish an automatic matching between the 3D model and the image. We will conclude with a comparison of all these methods using either simulated or real data
Hennebert, Christine. "Analyse d'une scène dynamique avec une caméra en mouvement : système de détection de cibles mobiles." Grenoble INPG, 1996. http://www.theses.fr/1996INPG0140.
Full textVestin, Albin, and Gustav Strandberg. "Evaluation of Target Tracking Using Multiple Sensors and Non-Causal Algorithms." Thesis, Linköpings universitet, Reglerteknik, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-160020.
Full text"Multiple camera pose estimation." Thesis, 2008. http://library.cuhk.edu.hk/record=b6074556.
Full textFurthermore, the process of estimating the rotation and translation parameters between a stereo pair from the essential matrix is investigated. This is an essential step for our multi-camera pose estimation method. We show that there are 16 solutions based on the singular value decomposition (not four or eight as previously thought). We also suggest a checking step to ascertain that the proposed algorithm will come up with accurate results. The checking step ensures the accuracy of the fundamental matrix calculated using the pose obtained. This provides a speedy way to calibrate a stereo rig. Our proposed theories are supported by the real and synthetic data experiments reported in this thesis.
In this thesis, we solve the pose estimation problem for robot motion by placing multiple cameras on the robot. In particular, we use four cameras arranged as two back-to-back stereo pairs combined with the Extended Kalman Filter (EKF). The EKF is used to provide a frame by frame recursive solution suitable for the real-time application at hand. The reason for using multiple cameras is that the pose estimation problem is more constrained for multiple cameras than for a single camera. Their use is further justified by the drop in price which is accompanied by the remarkable increase in accuracy. Back-to-back cameras are used since they are likely to have a larger field of view, provide more information, and capture more features. In this way, they are more likely to disambiguate the pose translation and rotation parameters. Stereo information is used in self-initialization and outlier rejection. Simple yet efficient methods have been proposed to tackle the problem of long-sequence drift. Our approaches have been compared, under different motion patterns, to other methods in the literature which use a single camera. Both the simulations and the real experiments show that our approaches are the most robust and accurate among them all as well as fast enough to realize the real-time requirement of robot navigation.
Mohammad Ehab Mohammad Ragab.
"April 2008."
Adviser: K. H. Wong.
Source: Dissertation Abstracts International, Volume: 70-03, Section: B, page: 1763.
Thesis (Ph.D.)--Chinese University of Hong Kong, 2008.
Includes bibliographical references (p. 138-148) and index.
Electronic reproduction. Hong Kong : Chinese University of Hong Kong, [2012] System requirements: Adobe Acrobat Reader. Available via World Wide Web.
Electronic reproduction. [Ann Arbor, MI] : ProQuest Information and Learning, [200-] System requirements: Adobe Acrobat Reader. Available via World Wide Web.
Abstracts in English and Chinese.
School code: 1307.
Li, Kun-Hung, and 李昆鴻. "Feature-Based Motion Estimation for Underwater Towed Camera." Thesis, 2013. http://ndltd.ncl.edu.tw/handle/20216010484193983786.
Full text國立中山大學
海下科技暨應用海洋物理研究所
102
When TowCam (Towed Camera System) is under the deep sea exploration, its moving track and heading always have different direction with degree of α caused by the current. This situation will make images shot by the camera mounted on the vehicle to difficultly be processed by the photo mosaicking method and get the deformed or unaccepted mosaic results. In this thesis, we used the SURF (speeded-up robust feature) algorithm to extract the critical point of the shot image to estimate the declination between the heading and the track for future development of visual servo control of TowCam. The images shot by the Twist-Pair TowCam (TP-TowCam) were also used to check the reliable of estimated algorithm in the filed experiment test and to find the best Hessian threshold value of SURF algorithm for the estimation. The results of algorithm showed the good estimation even in the turbid condition and the acceptable displacement of moving vehicle in the unit time was also increased, indicating the improvement of evaluation.
Lai, Chien-Ming, and 賴建明. "A Study on Camera Motion Estimation in MPEG Video." Thesis, 2007. http://ndltd.ncl.edu.tw/handle/97174028743945158542.
Full text國立成功大學
工程科學系碩博士班
95
Camera motion estimation is very important from a video processing view. Video application, including object segmentation and tracking, mosaic and content-based video indexing and retrieval, based on the video content should represent motion efficiently. Since, to decode MPEG video and estimate camera motion is extremely time-consuming. To obtain efficient estimates of camera motion by directly processing data easily obtained from MPEG-compressed video becomes more important. In this thesis, two main concerns of camera motion estimation from motion vector are addressed. Firstly, we introduce the connection from a GOP to next GOP by B-frame bidirectional reference. Secondly, we propose a 2-pass camera motion estimation to cope with outliers caused by the uncertainty of motion vectors directly from MPEG video format. Experiment results show that the 2-pass camera motion estimation method is reliable. Finally, we have developed a mosaic system by applying the proposed method.
Hsiao, Ching-Chun, and 蕭晴駿. "Model-Based Pose Estimation for Multi-Camera Motion Capture System." Thesis, 2008. http://ndltd.ncl.edu.tw/handle/79385950596603901992.
Full textLiu, Jen-Chu, and 劉仁竹. "Nonlinear filters for Single Camera Based Motion and Shape Estimation." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/15495303298361539559.
Full text國立臺灣海洋大學
輪機工程系
97
The accurate of estimating motion and shape of a moving object is a challenging task because of great variety of noise such as electronic component and influence of external environment etc. In order to alleviate the noise problem, this paper use kalman filter to reduces noise in the streaming video to get more accurate estimation in feature point on the object. Extended Kalman filter have to neglect higher-order derivatives in the nonlinear system to get linearization. However, it will cause unsteady condition. The Unscented Kalman filter uses a deterministic sampling approach to capture the mean and covariance estimates with a minimal set of sample points. UKF can be accurate to the second order for any nonlinearity, avoiding Jacobian’s computation. This paper uses a deterministic sampling of UKF with rigid body motion dynamics, shape dynamics, optical flow dynamics to estimate feature points on the moving object. Results obtained shows that UKF is accurate and reliable to get motion and shape of the object.
E, Yen-Chi, and 鄂彥齊. "Ego-motion Estimation Based on RGB-D Camera and Inertial Sensor." Thesis, 2015. http://ndltd.ncl.edu.tw/handle/01115649228044152260.
Full text國立臺灣大學
資訊網路與多媒體研究所
103
Ego-motion estimation has a wide variety of applications in robot control and automation. Proper local estimation of ego-motion benefits to recognize surrounding environment and recover the trajectory traversed for autonomous robot. In this thesis, we present a system that estimates ego-motion by fusing key frame based visual odometry and inertial measurements. The hardware of the system includes a RGB-D camera for capturing color and depth images and an Inertial Measurement Unit (IMU) for acquiring inertial measurements. Motion of camera between two consecutive images is estimated by finding correspondences of visual features. Rigidity constraints are used to efficiently remove outliers from a set of initial correspondence. Moreover, we apply random sample consensus (RANSAC) to handle the effect of the remaining outliers in the motion estimation step. These strategies are reasonable to insure that the remaining correspondences which involved in motion estimation almost contain inliers. Several experiments with different kind of camera movements are performed to show that the robustness and accuracy of the ego-motion estimation algorithm, and the ability of our system to handle the real scene data correctly.
Tsao, Tai-Chia, and 曹太嘉. "Omni-Directional Lane Detection and Vehicle Motion Estimation with Multi-Camera Vision." Thesis, 2008. http://ndltd.ncl.edu.tw/handle/vtbhp4.
Full text國立臺北科技大學
電機工程系研究所
96
Driver assistant system has been an active research topic in intelligent transportation systems. It could be used to provide surrounding dynamics. It can also be integrated into an adaptive cruise control system. This paper presents an omnidirectional lane detection and vehicle motion estimation system. The system employs a multi-camera vision and vehicle detection system to detect the surrounding vehicles. Based on the detected vehicles, the system uses inverse perspective projection and motion estimate techniques to estimate the motion of surrounding vehicles in Cartesian space and to build a bird’s eye view map. To make the vehicle motion estimation more accurate and reliable, Kalman filter, Extended Kalman filter, and particle filter have been used to estimate the motion of surrounding the vehicles. Meanwhile, the system integrates a lane detection technique to detect the lane lines and reconstruct them in Cartesian space to complete the bird’s eye view map with surrounding traffic. Our approach has been successfully validated in real traffic environments by performing experiments with 6 CCD cameras mounted onboard a highway vehicle.
Wan, Yi-Chen, and 萬宜蓁. "Algorithm and Architecture Design of RANSAC-Based Camera Motion Estimation for Action Recognition." Thesis, 2018. http://ndltd.ncl.edu.tw/handle/tj8vb7.
Full text國立臺灣大學
電子工程學研究所
106
Camera motion is one of the most challenges in action recognition in the real world. Improved Dense Trajectories introduced the concept of Warped Optical Flow which use RANSAC (Random Sample Consensus Algorithm) to approximate the camera motion and cancel out camera motion from the optical flow. The work significantly improved the performance of action recognition. This thesis proposes the real-time hardware architecture called Homography Estimation Processor using RANSAC for action recognition or other general systems. We use the SVD-based linear solver with the re-estimation method which provides the numerical stability and robust estimation. The system throughput is improved by hardware technique, including mesh-connected array architecture for SVD, pipelining and parallelization architectures. We also provide the flexibility using the folded technique in conjunction with parallelized architectures that can be easily configured to a different specification. The synthesis result of our proposed design using 40 nm TSMC technology is the gate count of 808.94 k and the SRAM size of 15.75 k Byte. The proposed Homography Estimation Processor can support 1000 matches and 2000 iterations with the throughput of at least 30 fps, and speeds up at least 73.17 times compared to software.
"Temporal Coding of Cortical Neural Signals and Camera Motion Estimation in Target Tracking." Doctoral diss., 2012. http://hdl.handle.net/2286/R.I.14715.
Full textDissertation/Thesis
Ph.D. Electrical Engineering 2012
Liou, Yun-Jung, and 劉允中. "An Improved CamShift Algorithm Based on Adaptive Motion Estimation for Multiple Camera Systems." Thesis, 2012. http://ndltd.ncl.edu.tw/handle/78721416340903036566.
Full text淡江大學
電機工程學系碩士班
100
Smart video surveillance has been developed for a long time, and many approaches to track moving objects have been proposed in recent years. The research of good tracking algorithms becomes one of the main streams for the smart video surveillance research. Multiple moving object tracking is a fundamental task on smart video surveillance systems, because it provides a focus of attention for further investigation. Video surveillance using multiple cameras system has attracted increasing interest in recent years. Moving objects occlusion is a key operation using correspondence between multiple cameras for surveillance system. In this thesis, the current state-of-the-art in moving objects tracker for multiple cameras surveillance has been surveyed. An efficient modified adaptive CamShift structure is proposed to further reduce the computational cost and increase the object tracking information in occlusion image. In this work, a new CamShift approach, directional prediction for adaptive CamShift algorithm, is proposed to improve the tracking accuracy rate. According to the characteristic of the center-based motion vector distribution for the real-world video sequence, this thesis employs an adaptive pattern (ASP) search to refine the central area search. Furthermore for estimation in large motion situations, the strategy of the adaptive CamShift search can preserve good performance. Experimental results indicate that the accuracy rate of the adaptive CamShift algorithm is better than that of the CamShift algorithm. Furthermore, the proposed method has given an average accuracy rate of 90%, and the operation speed can reach 12 FPS with frame size of 320
Diviš, Jiří. "Visual odometry from omnidirectional camera." Master's thesis, 2013. http://www.nusl.cz/ntk/nusl-328572.
Full textDiviš, Jiří. "Visual odometry from omnidirectional camera." Master's thesis, 2012. http://www.nusl.cz/ntk/nusl-305129.
Full textHong, Kai-Chen, and 洪楷宸. "The Study of Ego-motion Estimation for a Moving Object with Monocular Camera using Visual Odometry." Thesis, 2018. http://ndltd.ncl.edu.tw/handle/x64wny.
Full text國立交通大學
電控工程研究所
107
Visual odometry is the process of estimating the ego-motion of a moving object. In other words, visual odometry is the process of determining the position of a moving object. Then, the SLAM system is considered to be the best method for spatial positioning technology in the visual field. However, the SLAM system is quite large (the front-end: visual odometry, the back-end: optimization of the ego-motion estimation error), If the system need to perform other arithmetic processing at the same time, it will face challenges in terms of real-time. There are two contributions of this thesis. First, this thesis proposes an algorithm called image series from ego-motion estimation. Through the processing of the algorithm, even if the optimization of the ego-motion estimation error is not relied on by the back-end of the SLAM system, the estimation of the ego-motion of a moving object can be appropriately performed. Second, the system proposed in this paper can achieve a well balance between real-time, processing speed, lightness, and accuracy.
Chien, Shao-Yu, and 簡紹羽. "Background depth estimation based on camera motion and its application to 2D-to-3D stereoscopic video conversion." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/40125316147337511936.
Full text國立中正大學
電機工程所
97
The stereo display technology has become mature in recent years. The use of stereo display, in combination with 3D image/video contents, really makes the viewers immersed in reality. There are many methods for getting stereo video. A practical, low-cost, and fast way is to use the original 2-D video, in combination with depth conversion, to generate 3-D video. Therefore, in this thesis, we would like to develop a system to convert existing 2-D videos to their 3-D versions for stereo display. In this thesis, it is assumed that the camera has slow motion, which is used to estimate the disparity/depth information of static backgrounds. For the moving foreground objects, we combine several depth cues from the image to estimate their depth information. Finally, we fuse the background and foreground depth information to get the whole depth map. In our conversion system, we first detect region of moving objects for each image. For two images captured from two different time instants, we explore the spatial correlation to calculate the motion parallax information, and estimate a disparity plane to each segmented region. Based on the similarity of depth information in temporal domain, depth outliers can be corrected. For moving objects, we compute several depth cues, such as motion parallax, atmospheric perspective and texture gradients, by which the depth is assigned. Also we assume that the depth at the bottom of the foreground will be similar to the background depth at the same image position. Based on this, the depths of foreground and background can be fused to get the whole depth map. With the depth map, the stereo image pair can be synthesized for display. Experimental results show that the synthesized 3D video presents a good depth perception, but with slight depth discontinuity along vertical direction. Subject tests show that our synthesized 3D video and the real captured 3D video have similar depth perception.
Lou, Hui. "Acquiring 3D Full-body Motion from Noisy and Ambiguous Input." Thesis, 2012. http://hdl.handle.net/1969.1/ETD-TAMU-2012-05-11159.
Full textNetramai, Chayakorn [Verfasser]. "Using mobile multi-camera unit for real-time 3D motion estimation and map building of indoor environment / von Chayakorn Netramai." 2011. http://d-nb.info/1013088387/34.
Full textChiang, Yi, and 江懿. "A Prioritized Gauss-Seidel Method for Dense Correspondence Estimation and Motion Segmentation in Crowded Urban Areas with a Moving Depth Camera." Thesis, 2014. http://ndltd.ncl.edu.tw/handle/40104611508341776709.
Full text國立臺灣大學
資訊工程學研究所
102
Dense RGB-D video motion segmentation is an important preprocessing module in computer vision, image processing and robotics. A motion segmentation algorithm based on an optimization framework which utilizes depth information only is presented in this thesis. The proposed optimization framework segments and estimates rigid motion parameters of each locally rigid moving objects with coherent motion. The proposed method also calculates dense point correspondences while performing segmentation. An efficient numerical algorithm based on Constrained Block Nonlinear Gauss-Seidel (CNLGS) algorithm [1] and Prioritized Step Search [2] is proposed to solve the optimization problem. It classifies variables including point correspondences into groups and determines the ordering of variables to optimize. We prove the proposed numerical algorithm to converge to a theoretical bound. The proposed algorithm works well with a moving camera in highly dynamic urban scenes with non-rigid moving objects.
Tsao, An-Ting, and 曹安廷. "Ego-motion estimation using optical flow fields observed from multiple cameras." Thesis, 1996. http://ndltd.ncl.edu.tw/handle/53743692521793352486.
Full textChen, Yu-Liang, and 陳瑜靚. "Estimating Camera Position based on Structure from Motion and Image Retrieval Techniques." Thesis, 2016. http://ndltd.ncl.edu.tw/handle/3s496w.
Full text元智大學
資訊工程學系
104
Camera pose estimation is an important research topic in computer vision. It is a fundamental and crucial step in the applications of augmented reality. Knowing camera pose, we can edit the content of an image such as generating a virtual object in the image. Currently, many camera pose estimation techniques have been developed focusing on small and indoor environment. For large and outdoor environments, camera pose estimation faces some problems such as poor applicability and high computational complexity. Due to this, in this thesis we develop a two-stage camera position estimation system for large outdoor environments where the SFM (Structure from Motion) and an image retrieval technique are integrated in the system. In the stage of SFM, The SURF features in each image are extracted and the 3D structure of the scene are reconstructed. After that, a matching table which maps 2D image features to corresponding 3D points is created for subsequent camera pose estimation. In the stage of image retrieval, a method with high retrieval precision, the CNN (convolutional neural network), is employed to extract similar images from the dataset. In addition, the locality sensitive hashing algorithm is also included in the system to improve the retrieval efficiency. The features in the query image are extracted and matched with the images returned by the retrieval technique. A set of 2D and 3D matching points is established from the matching table and the camera pose is estimated. A series of experiments are conducted and results are encouraging. The camera pose of a query image in a large outdoor environment can be accurately estimated which demonstrates the feasibility of developed system.
Chen, Chih-Ting, and 陳芝婷. "Ego Motion Estimation in a Scene with Moving Objects Using Stereo Cameras." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/75105273156349053462.
Full textSchröder-Schetelig, Johannes. "Multimodal high-resolution mapping of contracting intact Langendorff-perfused hearts." Doctoral thesis, 2020. http://hdl.handle.net/21.11130/00-1735-0000-0005-1551-8.
Full textNdayikengurukiye, Didier. "Estimation de cartes d'énergie de hautes fréquences ou d'irrégularité de périodicité de la marche humaine par caméra de profondeur pour la détection de pathologies." Thèse, 2016. http://hdl.handle.net/1866/16178.
Full textThis work presents two new and simple human gait analysis systems based on a depth camera (Microsoft Kinect) placed in front of a subject walking on a conventional treadmill, capable of detecting a healthy gait from an impaired one. The first system presented relies on the fact that a normal walk typically exhibits a smooth motion (depth) signal, at each pixel with less high-frequency spectral energy content than an abnormal walk. This permits to estimate a map for that subject, showing the location and the amplitude of the high-frequency spectral energy (HFSE). The second system analyses the patient's body parts that have an irregular movement pattern, in terms of periodicity, during walking. Herein we assume that the gait of a healthy subject exhibits anywhere in the human body, during the walking cycles, a depth signal with a periodic pattern without noise. From each subject’s video sequence, we estimate a saliency color map showing the areas of strong gait irregularities also called aperiodic noise energy. Either the HFSE or aperiodic noise energy shown in the map can be used as a good indicator of possible pathology in an early, fast and reliable diagnostic tool or to provide information about the presence and extent of disease or (orthopedic, muscular or neurological) patient's problems. Even if the maps obtained are informative and highly discriminant for a direct visual classification, even for a non-specialist, the proposed systems allow us to automatically detect maps representing healthy individuals and those representing individuals with locomotor problems.