Dissertations / Theses on the topic 'Vision, Monocular'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Vision, Monocular.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Jama, Michal. "Monocular vision based localization and mapping." Diss., Kansas State University, 2011. http://hdl.handle.net/2097/8561.
Full textDepartment of Electrical and Computer Engineering
Balasubramaniam Natarajan
Dale E. Schinstock
In this dissertation, two applications related to vision-based localization and mapping are considered: (1) improving navigation system based satellite location estimates by using on-board camera images, and (2) deriving position information from video stream and using it to aid an auto-pilot of an unmanned aerial vehicle (UAV). In the first part of this dissertation, a method for analyzing a minimization process called bundle adjustment (BA) used in stereo imagery based 3D terrain reconstruction to refine estimates of camera poses (positions and orientations) is presented. In particular, imagery obtained with pushbroom cameras is of interest. This work proposes a method to identify cases in which BA does not work as intended, i.e., the cases in which the pose estimates returned by the BA are not more accurate than estimates provided by a satellite navigation systems due to the existence of degrees of freedom (DOF) in BA. Use of inaccurate pose estimates causes warping and scaling effects in the reconstructed terrain and prevents the terrain from being used in scientific analysis. Main contributions of this part of work include: 1) formulation of a method for detecting DOF in the BA; and 2) identifying that two camera geometries commonly used to obtain stereo imagery have DOF. Also, this part presents results demonstrating that avoidance of the DOF can give significant accuracy gains in aerial imagery. The second part of this dissertation proposes a vision based system for UAV navigation. This is a monocular vision based simultaneous localization and mapping (SLAM) system, which measures the position and orientation of the camera and builds a map of the environment using a video-stream from a single camera. This is different from common SLAM solutions that use sensors that measure depth, like LIDAR, stereoscopic cameras or depth cameras. The SLAM solution was built by significantly modifying and extending a recent open-source SLAM solution that is fundamentally different from a traditional approach to solving SLAM problem. The modifications made are those needed to provide the position measurements necessary for the navigation solution on a UAV while simultaneously building the map, all while maintaining control of the UAV. The main contributions of this part include: 1) extension of the map building algorithm to enable it to be used realistically while controlling a UAV and simultaneously building the map; 2) improved performance of the SLAM algorithm for lower camera frame rates; and 3) the first known demonstration of a monocular SLAM algorithm successfully controlling a UAV while simultaneously building the map. This work demonstrates that a fully autonomous UAV that uses monocular vision for navigation is feasible, and can be effective in Global Positioning System denied environments.
Cheda, Diego. "Monocular Depth Cues in Computer Vision Applications." Doctoral thesis, Universitat Autònoma de Barcelona, 2012. http://hdl.handle.net/10803/121644.
Full textDepth perception is a key aspect of human vision. It is a routine and essential visual task that the human do effortlessly in many daily activities. This has often been associated with stereo vision, but humans have an amazing ability to perceive depth relations even from a single image by using several monocular cues. In the computer vision field, if image depth information were available, many tasks could be posed from a different perspective for the sake of higher performance and robustness. Nevertheless, given a single image, this possibility is usually discarded, since obtaining depth information has frequently been performed by three-dimensional reconstruction techniques, requiring two or more images of the same scene taken from different viewpoints. Recently, some proposals have shown the feasibility of computing depth information from single images. In essence, the idea is to take advantage of a priori knowledge of the acquisition conditions and the observed scene to estimate depth from monocular pictorial cues. These approaches try to precisely estimate the scene depth maps by employing computationally demanding techniques. However, to assist many computer vision algorithms, it is not really necessary computing a costly and detailed depth map of the image. Indeed, just a rough depth description can be very valuable in many problems. In this thesis, we have demonstrated how coarse depth information can be integrated in different tasks following holistic and alternative strategies to obtain more precise and robustness results. In that sense, we have proposed a simple, but reliable enough technique, whereby image scene regions are categorized into discrete depth ranges to build a coarse depth map. Based on this representation, we have explored the potential usefulness of our method in three application domains from novel viewpoints: camera rotation parameters estimation, background estimation and pedestrian candidate generation. In the first case, we have computed camera rotation mounted in a moving vehicle from two novels methods that identify distant elements in the image, where the translation component of the image flow field is negligible. In background estimation, we have proposed a novel method to reconstruct the background by penalizing close regions in a cost function, which integrates color, motion, and depth terms. Finally, we have benefited of geometric and depth information available on single images for pedestrian candidate generation to significantly reduce the number of generated windows to be further processed by a pedestrian classifier. In all cases, results have shown that our depth-based approaches contribute to better performances.
Veldman, Kyle John. "Monocular vision for collision avoidance in vehicles." Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/101478.
Full textCataloged from PDF version of thesis.
Includes bibliographical references (page 21).
An experimental study facilitated by Ford Global Technologies, Inc. on the potential substitution of stereovision systems in car automation with monocular vision systems. The monocular system pairs a camera and passive lens with an active lens. Most active lenses require linear actuating systems to adjust the optical parameters of the system but this experiment employed an Optotune focus tunable lens adjusted by a Lorentz actuator for a much more reliable system. Tests were conducted in a lab environment to capture images of environmental objects at different distances from the system, pass those images through an image processing algorithm operating a high-pass filter to separate in-focus aspects of the image from out-of focus ones. Although the system is in the early phases of testing, monocular vision shows the ability to replace stereovision system. However, additional testing must be done to acclimate the apparatus to environmental factors, minimize the processing speed, and redesign the system for portability.
by Kyle John Veldman.
S.B.
Ng, Matthew James. "Corridor Navigation for Monocular Vision Mobile Robots." DigitalCommons@CalPoly, 2018. https://digitalcommons.calpoly.edu/theses/1856.
Full textPereira, Fabio Irigon. "High precision monocular visual odometry." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2018. http://hdl.handle.net/10183/183233.
Full textRecovering three-dimensional information from bi-dimensional images is an important problem in computer vision that finds several applications in our society. Robotics, entertainment industry, medical diagnose and prosthesis, and even interplanetary exploration benefit from vision based 3D estimation. The problem can be divided in two interdependent operations: estimating the camera position and orientation when each image was produced, and estimating the 3D scene structure. This work focuses on computer vision techniques, used to estimate the trajectory of a vehicle equipped camera, a problem known as visual odometry. In order to provide an objective measure of estimation efficiency and to compare the achieved results to the state-of-the-art works in visual odometry a high precision popular dataset was selected and used. In the course of this work new techniques for image feature tracking, camera pose estimation, point 3D position calculation and scale recovery are proposed. The achieved results outperform the best ranked results in the popular chosen dataset.
Goroshin, Rostislav. "Obstacle detection using a monocular camera." Thesis, Atlanta, Ga. : Georgia Institute of Technology, 2008. http://hdl.handle.net/1853/24697.
Full textBenoit, Stephen M. "Monocular optical flow for real-time vision systems." Thesis, McGill University, 1996. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=23862.
Full text李宏釗 and Wan-chiu Li. "Localization of a mobile robot by monocular vision." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2001. http://hub.hku.hk/bib/B31226371.
Full textMalan, Daniel Francois. "3D tracking between satellites using monocular computer vision." Thesis, Stellenbosch : University of Stellenbosch, 2005. http://hdl.handle.net/10019.1/3081.
Full textVisually estimating three-dimensional position, orientation and motion, between an observer and a target, is an important problem in computer vision. Solutions which compute threedimensional movement from two-dimensional intensity images, usually rely on stereoscopic vision. Some research has also been done in systems utilising a single (monocular) camera. This thesis investigates methods for estimating position and pose from monocular image sequences. The intended future application is of visual tracking between satellites flying in close formation. The ideas explored in this thesis build on methods developed for use in camera calibration, and structure from motion (SfM). All these methods rely heavily on the use of different variations of the Kalman Filter. After describing the problem from a mathematical perspective we develop different approaches to solving the estimation problem. The different approaches are successfully tested on simulated as well as real-world image sequences, and their performance analysed.
Li, Wan-chiu. "Localization of a mobile robot by monocular vision /." Hong Kong : University of Hong Kong, 2001. http://sunzi.lib.hku.hk/hkuto/record.jsp?B23765896.
Full textWatanabe, Yoko. "Stochastically optimized monocular vision-based navigation and guidance." Diss., Atlanta, Ga. : Georgia Institute of Technology, 2007. http://hdl.handle.net/1853/22545.
Full textCommittee Chair: Johnson, Eric; Committee Co-Chair: Calise, Anthony; Committee Member: Prasad, J.V.R.; Committee Member: Tannenbaum, Allen; Committee Member: Tsiotras, Panagiotis.
Salama, Gouda Ismail Mohamed. "Monocular and Binocular Visual Tracking." Diss., Virginia Tech, 1999. http://hdl.handle.net/10919/37179.
Full textPh. D.
Avanzini, Pierre. "Modélisation et commande d'un convoi de véhicules urbains par vision." Phd thesis, Université Blaise Pascal - Clermont-Ferrand II, 2010. http://tel.archives-ouvertes.fr/tel-00683626.
Full textFrost, Duncan. "Long range monocular SLAM." Thesis, University of Oxford, 2017. https://ora.ox.ac.uk/objects/uuid:af38cfa6-fc0a-48ab-b919-63c440ae8774.
Full textCheng, Kelvin. "Direct interaction with large displays through monocular computer vision." University of Sydney, 2009. http://hdl.handle.net/2123/5331.
Full textLarge displays are everywhere, and have been shown to provide higher productivity gain and user satisfaction compared to traditional desktop monitors. The computer mouse remains the most common input tool for users to interact with these larger displays. Much effort has been made on making this interaction more natural and more intuitive for the user. The use of computer vision for this purpose has been well researched as it provides freedom and mobility to the user and allows them to interact at a distance. Interaction that relies on monocular computer vision, however, has not been well researched, particularly when used for depth information recovery. This thesis aims to investigate the feasibility of using monocular computer vision to allow bare-hand interaction with large display systems from a distance. By taking into account the location of the user and the interaction area available, a dynamic virtual touchscreen can be estimated between the display and the user. In the process, theories and techniques that make interaction with computer display as easy as pointing to real world objects is explored. Studies were conducted to investigate the way human point at objects naturally with their hand and to examine the inadequacy in existing pointing systems. Models that underpin the pointing strategy used in many of the previous interactive systems were formalized. A proof-of-concept prototype is built and evaluated from various user studies. Results from this thesis suggested that it is possible to allow natural user interaction with large displays using low-cost monocular computer vision. Furthermore, models developed and lessons learnt in this research can assist designers to develop more accurate and natural interactive systems that make use of human’s natural pointing behaviours.
Kurdziel, Michael Scott. "A monocular color vision system for road intersection detection /." Online version of thesis, 2008. http://hdl.handle.net/1850/6208.
Full textMienie, Dewald. "Autonomous docking for a satellite pair using monocular vision." Thesis, Stellenbosch : University of Stellenbosch, 2009. http://hdl.handle.net/10019.1/2382.
Full textAutonomous rendezvouz and docking is seen as an enabling technology. It allows, among others, the construction of larger space platforms in-orbit and also provides a means for the in-orbit servicing of space vehicles. In this thesis a docking sequence is proposed and tested in both simulation and practice. This therefore also requires the design and construction of a test platform. A model hovercraft is used to emulate the chaser satellite in a 2-dimensional plane as it moves relatively frictionlessly. The hovercraft is also equipped with a single camera (monocular vision) that is used as the main sensor to estimate the target’s pose (relative position and orientation). An imitation of a target satellite was made and equipped with light markers that are used by the chaser’s camera sensor. The position of the target’s lights in the image is used to determine the target’s pose using a modified version ofMalan’s Extended Kalman Filter [20]. This information is then used during the docking sequence. This thesis successfully demonstrated the autonomous and reliable identification of the target’s lights in the image, and the autonomous docking of a satellite pair using monocular camera vision in both simulation and emulation.
Cheng, Kelvin. "0Direct interaction with large displays through monocular computer vision." Connect to full text, 2008. http://ses.library.usyd.edu.au/handle/2123/5331.
Full textTitle from title screen (viewed November 5, 2009). Submitted in fulfilment of the requirements for the degree of Doctor of Philosophy to the School of Information Technologies in the the Faculty of Engineering & Information Technologies. Degree awarded 2009; thesis submitted 2008. Includes bibliographical references. Also available in print form.
Schlachtman, Matthew Paul. "Monocular Vision and Image Correlation to Accomplish Autonomous Localization." DigitalCommons@CalPoly, 2010. https://digitalcommons.calpoly.edu/theses/320.
Full textKarlsson, Samuel. "Monocular vision-based obstacle avoidance for Micro Aerial Vehicles." Thesis, Luleå tekniska universitet, Institutionen för system- och rymdteknik, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:ltu:diva-80906.
Full textMagree, Daniel Paul. "Monocular vision-aided inertial navigation for unmanned aerial vehicles." Diss., Georgia Institute of Technology, 2015. http://hdl.handle.net/1853/53892.
Full textSpencer, Lisa. "REAL-TIME MONOCULAR VISION-BASED TRACKING FOR INTERACTIVE AUGMENTED REALITY." Doctoral diss., University of Central Florida, 2006. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/4289.
Full textPh.D.
School of Computer Science
Engineering and Computer Science
Computer Science
Murali, Vidya N. "Autonomous navigation and mapping using monocular low-resolution grayscale vision." Connect to this title online, 2008. http://etd.lib.clemson.edu/documents/1219852130/.
Full textKang, Changkoo. "Small UAV Trajcetory Prediction and Avoidance using Monocular Computer Vision." Thesis, Virginia Tech, 2017. http://hdl.handle.net/10919/79950.
Full textMaster of Science
Gampher, John Eric. "Perception of motion-in-depth induced motion effects on monocular and binocular cues /." Birmingham, Ala. : University of Alabama at Birmingham, 2008. https://www.mhsl.uab.edu/dt/2009r/gampher.pdf.
Full textTitle from PDF title page (viewed Mar. 30, 2010). Additional advisors: Franklin R. Amthor, James E. Cox, Timothy J. Gawne, Rosalyn E. Weller. Includes bibliographical references (p. 104-114).
Agarwal, Saurav. "Monocular vision based indoor simultaneous localisation and mapping for quadrotor platform." Thesis, Cranfield University, 2012. http://dspace.lib.cranfield.ac.uk/handle/1826/7210.
Full textTournier, Glenn P. (Glenn Paul). "Six degrees of freedom estimation using monocular vision and moiré patterns." Thesis, Massachusetts Institute of Technology, 2006. http://hdl.handle.net/1721.1/37951.
Full textIncludes bibliographical references (p. 105-107).
We present the vision-based estimation of the position and orientation of an object using a single camera relative to a novel target that incorporates the use of moire patterns. The objective is to acquire the six degree of freedom estimation that is essential for the operation of vehicles in close proximity to other craft and landing platforms. A target contains markers to determine relative orientation and locate two sets of orthogonal moire patterns at two different frequencies. A camera is mounted on a small vehicle with the target in the field of view. An algorithm processes the images extracting the attitude and position information of the camera relative to the target utilizing geometry and 4 single-point discrete Fourier transforms (DFTs) on the moire patterns. Manual and autonomous movement tests are conducted to determine the accuracy of the system relative to ground truth locations obtained through an external indoor positioning system. Position estimations with accompanying control techniques have been implemented including hovering, static platform landings, and dynamic platform landings to display the algorithm's ability to provide accurate information to precisely control the vehicle. The results confirm the moire target system's feasibility as a viable option for low-cost relative navigation for indoor and outdoor operations including landing on static and dynamic surfaces.
by Glenn P. Tournier.
S.M.
Mercado-Ravell, Diego Alberto. "Autonomous navigation and teleoperation of unmanned aerial vehicles using monocular vision." Thesis, Compiègne, 2015. http://www.theses.fr/2015COMP2239/document.
Full textThe present document addresses, theoretically and experimentally, the most relevant topics for Unmanned Aerial Vehicles (UAVs) in autonomous and semi-autonomous navigation. According with the multidisciplinary nature of the studied problems, a wide range of techniques and theories are covered in the fields of robotics, automatic control, computer science, computer vision and embedded systems, among others. As part of this thesis, two different experimental platforms were developed in order to explore and evaluate various theories and techniques of interest for autonomous navigation. The first prototype is a quadrotor specially designed for outdoor applications and was fully developed in our lab. The second testbed is composed by a non expensive commercial quadrotor kind AR. Drone, wireless connected to a ground station equipped with the Robot Operating System (ROS), and specially intended to test computer vision algorithms and automatic control strategies in an easy, fast and safe way. In addition, this work provides a study of data fusion techniques looking to enhance the UAVs pose estimation provided by commonly used sensors. Two strategies are evaluated in particular, an Extended Kalman Filter (EKF) and a Particle Filter (PF). Both estimators are adapted for the system under consideration, taking into account noisy measurements of the UAV position, velocity and orientation. Simulations show the performance of the developed algorithms while adding noise from real GPS (Global Positioning System) measurements. Safe and accurate navigation for either autonomous trajectory tracking or haptic teleoperation of quadrotors is presented as well. A second order Sliding Mode (2-SM) control algorithm is used to track trajectories while avoiding frontal collisions in autonomous flight. The time-scale separation of the translational and rotational dynamics allows us to design position controllers by giving desired references in the roll and pitch angles, which is suitable for quadrotors equipped with an internal attitude controller. The 2-SM control allows adding robustness to the closed-loop system. A Lyapunov based analysis probes the system stability. Vision algorithms are employed to estimate the pose of the vehicle using only a monocular SLAM (Simultaneous Localization and Mapping) fused with inertial measurements. Distance to potential obstacles is detected and computed using the sparse depth map from the vision algorithm. For teleoperation tests, a haptic device is employed to feedback information to the pilot about possible collisions, by exerting opposite forces. The proposed strategies are successfully tested in real-time experiments, using a low-cost commercial quadrotor. Also, conception and development of a Micro Aerial Vehicle (MAV) able to safely interact with human users by following them autonomously, is achieved in the present work. Once a face is detected by means of a Haar cascade classifier, it is tracked applying a Kalman Filter (KF), and an estimation of the relative position with respect to the face is obtained at a high rate. A linear Proportional Derivative (PD) controller regulates the UAV’s position in order to keep a constant distance to the face, employing as well the extra available information from the embedded UAV’s sensors. Several experiments were carried out through different conditions, showing good performance even under disadvantageous scenarios like outdoor flight, being robust against illumination changes, wind perturbations, image noise and the presence of several faces on the same image. Finally, this thesis deals with the problem of implementing a safe and fast transportation system using an UAV kind quadrotor with a cable suspended load. The objective consists in transporting the load from one place to another, in a fast way and with minimum swing in the cable
Ferrera, Maxime. "Monocular Visual-Inertial-Pressure fusion for Underwater localization and 3D mapping." Thesis, Montpellier, 2019. http://www.theses.fr/2019MONTS089.
Full textThis thesis addresses the problem of real-time 3D localization and mapping in underwater environments.In the underwater archaeology field, Remotely Operated Vehicles (ROVs) are used to conduct deep-seasurveys and excavations. Providing both accurate localization and mapping information in real-time iscrucial for manual or automated piloting of the robots. While many localization solutions already existfor underwater robots, most of them rely on very accurate sensors, such as Doppler velocity logs or fiberoptic gyroscopes, which are very expensive and may be too bulky for small ROVs. Acoustic positioningsystems are also commonly used for underwater positioning, but they provide low frequencymeasurements, with limited accuracy.In this thesis, we study the use of low-cost sensors for accurate underwater localization. Our studyinvestigates the use of a monocular camera, a pressure sensor and a low-cost MEMS-IMU as the onlymeans of performing localization and mapping in the context of underwater archaeology.We have conducted an evaluation of different features tracking methods on images affected by typicaldisturbances met in an underwater context. From the results obtained with this evaluation, we havedeveloped a monocular Visual SLAM (Simultaneous Localization and Mapping) method, robust to thespecific disturbances of underwater environments. Then, we propose an extension of this method totightly integrate the measurements of a pressure sensor and an IMU in the SLAM algorithm. The finalmethod provides a very accurate localization and runs in real-time. In addition, an online dense 3Dreconstruction module, compliant with a monocular setup, is also proposed. Two lightweight and compactprototypes of this system have been designed and used to record datasets that have been publiclyreleased. Furthermore, these prototypes have been successfully used to test and validate the proposedlocalization and mapping algorithms in real-case scenarios
Xie, Bingqian. "Lane Departure and Front Collision Warning System Using Monocular and Stereo Vision." Digital WPI, 2015. https://digitalcommons.wpi.edu/etd-theses/274.
Full textRandell, Charles James. "3D underwater monocular machine vision from 2D images in an attenuating medium." Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1997. http://www.collectionscanada.ca/obj/s4/f2/dsk2/ftp02/NQ32764.pdf.
Full textSwart, Andre Dewald. "Monocular vision assisted autonomous landing of a helicopter on a moving deck." Thesis, Stellenbosch : Stellenbosch University, 2013. http://hdl.handle.net/10019.1/80134.
Full textENGLISH ABSTRACT: The landing phase of any helicopter is the most critical part of the whole flight envelope, particularly on a moving flight deck. The flight deck is usually located at the stern of the ship, translating to large heave motions. This thesis focuses on the three fundamental components required for a successful landing: accurate, relative state-estimation between the helicopter and the flight deck; a prediction horizon to forecast suitable landing opportunities; and excellent control to safely unite the helicopter with the flight deck. A monocular-vision sensor node was developed to provide accurate, relative position and attitude information of the flight deck. The flight deck is identified by a distinct, geometric pattern. The relative states are combined with the onboard, kinematic state-estimates of the helicopter to provide an inertial estimate of the flight deck states. Onboard motion prediction is executed to forecast a possible safe landing time which is conveyed to the landing controller. Camera pose-estimation tests and hardware-in-the-loop simulations proved the system developed in this thesis viable for flight tests. The practical flight tests confirmed the success of the monocular-vision sensor node.
AFRIKAANSE OPSOMMING: Die mees kritiese deel van die hele vlug-duurte van ’n helikopter is die landings-fase, veral op ’n bewegende vlugdek. Die vlugdek is gewoonlik geleë aan die agterstewe-kant van die skip wat groot afgee bewegings mee bring. Hierdie tesis ondersoek die drie fundamentele komponente van ’n suksesvolle landing: akkurate, relatiewe toestand-beraming tussen die helikopter en die vlugdek; ’n vooruitskatting horison om geskikte landings geleenthede te voorspel; en uitstekended beheer om die helikopter en vlugdek veilig te verenig. ’n Monokulêre-visie sensor-nodus was ontwikkel om akkurate, relatiewe-posisie en oriëntasie informasie van die vlugdek te verwerf. Die vlugdek is geidentifiseer deur ’n kenmerkende, geometriese patroon. Die relatiewe toestande word met die aan-boord kinematiese toestandafskatter van die helikopter gekombineer, om ’n beraming van die inertiale vlugdek-toestande te verskaf. Aan-boord beweging-vooruitskatting is uitgevoer om moontlike, veilige landingstyd te voorspel en word teruggevoer na die landingsbeheerder. Kamera-orientasie afskat-toetse en hardeware-in-die-lus simulasies het die ontwikkelde sisteem van hierdie tesis lewensvatbaar vir vlug-toetse bewys. Praktiese vlug-toetse het die sukses van die monokulêre-visie sensor-nodus bevestig.
Diskin, Yakov. "Dense 3D Point Cloud Representation of a Scene Using Uncalibrated Monocular Vision." University of Dayton / OhioLINK, 2013. http://rave.ohiolink.edu/etdc/view?acc_num=dayton1366386933.
Full textKatramados, Ioannis. "Real-time object detection using monocular vision for low-cost automotive sensing systems." Thesis, Cranfield University, 2013. http://dspace.lib.cranfield.ac.uk/handle/1826/10386.
Full textGarg, Ravi. "Dense motion capture of deformable surfaces from monocular video." Thesis, Queen Mary, University of London, 2013. http://qmro.qmul.ac.uk/xmlui/handle/123456789/8823.
Full textNassir, Cesar. "Domain-Independent Moving Object Depth Estimation using Monocular Camera." Thesis, KTH, Robotik, perception och lärande, RPL, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-233519.
Full textI dag strävar bilföretag över hela världen för att skapa fordon med helt autonoma möjligheter. Det finns många fördelar med att utveckla autonoma fordon, såsom minskad trafikstockning, ökad säkerhet och minskad förorening, etc. För att kunna uppnå det målet finns det många utmaningar framåt, en av dem är visuell uppfattning. Att kunna uppskatta djupet från en 2D-bild har visat sig vara en nyckelkomponent för 3D-igenkännande, rekonstruktion och segmentering. Att kunna uppskatta djupet i en bild från en monokulär kamera är ett svårt problem eftersom det finns tvetydighet mellan kartläggningen från färgintensitet och djupvärde. Djupestimering från stereobilder har kommit långt jämfört med monokulär djupestimering och var ursprungligen den metod som man har förlitat sig på. Att kunna utnyttja monokulära bilder är dock nödvändig för scenarier när stereodjupuppskattning inte är möjligt. Vi har presenterat ett nytt nätverk, BiNet som är inspirerat av ENet, för att ta itu med djupestimering av rörliga objekt med endast en monokulär kamera i realtid. Det fungerar bättre än ENet med datasetet Cityscapes och lägger bara till en liten kostnad på komplexiteten.
Repo, T. (Tapio). "Modeling of structured 3-D environments from monocular image sequences." Doctoral thesis, University of Oulu, 2002. http://urn.fi/urn:isbn:9514268571.
Full textSköld, Jonas. "Estimating 3D-trajectories from Monocular Video Sequences." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-172919.
Full textAtt spåra ett objekt i rörelse och rekonstruera dess bana kan göras med ett stereokamerasystem, eftersom de två kamerorna möjliggör djupseende. Ett sådant system skulle dock inte fungera om en av kamerorna misslyckas med att detektera objektet. Om det händer skulle det vara fördelaktigt om systemet ändå kunde använda den fungerande kameran för att göra en approximativ rekonstruktion av banan. I den här studien har jag undersökt hur tidigare observationer från ett stereosystem kan användas för att rekonstruera banor när video från enbart en av kamerorna är tillgänglig. Ett flertal metoder har implementerats och testats, med varierande resultat. Den bästa metoden visade sig vara en närmaste-grannar-sökning optimerad med ett Kalman-filter. På en testmängd bestående av 10000 golfslag kunde algoritmen skapa uppskattningar som i genomsnitt skiljde sig 3.5 meter från den korrekta banan, med bättre resultat för banor som startat nära kameran.
Ekström, Marcus. "Road Surface Preview Estimation Using a Monocular Camera." Thesis, Linköpings universitet, Datorseende, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-151873.
Full textSchultz, Kevin P. "Exploration of the crosslinks between saccadic and vergence eye movement pathways using motor and visual perturbations." Thesis, Birmingham, Ala. : University of Alabama at Birmingham, 2010. https://www.mhsl.uab.edu/dt/2010p/schultz.pdf.
Full textBayle, Elodie. "Entre fusion et rivalité binoculaires : impact des caractéristiques des stimuli visuels lors de l’utilisation d’un système de réalité augmentée semi-transparent monoculaire." Electronic Thesis or Diss., université Paris-Saclay, 2021. http://www.theses.fr/2021UPASG029.
Full textMonocular augmented reality devices are used in the aeronautical field to enhance pilots' vision by providing access to essential information such as flight symbology. They are lighter and more adjustable than their binocular counterparts, can be integrated into any aircraft, and allow information to be retained regardless of gaze direction. However, they generate a particular perception since a monocular virtual image is superimposed on the real binocular environment. Different information is projected to corresponding regions of the two eyes creating an interocular conflict. The goal of this thesis is to evaluate the impact of the stimuli characteristics on the performance of tasks performed with this type of system to optimize its use. Two psychophysical studies and an ecological study in a flight simulator have been carried out. All of them showed a good comfort when exposed to interocular conflict. The performances were evaluated according to the characteristics of the binocular background, the display of the monocular image and the characteristics of events to be detected. The choice of the presenting eye is not insignificant given the differences between the performances achieved with the monocular on each of the two eyes. Our results from the three studies also show that, as with two fusible or two dichoptic images, performance is dependent on visual stimuli. They therefore suggest that an adaptive symbology should be considered, which cannot be summarized by the change in brightness currently available to pilots
Spaenlehauer, Ariane. "Decentralized monocular-inertial multi-UAV SLAM system." Thesis, Compiègne, 2019. http://www.theses.fr/2019COMP2494.
Full textIn this thesis, we provide a scheme for localization of a fleet of autonomous UAVs (unmanned autonomous vehicles) within a Technological System-of-Systems architecture. Specifically, we aim for a fleet of autonomous UAVs to localize themselves and to obtain a map of an unknown environment using a minimal set of sensors on each UAV: A front monocular camera and an Inertial Measurement Unit. This is a critically important problem for applications such as exploration of unknown areas, or search and rescue missions. The choices for designing such a system are supported by an extensive study of the scientific literature on two broad fronts: First, about the multi-robot systems performing localization, mapping, navigation and exploration, and second, about the monocular, real-time and inertial-monocular SLAM (Simultaneous Localization and Mapping) algorithms. Processing monocular camera frames suffers the drawback of lacking the capability of providing metric estimates as the depth dimension is lost when the frames are photographed by the camera. Although, it is usually not a critical problem for single-robot systems, having accurate metric estimates is required for multi-robot systems. This requirement becomes critical if the system is designed for control, navigation and exploration purposes. In this thesis, we provide a novel approach to make the outputs of monocular SLAM algorithms metric through a loosely-coupled fusion scheme by using the inertial measurements. This work also explores a design for a fleet of UAVs to localize each robot with minimal requirements: No a priori knowledge about the environment, information about neither the position nor the moment in time the UAV takes off and land is required. Moreover, the system presented in the thesis handles aggressive UAV trajectories having dramatic changes in speed and altitude. In multi-robot systems, the question of the coordinate frames require more attention than in single robot systems. In many studies, the coordinate frame problem is simplified to the representation of the fleet and the expression of the measurements in a global coordinate frame. However, this kind of hypothesis implies either the use of additional sensors to be able to measure the transformations to the global coordinate frame or additional experimental constraints, for example about the starting position of the robots. Our system does not require absolute measurements like GNSS positioning or knowledge about the coordinate frame of each UAV. As each UAV of the fleet estimates its location and produces a map in its own coordinate frame, relations between those coordinate frames are found by our scheme. For that purpose, we extend the well known concept of loop-closures in single-robot SLAM approaches, to multi-robot systems. In this research work, we also provide an overview of the new effects due to the extended definition of loop-closures we provide in comparison with the loop-closures scheme that can be found in single robot SLAM algorithms. In addition to the coordinate frame problem, we provide experimental results about the possibilities for improving the location estimate of a fleet by considering the places visited by several UAVs. By searching for similar places using each UAV imagery, using the 2-D information encapsulated in the images of the same sceneryfrom different view points, and the 3-D map locally estimated by each UAV, we add new constraints to the SLAM problem that is the main scheme that can be used to improve the UAV location estimates. We included experiments to assess the accuracy of the inter-UAV location estimation. The system was tested using datasets with measurements recorded on board UAVs in similar conditions as the ones we target
Herdtweck, Christian [Verfasser], and Heinrich [Akademischer Betreuer] Bülthoff. "Learning Data-Driven Representations for Robust Monocular Computer Vision Applications / Christian Herdtweck ; Betreuer: Heinrich Bülthoff." Tübingen : Universitätsbibliothek Tübingen, 2014. http://d-nb.info/1162897317/34.
Full textFormankiewicz, Monika Anna. "The psychophysics of lustre and the use of monocular filters to treat colour vision deficiencies." Thesis, University of Cambridge, 2005. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.615264.
Full textMunn, Susan M. "3D head motion, point-of-regard and encoded gaze fixations in real scenes : next-generation portable video-based monocular eye tracking /." Online version of thesis, 2009. http://hdl.handle.net/1850/11206.
Full textBarra, Roberto José Giordano. "Combinação de visão monocular e sonares esparsos para a localização de robôs móveis." Universidade de São Paulo, 2007. http://www.teses.usp.br/teses/disponiveis/3/3141/tde-13072007-164026/.
Full textA key component of a mobile robot system is the ability to localize itself accurately, which involves estimating its pose with respect to some global representation of space. The general specification of a sensor-based localization approach starts with an initial estimate of the robot\'s pose and uses sensor data in conjunction with a map to produce a refined pose estimate that has an increased confidence about the true pose of the robot. One of the main difficulties is that sensor data is corrupted by measurement errors. These errors can arise from noise, quantization, digitalization artifacts, wheel slippage, and other such sources. Different sensors measure different physical properties, which are corrupted by different sources of measurement errors. The use of data from multiple sensors provides redundant and complementary information that can be processed to obtain a combined estimate aiming at an increase in the confidence of the final pose estimate. In this work we propose ELViS, a system that estimates the localization of a mobile robot equipped with odometers, a video camera and a frontal semi-ring of 8 sonar sensors, and that operates successfully in stationary and structured indoor environments. It is assumed that the robot navigates on flat surfaces and that straight lines can be identified in the environment image acquired by the camera. To increase selectivity of the landmarks and reduce computational complexity in data processing and matching to the map, environment features are represented using minimalist models in the map. This allows the use of ELViS in a large number of applications where tight budget or execution time constraints exist. ELViS has been implemented and tested using two estimators based on the Kalman Filter. The results, obtained with the real robots and in series of simulation runs, indicate promising directions.
Aguilar-Gonzalez, Abiel. "Monocular-SLAM dense mapping algorithm and hardware architecture for FPGA acceleration." Thesis, Université Clermont Auvergne (2017-2020), 2019. http://www.theses.fr/2019CLFAC055.
Full textSimultaneous Localization and Mapping (SLAM) is the problem of constructing a 3D map while simultaneously keeping track of an agent location within the map. In recent years, work has focused on systems that use a single moving camera as the only sensing mechanism (monocular-SLAM). This choice was motivated because nowadays, it is possible to find inexpensive commercial cameras, smaller and lighter than other sensors previously used and, they provide visual environmental information that can be exploited to create complex 3D maps while camera poses can be simultaneously estimated. Unfortunately, previous monocular-SLAM systems are based on optimization techniques that limits the performance for real-time embedded applications. To solve this problem, in this work, we propose a new monocular SLAM formulation based on the hypothesis that it is possible to reach high efficiency for embedded applications, increasing the density of the point cloud map (and therefore, the 3D map density and the overall positioning and mapping) by reformulating the feature-tracking/feature-matching process to achieve high performance for embedded hardware architectures, such as FPGA or CUDA. In order to increase the point cloud map density, we propose new feature-tracking/feature-matching and depth-from-motion algorithms that consists of extensions of the stereo matching problem. Then, two different hardware architectures (based on FPGA and CUDA, respectively) fully compliant for real-time embedded applications are presented. Experimental results show that it is possible to obtain accurate camera pose estimations. Compared to previous monocular systems, we are ranked as the 5th place in the KITTI benchmark suite, with a higher processing speed (we are the fastest algorithm in the benchmark) and more than x10 the density of the point cloud from previous approaches
Spencer, Justina. "Peeping in, peering out : monocularity and early modern vision." Thesis, University of Oxford, 2014. https://ora.ox.ac.uk/objects/uuid:b8854565-ce57-4c83-9cdb-64249d171142.
Full textKalghatgi, Roshan Satish. "Reconstruction techniques for fixed 3-D lines and fixed 3-D points using the relative pose of one or two cameras." Thesis, Georgia Institute of Technology, 2012. http://hdl.handle.net/1853/43590.
Full textAlmanza-Ojeda, Dora Luz. "Détection et suivi d'objets mobiles perçus depuis un capteur visuel embarqué." Phd thesis, Université Paul Sabatier - Toulouse III, 2011. http://tel.archives-ouvertes.fr/tel-01017785.
Full text