Dissertations / Theses on the topic 'Computer vision technology'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Computer vision technology.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Johansson, Björn. "Multiscale Curvature Detection in Computer Vision." Licentiate thesis, Linköping University, Linköping University, Computer Vision, 2001. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-54966.
Full textThis thesis presents a new method for detection of complex curvatures such as corners, circles, and star patterns. The method is based on a second degree local polynomial model applied to a local orientation description in double angle representation. The theory of rotational symmetries is used to compute curvature responses from the parameters of the polynomial model. The responses are made more selective using a scheme of inhibition between different symmetry models. These symmetries can serve as feature points at a high abstraction level for use in hierarchical matching structures for 3D estimation, object recognition, image database search, etc.
A very efficient approximative algorithm for single and multiscale polynomial expansion is developed, which is used for detection of the complex curvatures in one or several scales. The algorithm is based on the simple observation that polynomial functions multiplied with a Gaussian function can be described in terms of partial derivatives of the Gaussian. The approximative polynomial expansion algorithm is evaluated in an experiment to estimate local orientation on 3D data, and the performance is comparable to previously tested algorithms which are more computationally expensive.
The curvature algorithm is demonstrated on natural images and in an object recognition experiment. Phase histograms based on the curvature features are developed and shown to be useful as an alternative compact image representation.
The importance of curvature is furthermore motivated by reviewing examples from biological and perceptual studies. The usefulness of local orientation information to detect curvature is also motivated by an experiment about learning a corner detector.
Bårman, Håkan. "Hierarchical curvature estimation in computer vision." Doctoral thesis, Linköpings universitet, Bildbehandling, 1991. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-54887.
Full textMoe, Anders. "Passive Aircraft Altitude Estimation using Computer Vision." Licentiate thesis, Linköping University, Linköping University, Computer Vision, 2000. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-53415.
Full textThis thesis presents a number of methods to estimate 3D structures with a single translating camera. The camera is assumed to be calibrated and to have a known translation and rotation.
Applications for aircraft altitude estimation and ground structure estimation ahead of the aircraft are discussed. The idea is to mount a camera on the aircraft and use the motion estimates obtained in the inertia navigation system. One reason for this arrangement is to make the aircraft more passive, in comparison to conventional radar based altitude estimation.
Two groups of methods are considered, optical flow based and region tracking based. Both groups have advantages and drawbacks.
Two methods to estimate the optical flow are presented. The accuracy of the estimated ground structure is increased by varying the temporal distance between the frames used in the optical flow estimation algorithms.
Four region tracking algorithms are presented. Two of them use canonical correlation and the other two are based on sum of squared difference and complex correlation respectively.
The depth estimates are then temporally filtered using weighted least squares or a Kalman filter.
A simple estimation of the computational complexity and memory requirements for the algorithms is presented to aid estimation of the hardware requirements.
Tests on real flight sequences are performed, showing that the aircraft altitude can be estimated with a good accuracy.
Söderkvist, Oskar. "Computer Vision Classification of Leaves from Swedish Trees." Thesis, Linköping University, Linköping University, Computer Vision, 2001. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-54366.
Full textThe aim of this master thesis is to classify the tree class from an image of a leaf with a computer vision classiffication system. We compare different descriptors that will describe the leaves different features. We will also look at different classiffication models and combine them with the descriptors to build a system hat could classify the different tree classes.
Johansson, Björn. "Low Level Operations and Learning in Computer Vision." Doctoral thesis, Linköpings universitet, Bildbehandling, 2004. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-24005.
Full textKlomark, Marcus. "Occupant Detection using Computer Vision." Thesis, Linköping University, Linköping University, Computer Vision, 2000. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-54363.
Full textThe purpose of this master’s thesis was to study the possibility to use computer vision methods to detect and classify objects in the front passenger seat in a car. This work presents different approaches to solve this problem and evaluates the usefulness of each technique. The classification information should later be used to modulate the speed and the force of the airbag, to be able to provide each occupant with optimal protection and safety.
This work shows that computer vision has a great potential in order to provide data, which may be used to perform reliable occupant classification. Future choice of method to use depends on many factors, for example costs and requirements on the system from laws and car manufacturers. Further, evaluation and tests of the methods in this thesis, other methods, the ABE approach and post-processing of the results should also be made before a reliable classification algorithm may be written.
Fang, Jian. "Optical Imaging and Computer Vision Technology for Corn Quality Measurement." OpenSIUC, 2011. https://opensiuc.lib.siu.edu/theses/733.
Full textAndersson, Mats T. "Controllable Multi-dimensional Filters and Models in Low-Level Computer Vision." Doctoral thesis, Linköpings universitet, Bildbehandling, 1992. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-54340.
Full textMöller, Sebastian. "Image Segmentation and Target Tracking using Computer Vision." Thesis, Linköpings universitet, Datorseende, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-68061.
Full textI detta examensarbete undersöks möjligheterna att detektera och spåra intressanta objekt i multispektrala infraröda videosekvenser. Den nuvarande metoden, som använder sig av rektanglar med fix storlek, har sina nackdelar. Dessa nackdelar kommer att lösas med hjälp av bildsegmentering för att uppskatta formen på önskade mål.Utöver detektering och spårning försöker vi också att hitta formen och konturen för intressanta objekt för att kunna använda den exaktare passformen vid kontrastberäkningar. Denna framsegmenterade kontur ersätter de gamla fixa rektanglarna som använts tidigare för att beräkna intensitetskontrasten för objekt i de infraröda våglängderna. Resultaten som presenteras visar att det för vissa objekt, som motmedel och facklor, är lättare att få fram en bra kontur samt målföljning än vad det är med helikoptrar, som var en annan önskad måltyp. De svårigheter som uppkommer med helikoptrar beror till stor del på att de är mycket svalare vilket gör att delar av helikoptern kan helt döljas i bruset från bildsensorn. För att kompensera för detta används metoder som utgår ifrån att objektet rör sig mycket i videon så att rörelsen kan användas som detekteringsparameter. Detta ger bra resultat för de videosekvenser där målet rör sig mycket i förhållande till sin storlek.
Lindvall, Victor. "A Computer Vision-Based Approach for Automated Inspection of Cable Connections." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-448446.
Full textALI, FAIZA, and MAKSIMS SVJATOHA. "Integration of Computer Vision Methods and Sensor Fusion Technologies for Precision Driving." Thesis, KTH, Skolan för industriell teknik och management (ITM), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-299793.
Full textMed ett ökat intresse för artificiell intelligens har avancerade sensorer som använder datorseende slagit igenom stort i teknikvärlden. Dessa sensorer är kända för att ha hög precision, ger mer konsekventa mätningar och har inga fel som ackumuleras med tiden. En nackdel med dessa sensorer är att de kräver mer datorkraft för processering, vilket kan orsaka fördröjningar i utsignalen. Därför är det intressant att kombinera dessa med mer traditionella sensorer för att försöka kompensera för deras svagheter. Trots att denna metod redan har implementerats, är det intressant att studera om det finns ett sätt att öka prestandan hos billigare sensorer till en precisionsnivå som är likvärdig med dyrare och industriella varianter. Detta kan möjliggöras genom att använda olika bildbehandlingsmetoder och algoritmer för datorseende. I detta examensarbete, har en Raspberry Pi-kamera monterats på ett Jetracer-fordon för att uppskatta avståndet till ett målobjekt. Med hjälp av sensorfusion, kommer dess utsignal att kombineras med utsignalen från en roterande pulsgivare. Denna sensorfusion sker med hjälp av ett Kalman-filter vilket eventuellt kommer att minska på de osäkerheter som både bildbehandlingsalgoritmen för kameran och utsignalen från pulsgivaren medför. Det finns ett samband mellan osäkerheten i utsignalerna från datorseendesystemet och den effektiva upplösningen. Sensorfusion har delvis implementerats i ett online-scenario, men fokuset låg på att sammanfoga data från inspelningar för att undvika latensproblem från datorseendesystemet. Fusion av sensordata minskade osäkerheten i position jämfört med att endast använda datorseendesystemet. Det är dock oklart om det är bättre än att endast använda pulsgivaren. Ökade latenser och minskade samplingsfrekvenser hade en negativ inverkan på osäkerheten i position. Latenser som ligger inom realistiska gränser visar har dock försumbar inverkan. Vidare finns det en avvägning att göra mellan precision och noggrannhet - högre samplingsfrekvenser ökar noggrannheten men minskar precisionen.
Martinsson, Jonas. "Examine vision technology for small object recognition in an industrial robotics application." Thesis, Mälardalens högskola, Akademin för innovation, design och teknik, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:mdh:diva-28218.
Full textJohnson, Abioseh Saeley. "Automatic number-plate recognition : an application of computer vision technology to automatic vehicle identification." Thesis, University of Bristol, 1990. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.300053.
Full textMuchaneta, Irikidzai Zorodzai. "Enhancing colour-coded poll sheets using computer vision as a viable Audience Response System (ARS) in Africa." Master's thesis, University of Cape Town, 2018. http://hdl.handle.net/11427/27854.
Full textAragon, Camarasa Gerardo. "A hierarchical active binocular robot vision architecture for scene exploration and object appearance learning." Thesis, University of Glasgow, 2012. http://theses.gla.ac.uk/3640/.
Full textNicander, Torun. "Indoor triangulation system using vision sensors." Thesis, Uppsala universitet, Signaler och system, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-429676.
Full textRingaby, Erik. "Geometric Computer Vision for Rolling-shutter and Push-broom Sensors." Licentiate thesis, Linköpings universitet, Datorseende, 2012. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-77391.
Full textVGS
Villaroman, Norman. "Face Tracking User Interfaces Using Vision-Based Consumer Devices." BYU ScholarsArchive, 2013. https://scholarsarchive.byu.edu/etd/3941.
Full textMi, Yongcui. "Novel beam shaping and computer vision methods for laser beam welding." Licentiate thesis, Högskolan Väst, Avdelningen för produktionssystem (PS), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:hv:diva-16970.
Full textLasersvetsning används i stor utsträckning i olika industrisektorer på grund av dess unika fördelar. Det finns emellertid fortfarande utmaningar, såsom rätt positionering av laserstrålen vid genomträngningssvetsning av T-fogar och hantering av varierande spaltbredd längs fogen vid svetsning av stumfogar. Sådana problem förväntas kunna lösas med avancerade metoder för automatisering, metoder som också förväntas ge fördjupade kunskaper om processen. Syftet med detta arbete är att ta itu med dessa problem med hjälp av en teknik för lasereffektens fördelning på arbetsstycket, s.k. beam shaping. Det sker med hjälp av en ny typ av i realtid deformerbar spegel tillsammans med bildbehandling av kamerabilder från processen. För- och nackdelar med detta tillvägagångssätt undersöks.Beam shaping åstadkoms med hjälp av ny typ av deformerbart spegelsystem som integreras i en industriell processoptik. Tillsammans med en vågfrontsensor bildas ett adaptivt system för beam shaping med en svarstid på 10 ms. Processen övervakas av en kamera linjerad koaxialt med laserstrålen. För att kunna ta bilder av svetspunkten belyses den med ljus av lämplig våglängd, och kameran är försedd med ett motsvarande optiskt filter. Försök har utförts med svetsning utan tillsatsmaterial, direkt på plåtar, svetsning utan s.k. nyckelhål, för att förstå effekten av beam shaping på svetssmältans geometri. Gauss fördelade cirkulära och elliptiska former, långsträckta både tvärs och längs svetsriktningen har studerats. Bilder från svetssmältan har analyserats och även mikrostrukturen i tvärsnitt från de svetsade plåtarna. Resultaten visar att svetssmältans geometri kan modifieras signifikant genom beam shaping med hjälp av det deformerbara spegelsystemet. Genomträngningssvetsning av T-fogar med avvikelser relativt foglinjens centrum genomfördes för att studera potentialen i att använda maskininlärning för att fånga processens tillstånd. Resultaten visade att maskininlärning kan nå tillräcklig prestanda för detektering och skattning av denna avvikelse. Något som också kan användas för återkopplad styrning. Flerdimensionell processdata har samlats i realtid och analyserats med hjälp av bildbehandlingsmetoder. Dessa data avslöjar brister i nuvarande simuleringsmodeller,vilket i sin tur hjälper till med att bättre förstå och styra lasersvetsning.Resultaten från detta arbete uppvisar en god potential i att använda de föreslagna metoderna för att lösa relevanta utmaningar inom lasersvetsning.
Till licentiatuppsats hör 2 inskickade artiklar, som visas inte nu.
CABAN, JESUS. "INFORMATION TECHNOLOGY FOR NEXT-GENERATION OF SURGICAL ENVIRONMENTS." UKnowledge, 2006. http://uknowledge.uky.edu/gradschool_theses/229.
Full textKutiyanawala, Aliasgar. "Eyes-Free Vision-Based Scanning of Aligned Barcodes and Information Extraction from Aligned Nutrition Tables." DigitalCommons@USU, 2013. http://digitalcommons.usu.edu/etd/1522.
Full textMollberg, Alexander. "A Resource-Efficient and High-Performance Implementation of Object Tracking on a Programmable System-on-Chip." Thesis, Linköpings universitet, Datorteknik, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-124044.
Full textLuwes, Nicolaas Johannes. "Artificial intelligence machine vision grading system." Thesis, Bloemfontein : Central University of Technology, Free State, 2014. http://hdl.handle.net/11462/35.
Full textEidehall, Andreas. "Tensor representation of 3D structures." Thesis, Linköping University, Department of Electrical Engineering, 2002. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-1241.
Full textThis is a thesis written for a master's degree at the Computer Vision Laboratory, University of Linköping. An abstract outer product is defined and used as a bridge to reach 2:nd and 4:th order tensors. Some applications of these in geometric analysis of range data are discussed and illustrated. In idealized setups, simple geometric objects, like spheres or polygons, are successfully detected. Finally, the generalization to n:th order tensors for storing and analysing geometric information is discussed.
Viljoen, Vernon. "Integration of a vision-guided robot into a reconfigurable component- handling platform." Thesis, [Bloemfontein?] : Central University of Technology, Free State, 2014. http://hdl.handle.net/11462/120.
Full textThe latest technological trend in manufacturing worldwide is automation. Reducing human labour by using robots to do the work is purely a business decision. The reasons for automating a plant include: Improving productivity Reducing labour and equipment costs Reducing product damage Monitoring system reliability Improving plant safety. The use of robots in the automation sector adds value to the production line because of their versatility. They can be programmed to follow specific paths when moving material from one point to another and their biggest advantage is that they can operate for twenty-four hours a day while delivering consistent quality and accuracy. Vision-Guided Robots (VGRs) are developed for many different applications and therefore many different combinations of VGR systems are available. All VGRs are equipped with vision sensors which are used to locate and inspect various objects. In this study a robot and a vision system were combined for a pick-and-place application. Research was done on the design of a robot for locating, inspecting and picking selected components from a moving conveyor system.
Johansson, Alexander. "Automated panorama sequence detection using the Narrative platform." Thesis, Linköpings universitet, Medie- och Informationsteknik, 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-108402.
Full textHüdig, Daniel H. "The vision of a future information and communication society computer-mediated communication and technology policy in the United States, the European Union and Japan /." [S.l. : s.n.], 2000. http://deposit.ddb.de/cgi-bin/dokserv?idn=961031476.
Full textAndreasson, Henrik. "Local visual feature based localisation and mapping by mobile robots." Doctoral thesis, Örebro : Örebro University, 2008. http://urn.kb.se/resolve?urn=urn:nbn:se:oru:diva-2444.
Full textAxelsson, Viktor. "Automatisk segmentering och maskering av implantat i mammografibilder." Thesis, Linköpings universitet, Medie- och Informationsteknik, 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-113459.
Full textHåkansson, Staffan. "Detektering av sprickor i vägytor med hjälp av Datorseende." Thesis, Linköping University, Department of Electrical Engineering, 2005. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-2818.
Full textThis thesis describes new methods for automatic crack detection in pavements. Cracks in pavements can be used as an early indication for the need of reparation.
Automatic crack detection is preferable compared to manual inventory; the repeatability can be better, the inventory can be done at a higher speed and can be done without interruption of the traffic.
The automatic and semi-automatic crack detection systems that exist today use Image Analysis methods. There are today powerful methods available in the area of Computer Vision. These methods work in higher dimensions with greater complexity and generate measures of local signal properties, while Image Analyses methods for crack detection use morphological operations on binary images.
Methods for digitalizing video data on VHS-cassettes and stitching images from nearby frames have been developed.
Four methods for crack detection have been evaluated, and two of them have been used to form a crack detection and classification program implemented in the calculation program Matlab.
One image set was used during the implementation and another image set was used for validation. The crack detection system did perform correct detection on 99.2 percent when analysing the images which were used during implementation. The result of the crack detection on the validation data was not very good. When the program is being used on data from other pavements than the one used during implementation, information about the surface texture is required to calibrate the crack detection.
Chen, Yiqiang. "Person re-identification in images with deep learning." Thesis, Lyon, 2018. http://www.theses.fr/2018LYSEI074/document.
Full textVideo surveillance systems are of a great value for public safety. As one of the most import surveillance applications, person re-identification is defined as the problem of identifying people across images that have been captured by different surveillance cameras without overlapping fields of view. With the increasing need for automated video analysis, this task is increasingly receiving attention. However, this problem is challenging due to the large variations of lighting, pose, viewpoint and background. To tackle these different difficulties, in this thesis, we propose several deep learning based approaches to obtain a better person re-identification performance in different ways. In the first proposed approach, we use pedestrian attributes to enhance the person re-identification. The attributes are defined as semantic mid-level descriptions of persons, such as gender, accessories, clothing etc. They could be helpful to extract characteristics that are invariant to the pose and viewpoint variations thanks to the descriptor being on a higher semantic level. In order to make use of the attributes, we propose a CNN-based person re-identification framework composed of an identity classification branch and of an attribute recognition branch. At a later stage, these two cues are combined to perform person re-identification. Secondly, among the challenges, one of the most difficult is the variation under different viewpoint. The same person shows very different appearances from different points of view. To deal with this issue, we consider that the images under various orientations are from different domains. We propose an orientation-specific CNN. This framework performs body orientation regression in a gating branch, and in another branch learns separate orientation-specific layers as local experts. The combined orientation-specific CNN feature representations are used for the person re-identification task. Thirdly, learning a similarity metric for person images is a crucial aspect of person re-identification. As the third contribution, we propose a novel listwise loss function taking into account the order in the ranking of gallery images with respect to different probe images. Further, an evaluation gain-based weighting is introduced in the loss function to optimize directly the evaluation measures of person re-identification. At the end, in a large gallery set, many people could have similar clothing. In this case, using only the appearance of single person leads to strong ambiguities. In realistic settings, people often walk in groups rather than alone. As the last contribution, we propose to learn a deep feature representation with displacement invariance for group context and introduce a method to combine the group context and single-person appearance. For all the four contributions of this thesis, we carry out extensive experiments on popular benchmarks and datasets to demonstrate the effectiveness of the proposed systems
Hemery, Edgar. "Modélisation, reconnaissance du geste des doigts et du haut du corps dans le design d’interaction musicale." Thesis, Paris Sciences et Lettres (ComUE), 2017. http://www.theses.fr/2017PSLEM075/document.
Full textThis thesis presents a novel musical instrument, named the Embodied Musical Instrument (EMI), which has been designed to answer two problems : how can we capture and model musical gestures and how can we use this model to control sound synthesis parameters expressively. The EMI is articulated around an explicit mapping strategy, which draws inspiration from the piano-playing techniques and other objects’ affordances. The system we propose makes use of 3D cameras and computer vision algorithms in order to free the gesture from intrusive devices and ease the process of capture and performance, while enabling precise and reactive tracking of the fingertips and upper-body. Having recourse to different 3D cameras tracking solutions, we fully exploit their potential by adding a transparent sheet, which serves as a detection threshold for fingerings as well as bringing a simple but essential haptic feedback. We examined finger movements while tapping on the surface of the EMI and decomposed their trajectories into essential phases, which enabled us to model and analyse piano-like gestures. A preliminary study of generic musical gestures directed our interest not only on the effective gestures operated by the fingers - in the case of keyboard instruments - but also on the accompanying and figurative gestures, which are mostly characterised by the arms and head movements. Consequently, we distinguish two level of interactions, delimited by two bounding volumes. The micro bounding volume includes the micro-gestures operated with the fingers, while the macro bounding volume includes larger movements with the upper-body. Building from this, we extend our piano-like model to a 3D interaction paradigm, where higher-level musical parameters, such as sound effects, can be controlled continuously by upper-body free movements. We explored a set of real-world scenarios for this instrument, namely practice, composition and performance. The EMI introduces a framework for capture and analysis, of specific musical gestures. An off-line analysis of gesture features can reveal trends, faults and musical specificities of an interpret. Several musical works have been created and performed live; either solo or accompanied by a string quartet, revealing the body gesture specificities through the sounds it synthesises. User experience feedback shows that the instrument can be easily taught - if not self-taught - thanks to the intuitive gesture paradigms drawn from piano-like gestures and other metaphorical gestures
Bihi, Thabo George. "Assembly-setup verification and quality control using machine vision within a reconfigurable assembly system." Thesis, [Bloemfontein?] : Central University of Technology, Free State, 2014. http://hdl.handle.net/11462/188.
Full textThe project is aimed at exploring the application of Machine Vision in a Reconfigurable Manufacturing System (RMS) Environment. The Machine Vision System interfaces with the RMS to verify the reconfiguration and positioning of devices within the assembly system, and inspects the product for defects that infringe on the quality of that product. The vision system interfaces to the Multi-agent System (MAS), which is in charge of scheduling and allocating resources of the RMS, in order to communicate and exchange data regarding the quality of the product. The vision system is comprised of a Compact Vision System (CVS) device with fire-wire cameras to aid in the image acquisition, inspection and verification process. Various hardware and software manufacturers offer a platform to implement this with a multiple array of vision equipment and software packages. The most appropriate devices and software platform were identified for the implementation of the project. An investigation into illumination was also undertaken in order to determine whether external lighting sources would be required at the point of inspection. Integration into the assembly system involved the establishment communication between the vision system and assembly system controller.
Johansson, Filip, and Ali Karabiber. "Övervakningssystem för inomhusmiljöer." Thesis, Malmö högskola, Fakulteten för teknik och samhälle (TS), 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-20874.
Full textSecurity systems have always had an important role in society for a long time. The purpose ofthese systems has always remained primarily focused on the security aspect. In most cases theefficiency of current security systems depend on human factors. The frames have to beprocessed by a person to reach a definitive conclusion. In recent years, security systems havebecome increasingly intelligent as technology continues to develop. Methods that implementautomatic analysis of any given frame in a video have been introduced and some of thesemethods are already standardized. The aim of this study is to develop a prototype and stresstest it to identify possible flaws. This study is primarily a feasibility study to show thepossibilities of what can be achieved with limited resources. With a fully functional prototypeavailable a series of stress tests will be specified to measure the system’s limits. The resultshave been positive overall but a number of critical flaws have been identified.
Park, Chung Hyuk. "Robot-based haptic perception and telepresence for the visually impaired." Diss., Georgia Institute of Technology, 2012. http://hdl.handle.net/1853/44848.
Full textSamanci, Ozge. "Embodying comics reinventing comics and animation for a digital performance /." Diss., Atlanta, Ga. : Georgia Institute of Technology, 2009. http://hdl.handle.net/1853/29630.
Full textCommittee Chair: Mazalek, Alexandra; Committee Member: Bolter, Jay; Committee Member: Knospel, Kenneth; Committee Member: Murray, Janet; Committee Member: Winegarden, Claudia Rebola. Part of the SMARTech Electronic Thesis and Dissertation Collection.
Filip, Mori. "A 2D video player for Virtual Reality and Mixed Reality." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-217359.
Full textMedan 360 graders video under senare tid varit föremål för studier, så verkar inte traditionella rektangulära 2D videos i virtuella miljöer ha fått samma uppmärksamhet. Mer specifikt, 2D videouppspelning i Virtual Reality (VR) och Mixed Reality (MR) verkar sakna utforskning i egenskaper som upplösning, ljud och interaktion, som slutligen bidrar till ”presence” i videon och den virtuella miljön. Det här pappret reflekterar över definitionerna VR och MR, samtidigt som den utökar de kända koncepten ”immersion” och ”presence” för 2D video i virtuella miljöer. Relevanta attribut till ”presence” som kan appliceras på 2D video utreddes sedan med hjälp av litteraturen. Det huvudsakliga problemet var att ta reda på komponenterna och processerna i den mjukvara som skall spela upp video i VR och MR med företagsönskemål och avgränsningar i åtanke, och möjligen, hur man kan justera dessa komponenter för att utöka närvaron i framförallt 2D video och sekundärt den virtuella miljön, även om dessa medium är relaterade och kan påverka varandra. Examensarbetet tog plats på Advrty, ett företag som utvecklar en annonseringsplattform för VR och MR. Utveckling och framtagande av komponenterna, var gjorda genom inkrementell utveckling där en enklare 2D videospelare skapades, sedan genom en andra inkrementell fas där videospelaren implementerades i VR och MR. Jämförelser med proof-of-concept-videospelaren i VR och MR samt den enklare videospelaren gjordes. I diskussionen om arbetet, gjordes reflektioner på användningen av open source-bibliotek i en kommersiell applikation, de tekniska begränsningarna i nuvarande VR och MR Head-mounted displays, relevanta ”presence” inducerande attribut samt val av metod för utvecklingen av videospelaren.
BALDHAGEN, FREDRIK, and ANTON HEDSTRÖM. "Chess Playing Robot : Robotic arm capable of playing chess." Thesis, KTH, Skolan för industriell teknik och management (ITM), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-279829.
Full textSyftet med det här examensarbetet var att skapa en robot som genom användning av bildigenkänning och robotik kunde spela schack. Idén till detta projekt kom från det faktum att det finns ett ökande behov av smarta robotar som kan fatta sina egna beslut i en förändring miljö och det faktum att schack nyligen har sett en ökning av nya spelare. Den optimala utformningen av armen som flyttar pjäserna beslutades vara av SCARA-typ, som är en vanlig robotarm som utmärker sig i ’pick-and-place’ operationer. Armens rörelse drivs av två stegmotorer anslutna till en Raspberry Pi och en extern strömkälla. Rörelse i Z-riktningen uppnåddes genom användning av en servomotor som drev en kuggstång vertikalt. En kamera placerades ovanför schackbrädet, och genom användning av flera program och funktioner konverterades bilder till schacknotation som sedan skickades till en schackmotor som körs på Raspberry Pi. Bildigenkänningen fungerade optimalt när schackbrädet var väl och jämnt upplyst. När belysningen var dålig kunde värden som definierade färger ändras för att möjliggöra korrekta utvärderingar av färgen, men när belysningsnivån sjönk under 15 lux blev de blå pjäserna oskiljbara från de svarta rutorna och programmet slutade därför att fungera
Blåwiik, Per. "Fusing Stereo Measurements into a Global 3D Representation." Thesis, Linköpings universitet, Medie- och Informationsteknik, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-177586.
Full textExamensarbetet är utfört vid Institutionen för teknik och naturvetenskap (ITN) vid Tekniska fakulteten, Linköpings universitet
Moujtahid, Salma. "Exploiting scene context for on-line object tracking in unconstrained environments." Thesis, Lyon, 2016. http://www.theses.fr/2016LYSEI110/document.
Full textWith the increasing need for automated video analysis, visual object tracking became an important task in computer vision. Object tracking is used in a wide range of applications such as surveillance, human-computer interaction, medical imaging or vehicle navigation. A tracking algorithm in unconstrained environments faces multiple challenges : potential changes in object shape and background, lighting, camera motion, and other adverse acquisition conditions. In this setting, classic methods of background subtraction are inadequate, and more discriminative methods of object detection are needed. Moreover, in generic tracking algorithms, the nature of the object is not known a priori. Thus, off-line learned appearance models for specific types of objects such as faces, or pedestrians can not be used. Further, the recent evolution of powerful machine learning techniques enabled the development of new tracking methods that learn the object appearance in an online manner and adapt to the varying constraints in real time, leading to very robust tracking algorithms that can operate in non-stationary environments to some extent. In this thesis, we start from the observation that different tracking algorithms have different strengths and weaknesses depending on the context. To overcome the varying challenges, we show that combining multiple modalities and tracking algorithms can considerably improve the overall tracking performance in unconstrained environments. More concretely, we first introduced a new tracker selection framework using a spatial and temporal coherence criterion. In this algorithm, multiple independent trackers are combined in a parallel manner, each of them using low-level features based on different complementary visual aspects like colour, texture and shape. By recurrently selecting the most suitable tracker, the overall system can switch rapidly between different tracking algorithms with specific appearance models depending on the changes in the video. In the second contribution, the scene context is introduced to the tracker selection. We designed effective visual features, extracted from the scene context to characterise the different image conditions and variations. At each point in time, a classifier is trained based on these features to predict the tracker that will perform best under the given scene conditions. We further improved this context-based framework and proposed an extended version, where the individual trackers are changed and the classifier training is optimised. Finally, we started exploring one interesting perspective that is the use of a Convolutional Neural Network to automatically learn to extract these scene features directly from the input image and predict the most suitable tracker
Ringaby, Erik. "Optical Flow Computation on Compute Unified Device Architecture." Thesis, Linköping University, Department of Electrical Engineering, 2008. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-15426.
Full textThere has been a rapid progress of the graphics processor the last years, much because of the demands from computer games on speed and image quality. Because of the graphics processor’s special architecture it is much faster at solving parallel problems than the normal processor. Due to its increasing programmability it is possible to use it for other tasks than it was originally designed for.
Even though graphics processors have been programmable for some time, it has been quite difficult to learn how to use them. CUDA enables the programmer to use C-code, with a few extensions, to program NVIDIA’s graphics processor and completely skip the traditional programming models. This thesis investigates if the graphics processor can be used for calculations without knowledge of how the hardware mechanisms work. An image processing algorithm calculating the optical flow has been implemented. The result shows that it is rather easy to implement programs using CUDA, but some knowledge of how the graphics processor works is required to achieve high performance.
Haskins, Bertram Peter. "A feasibility study on the use of agent-based image recognition on a desktop computer for the purpose of quality control in a production environment." Thesis, [Bloemfontein?] : Central University of Technology, Free State, 2006. http://hdl.handle.net/11462/66.
Full textA multi-threaded, multi-agent image recognition software application called RecMaster has been developed specifically for the purpose of quality control in a production environment. This entails using the system as a monitor to identify invalid objects moving on a conveyor belt and to pass on the relevant information to an attached device, such as a robotic arm, which will remove the invalid object. The main purpose of developing this system was to prove that a desktop computer could run an image recognition system efficiently, without the need for high-end, high-cost, specialised computer hardware. The programme operates by assigning each agent a task in the recognition process and then waiting for resources to become available. Tasks related to edge detection, colour inversion, image binarisation and perimeter determination were assigned to individual agents. Each agent is loaded onto its own processing thread, with some of the agents delegating their subtasks to other processing threads. This enables the application to utilise the available system resources more efficiently. The application is very limited in its scope, as it requires a uniform image background as well as little to no variance in camera zoom levels and object to lens distance. This study focused solely on the development of the application software, and not on the setting up of the actual imaging hardware. The imaging device, on which the system was tested, was a web cam capable of a 640 x 480 resolution. As such, all image capture and processing was done on images with a horizontal resolution of 640 pixels and a vertical resolution of 480 pixels, so as not to distort image quality. The application locates objects on an image feed - which can be in the format of a still image, a video file or a camera feed - and compares these objects to a model of the object that was created previously. The coordinates of the object are calculated and translated into coordinates on the conveyor system. These coordinates are then passed on to an external recipient, such as a robotic arm, via a serial link. The system has been applied to the model of a DVD, and tested against a variety of similar and dissimilar objects to determine its accuracy. The tests were run on both an AMD- and Intel-based desktop computer system, with the results indicating that both systems are capable of efficiently running the application. On average, the AMD-based system tended to be 81% faster at matching objects in still images, and 100% faster at matching objects in moving images. The system made matches within an average time frame of 250 ms, making the process fast enough to be used on an actual conveyor system. On still images, the results showed an 87% success rate for the AMD-based system, and 73% for Intel. For moving images, however, both systems showed a 100% success rate.
Tan, Jason, and Jake O'Donnell. "Hand gestures as a trigger for system initialisation." Thesis, Malmö universitet, Fakulteten för teknik och samhälle (TS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-20876.
Full textBiometric solutions for access control is a thriving concept, Precise Biometrics is a company that focuses on just that. YOUNiQ is a product that focuses on facial identification for access control, with it comes an issue in where every person's face is being identified. This means identifying people that do not want to use the facial identification module. This thesis focuses on implementing an intent-aware system, a system which uses a trigger to begin a process. This thesis was done in collaboration with engineers at Precise Biometrics. Instead of identifying faces without permission the intent-aware system uses a trigger based on different hand gestures to begin the process. This thesis does not focus on face identification but instead the trigger before a specific process begins. The development phase consisted of an iterative process in creating the prototype system. In order to evaluate the system, test cases were done to verify accuracy of each hand gesture. Thereafter, a scenario was created to simulate an activation of the prototype system. The evaluation was used to determine the convenience and guidance when implementing intent-aware systems. Furthermore, the system can be seen as a form of trigger to allow for extracting biometric data in for example face identification.
Husseini, Orabi Ahmed. "Multi-Modal Technology for User Interface Analysis including Mental State Detection and Eye Tracking Analysis." Thesis, Université d'Ottawa / University of Ottawa, 2017. http://hdl.handle.net/10393/36451.
Full textForssén, Per-Erik. "Detection of Man-made Objects in Satellite Images." Thesis, Linköping University, Linköping University, Computer Vision, 1997. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-54356.
Full textIn this report, the principles of man-made object detection in satellite images is investigated. An overview of terminology and of how the detection problem is usually solved today is given. A three level system to solve the detection problem is proposed. The main branches of this system handle road, and city detection respectively. To achieve data source flexibility, the Logical Sensor notion is used to model the low level system components. Three Logical Sensors have been implemented and tested on Landsat TM and SPOT XS scenes. These are: BDT (Background Discriminant Transformation) to construct a man-made object property field; Local-orientation for texture estimation and road tracking; Texture estimation using local variance and variance of local orientation. A gradient magnitude measure for road seed generation has also been tested.
Casserfelt, Karl. "A Deep Learning Approach to Video Processing for Scene Recognition in Smart Office Environments." Thesis, Malmö universitet, Fakulteten för teknik och samhälle (TS), 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-20429.
Full textAndersson, Anna, and Klara Eklund. "A Study of Oriented Mottle in Halftone Print." Thesis, Linköping University, Department of Science and Technology, 2007. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-9233.
Full textCoated solid bleached board belongs to the top-segment of paperboards. One important property of paperboard is the printability. In this diploma work a specific print defect, oriented mottle, has been studied in association with Iggesund Paperboard. The objectives of the work were to develop a method for analysis of the dark and light areas of oriented mottle, to analyse these areas, and to clarify the effect from the print, coating and paperboard surface related factors. This would clarify the origin of oriented mottle and predict oriented mottle on unprinted paperboard. The objectives were fulfilled by analysing the areas between the dark halftone dots, the amount of coating and the ink penetration, the micro roughness and the topography. The analysis of the areas between the dark halftone dots was performed on several samples and the results were compared regarding different properties. The other methods were only applied on a limited selection of samples. The results from the study showed that the intensity differences between the dark halftone dots were enhanced in the dark areas, the coating amount was lower in the dark areas and the ink did not penetrate into the paperboard. The other results showed that areas with high transmission corresponded to dark areas, smoother micro roughness, lower coating amount and high topography. A combination of the information from these properties might be used to predict oriented mottle. The oriented mottle is probably an optical phenomenon in half tone prints, and originates from variations in the coating and other paperboard properties.
Pistori, Hemerson. "Tecnologia adaptativa em engenharia de computação: estado da arte e aplicações." Universidade de São Paulo, 2003. http://www.teses.usp.br/teses/disponiveis/3/3141/tde-02032004-145107/.
Full textThis work presents a practical and theoretical assembly of contributions that consolidates some concepts from the rule-driven adaptive devices theory, emphasizing their high applicability. A supporting tool for the development of adaptive automata, which includes graphical animation resources, has been implemented, in agreement with our proposal of formalization. This proposal aims to complement and simplify the original proposal by including an in-depth analysis and formalization of adaptive functions implementation, in their most general form: with elementary query actions being able to return multiple results. The new formalization of adaptive functions, which includes an algorithm for adaptive function execution, is an important tool for determining the impact of an adaptive layer on the complexity analysis of general adaptive automata. The thesis also presents a new technique for the integration of adaptive automata with mechanisms for the manipulation of continuous values. Finally, the application of these theoretical results and the tools developed, to the solution of problems in the area of machine learning, compiler construction, man-machine interface, computational vision and medical diagnosis, is demonstrated.
Asif, Muhammad Salman. "Dynamic compressive sensing: sparse recovery algorithms for streaming signals and video." Diss., Georgia Institute of Technology, 2013. http://hdl.handle.net/1853/49106.
Full textMELLO, SIMON. "VATS : Voice-Activated Targeting System." Thesis, KTH, Skolan för industriell teknik och management (ITM), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-279837.
Full textMaskininlärning är viktigt inom röstigenkänning och datorseende, för både små såväl som stora applikationer. Syftet med det här projektet är att titta på om enkla implementationer av maskininlärning duger för den verkligen världen. Ett enkelt artificiellt neuronnät kodat i Python, samt existerande programbibliotek för Python, används för att kontrollera en laserpekare via en servomotor och en Arduino, för att skapa ett röstaktiverat identifieringssystem. Neuronnätet tränat på MNIST data når en precision på 0.95 ± 0.01 när den försöker klassificera MNIST test data, men lyckas även klassificera inspelade bilder korrekt om störningen är låg. Detta gäller även för röstigenkänningen, då den sällan ger fel avläsningar. Den slutliga prototypen lyckas i alla domäner förutom att förvandla bilder som klassificerats korrekt till mål som Arduinon kan läsa av och sikta på, vilket betyder att prototypen inte lyckas sammanfoga röstigenkänningen och datorseendet.