CINXE.COM
Search results for: omnidirectional camera
<!DOCTYPE html> <html lang="en" dir="ltr"> <head> <!-- Google tag (gtag.js) --> <script async src="https://www.googletagmanager.com/gtag/js?id=G-P63WKM1TM1"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('js', new Date()); gtag('config', 'G-P63WKM1TM1'); </script> <!-- Yandex.Metrika counter --> <script type="text/javascript" > (function(m,e,t,r,i,k,a){m[i]=m[i]||function(){(m[i].a=m[i].a||[]).push(arguments)}; m[i].l=1*new Date(); for (var j = 0; j < document.scripts.length; j++) {if (document.scripts[j].src === r) { return; }} k=e.createElement(t),a=e.getElementsByTagName(t)[0],k.async=1,k.src=r,a.parentNode.insertBefore(k,a)}) (window, document, "script", "https://mc.yandex.ru/metrika/tag.js", "ym"); ym(55165297, "init", { clickmap:false, trackLinks:true, accurateTrackBounce:true, webvisor:false }); </script> <noscript><div><img src="https://mc.yandex.ru/watch/55165297" style="position:absolute; left:-9999px;" alt="" /></div></noscript> <!-- /Yandex.Metrika counter --> <!-- Matomo --> <!-- End Matomo Code --> <title>Search results for: omnidirectional camera</title> <meta name="description" content="Search results for: omnidirectional camera"> <meta name="keywords" content="omnidirectional camera"> <meta name="viewport" content="width=device-width, initial-scale=1, minimum-scale=1, maximum-scale=1, user-scalable=no"> <meta charset="utf-8"> <link href="https://cdn.waset.org/favicon.ico" type="image/x-icon" rel="shortcut icon"> <link href="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/css/bootstrap.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/plugins/fontawesome/css/all.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/css/site.css?v=150220211555" rel="stylesheet"> </head> <body> <header> <div class="container"> <nav class="navbar navbar-expand-lg navbar-light"> <a class="navbar-brand" href="https://waset.org"> <img src="https://cdn.waset.org/static/images/wasetc.png" alt="Open Science Research Excellence" title="Open Science Research Excellence" /> </a> <button class="d-block d-lg-none navbar-toggler ml-auto" type="button" data-toggle="collapse" data-target="#navbarMenu" aria-controls="navbarMenu" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <div class="w-100"> <div class="d-none d-lg-flex flex-row-reverse"> <form method="get" action="https://waset.org/search" class="form-inline my-2 my-lg-0"> <input class="form-control mr-sm-2" type="search" placeholder="Search Conferences" value="omnidirectional camera" name="q" aria-label="Search"> <button class="btn btn-light my-2 my-sm-0" type="submit"><i class="fas fa-search"></i></button> </form> </div> <div class="collapse navbar-collapse mt-1" id="navbarMenu"> <ul class="navbar-nav ml-auto align-items-center" id="mainNavMenu"> <li class="nav-item"> <a class="nav-link" href="https://waset.org/conferences" title="Conferences in 2024/2025/2026">Conferences</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/disciplines" title="Disciplines">Disciplines</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/committees" rel="nofollow">Committees</a> </li> <li class="nav-item dropdown"> <a class="nav-link dropdown-toggle" href="#" id="navbarDropdownPublications" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Publications </a> <div class="dropdown-menu" aria-labelledby="navbarDropdownPublications"> <a class="dropdown-item" href="https://publications.waset.org/abstracts">Abstracts</a> <a class="dropdown-item" href="https://publications.waset.org">Periodicals</a> <a class="dropdown-item" href="https://publications.waset.org/archive">Archive</a> </div> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/page/support" title="Support">Support</a> </li> </ul> </div> </div> </nav> </div> </header> <main> <div class="container mt-4"> <div class="row"> <div class="col-md-9 mx-auto"> <form method="get" action="https://publications.waset.org/abstracts/search"> <div id="custom-search-input"> <div class="input-group"> <i class="fas fa-search"></i> <input type="text" class="search-query" name="q" placeholder="Author, Title, Abstract, Keywords" value="omnidirectional camera"> <input type="submit" class="btn_search" value="Search"> </div> </div> </form> </div> </div> <div class="row mt-3"> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Commenced</strong> in January 2007</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Frequency:</strong> Monthly</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Edition:</strong> International</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Paper Count:</strong> 615</div> </div> </div> </div> <h1 class="mt-3 mb-3 text-center" style="font-size:1.6rem;">Search results for: omnidirectional camera</h1> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">615</span> Localization of Mobile Robots with Omnidirectional Cameras</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Tatsuya%20Kato">Tatsuya Kato</a>, <a href="https://publications.waset.org/abstracts/search?q=Masanobu%20Nagata"> Masanobu Nagata</a>, <a href="https://publications.waset.org/abstracts/search?q=Hidetoshi%20Nakashima"> Hidetoshi Nakashima</a>, <a href="https://publications.waset.org/abstracts/search?q=Kazunori%20Matsuo"> Kazunori Matsuo</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Localization of mobile robots are important tasks for developing autonomous mobile robots. This paper proposes a method to estimate positions of a mobile robot using an omnidirectional camera on the robot. Landmarks for points of references are set up on a field where the robot works. The omnidirectional camera which can obtain 360 [deg] around images takes photographs of these landmarks. The positions of the robots are estimated from directions of these landmarks that are extracted from the images by image processing. This method can obtain the robot positions without accumulative position errors. Accuracy of the estimated robot positions by the proposed method are evaluated through some experiments. The results show that it can obtain the positions with small standard deviations. Therefore the method has possibilities of more accurate localization by tuning of appropriate offset parameters. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=mobile%20robots" title="mobile robots">mobile robots</a>, <a href="https://publications.waset.org/abstracts/search?q=localization" title=" localization"> localization</a>, <a href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera" title=" omnidirectional camera"> omnidirectional camera</a>, <a href="https://publications.waset.org/abstracts/search?q=estimating%20positions" title=" estimating positions"> estimating positions</a> </p> <a href="https://publications.waset.org/abstracts/11803/localization-of-mobile-robots-with-omnidirectional-cameras" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/11803.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">442</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">614</span> Development and Test of an Open Source PX4 Controler for omnidirectional Unmanned Surface Vehicle</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Norbert%20Szulc">Norbert Szulc</a>, <a href="https://publications.waset.org/abstracts/search?q=Cezary%20Wieczorkowski"> Cezary Wieczorkowski</a>, <a href="https://publications.waset.org/abstracts/search?q=Igor%20Baranowski"> Igor Baranowski</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, a control system that bridges the gap in support for Unmanned Surface Vessels in the PX4 Opensource Autopilot was developed. The system is designed for an omnidirectional water craft with four motors. A modular autopilot architecture design centred around publish-subscribe interprocess communication was used. The paper presents the implementation and integration process of a generic surface vehicle controller capable of driving any configuration of motors through the recently introduced in control allocator in PX4 autopilot. The proposed approach was successfully tested in a case study through implementation on the ASV Perkoz. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=control%20system" title="control system">control system</a>, <a href="https://publications.waset.org/abstracts/search?q=PX4" title=" PX4"> PX4</a>, <a href="https://publications.waset.org/abstracts/search?q=drones" title=" drones"> drones</a>, <a href="https://publications.waset.org/abstracts/search?q=rovers" title=" rovers"> rovers</a>, <a href="https://publications.waset.org/abstracts/search?q=surface%20vessels" title=" surface vessels"> surface vessels</a>, <a href="https://publications.waset.org/abstracts/search?q=omnidirectional" title=" omnidirectional"> omnidirectional</a> </p> <a href="https://publications.waset.org/abstracts/163531/development-and-test-of-an-open-source-px4-controler-for-omnidirectional-unmanned-surface-vehicle" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/163531.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">87</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">613</span> Video Sharing System Based On Wi-fi Camera</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Qidi%20Lin">Qidi Lin</a>, <a href="https://publications.waset.org/abstracts/search?q=Jinbin%20Huang"> Jinbin Huang</a>, <a href="https://publications.waset.org/abstracts/search?q=Weile%20Liang"> Weile Liang</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper introduces a video sharing platform based on WiFi, which consists of camera, mobile phone and PC server. This platform can receive wireless signal from the camera and show the live video on the mobile phone captured by camera. In addition that, it is able to send commands to camera and control the camera鈥檚 holder to rotate. The platform can be applied to interactive teaching and dangerous area鈥檚 monitoring and so on. Testing results show that the platform can share the live video of mobile phone. Furthermore, if the system鈥檚 PC sever and the camera and many mobile phones are connected together, it can transfer photos concurrently. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Wifi%20Camera" title="Wifi Camera">Wifi Camera</a>, <a href="https://publications.waset.org/abstracts/search?q=socket%20mobile" title=" socket mobile"> socket mobile</a>, <a href="https://publications.waset.org/abstracts/search?q=platform%20video%20monitoring" title=" platform video monitoring"> platform video monitoring</a>, <a href="https://publications.waset.org/abstracts/search?q=remote%20control" title=" remote control"> remote control</a> </p> <a href="https://publications.waset.org/abstracts/31912/video-sharing-system-based-on-wi-fi-camera" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/31912.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">337</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">612</span> Objects Tracking in Catadioptric Images Using Spherical Snake</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Khald%20Anisse">Khald Anisse</a>, <a href="https://publications.waset.org/abstracts/search?q=Amina%20Radgui"> Amina Radgui</a>, <a href="https://publications.waset.org/abstracts/search?q=Mohammed%20Rziza"> Mohammed Rziza</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Tracking objects on video sequences is a very challenging task in many works in computer vision applications. However, there is no article that treats this topic in catadioptric vision. This paper is an attempt that tries to describe a new approach of omnidirectional images processing based on inverse stereographic projection in the half-sphere. We used the spherical model proposed by Gayer and al. For object tracking, our work is based on snake method, with optimization using the Greedy algorithm, by adapting its different operators. The algorithm will respect the deformed geometries of omnidirectional images such as spherical neighborhood, spherical gradient and reformulation of optimization algorithm on the spherical domain. This tracking method that we call "spherical snake" permitted to know the change of the shape and the size of object in different replacements in the spherical image. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=computer%20vision" title="computer vision">computer vision</a>, <a href="https://publications.waset.org/abstracts/search?q=spherical%20snake" title=" spherical snake"> spherical snake</a>, <a href="https://publications.waset.org/abstracts/search?q=omnidirectional%20image" title=" omnidirectional image"> omnidirectional image</a>, <a href="https://publications.waset.org/abstracts/search?q=object%20tracking" title=" object tracking"> object tracking</a>, <a href="https://publications.waset.org/abstracts/search?q=inverse%20stereographic%20projection" title=" inverse stereographic projection"> inverse stereographic projection</a> </p> <a href="https://publications.waset.org/abstracts/2285/objects-tracking-in-catadioptric-images-using-spherical-snake" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/2285.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">402</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">611</span> A Development of Holonomic Mobile Robot Using Fuzzy Multi-Layered Controller</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Seungwoo%20Kim">Seungwoo Kim</a>, <a href="https://publications.waset.org/abstracts/search?q=Yeongcheol%20Cho"> Yeongcheol Cho</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, a holonomic mobile robot is designed in omnidirectional wheels and an adaptive fuzzy controller is presented for its precise trajectories. A kind of adaptive controller based on fuzzy multi-layered algorithm is used to solve the big parametric uncertainty of motor-controlled dynamic system of 3-wheels omnidirectional mobile robot. The system parameters such as a tracking force are so time-varying due to the kinematic structure of omnidirectional wheels. The fuzzy adaptive control method is able to solve the problems of classical adaptive controller and conventional fuzzy adaptive controllers. The basic idea of new adaptive control scheme is that an adaptive controller can be constructed with parallel combination of robust controllers. This new adaptive controller uses a fuzzy multi-layered architecture which has several independent fuzzy controllers in parallel, each with different robust stability area. Out of several independent fuzzy controllers, the most suited one is selected by a system identifier which observes variations in the controlled system parameter. This paper proposes a design procedure which can be carried out mathematically and systematically from the model of a controlled system. Finally, the good performance of a holonomic mobile robot is confirmed through live tests of the tracking control task. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=fuzzy%20adaptive%20control" title="fuzzy adaptive control">fuzzy adaptive control</a>, <a href="https://publications.waset.org/abstracts/search?q=fuzzy%20multi-layered%20controller" title=" fuzzy multi-layered controller"> fuzzy multi-layered controller</a>, <a href="https://publications.waset.org/abstracts/search?q=holonomic%20mobile%20robot" title=" holonomic mobile robot"> holonomic mobile robot</a>, <a href="https://publications.waset.org/abstracts/search?q=omnidirectional%20wheels" title=" omnidirectional wheels"> omnidirectional wheels</a>, <a href="https://publications.waset.org/abstracts/search?q=robustness%20and%20stability." title=" robustness and stability. "> robustness and stability. </a> </p> <a href="https://publications.waset.org/abstracts/43824/a-development-of-holonomic-mobile-robot-using-fuzzy-multi-layered-controller" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/43824.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">359</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">610</span> Embedded Hardware and Software Design of Omnidirectional Autonomous Robotic Platform Suitable for Advanced Driver Assistance Systems Testing with Focus on Modularity and Safety</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ondrej%20Lufinka">Ondrej Lufinka</a>, <a href="https://publications.waset.org/abstracts/search?q=Jan%20Kaderabek"> Jan Kaderabek</a>, <a href="https://publications.waset.org/abstracts/search?q=Juraj%20Prstek"> Juraj Prstek</a>, <a href="https://publications.waset.org/abstracts/search?q=Jiri%20Skala"> Jiri Skala</a>, <a href="https://publications.waset.org/abstracts/search?q=Kamil%20Kosturik"> Kamil Kosturik</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper deals with the problem of using Autonomous Robotic Platforms (ARP) for the ADAS (Advanced Driver Assistance Systems) testing in automotive. There are different possibilities of the testing already in development, and lately, the autonomous robotic platforms are beginning to be used more and more widely. Autonomous Robotic Platform discussed in this paper explores the hardware and software design possibilities related to the field of embedded systems. The paper focuses on its chapters on the introduction of the problem in general; then, it describes the proposed prototype concept and its principles from the embedded HW and SW point of view. It talks about the key features that can be used for the innovation of these platforms (e.g., modularity, omnidirectional movement, common and non-traditional sensors used for localization, synchronization of more platforms and cars together, or safety mechanisms). In the end, the future possible development of the project is discussed as well. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=advanced%20driver%20assistance%20systems" title="advanced driver assistance systems">advanced driver assistance systems</a>, <a href="https://publications.waset.org/abstracts/search?q=ADAS" title=" ADAS"> ADAS</a>, <a href="https://publications.waset.org/abstracts/search?q=autonomous%20robotic%20platform" title=" autonomous robotic platform"> autonomous robotic platform</a>, <a href="https://publications.waset.org/abstracts/search?q=embedded%20systems" title=" embedded systems"> embedded systems</a>, <a href="https://publications.waset.org/abstracts/search?q=hardware" title=" hardware"> hardware</a>, <a href="https://publications.waset.org/abstracts/search?q=localization" title=" localization"> localization</a>, <a href="https://publications.waset.org/abstracts/search?q=modularity" title=" modularity"> modularity</a>, <a href="https://publications.waset.org/abstracts/search?q=multiple%20robots%20synchronization" title=" multiple robots synchronization"> multiple robots synchronization</a>, <a href="https://publications.waset.org/abstracts/search?q=omnidirectional%20movement" title=" omnidirectional movement"> omnidirectional movement</a>, <a href="https://publications.waset.org/abstracts/search?q=safety%20mechanisms" title=" safety mechanisms"> safety mechanisms</a>, <a href="https://publications.waset.org/abstracts/search?q=software" title=" software"> software</a> </p> <a href="https://publications.waset.org/abstracts/130591/embedded-hardware-and-software-design-of-omnidirectional-autonomous-robotic-platform-suitable-for-advanced-driver-assistance-systems-testing-with-focus-on-modularity-and-safety" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/130591.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">143</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">609</span> The Design of Broadband 8x2 Phased Array 5G Antenna MIMO 28 GHz for Base Station</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Muhammad%20Saiful%20Fadhil%20Reyhan">Muhammad Saiful Fadhil Reyhan</a>, <a href="https://publications.waset.org/abstracts/search?q=Yusnita%20Rahayu"> Yusnita Rahayu</a>, <a href="https://publications.waset.org/abstracts/search?q=Fadhel%20Muhammadsyah"> Fadhel Muhammadsyah</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper proposed a design of 16 elements, 8x2 linear fed patch antenna array with 16 ports, for 28 GHz, mm-wave band 5G for base station. The phased array covers along the azimuth plane to provide the coverage to the users in omnidirectional. The proposed antenna is designed RT Duroid 5880 substrate with the overall size of 85x35.6x0.787 mm<sup>3</sup>. The array is operating from 27.43 GHz to 28.34 GHz with a 910 MHz impedance bandwidth. The gain of the array is 18.3 dB, while the suppression of the side lobes is -1.0 dB. The main lobe direction of the array is 15 deg. The array shows a high array gain throughout the impedance bandwidth with overall of VSWR is below 1.12. The design will be proposed in single element and 16 elements antenna. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=5G%20antenna" title="5G antenna">5G antenna</a>, <a href="https://publications.waset.org/abstracts/search?q=28%20GHz" title=" 28 GHz"> 28 GHz</a>, <a href="https://publications.waset.org/abstracts/search?q=MIMO" title=" MIMO"> MIMO</a>, <a href="https://publications.waset.org/abstracts/search?q=omnidirectional" title=" omnidirectional"> omnidirectional</a>, <a href="https://publications.waset.org/abstracts/search?q=phased%20array" title=" phased array"> phased array</a>, <a href="https://publications.waset.org/abstracts/search?q=base%20station" title=" base station"> base station</a>, <a href="https://publications.waset.org/abstracts/search?q=broadband" title=" broadband"> broadband</a> </p> <a href="https://publications.waset.org/abstracts/85986/the-design-of-broadband-8x2-phased-array-5g-antenna-mimo-28-ghz-for-base-station" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/85986.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">249</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">608</span> Reductive Control in the Management of Redundant Actuation</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Mkhinini%20Maher">Mkhinini Maher</a>, <a href="https://publications.waset.org/abstracts/search?q=Knani%20Jilani"> Knani Jilani</a> </p> <p class="card-text"><strong>Abstract:</strong></p> We present in this work the performances of a mobile omnidirectional robot through evaluating its management of the redundancy of actuation. Thus we come to the predictive control implemented. The distribution of the wringer on the robot actions, through the inverse pseudo of Moore-Penrose, corresponds to a -geometric- distribution of efforts. We will show that the load on vehicle wheels would not be equi-distributed in terms of wheels configuration and of robot movement. Thus, the threshold of sliding is not the same for the three wheels of the vehicle. We suggest exploiting the redundancy of actuation to reduce the risk of wheels sliding and to ameliorate, thereby, its accuracy of displacement. This kind of approach was the subject of study for the legged robots. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=mobile%20robot" title="mobile robot">mobile robot</a>, <a href="https://publications.waset.org/abstracts/search?q=actuation" title=" actuation"> actuation</a>, <a href="https://publications.waset.org/abstracts/search?q=redundancy" title=" redundancy"> redundancy</a>, <a href="https://publications.waset.org/abstracts/search?q=omnidirectional" title=" omnidirectional"> omnidirectional</a>, <a href="https://publications.waset.org/abstracts/search?q=inverse%20pseudo%20moore-penrose" title=" inverse pseudo moore-penrose"> inverse pseudo moore-penrose</a>, <a href="https://publications.waset.org/abstracts/search?q=reductive%20control" title=" reductive control"> reductive control</a> </p> <a href="https://publications.waset.org/abstracts/3204/reductive-control-in-the-management-of-redundant-actuation" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/3204.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">510</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">607</span> Low-Cost Mechatronic Design of an Omnidirectional Mobile Robot </h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=S.%20Cobos-Guzman">S. Cobos-Guzman</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents the results of a mechatronic design based on a 4-wheel omnidirectional mobile robot that can be used in indoor logistic applications. The low-level control has been selected using two open-source hardware (Raspberry Pi 3 Model B+ and Arduino Mega 2560) that control four industrial motors, four ultrasound sensors, four optical encoders, a vision system of two cameras, and a Hokuyo URG-04LX-UG01 laser scanner. Moreover, the system is powered with a lithium battery that can supply 24 V DC and a maximum current-hour of 20Ah.The Robot Operating System (ROS) has been implemented in the Raspberry Pi and the performance is evaluated with the selection of the sensors and hardware selected. The mechatronic system is evaluated and proposed safe modes of power distribution for controlling all the electronic devices based on different tests. Therefore, based on different performance results, some recommendations are indicated for using the Raspberry Pi and Arduino in terms of power, communication, and distribution of control for different devices. According to these recommendations, the selection of sensors is distributed in both real-time controllers (Arduino and Raspberry Pi). On the other hand, the drivers of the cameras have been implemented in Linux and a python program has been implemented to access the cameras. These cameras will be used for implementing a deep learning algorithm to recognize people and objects. In this way, the level of intelligence can be increased in combination with the maps that can be obtained from the laser scanner. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=autonomous" title="autonomous">autonomous</a>, <a href="https://publications.waset.org/abstracts/search?q=indoor%20robot" title=" indoor robot"> indoor robot</a>, <a href="https://publications.waset.org/abstracts/search?q=mechatronic" title=" mechatronic"> mechatronic</a>, <a href="https://publications.waset.org/abstracts/search?q=omnidirectional%20robot" title=" omnidirectional robot"> omnidirectional robot</a> </p> <a href="https://publications.waset.org/abstracts/123510/low-cost-mechatronic-design-of-an-omnidirectional-mobile-robot" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/123510.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">175</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">606</span> OmniDrive Model of a Holonomic Mobile Robot</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hussein%20Altartouri">Hussein Altartouri</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper the kinematic and kinetic models of an omnidirectional holonomic mobile robot is presented. The kinematic and kinetic models form the OmniDrive model. Therefore, a mathematical model for the robot equipped with three- omnidirectional wheels is derived. This model which takes into consideration the kinematics and kinetics of the robot, is developed to state space representation. Relative analysis of the velocities and displacements is used for the kinematics of the robot. Lagrange鈥檚 approach is considered in this study for deriving the equation of motion. The drive train and the mechanical assembly only of the Festo Robotino庐 is considered in this model. Mainly the model is developed for motion control. Furthermore, the model can be used for simulation purposes in different virtual environments not only Robotino庐 View. Further use of the model is in the mechatronics research fields with the aim of teaching and learning the advanced control theories. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=mobile%20robot" title="mobile robot">mobile robot</a>, <a href="https://publications.waset.org/abstracts/search?q=omni-direction%20wheel" title=" omni-direction wheel"> omni-direction wheel</a>, <a href="https://publications.waset.org/abstracts/search?q=mathematical%20model" title=" mathematical model"> mathematical model</a>, <a href="https://publications.waset.org/abstracts/search?q=holonomic%20mobile%20robot" title=" holonomic mobile robot"> holonomic mobile robot</a> </p> <a href="https://publications.waset.org/abstracts/11200/omnidrive-model-of-a-holonomic-mobile-robot" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/11200.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">608</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">605</span> A Study of Effective Stereo Matching Method for Long-Wave Infrared Camera Module</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hyun-Koo%20Kim">Hyun-Koo Kim</a>, <a href="https://publications.waset.org/abstracts/search?q=Yonghun%20Kim"> Yonghun Kim</a>, <a href="https://publications.waset.org/abstracts/search?q=Yong-Hoon%20Kim"> Yong-Hoon Kim</a>, <a href="https://publications.waset.org/abstracts/search?q=Ju%20Hee%20Lee"> Ju Hee Lee</a>, <a href="https://publications.waset.org/abstracts/search?q=Myungho%20Song"> Myungho Song</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, we have described an efficient stereo matching method and pedestrian detection method using stereo types LWIR camera. We compared with three types stereo camera algorithm as block matching, ELAS, and SGM. For pedestrian detection using stereo LWIR camera, we used that SGM stereo matching method, free space detection method using u/v-disparity, and HOG feature based pedestrian detection. According to testing result, SGM method has better performance than block matching and ELAS algorithm. Combination of SGM, free space detection, and pedestrian detection using HOG features and SVM classification can detect pedestrian of 30m distance and has a distance error about 30 cm. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=advanced%20driver%20assistance%20system" title="advanced driver assistance system">advanced driver assistance system</a>, <a href="https://publications.waset.org/abstracts/search?q=pedestrian%20detection" title=" pedestrian detection"> pedestrian detection</a>, <a href="https://publications.waset.org/abstracts/search?q=stereo%20matching%20method" title=" stereo matching method"> stereo matching method</a>, <a href="https://publications.waset.org/abstracts/search?q=stereo%20long-wave%20IR%20camera" title=" stereo long-wave IR camera"> stereo long-wave IR camera</a> </p> <a href="https://publications.waset.org/abstracts/58413/a-study-of-effective-stereo-matching-method-for-long-wave-infrared-camera-module" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/58413.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">413</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">604</span> Image Features Comparison-Based Position Estimation Method Using a Camera Sensor</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Jinseon%20Song">Jinseon Song</a>, <a href="https://publications.waset.org/abstracts/search?q=Yongwan%20Park"> Yongwan Park</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, propose method that can user’s position that based on database is built from single camera. Previous positioning calculate distance by arrival-time of signal like GPS (Global Positioning System), RF(Radio Frequency). However, these previous method have weakness because these have large error range according to signal interference. Method for solution estimate position by camera sensor. But, signal camera is difficult to obtain relative position data and stereo camera is difficult to provide real-time position data because of a lot of image data, too. First of all, in this research we build image database at space that able to provide positioning service with single camera. Next, we judge similarity through image matching of database image and transmission image from user. Finally, we decide position of user through position of most similar database image. For verification of propose method, we experiment at real-environment like indoor and outdoor. Propose method is wide positioning range and this method can verify not only position of user but also direction. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=positioning" title="positioning">positioning</a>, <a href="https://publications.waset.org/abstracts/search?q=distance" title=" distance"> distance</a>, <a href="https://publications.waset.org/abstracts/search?q=camera" title=" camera"> camera</a>, <a href="https://publications.waset.org/abstracts/search?q=features" title=" features"> features</a>, <a href="https://publications.waset.org/abstracts/search?q=SURF%28Speed-Up%20Robust%20Features%29" title=" SURF(Speed-Up Robust Features)"> SURF(Speed-Up Robust Features)</a>, <a href="https://publications.waset.org/abstracts/search?q=database" title=" database"> database</a>, <a href="https://publications.waset.org/abstracts/search?q=estimation" title=" estimation"> estimation</a> </p> <a href="https://publications.waset.org/abstracts/11844/image-features-comparison-based-position-estimation-method-using-a-camera-sensor" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/11844.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">349</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">603</span> Multichannel Object Detection with Event Camera</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Rafael%20Iliasov">Rafael Iliasov</a>, <a href="https://publications.waset.org/abstracts/search?q=Alessandro%20Golkar"> Alessandro Golkar</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Object detection based on event vision has been a dynamically growing field in computer vision for the last 16 years. In this work, we create multiple channels from a single event camera and propose an event fusion method (EFM) to enhance object detection in event-based vision systems. Each channel uses a different accumulation buffer to collect events from the event camera. We implement YOLOv7 for object detection, followed by a fusion algorithm. Our multichannel approach outperforms single-channel-based object detection by 0.7% in mean Average Precision (mAP) for detection overlapping ground truth with IOU = 0.5. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=event%20camera" title="event camera">event camera</a>, <a href="https://publications.waset.org/abstracts/search?q=object%20detection%20with%20multimodal%20inputs" title=" object detection with multimodal inputs"> object detection with multimodal inputs</a>, <a href="https://publications.waset.org/abstracts/search?q=multichannel%20fusion" title=" multichannel fusion"> multichannel fusion</a>, <a href="https://publications.waset.org/abstracts/search?q=computer%20vision" title=" computer vision"> computer vision</a> </p> <a href="https://publications.waset.org/abstracts/190247/multichannel-object-detection-with-event-camera" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/190247.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">27</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">602</span> Subpixel Corner Detection for Monocular Camera Linear Model Research</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Guorong%20Sui">Guorong Sui</a>, <a href="https://publications.waset.org/abstracts/search?q=Xingwei%20Jia"> Xingwei Jia</a>, <a href="https://publications.waset.org/abstracts/search?q=Fei%20Tong"> Fei Tong</a>, <a href="https://publications.waset.org/abstracts/search?q=Xiumin%20Gao"> Xiumin Gao</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Camera calibration is a fundamental issue of high precision noncontact measurement. And it is necessary to analyze and study the reliability and application range of its linear model which is often used in the camera calibration. According to the imaging features of monocular cameras, a camera model which is based on the image pixel coordinates and three dimensional space coordinates is built. Using our own customized template, the image pixel coordinate is obtained by the subpixel corner detection method. Without considering the aberration of the optical system, the feature extraction and linearity analysis of the line segment in the template are performed. Moreover, the experiment is repeated 11 times by constantly varying the measuring distance. At last, the linearity of the camera is achieved by fitting 11 groups of data. The camera model measurement results show that the relative error does not exceed 1%, and the repeated measurement error is not more than 0.1 mm magnitude. Meanwhile, it is found that the model has some measurement differences in the different region and object distance. The experiment results show this linear model is simple and practical, and have good linearity within a certain object distance. These experiment results provide a powerful basis for establishment of the linear model of camera. These works will have potential value to the actual engineering measurement. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=camera%20linear%20model" title="camera linear model">camera linear model</a>, <a href="https://publications.waset.org/abstracts/search?q=geometric%20imaging%20relationship" title=" geometric imaging relationship"> geometric imaging relationship</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20pixel%20coordinates" title=" image pixel coordinates"> image pixel coordinates</a>, <a href="https://publications.waset.org/abstracts/search?q=three%20dimensional%20space%20coordinates" title=" three dimensional space coordinates"> three dimensional space coordinates</a>, <a href="https://publications.waset.org/abstracts/search?q=sub-pixel%20corner%20detection" title=" sub-pixel corner detection"> sub-pixel corner detection</a> </p> <a href="https://publications.waset.org/abstracts/77747/subpixel-corner-detection-for-monocular-camera-linear-model-research" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/77747.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">277</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">601</span> X-Corner Detection for Camera Calibration Using Saddle Points</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Abdulrahman%20S.%20Alturki">Abdulrahman S. Alturki</a>, <a href="https://publications.waset.org/abstracts/search?q=John%20S.%20Loomis"> John S. Loomis</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper discusses a corner detection algorithm for camera calibration. Calibration is a necessary step in many computer vision and image processing applications. Robust corner detection for an image of a checkerboard is required to determine intrinsic and extrinsic parameters. In this paper, an algorithm for fully automatic and robust X-corner detection is presented. Checkerboard corner points are automatically found in each image without user interaction or any prior information regarding the number of rows or columns. The approach represents each X-corner with a quadratic fitting function. Using the fact that the X-corners are saddle points, the coefficients in the fitting function are used to identify each corner location. The automation of this process greatly simplifies calibration. Our method is robust against noise and different camera orientations. Experimental analysis shows the accuracy of our method using actual images acquired at different camera locations and orientations. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=camera%20calibration" title="camera calibration">camera calibration</a>, <a href="https://publications.waset.org/abstracts/search?q=corner%20detector" title=" corner detector"> corner detector</a>, <a href="https://publications.waset.org/abstracts/search?q=edge%20detector" title=" edge detector"> edge detector</a>, <a href="https://publications.waset.org/abstracts/search?q=saddle%20points" title=" saddle points"> saddle points</a> </p> <a href="https://publications.waset.org/abstracts/40538/x-corner-detection-for-camera-calibration-using-saddle-points" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/40538.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">406</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">600</span> Frame Camera and Event Camera in Stereo Pair for High-Resolution Sensing</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Khen%20Cohen">Khen Cohen</a>, <a href="https://publications.waset.org/abstracts/search?q=Daniel%20Yankelevich"> Daniel Yankelevich</a>, <a href="https://publications.waset.org/abstracts/search?q=David%20Mendlovic"> David Mendlovic</a>, <a href="https://publications.waset.org/abstracts/search?q=Dan%20Raviv"> Dan Raviv</a> </p> <p class="card-text"><strong>Abstract:</strong></p> We present a 3D stereo system for high-resolution sensing in both the spatial and the temporal domains by combining a frame-based camera and an event-based camera. We establish a method to merge both devices into one unite system and introduce a calibration process, followed by a correspondence technique and interpolation algorithm for 3D reconstruction. We further provide quantitative analysis about our system in terms of depth resolution and additional parameter analysis. We show experimentally how our system performs temporal super-resolution up to effectively 1ms and can detect fast-moving objects and human micro-movements that can be used for micro-expression analysis. We also demonstrate how our method can extract colored events for an event-based camera without any degradation in the spatial resolution, compared to a colored filter array. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=DVS-CIS%20stereo%20vision" title="DVS-CIS stereo vision">DVS-CIS stereo vision</a>, <a href="https://publications.waset.org/abstracts/search?q=micro-movements" title=" micro-movements"> micro-movements</a>, <a href="https://publications.waset.org/abstracts/search?q=temporal%20super-resolution" title=" temporal super-resolution"> temporal super-resolution</a>, <a href="https://publications.waset.org/abstracts/search?q=3D%20reconstruction" title=" 3D reconstruction"> 3D reconstruction</a> </p> <a href="https://publications.waset.org/abstracts/143524/frame-camera-and-event-camera-in-stereo-pair-for-high-resolution-sensing" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/143524.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">297</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">599</span> H.263 Based Video Transceiver for Wireless Camera System</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Won-Ho%20Kim">Won-Ho Kim</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, a design of H.263 based wireless video transceiver is presented for wireless camera system. It uses standard WIFI transceiver and the covering area is up to 100m. Furthermore the standard H.263 video encoding technique is used for video compression since wireless video transmitter is unable to transmit high capacity raw data in real time and the implemented system is capable of streaming at speed of less than 1Mbps using NTSC 720x480 video. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=wireless%20video%20transceiver" title="wireless video transceiver">wireless video transceiver</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20surveillance%20camera" title=" video surveillance camera"> video surveillance camera</a>, <a href="https://publications.waset.org/abstracts/search?q=H.263%20video%20encoding%20digital%20signal%20processing" title=" H.263 video encoding digital signal processing"> H.263 video encoding digital signal processing</a> </p> <a href="https://publications.waset.org/abstracts/12951/h263-based-video-transceiver-for-wireless-camera-system" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/12951.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">364</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">598</span> A Wide View Scheme for Automobile's Black Box</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Jaemyoung%20Lee">Jaemyoung Lee</a> </p> <p class="card-text"><strong>Abstract:</strong></p> We propose a wide view camera scheme for automobile's black box. The proposed scheme uses the commercially available camera lenses of which view angles are about 120掳}^{\circ}掳. In the proposed scheme, we extend the view angle to approximately 200掳 ^{\circ}掳 using two cameras at the front side instead of three lenses with conventional black boxes. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=camera" title="camera">camera</a>, <a href="https://publications.waset.org/abstracts/search?q=black%20box" title=" black box"> black box</a>, <a href="https://publications.waset.org/abstracts/search?q=view%20angle" title=" view angle"> view angle</a>, <a href="https://publications.waset.org/abstracts/search?q=automobile" title=" automobile"> automobile</a> </p> <a href="https://publications.waset.org/abstracts/2582/a-wide-view-scheme-for-automobiles-black-box" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/2582.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">413</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">597</span> Modal Analysis of a Cantilever Beam Using an Inexpensive Smartphone Camera: Motion Magnification Technique</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hasan%20Hassoun">Hasan Hassoun</a>, <a href="https://publications.waset.org/abstracts/search?q=Jaafar%20Hallal"> Jaafar Hallal</a>, <a href="https://publications.waset.org/abstracts/search?q=Denis%20Duhamel"> Denis Duhamel</a>, <a href="https://publications.waset.org/abstracts/search?q=Mohammad%20Hammoud"> Mohammad Hammoud</a>, <a href="https://publications.waset.org/abstracts/search?q=Ali%20Hage%20Diab"> Ali Hage Diab</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper aims to prove the accuracy of an inexpensive smartphone camera as a non-contact vibration sensor to recover the vibration modes of a vibrating structure such as a cantilever beam. A video of a vibrating beam is filmed using a smartphone camera and then processed by the motion magnification technique. Based on this method, the first two natural frequencies and their associated mode shapes are estimated experimentally and compared to the analytical ones. Results show a relative error of less than 4% between the experimental and analytical approaches for the first two natural frequencies of the beam. Also, for the first two-mode shapes, a Modal Assurance Criterion (MAC) value of above 0.9 between the two approaches is obtained. This slight error between the different techniques ensures the viability of a cheap smartphone camera as a non-contact vibration sensor, particularly for structures vibrating at relatively low natural frequencies. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=modal%20analysis" title="modal analysis">modal analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=motion%20magnification" title=" motion magnification"> motion magnification</a>, <a href="https://publications.waset.org/abstracts/search?q=smartphone%20camera" title=" smartphone camera"> smartphone camera</a>, <a href="https://publications.waset.org/abstracts/search?q=structural%20vibration" title=" structural vibration"> structural vibration</a>, <a href="https://publications.waset.org/abstracts/search?q=vibration%20modes" title=" vibration modes"> vibration modes</a> </p> <a href="https://publications.waset.org/abstracts/127525/modal-analysis-of-a-cantilever-beam-using-an-inexpensive-smartphone-camera-motion-magnification-technique" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/127525.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">148</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">596</span> GIS-Based Automatic Flight Planning of Camera-Equipped UAVs for Fire Emergency Response</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Mohammed%20Sulaiman">Mohammed Sulaiman</a>, <a href="https://publications.waset.org/abstracts/search?q=Hexu%20Liu"> Hexu Liu</a>, <a href="https://publications.waset.org/abstracts/search?q=Mohamed%20Binalhaj"> Mohamed Binalhaj</a>, <a href="https://publications.waset.org/abstracts/search?q=William%20W.%20Liou"> William W. Liou</a>, <a href="https://publications.waset.org/abstracts/search?q=Osama%20Abudayyeh"> Osama Abudayyeh</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Emerging technologies such as camera-equipped unmanned aerial vehicles (UAVs) are increasingly being applied in building fire rescue to provide real-time visualization and 3D reconstruction of the entire fireground. However, flight planning of camera-equipped UAVs is usually a manual process, which is not sufficient to fulfill the needs of emergency management. This research proposes a Geographic Information System (GIS)-based approach to automatic flight planning of camera-equipped UAVs for building fire emergency response. In this research, Haversine formula and lawn mowing patterns are employed to automate flight planning based on geometrical and spatial information from GIS. The resulting flight mission satisfies the requirements of 3D reconstruction purposes of the fireground, in consideration of flight execution safety and visibility of camera frames. The proposed approach is implemented within a GIS environment through an application programming interface. A case study is used to demonstrate the effectiveness of the proposed approach. The result shows that flight mission can be generated in a timely manner for application to fire emergency response. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=GIS" title="GIS">GIS</a>, <a href="https://publications.waset.org/abstracts/search?q=camera-equipped%20UAVs" title=" camera-equipped UAVs"> camera-equipped UAVs</a>, <a href="https://publications.waset.org/abstracts/search?q=automatic%20flight%20planning" title=" automatic flight planning"> automatic flight planning</a>, <a href="https://publications.waset.org/abstracts/search?q=fire%20emergency%20response" title=" fire emergency response"> fire emergency response</a> </p> <a href="https://publications.waset.org/abstracts/125166/gis-based-automatic-flight-planning-of-camera-equipped-uavs-for-fire-emergency-response" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/125166.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">125</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">595</span> Object Recognition System Operating from Different Type Vehicles Using Raspberry and OpenCV</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Maria%20Pavlova">Maria Pavlova</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In our days, it is possible to put the camera on different vehicles like quadcopter, train, airplane and etc. The camera also can be the input sensor in many different systems. That means the object recognition like non separate part of monitoring control can be key part of the most intelligent systems. The aim of this paper is to focus of the object recognition process during vehicles movement. During the vehicle鈥檚 movement the camera takes pictures from the environment without storage in Data Base. In case the camera detects a special object (for example human or animal), the system saves the picture and sends it to the work station in real time. This functionality will be very useful in emergency or security situations where is necessary to find a specific object. In another application, the camera can be mounted on crossroad where do not have many people and if one or more persons come on the road, the traffic lights became the green and they can cross the road. In this papers is presented the system has solved the aforementioned problems. It is presented architecture of the object recognition system includes the camera, Raspberry platform, GPS system, neural network, software and Data Base. The camera in the system takes the pictures. The object recognition is done in real time using the OpenCV library and Raspberry microcontroller. An additional feature of this library is the ability to display the GPS coordinates of the captured objects position. The results from this processes will be sent to remote station. So, in this case, we can know the location of the specific object. By neural network, we can learn the module to solve the problems using incoming data and to be part in bigger intelligent system. The present paper focuses on the design and integration of the image recognition like a part of smart systems. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=camera" title="camera">camera</a>, <a href="https://publications.waset.org/abstracts/search?q=object%20recognition" title=" object recognition"> object recognition</a>, <a href="https://publications.waset.org/abstracts/search?q=OpenCV" title=" OpenCV"> OpenCV</a>, <a href="https://publications.waset.org/abstracts/search?q=Raspberry" title=" Raspberry"> Raspberry</a> </p> <a href="https://publications.waset.org/abstracts/81695/object-recognition-system-operating-from-different-type-vehicles-using-raspberry-and-opencv" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/81695.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">218</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">594</span> Person Re-Identification using Siamese Convolutional Neural Network</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Sello%20Mokwena">Sello Mokwena</a>, <a href="https://publications.waset.org/abstracts/search?q=Monyepao%20Thabang"> Monyepao Thabang</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this study, we propose a comprehensive approach to address the challenges in person re-identification models. By combining a centroid tracking algorithm with a Siamese convolutional neural network model, our method excels in detecting, tracking, and capturing robust person features across non-overlapping camera views. The algorithm efficiently identifies individuals in the camera network, while the neural network extracts fine-grained global features for precise cross-image comparisons. The approach's effectiveness is further accentuated by leveraging the camera network topology for guidance. Our empirical analysis on benchmark datasets highlights its competitive performance, particularly evident when background subtraction techniques are selectively applied, underscoring its potential in advancing person re-identification techniques. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=camera%20network" title="camera network">camera network</a>, <a href="https://publications.waset.org/abstracts/search?q=convolutional%20neural%20network%20topology" title=" convolutional neural network topology"> convolutional neural network topology</a>, <a href="https://publications.waset.org/abstracts/search?q=person%20tracking" title=" person tracking"> person tracking</a>, <a href="https://publications.waset.org/abstracts/search?q=person%20re-identification" title=" person re-identification"> person re-identification</a>, <a href="https://publications.waset.org/abstracts/search?q=siamese" title=" siamese"> siamese</a> </p> <a href="https://publications.waset.org/abstracts/171989/person-re-identification-using-siamese-convolutional-neural-network" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/171989.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">72</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">593</span> Hand Gesture Recognition Interface Based on IR Camera</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Yang-Keun%20Ahn">Yang-Keun Ahn</a>, <a href="https://publications.waset.org/abstracts/search?q=Kwang-Soon%20Choi"> Kwang-Soon Choi</a>, <a href="https://publications.waset.org/abstracts/search?q=Young-Choong%20Park"> Young-Choong Park</a>, <a href="https://publications.waset.org/abstracts/search?q=Kwang-Mo%20Jung"> Kwang-Mo Jung</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Vision based user interfaces to control TVs and PCs have the advantage of being able to perform natural control without being limited to a specific device. Accordingly, various studies on hand gesture recognition using RGB cameras or depth cameras have been conducted. However, such cameras have the disadvantage of lacking in accuracy or the construction cost being large. The proposed method uses a low cost IR camera to accurately differentiate between the hand and the background. Also, complicated learning and template matching methodologies are not used, and the correlation between the fingertips extracted through curvatures is utilized to recognize Click and Move gestures. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=recognition" title="recognition">recognition</a>, <a href="https://publications.waset.org/abstracts/search?q=hand%20gestures" title=" hand gestures"> hand gestures</a>, <a href="https://publications.waset.org/abstracts/search?q=infrared%20camera" title=" infrared camera"> infrared camera</a>, <a href="https://publications.waset.org/abstracts/search?q=RGB%20cameras" title=" RGB cameras"> RGB cameras</a> </p> <a href="https://publications.waset.org/abstracts/13373/hand-gesture-recognition-interface-based-on-ir-camera" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/13373.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">406</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">592</span> An Investigation of Direct and Indirect Geo-Referencing Techniques on the Accuracy of Points in Photogrammetry</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=F.%20Yildiz">F. Yildiz</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20Y.%20Oturanc"> S. Y. Oturanc</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Advances technology in the field of photogrammetry replaces analog cameras with reflection on aircraft GPS/IMU system with a digital aerial camera. In this system, when determining the position of the camera with the GPS, camera rotations are also determined by the IMU systems. All around the world, digital aerial cameras have been used for the photogrammetry applications in the last ten years. In this way, in terms of the work done in photogrammetry it is possible to use time effectively, costs to be reduced to a minimum level, the opportunity to make fast and accurate. Geo-referencing techniques that are the cornerstone of the GPS / INS systems, photogrammetric triangulation of images required for balancing (interior and exterior orientation) brings flexibility to the process. Also geo-referencing process; needed in the application of photogrammetry targets to help to reduce the number of ground control points. In this study, the use of direct and indirect geo-referencing techniques on the accuracy of the points was investigated in the production of photogrammetric mapping. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=photogrammetry" title="photogrammetry">photogrammetry</a>, <a href="https://publications.waset.org/abstracts/search?q=GPS%2FIMU%20systems" title=" GPS/IMU systems"> GPS/IMU systems</a>, <a href="https://publications.waset.org/abstracts/search?q=geo-referecing" title=" geo-referecing"> geo-referecing</a>, <a href="https://publications.waset.org/abstracts/search?q=digital%20aerial%20camera" title=" digital aerial camera"> digital aerial camera</a> </p> <a href="https://publications.waset.org/abstracts/13852/an-investigation-of-direct-and-indirect-geo-referencing-techniques-on-the-accuracy-of-points-in-photogrammetry" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/13852.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">411</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">591</span> Self-Calibration of Fish-Eye Camera for Advanced Driver Assistance Systems</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Atef%20Alaaeddine%20Sarraj">Atef Alaaeddine Sarraj</a>, <a href="https://publications.waset.org/abstracts/search?q=Brendan%20Jackman"> Brendan Jackman</a>, <a href="https://publications.waset.org/abstracts/search?q=Frank%20Walsh"> Frank Walsh</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Tomorrow鈥檚 car will be more automated and increasingly connected. Innovative and intuitive interfaces are essential to accompany this functional enrichment. For that, today the automotive companies are competing to offer an advanced driver assistance system (ADAS) which will be able to provide enhanced navigation, collision avoidance, intersection support and lane keeping. These vision-based functions require an accurately calibrated camera. To achieve such differentiation in ADAS requires sophisticated sensors and efficient algorithms. This paper explores the different calibration methods applicable to vehicle-mounted fish-eye cameras with arbitrary fields of view and defines the first steps towards a self-calibration method that adequately addresses ADAS requirements. In particular, we present a self-calibration method after comparing different camera calibration algorithms in the context of ADAS requirements. Our method gathers data from unknown scenes while the car is moving, estimates the camera intrinsic and extrinsic parameters and corrects the wide-angle distortion. Our solution enables continuous and real-time detection of objects, pedestrians, road markings and other cars. In contrast, other camera calibration algorithms for ADAS need pre-calibration, while the presented method calibrates the camera without prior knowledge of the scene and in real-time. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=advanced%20driver%20assistance%20system%20%28ADAS%29" title="advanced driver assistance system (ADAS)">advanced driver assistance system (ADAS)</a>, <a href="https://publications.waset.org/abstracts/search?q=fish-eye" title=" fish-eye"> fish-eye</a>, <a href="https://publications.waset.org/abstracts/search?q=real-time" title=" real-time"> real-time</a>, <a href="https://publications.waset.org/abstracts/search?q=self-calibration" title=" self-calibration"> self-calibration</a> </p> <a href="https://publications.waset.org/abstracts/70853/self-calibration-of-fish-eye-camera-for-advanced-driver-assistance-systems" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/70853.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">252</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">590</span> A Simple Autonomous Hovering and Operating Control of Multicopter Using Only Web Camera</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Kazuya%20Sato">Kazuya Sato</a>, <a href="https://publications.waset.org/abstracts/search?q=Toru%20Kasahara"> Toru Kasahara</a>, <a href="https://publications.waset.org/abstracts/search?q=Junji%20Kuroda"> Junji Kuroda</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, an autonomous hovering control method of multicopter using only Web camera is proposed. Recently, various control method of an autonomous flight for multicopter are proposed. But, in the previously proposed methods, a motion capture system (i.e., OptiTrack) and laser range finder are often used to measure the position and posture of multicopter. To achieve an autonomous flight control of multicopter with simple equipment, we propose an autonomous flight control method using AR marker and Web camera. AR marker can measure the position of multicopter with Cartesian coordinate in three dimensional, then its position connects with aileron, elevator, and accelerator throttle operation. A simple PID control method is applied to the each operation and adjust the controller gains. Experimental result are given to show the effectiveness of our proposed method. Moreover, another simple operation method for autonomous flight control multicopter is also proposed. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=autonomous%20hovering%20control" title="autonomous hovering control">autonomous hovering control</a>, <a href="https://publications.waset.org/abstracts/search?q=multicopter" title=" multicopter"> multicopter</a>, <a href="https://publications.waset.org/abstracts/search?q=Web%20camera" title=" Web camera"> Web camera</a>, <a href="https://publications.waset.org/abstracts/search?q=operation" title=" operation "> operation </a> </p> <a href="https://publications.waset.org/abstracts/20333/a-simple-autonomous-hovering-and-operating-control-of-multicopter-using-only-web-camera" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/20333.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">562</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">589</span> An Automated Procedure for Estimating the Glomerular Filtration Rate and Determining the Normality or Abnormality of the Kidney Stages Using an Artificial Neural Network</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hossain%20A.">Hossain A.</a>, <a href="https://publications.waset.org/abstracts/search?q=Chowdhury%20S.%20I."> Chowdhury S. I.</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Introduction: The use of a gamma camera is a standard procedure in nuclear medicine facilities or hospitals to diagnose chronic kidney disease (CKD), but the gamma camera does not precisely stage the disease. The authors sought to determine whether they could use an artificial neural network to determine whether CKD was in normal or abnormal stages based on GFR values (ANN). Method: The 250 kidney patients (Training 188, Testing 62) who underwent an ultrasonography test to diagnose a renal test in our nuclear medical center were scanned using a gamma camera. Before the scanning procedure, the patients received an injection of 鈦光伖岬怲c-DTPA. The gamma camera computes the pre- and post-syringe radioactive counts after the injection has been pushed into the patient's vein. The artificial neural network uses the softmax function with cross-entropy loss to determine whether CKD is normal or abnormal based on the GFR value in the output layer. Results: The proposed ANN model had a 99.20 % accuracy according to K-fold cross-validation. The sensitivity and specificity were 99.10 and 99.20 %, respectively. AUC was 0.994. Conclusion: The proposed model can distinguish between normal and abnormal stages of CKD by using an artificial neural network. The gamma camera could be upgraded to diagnose normal or abnormal stages of CKD with an appropriate GFR value following the clinical application of the proposed model. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=artificial%20neural%20network" title="artificial neural network">artificial neural network</a>, <a href="https://publications.waset.org/abstracts/search?q=glomerular%20filtration%20rate" title=" glomerular filtration rate"> glomerular filtration rate</a>, <a href="https://publications.waset.org/abstracts/search?q=stages%20of%20the%20kidney" title=" stages of the kidney"> stages of the kidney</a>, <a href="https://publications.waset.org/abstracts/search?q=gamma%20camera" title=" gamma camera"> gamma camera</a> </p> <a href="https://publications.waset.org/abstracts/153994/an-automated-procedure-for-estimating-the-glomerular-filtration-rate-and-determining-the-normality-or-abnormality-of-the-kidney-stages-using-an-artificial-neural-network" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/153994.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">103</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">588</span> Smart Side View Mirror Camera for Real Time System</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Nunziata%20Ivana%20Guarneri">Nunziata Ivana Guarneri</a>, <a href="https://publications.waset.org/abstracts/search?q=Arcangelo%20Bruna"> Arcangelo Bruna</a>, <a href="https://publications.waset.org/abstracts/search?q=Giuseppe%20Spampinato"> Giuseppe Spampinato</a>, <a href="https://publications.waset.org/abstracts/search?q=Antonio%20Buemi"> Antonio Buemi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In the last decade, automotive companies have invested a lot in terms of innovation about many aspects regarding the automatic driver assistance systems. One innovation regards the usage of a smart camera placed on the car’s side mirror for monitoring the back and lateral road situation. A common road scenario is the overtaking of the preceding car and, in this case, a brief distraction or a loss of concentration can lead the driver to undertake this action, even if there is an already overtaking vehicle, leading to serious accidents. A valid support for a secure drive can be a smart camera system, which is able to automatically analyze the road scenario and consequentially to warn the driver when another vehicle is overtaking. This paper describes a method for monitoring the side view of a vehicle by using camera optical flow motion vectors. The proposed solution detects the presence of incoming vehicles, assesses their distance from the host car, and warns the driver through different levels of alert according to the estimated distance. Due to the low complexity and computational cost, the proposed system ensures real time performances. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=camera%20calibration" title="camera calibration">camera calibration</a>, <a href="https://publications.waset.org/abstracts/search?q=ego-motion" title=" ego-motion"> ego-motion</a>, <a href="https://publications.waset.org/abstracts/search?q=Kalman%20filters" title=" Kalman filters"> Kalman filters</a>, <a href="https://publications.waset.org/abstracts/search?q=object%20tracking" title=" object tracking"> object tracking</a>, <a href="https://publications.waset.org/abstracts/search?q=real%20time%20systems" title=" real time systems"> real time systems</a> </p> <a href="https://publications.waset.org/abstracts/79998/smart-side-view-mirror-camera-for-real-time-system" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/79998.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">228</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">587</span> Multiplayer RC-car Driving System in a Collaborative Augmented Reality Environment</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Kikuo%20Asai">Kikuo Asai</a>, <a href="https://publications.waset.org/abstracts/search?q=Yuji%20Sugimoto"> Yuji Sugimoto</a> </p> <p class="card-text"><strong>Abstract:</strong></p> We developed a prototype system for multiplayer RC-car driving in a collaborative Augmented Reality (AR) environment. The tele-existence environment is constructed by superimposing digital data onto images captured by a camera on an RC-car, enabling players to experience an augmented coexistence of the digital content and the real world. Marker-based tracking was used for estimating position and orientation of the camera. The plural RC-cars can be operated in a field where square markers are arranged. The video images captured by the camera are transmitted to a PC for visual tracking. The RC-cars are also tracked by using an infrared camera attached to the ceiling, so that the instability is reduced in the visual tracking. Multimedia data such as texts and graphics are visualized to be overlaid onto the video images in the geometrically correct manner. The prototype system allows a tele-existence sensation to be augmented in a collaborative AR environment. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=multiplayer" title="multiplayer">multiplayer</a>, <a href="https://publications.waset.org/abstracts/search?q=RC-car" title=" RC-car"> RC-car</a>, <a href="https://publications.waset.org/abstracts/search?q=collaborative%20environment" title=" collaborative environment"> collaborative environment</a>, <a href="https://publications.waset.org/abstracts/search?q=augmented%20reality" title=" augmented reality"> augmented reality</a> </p> <a href="https://publications.waset.org/abstracts/4359/multiplayer-rc-car-driving-system-in-a-collaborative-augmented-reality-environment" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/4359.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">289</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">586</span> Improvement of Camera Calibration Based on the Relationship between Focal Length and Aberration Coefficient</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Guorong%20Sui">Guorong Sui</a>, <a href="https://publications.waset.org/abstracts/search?q=Xingwei%20Jia"> Xingwei Jia</a>, <a href="https://publications.waset.org/abstracts/search?q=Chenhui%20Yin"> Chenhui Yin</a>, <a href="https://publications.waset.org/abstracts/search?q=Xiumin%20Gao"> Xiumin Gao</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In the processing of camera-based high precision and non-contact measurement, the geometric-optical aberration is always inevitably disturbing the measuring system. Moreover, the aberration is different with the different focal length, which will increase the difficulties of the system鈥檚 calibration. Therefore, to understand the relationship between the focal length as a function of aberration properties is a very important issue to the calibration of the measuring systems. In this study, we propose a new mathematics model, which is based on the plane calibration method by Zhang Zhengyou, and establish a relationship between the focal length and aberration coefficient. By using the mathematics model and carefully modified compensation templates, the calibration precision of the system can be dramatically improved. The experiment results show that the relative error is less than 1%. It is important for optoelectronic imaging systems that apply to measure, track and position by changing the camera鈥檚 focal length. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=camera%20calibration" title="camera calibration">camera calibration</a>, <a href="https://publications.waset.org/abstracts/search?q=aberration%20coefficient" title=" aberration coefficient"> aberration coefficient</a>, <a href="https://publications.waset.org/abstracts/search?q=vision%20measurement" title=" vision measurement"> vision measurement</a>, <a href="https://publications.waset.org/abstracts/search?q=focal%20length" title=" focal length"> focal length</a>, <a href="https://publications.waset.org/abstracts/search?q=mathematics%20model" title=" mathematics model"> mathematics model</a> </p> <a href="https://publications.waset.org/abstracts/77749/improvement-of-camera-calibration-based-on-the-relationship-between-focal-length-and-aberration-coefficient" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/77749.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">364</span> </span> </div> </div> <ul class="pagination"> <li class="page-item disabled"><span class="page-link">‹</span></li> <li class="page-item active"><span class="page-link">1</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=2">2</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=3">3</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=4">4</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=5">5</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=6">6</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=7">7</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=8">8</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=9">9</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=10">10</a></li> <li class="page-item disabled"><span class="page-link">...</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=20">20</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=21">21</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=omnidirectional%20camera&page=2" rel="next">›</a></li> </ul> </div> </main> <footer> <div id="infolinks" class="pt-3 pb-2"> <div class="container"> <div style="background-color:#f5f5f5;" class="p-3"> <div class="row"> <div class="col-md-2"> <ul class="list-unstyled"> About <li><a href="https://waset.org/page/support">About Us</a></li> <li><a href="https://waset.org/page/support#legal-information">Legal</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/WASET-16th-foundational-anniversary.pdf">WASET celebrates its 16th foundational anniversary</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Account <li><a href="https://waset.org/profile">My Account</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Explore <li><a href="https://waset.org/disciplines">Disciplines</a></li> <li><a href="https://waset.org/conferences">Conferences</a></li> <li><a href="https://waset.org/conference-programs">Conference Program</a></li> <li><a href="https://waset.org/committees">Committees</a></li> <li><a href="https://publications.waset.org">Publications</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Research <li><a href="https://publications.waset.org/abstracts">Abstracts</a></li> <li><a href="https://publications.waset.org">Periodicals</a></li> <li><a href="https://publications.waset.org/archive">Archive</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Open Science <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Philosophy.pdf">Open Science Philosophy</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Award.pdf">Open Science Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Society-Open-Science-and-Open-Innovation.pdf">Open Innovation</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Postdoctoral-Fellowship-Award.pdf">Postdoctoral Fellowship Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Scholarly-Research-Review.pdf">Scholarly Research Review</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Support <li><a href="https://waset.org/page/support">Support</a></li> <li><a href="https://waset.org/profile/messages/create">Contact Us</a></li> <li><a href="https://waset.org/profile/messages/create">Report Abuse</a></li> </ul> </div> </div> </div> </div> </div> <div class="container text-center"> <hr style="margin-top:0;margin-bottom:.3rem;"> <a href="https://creativecommons.org/licenses/by/4.0/" target="_blank" class="text-muted small">Creative Commons Attribution 4.0 International License</a> <div id="copy" class="mt-2">© 2024 World Academy of Science, Engineering and Technology</div> </div> </footer> <a href="javascript:" id="return-to-top"><i class="fas fa-arrow-up"></i></a> <div class="modal" id="modal-template"> <div class="modal-dialog"> <div class="modal-content"> <div class="row m-0 mt-1"> <div class="col-md-12"> <button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">×</span></button> </div> </div> <div class="modal-body"></div> </div> </div> </div> <script src="https://cdn.waset.org/static/plugins/jquery-3.3.1.min.js"></script> <script src="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/js/bootstrap.bundle.min.js"></script> <script src="https://cdn.waset.org/static/js/site.js?v=150220211556"></script> <script> jQuery(document).ready(function() { /*jQuery.get("https://publications.waset.org/xhr/user-menu", function (response) { jQuery('#mainNavMenu').append(response); });*/ jQuery.get({ url: "https://publications.waset.org/xhr/user-menu", cache: false }).then(function(response){ jQuery('#mainNavMenu').append(response); }); }); </script> </body> </html>