CINXE.COM
Search results for: text localization
<!DOCTYPE html> <html lang="en" dir="ltr"> <head> <!-- Google tag (gtag.js) --> <script async src="https://www.googletagmanager.com/gtag/js?id=G-P63WKM1TM1"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('js', new Date()); gtag('config', 'G-P63WKM1TM1'); </script> <!-- Yandex.Metrika counter --> <script type="text/javascript" > (function(m,e,t,r,i,k,a){m[i]=m[i]||function(){(m[i].a=m[i].a||[]).push(arguments)}; m[i].l=1*new Date(); for (var j = 0; j < document.scripts.length; j++) {if (document.scripts[j].src === r) { return; }} k=e.createElement(t),a=e.getElementsByTagName(t)[0],k.async=1,k.src=r,a.parentNode.insertBefore(k,a)}) (window, document, "script", "https://mc.yandex.ru/metrika/tag.js", "ym"); ym(55165297, "init", { clickmap:false, trackLinks:true, accurateTrackBounce:true, webvisor:false }); </script> <noscript><div><img src="https://mc.yandex.ru/watch/55165297" style="position:absolute; left:-9999px;" alt="" /></div></noscript> <!-- /Yandex.Metrika counter --> <!-- Matomo --> <!-- End Matomo Code --> <title>Search results for: text localization</title> <meta name="description" content="Search results for: text localization"> <meta name="keywords" content="text localization"> <meta name="viewport" content="width=device-width, initial-scale=1, minimum-scale=1, maximum-scale=1, user-scalable=no"> <meta charset="utf-8"> <link href="https://cdn.waset.org/favicon.ico" type="image/x-icon" rel="shortcut icon"> <link href="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/css/bootstrap.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/plugins/fontawesome/css/all.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/css/site.css?v=150220211555" rel="stylesheet"> </head> <body> <header> <div class="container"> <nav class="navbar navbar-expand-lg navbar-light"> <a class="navbar-brand" href="https://waset.org"> <img src="https://cdn.waset.org/static/images/wasetc.png" alt="Open Science Research Excellence" title="Open Science Research Excellence" /> </a> <button class="d-block d-lg-none navbar-toggler ml-auto" type="button" data-toggle="collapse" data-target="#navbarMenu" aria-controls="navbarMenu" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <div class="w-100"> <div class="d-none d-lg-flex flex-row-reverse"> <form method="get" action="https://waset.org/search" class="form-inline my-2 my-lg-0"> <input class="form-control mr-sm-2" type="search" placeholder="Search Conferences" value="text localization" name="q" aria-label="Search"> <button class="btn btn-light my-2 my-sm-0" type="submit"><i class="fas fa-search"></i></button> </form> </div> <div class="collapse navbar-collapse mt-1" id="navbarMenu"> <ul class="navbar-nav ml-auto align-items-center" id="mainNavMenu"> <li class="nav-item"> <a class="nav-link" href="https://waset.org/conferences" title="Conferences in 2024/2025/2026">Conferences</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/disciplines" title="Disciplines">Disciplines</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/committees" rel="nofollow">Committees</a> </li> <li class="nav-item dropdown"> <a class="nav-link dropdown-toggle" href="#" id="navbarDropdownPublications" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Publications </a> <div class="dropdown-menu" aria-labelledby="navbarDropdownPublications"> <a class="dropdown-item" href="https://publications.waset.org/abstracts">Abstracts</a> <a class="dropdown-item" href="https://publications.waset.org">Periodicals</a> <a class="dropdown-item" href="https://publications.waset.org/archive">Archive</a> </div> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/page/support" title="Support">Support</a> </li> </ul> </div> </div> </nav> </div> </header> <main> <div class="container mt-4"> <div class="row"> <div class="col-md-9 mx-auto"> <form method="get" action="https://publications.waset.org/abstracts/search"> <div id="custom-search-input"> <div class="input-group"> <i class="fas fa-search"></i> <input type="text" class="search-query" name="q" placeholder="Author, Title, Abstract, Keywords" value="text localization"> <input type="submit" class="btn_search" value="Search"> </div> </div> </form> </div> </div> <div class="row mt-3"> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Commenced</strong> in January 2007</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Frequency:</strong> Monthly</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Edition:</strong> International</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Paper Count:</strong> 1681</div> </div> </div> </div> <h1 class="mt-3 mb-3 text-center" style="font-size:1.6rem;">Search results for: text localization</h1> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1681</span> Text Localization in Fixed-Layout Documents Using Convolutional Networks in a Coarse-to-Fine Manner</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Beier%20Zhu">Beier Zhu</a>, <a href="https://publications.waset.org/abstracts/search?q=Rui%20Zhang"> Rui Zhang</a>, <a href="https://publications.waset.org/abstracts/search?q=Qi%20Song"> Qi Song</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Text contained within fixed-layout documents can be of great semantic value and so requires a high localization accuracy, such as ID cards, invoices, cheques, and passports. Recently, algorithms based on deep convolutional networks achieve high performance on text detection tasks. However, for text localization in fixed-layout documents, such algorithms detect word bounding boxes individually, which ignores the layout information. This paper presents a novel architecture built on convolutional neural networks (CNNs). A global text localization network and a regional bounding-box regression network are introduced to tackle the problem in a coarse-to-fine manner. The text localization network simultaneously locates word bounding points, which takes the layout information into account. The bounding-box regression network inputs the features pooled from arbitrarily sized RoIs and refine the localizations. These two networks share their convolutional features and are trained jointly. A typical type of fixed-layout documents: ID cards, is selected to evaluate the effectiveness of the proposed system. These networks are trained on data cropped from nature scene images, and synthetic data produced by a synthetic text generation engine. Experiments show that our approach locates high accuracy word bounding boxes and achieves state-of-the-art performance. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=bounding%20box%20regression" title="bounding box regression">bounding box regression</a>, <a href="https://publications.waset.org/abstracts/search?q=convolutional%20networks" title=" convolutional networks"> convolutional networks</a>, <a href="https://publications.waset.org/abstracts/search?q=fixed-layout%20documents" title=" fixed-layout documents"> fixed-layout documents</a>, <a href="https://publications.waset.org/abstracts/search?q=text%20localization" title=" text localization"> text localization</a> </p> <a href="https://publications.waset.org/abstracts/85636/text-localization-in-fixed-layout-documents-using-convolutional-networks-in-a-coarse-to-fine-manner" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/85636.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">194</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1680</span> Video Text Information Detection and Localization in Lecture Videos Using Moments </h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Belkacem%20Soundes">Belkacem Soundes</a>, <a href="https://publications.waset.org/abstracts/search?q=Guezouli%20Larbi"> Guezouli Larbi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents a robust and accurate method for text detection and localization over lecture videos. Frame regions are classified into text or background based on visual feature analysis. However, lecture video shows significant degradation mainly related to acquisition conditions, camera motion and environmental changes resulting in low quality videos. Hence, affecting feature extraction and description efficiency. Moreover, traditional text detection methods cannot be directly applied to lecture videos. Therefore, robust feature extraction methods dedicated to this specific video genre are required for robust and accurate text detection and extraction. Method consists of a three-step process: Slide region detection and segmentation; Feature extraction and non-text filtering. For robust and effective features extraction moment functions are used. Two distinct types of moments are used: orthogonal and non-orthogonal. For orthogonal Zernike Moments, both Pseudo Zernike moments are used, whereas for non-orthogonal ones Hu moments are used. Expressivity and description efficiency are given and discussed. Proposed approach shows that in general, orthogonal moments show high accuracy in comparison to the non-orthogonal one. Pseudo Zernike moments are more effective than Zernike with better computation time. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=text%20detection" title="text detection">text detection</a>, <a href="https://publications.waset.org/abstracts/search?q=text%20localization" title=" text localization"> text localization</a>, <a href="https://publications.waset.org/abstracts/search?q=lecture%20videos" title=" lecture videos"> lecture videos</a>, <a href="https://publications.waset.org/abstracts/search?q=pseudo%20zernike%20moments" title=" pseudo zernike moments"> pseudo zernike moments</a> </p> <a href="https://publications.waset.org/abstracts/109549/video-text-information-detection-and-localization-in-lecture-videos-using-moments" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/109549.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">151</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1679</span> Extraction of Text Subtitles in Multimedia Systems</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Amarjit%20Singh">Amarjit Singh</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, a method for extraction of text subtitles in large video is proposed. The video data needs to be annotated for many multimedia applications. Text is incorporated in digital video for the motive of providing useful information about that video. So need arises to detect text present in video to understanding and video indexing. This is achieved in two steps. First step is text localization and the second step is text verification. The method of text detection can be extended to text recognition which finds applications in automatic video indexing; video annotation and content based video retrieval. The method has been tested on various types of videos. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video" title="video">video</a>, <a href="https://publications.waset.org/abstracts/search?q=subtitles" title=" subtitles"> subtitles</a>, <a href="https://publications.waset.org/abstracts/search?q=extraction" title=" extraction"> extraction</a>, <a href="https://publications.waset.org/abstracts/search?q=annotation" title=" annotation"> annotation</a>, <a href="https://publications.waset.org/abstracts/search?q=frames" title=" frames"> frames</a> </p> <a href="https://publications.waset.org/abstracts/24441/extraction-of-text-subtitles-in-multimedia-systems" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/24441.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">601</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1678</span> On-Road Text Detection Platform for Driver Assistance Systems</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Guezouli%20Larbi">Guezouli Larbi</a>, <a href="https://publications.waset.org/abstracts/search?q=Belkacem%20Soundes"> Belkacem Soundes</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The automation of the text detection process can help the human in his driving task. Its application can be very useful to help drivers to have more information about their environment by facilitating the reading of road signs such as directional signs, events, stores, etc. In this paper, a system consisting of two stages has been proposed. In the first one, we used pseudo-Zernike moments to pinpoint areas of the image that may contain text. The architecture of this part is based on three main steps, region of interest (ROI) detection, text localization, and non-text region filtering. Then, in the second step, we present a convolutional neural network architecture (On-Road Text Detection Network - ORTDN) which is considered a classification phase. The results show that the proposed framework achieved ≈ 35 fps and an mAP of ≈ 90%, thus a low computational time with competitive accuracy. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=text%20detection" title="text detection">text detection</a>, <a href="https://publications.waset.org/abstracts/search?q=CNN" title=" CNN"> CNN</a>, <a href="https://publications.waset.org/abstracts/search?q=PZM" title=" PZM"> PZM</a>, <a href="https://publications.waset.org/abstracts/search?q=deep%20learning" title=" deep learning"> deep learning</a> </p> <a href="https://publications.waset.org/abstracts/161507/on-road-text-detection-platform-for-driver-assistance-systems" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/161507.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">83</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1677</span> Enhanced Weighted Centroid Localization Algorithm for Indoor Environments</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=I.%20Ni%C5%BEeti%C4%87%20Kosovi%C4%87">I. Nižetić Kosović</a>, <a href="https://publications.waset.org/abstracts/search?q=T.%20Jagu%C5%A1t"> T. Jagušt</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Lately, with the increasing number of location-based applications, demand for highly accurate and reliable indoor localization became urgent. This is a challenging problem, due to the measurement variance which is the consequence of various factors like obstacles, equipment properties and environmental changes in complex nature of indoor environments. In this paper we propose low-cost custom-setup infrastructure solution and localization algorithm based on the Weighted Centroid Localization (WCL) method. Localization accuracy is increased by several enhancements: calibration of RSSI values gained from wireless nodes, repetitive measurements of RSSI to exclude deviating values from the position estimation, and by considering orientation of the device according to the wireless nodes. We conducted several experiments to evaluate the proposed algorithm. High accuracy of ~1m was achieved. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=indoor%20environment" title="indoor environment">indoor environment</a>, <a href="https://publications.waset.org/abstracts/search?q=received%20signal%20strength%20indicator" title=" received signal strength indicator"> received signal strength indicator</a>, <a href="https://publications.waset.org/abstracts/search?q=weighted%20centroid%20localization" title=" weighted centroid localization"> weighted centroid localization</a>, <a href="https://publications.waset.org/abstracts/search?q=wireless%20localization" title=" wireless localization"> wireless localization</a> </p> <a href="https://publications.waset.org/abstracts/11878/enhanced-weighted-centroid-localization-algorithm-for-indoor-environments" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/11878.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">232</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1676</span> Bug Localization on Single-Line Bugs of Apache Commons Math Library</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Cherry%20Oo">Cherry Oo</a>, <a href="https://publications.waset.org/abstracts/search?q=Hnin%20Min%20Oo"> Hnin Min Oo</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Software bug localization is one of the most costly tasks in program repair technique. Therefore, there is a high claim for automated bug localization techniques that can monitor programmers to the locations of bugs, with slight human arbitration. Spectrum-based bug localization aims to help software developers to discover bugs rapidly by investigating abstractions of the program traces to make a ranking list of most possible buggy modules. Using the Apache Commons Math library project, we study the diagnostic accuracy using our spectrum-based bug localization metric. Our outcomes show that the greater performance of a specific similarity coefficient, used to inspect the program spectra, is mostly effective on localizing of single line bugs. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=software%20testing" title="software testing">software testing</a>, <a href="https://publications.waset.org/abstracts/search?q=bug%20localization" title=" bug localization"> bug localization</a>, <a href="https://publications.waset.org/abstracts/search?q=program%20spectra" title=" program spectra"> program spectra</a>, <a href="https://publications.waset.org/abstracts/search?q=bug" title=" bug"> bug</a> </p> <a href="https://publications.waset.org/abstracts/104890/bug-localization-on-single-line-bugs-of-apache-commons-math-library" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/104890.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">143</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1675</span> Influence of Scalable Energy-Related Sensor Parameters on Acoustic Localization Accuracy in Wireless Sensor Swarms</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Joyraj%20Chakraborty">Joyraj Chakraborty</a>, <a href="https://publications.waset.org/abstracts/search?q=Geoffrey%20Ottoy"> Geoffrey Ottoy</a>, <a href="https://publications.waset.org/abstracts/search?q=Jean-Pierre%20Goemaere"> Jean-Pierre Goemaere</a>, <a href="https://publications.waset.org/abstracts/search?q=Lieven%20De%20Strycker"> Lieven De Strycker</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Sensor swarms can be a cost-effectieve and more user-friendly alternative for location based service systems in different application like health-care. To increase the lifetime of such swarm networks, the energy consumption should be scaled to the required localization accuracy. In this paper we have investigated some parameter for energy model that couples localization accuracy to energy-related sensor parameters such as signal length,Bandwidth and sample frequency. The goal is to use the model for the localization of undetermined environmental sounds, by means of wireless acoustic sensors. we first give an overview of TDOA-based localization together with the primary sources of TDOA error (including reverberation effects, Noise). Then we show that in localization, the signal sample rate can be under the Nyquist frequency, provided that enough frequency components remain present in the undersampled signal. The resulting localization error is comparable with that of similar localization systems. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=sensor%20swarms" title="sensor swarms">sensor swarms</a>, <a href="https://publications.waset.org/abstracts/search?q=localization" title=" localization"> localization</a>, <a href="https://publications.waset.org/abstracts/search?q=wireless%20sensor%20swarms" title=" wireless sensor swarms"> wireless sensor swarms</a>, <a href="https://publications.waset.org/abstracts/search?q=scalable%20energy" title=" scalable energy"> scalable energy</a> </p> <a href="https://publications.waset.org/abstracts/29900/influence-of-scalable-energy-related-sensor-parameters-on-acoustic-localization-accuracy-in-wireless-sensor-swarms" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/29900.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">422</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1674</span> The Making of a Yijing (Classic of Changes) Cultural Sphere in Asia</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ng%20Wai%20Ming">Ng Wai Ming</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The Yijing (Classic of Changes) is one of the most influential Chinese classics, and its text, images and divination have been widely studied and used by different people in the world from past to present. Its impact in Asia has been particularly strong due to cultural and geographical proximity. Based on many years of textual study of the history of the Yijing in the Sinosphere, the author attempts to identify various levels of acceptance and localization of the Yijing in different Asian regions, including Japan, Korea, the Ryukyu Kingdom, Vietnam, Mongolia and Tibet. It will create a new concept of “Yijing cultural sphere” to explain the popularization and indigenization of the Yijing in Asia. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=classic%20of%20changes" title="classic of changes">classic of changes</a>, <a href="https://publications.waset.org/abstracts/search?q=asia" title=" asia"> asia</a>, <a href="https://publications.waset.org/abstracts/search?q=sinosphere" title=" sinosphere"> sinosphere</a>, <a href="https://publications.waset.org/abstracts/search?q=localization" title=" localization"> localization</a> </p> <a href="https://publications.waset.org/abstracts/175084/the-making-of-a-yijing-classic-of-changes-cultural-sphere-in-asia" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/175084.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">62</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1673</span> The Selection of the Nearest Anchor Using Received Signal Strength Indication (RSSI)</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hichem%20Sassi">Hichem Sassi</a>, <a href="https://publications.waset.org/abstracts/search?q=Tawfik%20Najeh"> Tawfik Najeh</a>, <a href="https://publications.waset.org/abstracts/search?q=Noureddine%20Liouane"> Noureddine Liouane</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The localization information is crucial for the operation of WSN. There are principally two types of localization algorithms. The Range-based localization algorithm has strict requirements on hardware; thus, it is expensive to be implemented in practice. The Range-free localization algorithm reduces the hardware cost. However, it can only achieve high accuracy in ideal scenarios. In this paper, we locate unknown nodes by incorporating the advantages of these two types of methods. The proposed algorithm makes the unknown nodes select the nearest anchor using the Received Signal Strength Indicator (RSSI) and choose two other anchors which are the most accurate to achieve the estimated location. Our algorithm improves the localization accuracy compared with previous algorithms, which has been demonstrated by the simulating results. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=WSN" title="WSN">WSN</a>, <a href="https://publications.waset.org/abstracts/search?q=localization" title=" localization"> localization</a>, <a href="https://publications.waset.org/abstracts/search?q=DV-Hop" title=" DV-Hop"> DV-Hop</a>, <a href="https://publications.waset.org/abstracts/search?q=RSSI" title=" RSSI"> RSSI</a> </p> <a href="https://publications.waset.org/abstracts/36091/the-selection-of-the-nearest-anchor-using-received-signal-strength-indication-rssi" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/36091.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">360</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1672</span> An Introductory Study on Optimization Algorithm for Movable Sensor Network-Based Odor Source Localization</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Yossiri%20Ariyakul">Yossiri Ariyakul</a>, <a href="https://publications.waset.org/abstracts/search?q=Piyakiat%20Insom"> Piyakiat Insom</a>, <a href="https://publications.waset.org/abstracts/search?q=Poonyawat%20Sangiamkulthavorn"> Poonyawat Sangiamkulthavorn</a>, <a href="https://publications.waset.org/abstracts/search?q=Takamichi%20Nakamoto"> Takamichi Nakamoto</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, the method of optimization algorithm for sensor network comprised of movable sensor nodes which can be used for odor source localization was proposed. A sensor node is composed of an odor sensor, an anemometer, and a wireless communication module. The odor intensity measured from the sensor nodes are sent to the processor to perform the localization based on optimization algorithm by which the odor source localization map is obtained as a result. The map can represent the exact position of the odor source or show the direction toward it remotely. The proposed method was experimentally validated by creating the odor source localization map using three, four, and five sensor nodes in which the accuracy to predict the position of the odor source can be observed. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=odor%20sensor" title="odor sensor">odor sensor</a>, <a href="https://publications.waset.org/abstracts/search?q=odor%20source%20localization" title=" odor source localization"> odor source localization</a>, <a href="https://publications.waset.org/abstracts/search?q=optimization" title=" optimization"> optimization</a>, <a href="https://publications.waset.org/abstracts/search?q=sensor%20network" title=" sensor network"> sensor network</a> </p> <a href="https://publications.waset.org/abstracts/76005/an-introductory-study-on-optimization-algorithm-for-movable-sensor-network-based-odor-source-localization" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/76005.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">299</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1671</span> Hybrid Localization Schemes for Wireless Sensor Networks </h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Fatima%20Babar">Fatima Babar</a>, <a href="https://publications.waset.org/abstracts/search?q=Majid%20I.%20Khan"> Majid I. Khan</a>, <a href="https://publications.waset.org/abstracts/search?q=Malik%20Najmus%20Saqib"> Malik Najmus Saqib</a>, <a href="https://publications.waset.org/abstracts/search?q=Muhammad%20Tahir"> Muhammad Tahir</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This article provides range based improvements over a well-known single-hop range free localization scheme, Approximate Point in Triangulation (APIT) by proposing an energy efficient Barycentric coordinate based Point-In-Triangulation (PIT) test along with PIT based trilateration. These improvements result in energy efficiency, reduced localization error and improved localization coverage compared to APIT and its variants. Moreover, we propose to embed Received signal strength indication (RSSI) based distance estimation in DV-Hop which is a multi-hop localization scheme. The proposed localization algorithm achieves energy efficiency and reduced localization error compared to DV-Hop and its available improvements. Furthermore, a hybrid multi-hop localization scheme is also proposed that utilize Barycentric coordinate based PIT test and both range based (Received signal strength indicator) and range free (hop count) techniques for distance estimation. Our experimental results provide evidence that proposed hybrid multi-hop localization scheme results in two to five times reduction in the localization error compare to DV-Hop and its variants, at reduced energy requirements. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Localization" title="Localization">Localization</a>, <a href="https://publications.waset.org/abstracts/search?q=Trilateration" title=" Trilateration"> Trilateration</a>, <a href="https://publications.waset.org/abstracts/search?q=Triangulation" title=" Triangulation"> Triangulation</a>, <a href="https://publications.waset.org/abstracts/search?q=Wireless%20Sensor%20Networks" title=" Wireless Sensor Networks"> Wireless Sensor Networks</a> </p> <a href="https://publications.waset.org/abstracts/35863/hybrid-localization-schemes-for-wireless-sensor-networks" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/35863.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">467</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1670</span> Factor Analysis on Localization of Human Resources of Japanese Firms in Taiwan</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Nana%20Weng">Nana Weng</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Localization in the aspect of human resource means more diversity and more opportunities. The main purpose of this article is to identify the perception of local employees and intermediate managers (non-Japanese) and figure out exploratory factors which have been contributing and blocking the level of localization in the aspect of human resource management by using EFA (Exploratory Factors Analysis). Questionnaires will be designed for local employees and managers to inquire about the perceptions of regulations and implementation regarding recruitment, training and development, promotion and rewarding. The study finds that Japanese firms have worked well in the process of localization, especially in hiring and training local staffs in Taiwan. The significance of this study lies in paying more attention to the perception of local employees and intermediate managers regarding localization rather than interviews results from Japanese expatriates or top HR managers who are in charging of localization policy-making. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Japanese%20firms%20in%20Taiwan" title="Japanese firms in Taiwan">Japanese firms in Taiwan</a>, <a href="https://publications.waset.org/abstracts/search?q=localization%20of%20human%20resources" title=" localization of human resources"> localization of human resources</a>, <a href="https://publications.waset.org/abstracts/search?q=exploratory%20factors%20analysis" title=" exploratory factors analysis"> exploratory factors analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=local%20employees%20and%20intermediate%20managers" title=" local employees and intermediate managers"> local employees and intermediate managers</a> </p> <a href="https://publications.waset.org/abstracts/60959/factor-analysis-on-localization-of-human-resources-of-japanese-firms-in-taiwan" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/60959.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">312</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1669</span> A Summary-Based Text Classification Model for Graph Attention Networks</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Shuo%20Liu">Shuo Liu</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In Chinese text classification tasks, redundant words and phrases can interfere with the formation of extracted and analyzed text information, leading to a decrease in the accuracy of the classification model. To reduce irrelevant elements, extract and utilize text content information more efficiently and improve the accuracy of text classification models. In this paper, the text in the corpus is first extracted using the TextRank algorithm for abstraction, the words in the abstract are used as nodes to construct a text graph, and then the graph attention network (GAT) is used to complete the task of classifying the text. Testing on a Chinese dataset from the network, the classification accuracy was improved over the direct method of generating graph structures using text. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Chinese%20natural%20language%20processing" title="Chinese natural language processing">Chinese natural language processing</a>, <a href="https://publications.waset.org/abstracts/search?q=text%20classification" title=" text classification"> text classification</a>, <a href="https://publications.waset.org/abstracts/search?q=abstract%20extraction" title=" abstract extraction"> abstract extraction</a>, <a href="https://publications.waset.org/abstracts/search?q=graph%20attention%20network" title=" graph attention network"> graph attention network</a> </p> <a href="https://publications.waset.org/abstracts/158060/a-summary-based-text-classification-model-for-graph-attention-networks" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/158060.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">100</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1668</span> Urdu Text Extraction Method from Images</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Samabia%20Tehsin">Samabia Tehsin</a>, <a href="https://publications.waset.org/abstracts/search?q=Sumaira%20Kausar"> Sumaira Kausar</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Due to the vast increase in the multimedia data in recent years, efficient and robust retrieval techniques are needed to retrieve and index images/ videos. Text embedded in the images can serve as the strong retrieval tool for images. This is the reason that text extraction is an area of research with increasing attention. English text extraction is the focus of many researchers but very less work has been done on other languages like Urdu. This paper is focusing on Urdu text extraction from video frames. This paper presents a text detection feature set, which has the ability to deal up with most of the problems connected with the text extraction process. To test the validity of the method, it is tested on Urdu news dataset, which gives promising results. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=caption%20text" title="caption text">caption text</a>, <a href="https://publications.waset.org/abstracts/search?q=content-based%20image%20retrieval" title=" content-based image retrieval"> content-based image retrieval</a>, <a href="https://publications.waset.org/abstracts/search?q=document%20analysis" title=" document analysis"> document analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=text%20extraction" title=" text extraction"> text extraction</a> </p> <a href="https://publications.waset.org/abstracts/9566/urdu-text-extraction-method-from-images" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/9566.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">516</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1667</span> Indoor Localization by Pattern Matching Method Based on Extended Database</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Gyumin%20Hwang">Gyumin Hwang</a>, <a href="https://publications.waset.org/abstracts/search?q=Jihong%20Lee"> Jihong Lee</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper studied the CSS-based indoor localization system which is easy to implement, inexpensive to compose the systems, additionally CSS-based indoor localization system covers larger area than other system. However, this system has problem which is affected by reflected distance data. This problem in localization is caused by the multi-path effect. Error caused by multi-path is difficult to be corrected because the indoor environment cannot be described. In this paper, in order to solve the problem by multi-path, we have supplemented the localization system by using pattern matching method based on extended database. Thereby, this method improves precision of estimated. Also this method is verified by experiments in gymnasium. Database was constructed by 1 m intervals, and 16 sample data were collected from random position inside the region of DB points. As a result, this paper shows higher accuracy than existing method through graph and table. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=chirp%20spread%20spectrum" title="chirp spread spectrum">chirp spread spectrum</a>, <a href="https://publications.waset.org/abstracts/search?q=indoor%20localization" title=" indoor localization"> indoor localization</a>, <a href="https://publications.waset.org/abstracts/search?q=pattern-matching" title=" pattern-matching"> pattern-matching</a>, <a href="https://publications.waset.org/abstracts/search?q=time%20of%20arrival" title=" time of arrival"> time of arrival</a>, <a href="https://publications.waset.org/abstracts/search?q=multi-path" title=" multi-path"> multi-path</a>, <a href="https://publications.waset.org/abstracts/search?q=mahalanobis%20distance" title=" mahalanobis distance"> mahalanobis distance</a>, <a href="https://publications.waset.org/abstracts/search?q=reception%20rate" title=" reception rate"> reception rate</a>, <a href="https://publications.waset.org/abstracts/search?q=simultaneous%20localization%20and%20mapping" title=" simultaneous localization and mapping"> simultaneous localization and mapping</a>, <a href="https://publications.waset.org/abstracts/search?q=laser%20range%20finder" title=" laser range finder"> laser range finder</a> </p> <a href="https://publications.waset.org/abstracts/4353/indoor-localization-by-pattern-matching-method-based-on-extended-database" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/4353.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">244</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1666</span> Improved Acoustic Source Sensing and Localization Based On Robot Locomotion</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=V.%20Ramu%20Reddy">V. Ramu Reddy</a>, <a href="https://publications.waset.org/abstracts/search?q=Parijat%20Deshpande"> Parijat Deshpande</a>, <a href="https://publications.waset.org/abstracts/search?q=Ranjan%20Dasgupta"> Ranjan Dasgupta</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents different methodology for an acoustic source sensing and localization in an unknown environment. The developed methodology includes an acoustic based sensing and localization system, a converging target localization based on the recursive direction of arrival (DOA) error minimization, and a regressive obstacle avoidance function. Our method is able to augment the existing proven localization techniques and improve results incrementally by utilizing robot locomotion and is capable of converging to a position estimate with greater accuracy using fewer measurements. The results also evinced the DOA error minimization at each iteration, improvement in time for reaching the destination and the efficiency of this target localization method as gradually converging to the real target position. Initially, the system is tested using Kinect mounted on turntable with DOA markings which serve as a ground truth and then our approach is validated using a FireBird VI (FBVI) mobile robot on which Kinect is used to obtain bearing information. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=acoustic%20source%20localization" title="acoustic source localization">acoustic source localization</a>, <a href="https://publications.waset.org/abstracts/search?q=acoustic%20sensing" title=" acoustic sensing"> acoustic sensing</a>, <a href="https://publications.waset.org/abstracts/search?q=recursive%20direction%20of%20arrival" title=" recursive direction of arrival"> recursive direction of arrival</a>, <a href="https://publications.waset.org/abstracts/search?q=robot%20locomotion" title=" robot locomotion"> robot locomotion</a> </p> <a href="https://publications.waset.org/abstracts/43889/improved-acoustic-source-sensing-and-localization-based-on-robot-locomotion" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/43889.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">492</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1665</span> Quantum Localization of Vibrational Mirror in Cavity Optomechanics</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Madiha%20Tariq">Madiha Tariq</a>, <a href="https://publications.waset.org/abstracts/search?q=Hena%20Rabbani"> Hena Rabbani</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Recently, cavity-optomechanics becomes an extensive research field that has manipulated the mechanical effects of light for coupling of the optical field with other physical objects specifically with regards to dynamical localization. We investigate the dynamical localization (both in momentum and position space) for a vibrational mirror in a Fabry-Pérot cavity driven by a single mode optical field and a transverse probe field. The weak probe field phenomenon results in classical chaos in phase space and spatio temporal dynamics in position |ψ(x)²| and momentum space |ψ(p)²| versus time show quantum localization in both momentum and position space. Also, we discuss the parametric dependencies of dynamical localization for a designated set of parameters to be experimentally feasible. Our work opens an avenue to manipulate the other optical phenomena and applicability of proposed work can be prolonged to turn-able laser sources in the future. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=dynamical%20localization" title="dynamical localization">dynamical localization</a>, <a href="https://publications.waset.org/abstracts/search?q=cavity%20optomechanics" title=" cavity optomechanics"> cavity optomechanics</a>, <a href="https://publications.waset.org/abstracts/search?q=Hamiltonian%20chaos" title=" Hamiltonian chaos"> Hamiltonian chaos</a>, <a href="https://publications.waset.org/abstracts/search?q=probe%20field" title=" probe field"> probe field</a> </p> <a href="https://publications.waset.org/abstracts/108127/quantum-localization-of-vibrational-mirror-in-cavity-optomechanics" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/108127.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">150</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1664</span> Small Text Extraction from Documents and Chart Images</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Rominkumar%20Busa">Rominkumar Busa</a>, <a href="https://publications.waset.org/abstracts/search?q=Shahira%20K.%20C."> Shahira K. C.</a>, <a href="https://publications.waset.org/abstracts/search?q=Lijiya%20A."> Lijiya A.</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Text recognition is an important area in computer vision which deals with detecting and recognising text from an image. The Optical Character Recognition (OCR) is a saturated area these days and with very good text recognition accuracy. However the same OCR methods when applied on text with small font sizes like the text data of chart images, the recognition rate is less than 30%. In this work, aims to extract small text in images using the deep learning model, CRNN with CTC loss. The text recognition accuracy is found to improve by applying image enhancement by super resolution prior to CRNN model. We also observe the text recognition rate further increases by 18% by applying the proposed method, which involves super resolution and character segmentation followed by CRNN with CTC loss. The efficiency of the proposed method shows that further pre-processing on chart image text and other small text images will improve the accuracy further, thereby helping text extraction from chart images. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=small%20text%20extraction" title="small text extraction">small text extraction</a>, <a href="https://publications.waset.org/abstracts/search?q=OCR" title=" OCR"> OCR</a>, <a href="https://publications.waset.org/abstracts/search?q=scene%20text%20recognition" title=" scene text recognition"> scene text recognition</a>, <a href="https://publications.waset.org/abstracts/search?q=CRNN" title=" CRNN"> CRNN</a> </p> <a href="https://publications.waset.org/abstracts/150310/small-text-extraction-from-documents-and-chart-images" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/150310.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">125</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1663</span> Text Data Preprocessing Library: Bilingual Approach</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Kabil%20Boukhari">Kabil Boukhari</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In the context of information retrieval, the selection of the most relevant words is a very important step. In fact, the text cleaning allows keeping only the most representative words for a better use. In this paper, we propose a library for the purpose text preprocessing within an implemented application to facilitate this task. This study has two purposes. The first, is to present the related work of the various steps involved in text preprocessing, presenting the segmentation, stemming and lemmatization algorithms that could be efficient in the rest of study. The second, is to implement a developed tool for text preprocessing in French and English. This library accepts unstructured text as input and provides the preprocessed text as output, based on a set of rules and on a base of stop words for both languages. The proposed library has been made on different corpora and gave an interesting result. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=text%20preprocessing" title="text preprocessing">text preprocessing</a>, <a href="https://publications.waset.org/abstracts/search?q=segmentation" title=" segmentation"> segmentation</a>, <a href="https://publications.waset.org/abstracts/search?q=knowledge%20extraction" title=" knowledge extraction"> knowledge extraction</a>, <a href="https://publications.waset.org/abstracts/search?q=normalization" title=" normalization"> normalization</a>, <a href="https://publications.waset.org/abstracts/search?q=text%20generation" title=" text generation"> text generation</a>, <a href="https://publications.waset.org/abstracts/search?q=information%20retrieval" title=" information retrieval"> information retrieval</a> </p> <a href="https://publications.waset.org/abstracts/150846/text-data-preprocessing-library-bilingual-approach" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/150846.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">94</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1662</span> Local Food Movements and Community Building in Turkey</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Derya%20Nizam">Derya Nizam</a> </p> <p class="card-text"><strong>Abstract:</strong></p> An alternative understanding of "localization" has gained significance as the ecological and social issues associated with the growing pressure of agricultural homogeneity and standardization become more apparent. Through an analysis of a case study on an alternative food networks in Turkey, this research seeks to critically examine the localization movement. The results indicate that the idea of localization helps to create new niche markets by creating place-based labels, but it also strengthens local identities through social networks that connect rural and urban areas. In that context, localization manifests as a commodification movement that appropriates local and cultural values to generate capitalist profit, as well as a grassroots movement that strengthens the resilience of local communities. This research addresses the potential of community development approaches in the democratization of global agro-food networks. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=community%20building" title="community building">community building</a>, <a href="https://publications.waset.org/abstracts/search?q=local%20food" title=" local food"> local food</a>, <a href="https://publications.waset.org/abstracts/search?q=alternative%20food%20movements" title=" alternative food movements"> alternative food movements</a>, <a href="https://publications.waset.org/abstracts/search?q=localization" title=" localization"> localization</a> </p> <a href="https://publications.waset.org/abstracts/177916/local-food-movements-and-community-building-in-turkey" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/177916.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">79</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1661</span> Visual Search Based Indoor Localization in Low Light via RGB-D Camera</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Yali%20Zheng">Yali Zheng</a>, <a href="https://publications.waset.org/abstracts/search?q=Peipei%20Luo"> Peipei Luo</a>, <a href="https://publications.waset.org/abstracts/search?q=Shinan%20Chen"> Shinan Chen</a>, <a href="https://publications.waset.org/abstracts/search?q=Jiasheng%20Hao"> Jiasheng Hao</a>, <a href="https://publications.waset.org/abstracts/search?q=Hong%20Cheng"> Hong Cheng</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Most of traditional visual indoor navigation algorithms and methods only consider the localization in ordinary daytime, while we focus on the indoor re-localization in low light in the paper. As RGB images are degraded in low light, less discriminative infrared and depth image pairs are taken, as the input, by RGB-D cameras, the most similar candidates, as the output, are searched from databases which is built in the bag-of-word framework. Epipolar constraints can be used to relocalize the query infrared and depth image sequence. We evaluate our method in two datasets captured by Kinect2. The results demonstrate very promising re-localization results for indoor navigation system in low light environments. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=indoor%20navigation" title="indoor navigation">indoor navigation</a>, <a href="https://publications.waset.org/abstracts/search?q=low%20light" title=" low light"> low light</a>, <a href="https://publications.waset.org/abstracts/search?q=RGB-D%20camera" title=" RGB-D camera"> RGB-D camera</a>, <a href="https://publications.waset.org/abstracts/search?q=vision%20based" title=" vision based"> vision based</a> </p> <a href="https://publications.waset.org/abstracts/66057/visual-search-based-indoor-localization-in-low-light-via-rgb-d-camera" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/66057.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">460</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1660</span> Room Level Indoor Localization Using Relevant Channel Impulse Response Parameters</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Raida%20Zouari">Raida Zouari</a>, <a href="https://publications.waset.org/abstracts/search?q=Iness%20Ahriz"> Iness Ahriz</a>, <a href="https://publications.waset.org/abstracts/search?q=Rafik%20Zayani"> Rafik Zayani</a>, <a href="https://publications.waset.org/abstracts/search?q=Ali%20Dziri"> Ali Dziri</a>, <a href="https://publications.waset.org/abstracts/search?q=Ridha%20Bouallegue"> Ridha Bouallegue</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper proposes a room level indoor localization algorithm based on the use Multi-Layer Neural Network (MLNN) classifiers and one versus one strategy. Seven parameters of the Channel Impulse Response (CIR) were used and Gram-Shmidt Orthogonalization was performed to study the relevance of the extracted parameters. Simulation results show that when relevant CIR parameters are used as position fingerprint and when optimal MLNN architecture is selected good room level localization score can be achieved. The current study showed also that some of the CIR parameters are not correlated to the location and can decrease the localization performance of the system. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=mobile%20indoor%20localization" title="mobile indoor localization">mobile indoor localization</a>, <a href="https://publications.waset.org/abstracts/search?q=multi-layer%20neural%20network%20%28MLNN%29" title=" multi-layer neural network (MLNN)"> multi-layer neural network (MLNN)</a>, <a href="https://publications.waset.org/abstracts/search?q=channel%20impulse%20response%20%28CIR%29" title=" channel impulse response (CIR)"> channel impulse response (CIR)</a>, <a href="https://publications.waset.org/abstracts/search?q=Gram-Shmidt%20orthogonalization" title=" Gram-Shmidt orthogonalization"> Gram-Shmidt orthogonalization</a> </p> <a href="https://publications.waset.org/abstracts/40068/room-level-indoor-localization-using-relevant-channel-impulse-response-parameters" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/40068.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">357</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1659</span> OCR/ICR Text Recognition Using ABBYY FineReader as an Example Text</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=A.%20R.%20Bagirzade">A. R. Bagirzade</a>, <a href="https://publications.waset.org/abstracts/search?q=A.%20Sh.%20Najafova"> A. Sh. Najafova</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20M.%20Yessirkepova"> S. M. Yessirkepova</a>, <a href="https://publications.waset.org/abstracts/search?q=E.%20S.%20Albert"> E. S. Albert</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This article describes a text recognition method based on Optical Character Recognition (OCR). The features of the OCR method were examined using the ABBYY FineReader program. It describes automatic text recognition in images. OCR is necessary because optical input devices can only transmit raster graphics as a result. Text recognition describes the task of recognizing letters shown as such, to identify and assign them an assigned numerical value in accordance with the usual text encoding (ASCII, Unicode). The peculiarity of this study conducted by the authors using the example of the ABBYY FineReader, was confirmed and shown in practice, the improvement of digital text recognition platforms developed by Electronic Publication. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=ABBYY%20FineReader%20system" title="ABBYY FineReader system">ABBYY FineReader system</a>, <a href="https://publications.waset.org/abstracts/search?q=algorithm%20symbol%20recognition" title=" algorithm symbol recognition"> algorithm symbol recognition</a>, <a href="https://publications.waset.org/abstracts/search?q=OCR%2FICR%20techniques" title=" OCR/ICR techniques"> OCR/ICR techniques</a>, <a href="https://publications.waset.org/abstracts/search?q=recognition%20technologies" title=" recognition technologies"> recognition technologies</a> </p> <a href="https://publications.waset.org/abstracts/130255/ocricr-text-recognition-using-abbyy-finereader-as-an-example-text" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/130255.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">168</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1658</span> Savi Scout versus Wire-Guided Localization in Non-palpable Breast Lesions – Comparison of Breast Tissue Volume and Weight and Excision Safety Margin</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Walid%20Ibrahim">Walid Ibrahim</a>, <a href="https://publications.waset.org/abstracts/search?q=Abdul%20Kasem"> Abdul Kasem</a>, <a href="https://publications.waset.org/abstracts/search?q=Sudeendra%20Doddi"> Sudeendra Doddi</a>, <a href="https://publications.waset.org/abstracts/search?q=Ilaria%20Giono"> Ilaria Giono</a>, <a href="https://publications.waset.org/abstracts/search?q=Tareq%20Sabagh"> Tareq Sabagh</a>, <a href="https://publications.waset.org/abstracts/search?q=Muhammad%20Ammar"> Muhammad Ammar</a>, <a href="https://publications.waset.org/abstracts/search?q=Nermin%20Osman"> Nermin Osman</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Background: wire-guided localization (WL) is the most widely used method for the localization of non-palpable breast lesions. SAVI SCOUT occult lesion localization (SSL) is a new technique in breast-conservative surgery. SSL has the potential benefit of improving radiology workflow as well as accurate localization. Purpose: The purpose of this study is to compare the breast tissue specimen volume and weight and margin excision between WL and SSL. Materials and methods: A single institution retrospective analysis of 377 female patients who underwent wide local breast excision with SAVI SCOUT and or wire-guided technique between 2018 and 2021 in a UK University teaching hospital. Breast department. Breast tissue specimen volume and weight, and margin excision have been evaluated in the three groups of different localization. Results: Three hundred and seventy-seven patients were studied. Of these, 261 had wire localization, 88 had SCOUT and 28 had dual localization techniques. Tumor size ranged from 1 to 75mm (Median 20mm). The pathology specimen weight ranged from 1 to 466gm (Median 46.8) and the volume ranged from 1.305 to 1560cm³ (Median 106.32 cm³). SCOUT localization was associated with a significantly low specimen weight than wire or the dual technique localization (Median 41gm vs 47.3gm and 47gm, p = 0.029). SCOUT was not associated with better specimen volume with a borderline significance in comparison to wire and combined techniques (Median 108cm³ vs 105cm³ and 105cm³, p = 0.047). There was a significant correlation between tumor size and pathology specimen weight in the three groups. SCOUT showed a better >2mm safety margin in comparison to the other 2 techniques (p = 0.031). Conclusion: Preoperative SCOUT localization is associated with better specimen weight and better specimen margin. SCOUT did not show any benefits in terms of specimen volume which may be due to difficulty in getting the accurate specimen volume due to the irregularity of the soft tissue specimen. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=scout" title="scout">scout</a>, <a href="https://publications.waset.org/abstracts/search?q=wire" title=" wire"> wire</a>, <a href="https://publications.waset.org/abstracts/search?q=localization" title=" localization"> localization</a>, <a href="https://publications.waset.org/abstracts/search?q=breast" title=" breast"> breast</a> </p> <a href="https://publications.waset.org/abstracts/166888/savi-scout-versus-wire-guided-localization-in-non-palpable-breast-lesions-comparison-of-breast-tissue-volume-and-weight-and-excision-safety-margin" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/166888.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">110</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1657</span> Localization of Near Field Radio Controlled Unintended Emitting Sources </h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Nurbanu%20Guzey">Nurbanu Guzey</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20Jagannathan"> S. Jagannathan</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Locating radio controlled (RC) devices using their unintended emissions has a great interest considering security concerns. Weak nature of these emissions requires near field localization approach since it is hard to detect these signals in far field region of array. Instead of only angle estimation, near field localization also requires range estimation of the source which makes this method more complicated than far field models. Challenges of locating such devices in a near field region and real time environment are analyzed in this paper. An ESPRIT like near field localization scheme is utilized for both angle and range estimation. 1-D search with symmetric subarrays is provided. Two 7 element uniform linear antenna arrays (ULA) are employed for locating RC source. Experiment results of location estimation for one unintended emitting walkie-talkie for different positions are given. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=localization" title="localization">localization</a>, <a href="https://publications.waset.org/abstracts/search?q=angle%20of%20arrival%20%28AoA%29" title=" angle of arrival (AoA)"> angle of arrival (AoA)</a>, <a href="https://publications.waset.org/abstracts/search?q=range%20estimation" title=" range estimation"> range estimation</a>, <a href="https://publications.waset.org/abstracts/search?q=array%20signal%20processing" title=" array signal processing"> array signal processing</a>, <a href="https://publications.waset.org/abstracts/search?q=ESPRIT" title=" ESPRIT"> ESPRIT</a>, <a href="https://publications.waset.org/abstracts/search?q=Uniform%20Linear%20Array%20%28ULA%29" title=" Uniform Linear Array (ULA)"> Uniform Linear Array (ULA)</a> </p> <a href="https://publications.waset.org/abstracts/22966/localization-of-near-field-radio-controlled-unintended-emitting-sources" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/22966.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">526</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1656</span> Failure Localization of Bipolar Integrated Circuits by Implementing Active Voltage Contrast</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Yiqiang%20Ni">Yiqiang Ni</a>, <a href="https://publications.waset.org/abstracts/search?q=Xuanlong%20Chen"> Xuanlong Chen</a>, <a href="https://publications.waset.org/abstracts/search?q=Enliang%20Li"> Enliang Li</a>, <a href="https://publications.waset.org/abstracts/search?q=Linting%20Zheng"> Linting Zheng</a>, <a href="https://publications.waset.org/abstracts/search?q=Shizheng%20Yang"> Shizheng Yang</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Bipolar ICs are playing an important role in military applications, mainly used in logic gates, such as inverter and NAND gate. The defect of metal break located on the step is one of the main failure mechanisms of bipolar ICs, resulting in open-circuit or functional failure. In this situation, general failure localization methods like optical beam-induced resistance change (OBIRCH) and photon emission microscopy (PEM) might not be fully effective. However, active voltage contrast (AVC) can be used as a voltage probe, which may pinpoint the incorrect potential and thus locate the failure position. Two case studies will be present in this paper on how to implement AVC for failure localization, and the detailed failure mechanism will be discussed. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=bipolar%20IC" title="bipolar IC">bipolar IC</a>, <a href="https://publications.waset.org/abstracts/search?q=failure%20localization" title=" failure localization"> failure localization</a>, <a href="https://publications.waset.org/abstracts/search?q=metal%20break" title=" metal break"> metal break</a>, <a href="https://publications.waset.org/abstracts/search?q=open%20failure" title=" open failure"> open failure</a>, <a href="https://publications.waset.org/abstracts/search?q=voltage%20contrast" title=" voltage contrast"> voltage contrast</a> </p> <a href="https://publications.waset.org/abstracts/132527/failure-localization-of-bipolar-integrated-circuits-by-implementing-active-voltage-contrast" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/132527.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">291</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1655</span> Programmed Speech to Text Summarization Using Graph-Based Algorithm</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hamsini%20Pulugurtha">Hamsini Pulugurtha</a>, <a href="https://publications.waset.org/abstracts/search?q=P.%20V.%20S.%20L.%20Jagadamba"> P. V. S. L. Jagadamba</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Programmed Speech to Text and Text Summarization Using Graph-based Algorithms can be utilized in gatherings to get the short depiction of the gathering for future reference. This gives signature check utilizing Siamese neural organization to confirm the personality of the client and convert the client gave sound record which is in English into English text utilizing the discourse acknowledgment bundle given in python. At times just the outline of the gathering is required, the answer for this text rundown. Thus, the record is then summed up utilizing the regular language preparing approaches, for example, solo extractive text outline calculations <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Siamese%20neural%20network" title="Siamese neural network">Siamese neural network</a>, <a href="https://publications.waset.org/abstracts/search?q=English%20speech" title=" English speech"> English speech</a>, <a href="https://publications.waset.org/abstracts/search?q=English%20text" title=" English text"> English text</a>, <a href="https://publications.waset.org/abstracts/search?q=natural%20language%20processing" title=" natural language processing"> natural language processing</a>, <a href="https://publications.waset.org/abstracts/search?q=unsupervised%20extractive%20text%20summarization" title=" unsupervised extractive text summarization"> unsupervised extractive text summarization</a> </p> <a href="https://publications.waset.org/abstracts/143079/programmed-speech-to-text-summarization-using-graph-based-algorithm" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/143079.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">217</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1654</span> In-door Localization Algorithm and Appropriate Implementation Using Wireless Sensor Networks</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Adeniran%20K.%20Ademuwagun">Adeniran K. Ademuwagun</a>, <a href="https://publications.waset.org/abstracts/search?q=Alastair%20Allen"> Alastair Allen</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The relationship dependence between RSS and distance in an enclosed environment is an important consideration because it is a factor that can influence the reliability of any localization algorithm founded on RSS. Several algorithms effectively reduce the variance of RSS to improve localization or accuracy performance. Our proposed algorithm essentially avoids this pitfall and consequently, its high adaptability in the face of erratic radio signal. Using 3 anchors in close proximity of each other, we are able to establish that RSS can be used as reliable indicator for localization with an acceptable degree of accuracy. Inherent in this concept, is the ability for each prospective anchor to validate (guarantee) the position or the proximity of the other 2 anchors involved in the localization and vice versa. This procedure ensures that the uncertainties of radio signals due to multipath effects in enclosed environments are minimized. A major driver of this idea is the implicit topological relationship among sensors due to raw radio signal strength. The algorithm is an area based algorithm; however, it does not trade accuracy for precision (i.e the size of the returned area). <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=anchor%20nodes" title="anchor nodes">anchor nodes</a>, <a href="https://publications.waset.org/abstracts/search?q=centroid%20algorithm" title=" centroid algorithm"> centroid algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=communication%20graph" title=" communication graph"> communication graph</a>, <a href="https://publications.waset.org/abstracts/search?q=radio%20signal%20strength" title=" radio signal strength"> radio signal strength</a> </p> <a href="https://publications.waset.org/abstracts/34186/in-door-localization-algorithm-and-appropriate-implementation-using-wireless-sensor-networks" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/34186.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">508</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1653</span> Reducing Accidents Using Text Stops</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Benish%20Chaudhry">Benish Chaudhry</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Most of the accidents these days are occurring because of the ‘text-and-drive’ concept. If we look at the structure of cities in UAE, there are great distances, because of which it is impossible to drive without using or merely checking the cellphone. Moreover, if we look at the road structure, it is almost impossible to stop at a point and text. With the introduction of TEXT STOPs, drivers will be able to stop different stops for a maximum of 1 and a half-minute in order to reply or write a message. They can be introduced at a distance of 10 minutes of driving on the average speed of the road, so the drivers can look forward to a stop and can reply to a text when needed. A user survey indicates that drivers are willing to NOT text-and-drive if they have such a facility available. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=transport" title="transport">transport</a>, <a href="https://publications.waset.org/abstracts/search?q=accidents" title=" accidents"> accidents</a>, <a href="https://publications.waset.org/abstracts/search?q=urban%20planning" title=" urban planning"> urban planning</a>, <a href="https://publications.waset.org/abstracts/search?q=road%20planning" title=" road planning"> road planning</a> </p> <a href="https://publications.waset.org/abstracts/44563/reducing-accidents-using-text-stops" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/44563.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">394</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1652</span> Structure Analysis of Text-Image Connection in Jalayrid Period Illustrated Manuscripts</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Mahsa%20Khani%20Oushani">Mahsa Khani Oushani</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Text and image are two important elements in the field of Iranian art, the text component and the image component have always been manifested together. The image narrates the text and the text is the factor in the formation of the image and they are closely related to each other. The connection between text and image is an interactive and two-way connection in the tradition of Iranian manuscript arrangement. The interaction between the narrative description and the image scene is the result of a direct and close connection between the text and the image, which in addition to the decorative aspect, also has a descriptive aspect. In this article the connection between the text element and the image element and its adaptation to the theory of Roland Barthes, the structuralism theorist, in this regard will be discussed. This study tends to investigate the question of how the connection between text and image in illustrated manuscripts of the Jalayrid period is defined according to Barthes’ theory. And what kind of proportion has the artist created in the composition between text and image. Based on the results of reviewing the data of this study, it can be inferred that in the Jalayrid period, the image has a reference connection and although it is of major importance on the page, it also maintains a close connection with the text and is placed in a special proportion. It is not necessarily balanced and symmetrical and sometimes uses imbalance for composition. This research has been done by descriptive-analytical method, which has been done by library collection method. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=structure" title="structure">structure</a>, <a href="https://publications.waset.org/abstracts/search?q=text" title=" text"> text</a>, <a href="https://publications.waset.org/abstracts/search?q=image" title=" image"> image</a>, <a href="https://publications.waset.org/abstracts/search?q=Jalayrid" title=" Jalayrid"> Jalayrid</a>, <a href="https://publications.waset.org/abstracts/search?q=painter" title=" painter"> painter</a> </p> <a href="https://publications.waset.org/abstracts/138869/structure-analysis-of-text-image-connection-in-jalayrid-period-illustrated-manuscripts" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/138869.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">233</span> </span> </div> </div> <ul class="pagination"> <li class="page-item disabled"><span class="page-link">‹</span></li> <li class="page-item active"><span class="page-link">1</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=2">2</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=3">3</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=4">4</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=5">5</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=6">6</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=7">7</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=8">8</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=9">9</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=10">10</a></li> <li class="page-item disabled"><span class="page-link">...</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=56">56</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=57">57</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=text%20localization&page=2" rel="next">›</a></li> </ul> </div> </main> <footer> <div id="infolinks" class="pt-3 pb-2"> <div class="container"> <div style="background-color:#f5f5f5;" class="p-3"> <div class="row"> <div class="col-md-2"> <ul class="list-unstyled"> About <li><a href="https://waset.org/page/support">About Us</a></li> <li><a href="https://waset.org/page/support#legal-information">Legal</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/WASET-16th-foundational-anniversary.pdf">WASET celebrates its 16th foundational anniversary</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Account <li><a href="https://waset.org/profile">My Account</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Explore <li><a href="https://waset.org/disciplines">Disciplines</a></li> <li><a href="https://waset.org/conferences">Conferences</a></li> <li><a href="https://waset.org/conference-programs">Conference Program</a></li> <li><a href="https://waset.org/committees">Committees</a></li> <li><a href="https://publications.waset.org">Publications</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Research <li><a href="https://publications.waset.org/abstracts">Abstracts</a></li> <li><a href="https://publications.waset.org">Periodicals</a></li> <li><a href="https://publications.waset.org/archive">Archive</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Open Science <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Philosophy.pdf">Open Science Philosophy</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Award.pdf">Open Science Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Society-Open-Science-and-Open-Innovation.pdf">Open Innovation</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Postdoctoral-Fellowship-Award.pdf">Postdoctoral Fellowship Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Scholarly-Research-Review.pdf">Scholarly Research Review</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Support <li><a href="https://waset.org/page/support">Support</a></li> <li><a href="https://waset.org/profile/messages/create">Contact Us</a></li> <li><a href="https://waset.org/profile/messages/create">Report Abuse</a></li> </ul> </div> </div> </div> </div> </div> <div class="container text-center"> <hr style="margin-top:0;margin-bottom:.3rem;"> <a href="https://creativecommons.org/licenses/by/4.0/" target="_blank" class="text-muted small">Creative Commons Attribution 4.0 International License</a> <div id="copy" class="mt-2">© 2024 World Academy of Science, Engineering and Technology</div> </div> </footer> <a href="javascript:" id="return-to-top"><i class="fas fa-arrow-up"></i></a> <div class="modal" id="modal-template"> <div class="modal-dialog"> <div class="modal-content"> <div class="row m-0 mt-1"> <div class="col-md-12"> <button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">×</span></button> </div> </div> <div class="modal-body"></div> </div> </div> </div> <script src="https://cdn.waset.org/static/plugins/jquery-3.3.1.min.js"></script> <script src="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/js/bootstrap.bundle.min.js"></script> <script src="https://cdn.waset.org/static/js/site.js?v=150220211556"></script> <script> jQuery(document).ready(function() { /*jQuery.get("https://publications.waset.org/xhr/user-menu", function (response) { jQuery('#mainNavMenu').append(response); });*/ jQuery.get({ url: "https://publications.waset.org/xhr/user-menu", cache: false }).then(function(response){ jQuery('#mainNavMenu').append(response); }); }); </script> </body> </html>