CINXE.COM

Search results for: Blur

<!DOCTYPE html> <html lang="en" dir="ltr"> <head> <!-- Google tag (gtag.js) --> <script async src="https://www.googletagmanager.com/gtag/js?id=G-P63WKM1TM1"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('js', new Date()); gtag('config', 'G-P63WKM1TM1'); </script> <!-- Yandex.Metrika counter --> <script type="text/javascript" > (function(m,e,t,r,i,k,a){m[i]=m[i]||function(){(m[i].a=m[i].a||[]).push(arguments)}; m[i].l=1*new Date(); for (var j = 0; j < document.scripts.length; j++) {if (document.scripts[j].src === r) { return; }} k=e.createElement(t),a=e.getElementsByTagName(t)[0],k.async=1,k.src=r,a.parentNode.insertBefore(k,a)}) (window, document, "script", "https://mc.yandex.ru/metrika/tag.js", "ym"); ym(55165297, "init", { clickmap:false, trackLinks:true, accurateTrackBounce:true, webvisor:false }); </script> <noscript><div><img src="https://mc.yandex.ru/watch/55165297" style="position:absolute; left:-9999px;" alt="" /></div></noscript> <!-- /Yandex.Metrika counter --> <!-- Matomo --> <!-- End Matomo Code --> <title>Search results for: Blur</title> <meta name="description" content="Search results for: Blur"> <meta name="keywords" content="Blur"> <meta name="viewport" content="width=device-width, initial-scale=1, minimum-scale=1, maximum-scale=1, user-scalable=no"> <meta charset="utf-8"> <link href="https://cdn.waset.org/favicon.ico" type="image/x-icon" rel="shortcut icon"> <link href="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/css/bootstrap.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/plugins/fontawesome/css/all.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/css/site.css?v=150220211555" rel="stylesheet"> </head> <body> <header> <div class="container"> <nav class="navbar navbar-expand-lg navbar-light"> <a class="navbar-brand" href="https://waset.org"> <img src="https://cdn.waset.org/static/images/wasetc.png" alt="Open Science Research Excellence" title="Open Science Research Excellence" /> </a> <button class="d-block d-lg-none navbar-toggler ml-auto" type="button" data-toggle="collapse" data-target="#navbarMenu" aria-controls="navbarMenu" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <div class="w-100"> <div class="d-none d-lg-flex flex-row-reverse"> <form method="get" action="https://waset.org/search" class="form-inline my-2 my-lg-0"> <input class="form-control mr-sm-2" type="search" placeholder="Search Conferences" value="Blur" name="q" aria-label="Search"> <button class="btn btn-light my-2 my-sm-0" type="submit"><i class="fas fa-search"></i></button> </form> </div> <div class="collapse navbar-collapse mt-1" id="navbarMenu"> <ul class="navbar-nav ml-auto align-items-center" id="mainNavMenu"> <li class="nav-item"> <a class="nav-link" href="https://waset.org/conferences" title="Conferences in 2024/2025/2026">Conferences</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/disciplines" title="Disciplines">Disciplines</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/committees" rel="nofollow">Committees</a> </li> <li class="nav-item dropdown"> <a class="nav-link dropdown-toggle" href="#" id="navbarDropdownPublications" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Publications </a> <div class="dropdown-menu" aria-labelledby="navbarDropdownPublications"> <a class="dropdown-item" href="https://publications.waset.org/abstracts">Abstracts</a> <a class="dropdown-item" href="https://publications.waset.org">Periodicals</a> <a class="dropdown-item" href="https://publications.waset.org/archive">Archive</a> </div> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/page/support" title="Support">Support</a> </li> </ul> </div> </div> </nav> </div> </header> <main> <div class="container mt-4"> <div class="row"> <div class="col-md-9 mx-auto"> <form method="get" action="https://publications.waset.org/search"> <div id="custom-search-input"> <div class="input-group"> <i class="fas fa-search"></i> <input type="text" class="search-query" name="q" placeholder="Author, Title, Abstract, Keywords" value="Blur"> <input type="submit" class="btn_search" value="Search"> </div> </div> </form> </div> </div> <div class="row mt-3"> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Commenced</strong> in January 2007</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Frequency:</strong> Monthly</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Edition:</strong> International</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Paper Count:</strong> 25</div> </div> </div> </div> <h1 class="mt-3 mb-3 text-center" style="font-size:1.6rem;">Search results for: Blur</h1> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">25</span> Blur and Ringing Artifact Measurement in Image Compression using Wavelet Transform </h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Madhuri%20Khambete">Madhuri Khambete</a>, <a href="https://publications.waset.org/search?q=Madhuri%20Joshi"> Madhuri Joshi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>Quality evaluation of an image is an important task in image processing applications. In case of image compression, quality of decompressed image is also the criterion for evaluation of given coding scheme. In the process of compression -decompression various artifacts such as blocking artifacts, blur artifact, ringing or edge artifact are observed. However quantification of these artifacts is a difficult task. We propose here novel method to quantify blur and ringing artifact in an image.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Blur" title="Blur">Blur</a>, <a href="https://publications.waset.org/search?q=Compression" title=" Compression"> Compression</a>, <a href="https://publications.waset.org/search?q=Objective%20Quality%20assessment" title=" Objective Quality assessment"> Objective Quality assessment</a>, <a href="https://publications.waset.org/search?q=Ringing%20artifact." title="Ringing artifact.">Ringing artifact.</a> </p> <a href="https://publications.waset.org/2394/blur-and-ringing-artifact-measurement-in-image-compression-using-wavelet-transform" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/2394/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/2394/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/2394/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/2394/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/2394/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/2394/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/2394/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/2394/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/2394/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/2394/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/2394.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">4853</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">24</span> Object Tracking in Motion Blurred Images with Adaptive Mean Shift and Wavelet Feature</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Iman%20Iraei">Iman Iraei</a>, <a href="https://publications.waset.org/search?q=Mina%20Sharifi"> Mina Sharifi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> A method for object tracking in motion blurred images is proposed in this article. This paper shows that object tracking could be improved with this approach. We use mean shift algorithm to track different objects as a main tracker. But, the problem is that mean shift could not track the selected object accurately in blurred scenes. So, for better tracking result, and increasing the accuracy of tracking, wavelet transform is used. We use a feature named as blur extent, which could help us to get better results in tracking. For calculating of this feature, we should use Harr wavelet. We can look at this matter from two different angles which lead to determine whether an image is blurred or not and to what extent an image is blur. In fact, this feature left an impact on the covariance matrix of mean shift algorithm and cause to better performance of tracking. This method has been concentrated mostly on motion blur parameter. transform. The results reveal the ability of our method in order to reach more accurately tracking. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Mean%20shift" title="Mean shift">Mean shift</a>, <a href="https://publications.waset.org/search?q=object%20tracking" title=" object tracking"> object tracking</a>, <a href="https://publications.waset.org/search?q=blur%20extent" title=" blur extent"> blur extent</a>, <a href="https://publications.waset.org/search?q=wavelet%20transform" title=" wavelet transform"> wavelet transform</a>, <a href="https://publications.waset.org/search?q=motion%20blur." title=" motion blur."> motion blur.</a> </p> <a href="https://publications.waset.org/10008684/object-tracking-in-motion-blurred-images-with-adaptive-mean-shift-and-wavelet-feature" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/10008684/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/10008684/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/10008684/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/10008684/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/10008684/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/10008684/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/10008684/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/10008684/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/10008684/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/10008684/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/10008684.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">811</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">23</span> No-Reference Image Quality Assessment using Blur and Noise</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Min%20Goo%20Choi">Min Goo Choi</a>, <a href="https://publications.waset.org/search?q=Jung%20Hoon%20Jung"> Jung Hoon Jung</a>, <a href="https://publications.waset.org/search?q=Jae%20Wook%20Jeon"> Jae Wook Jeon</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Assessment for image quality traditionally needs its original image as a reference. The conventional method for assessment like Mean Square Error (MSE) or Peak Signal to Noise Ratio (PSNR) is invalid when there is no reference. In this paper, we present a new No-Reference (NR) assessment of image quality using blur and noise. The recent camera applications provide high quality images by help of digital Image Signal Processor (ISP). Since the images taken by the high performance of digital camera have few blocking and ringing artifacts, we only focus on the blur and noise for predicting the objective image quality. The experimental results show that the proposed assessment method gives high correlation with subjective Difference Mean Opinion Score (DMOS). Furthermore, the proposed method provides very low computational load in spatial domain and similar extraction of characteristics to human perceptional assessment. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=No%20Reference" title="No Reference">No Reference</a>, <a href="https://publications.waset.org/search?q=Image%20Quality%20Assessment" title=" Image Quality Assessment"> Image Quality Assessment</a>, <a href="https://publications.waset.org/search?q=blur" title=" blur"> blur</a>, <a href="https://publications.waset.org/search?q=noise." title="noise.">noise.</a> </p> <a href="https://publications.waset.org/12066/no-reference-image-quality-assessment-using-blur-and-noise" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/12066/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/12066/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/12066/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/12066/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/12066/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/12066/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/12066/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/12066/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/12066/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/12066/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/12066.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">3879</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">22</span> Multiscale Blind Image Restoration with a New Method</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Alireza%20Mallahzadeh">Alireza Mallahzadeh</a>, <a href="https://publications.waset.org/search?q=Hamid%20Dehghani"> Hamid Dehghani</a>, <a href="https://publications.waset.org/search?q=Iman%20Elyasi"> Iman Elyasi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> A new method, based on the normal shrink and modified version of Katssagelous and Lay, is proposed for multiscale blind image restoration. The method deals with the noise and blur in the images. It is shown that the normal shrink gives the highest S/N (signal to noise ratio) for image denoising process. The multiscale blind image restoration is divided in two sections. The first part of this paper proposes normal shrink for image denoising and the second part of paper proposes modified version of katssagelous and Lay for blur estimation and the combination of both methods to reach a multiscale blind image restoration. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Multiscale%20blind%20image%20restoration" title="Multiscale blind image restoration">Multiscale blind image restoration</a>, <a href="https://publications.waset.org/search?q=image%20denoising" title=" image denoising"> image denoising</a>, <a href="https://publications.waset.org/search?q=blur%20estimation." title=" blur estimation."> blur estimation.</a> </p> <a href="https://publications.waset.org/11297/multiscale-blind-image-restoration-with-a-new-method" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/11297/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/11297/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/11297/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/11297/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/11297/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/11297/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/11297/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/11297/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/11297/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/11297/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/11297.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1722</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">21</span> Object Speed Estimation by using Fuzzy Set</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Hossein%20Pazhoumand-Dar">Hossein Pazhoumand-Dar</a>, <a href="https://publications.waset.org/search?q=Amir%20Mohsen%20Toliyat%20Abolhassani"> Amir Mohsen Toliyat Abolhassani</a>, <a href="https://publications.waset.org/search?q=Ehsan%20Saeedi"> Ehsan Saeedi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>Speed estimation is one of the important and practical tasks in machine vision, Robotic and Mechatronic. the availability of high quality and inexpensive video cameras, and the increasing need for automated video analysis has generated a great deal of interest in machine vision algorithms. Numerous approaches for speed estimation have been proposed. So classification and survey of the proposed methods can be very useful. The goal of this paper is first to review and verify these methods. Then we will propose a novel algorithm to estimate the speed of moving object by using fuzzy concept. There is a direct relation between motion blur parameters and object speed. In our new approach we will use Radon transform to find direction of blurred image, and Fuzzy sets to estimate motion blur length. The most benefit of this algorithm is its robustness and precision in noisy images. Our method was tested on many images with different range of SNR and is satisfiable.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Blur%20Analysis" title="Blur Analysis">Blur Analysis</a>, <a href="https://publications.waset.org/search?q=Fuzzy%20sets" title=" Fuzzy sets"> Fuzzy sets</a>, <a href="https://publications.waset.org/search?q=Speed%20estimation." title=" Speed estimation."> Speed estimation.</a> </p> <a href="https://publications.waset.org/5044/object-speed-estimation-by-using-fuzzy-set" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/5044/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/5044/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/5044/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/5044/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/5044/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/5044/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/5044/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/5044/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/5044/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/5044/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/5044.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1879</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">20</span> Contrast Enhancement in Digital Images Using an Adaptive Unsharp Masking Method</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Z.%20Mortezaie">Z. Mortezaie</a>, <a href="https://publications.waset.org/search?q=H.%20Hassanpour"> H. Hassanpour</a>, <a href="https://publications.waset.org/search?q=S.%20Asadi%20Amiri"> S. Asadi Amiri</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Captured images may suffer from Gaussian blur due to poor lens focus or camera motion. Unsharp masking is a simple and effective technique to boost the image contrast and to improve digital images suffering from Gaussian blur. The technique is based on sharpening object edges by appending the scaled high-frequency components of the image to the original. The quality of the enhanced image is highly dependent on the characteristics of both the high-frequency components and the scaling/gain factor. Since the quality of an image may not be the same throughout, we propose an adaptive unsharp masking method in this paper. In this method, the gain factor is computed, considering the gradient variations, for individual pixels of the image. Subjective and objective image quality assessments are used to compare the performance of the proposed method both with the classic and the recently developed unsharp masking methods. The experimental results show that the proposed method has a better performance in comparison to the other existing methods. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Unsharp%20masking" title="Unsharp masking">Unsharp masking</a>, <a href="https://publications.waset.org/search?q=blur%20image" title=" blur image"> blur image</a>, <a href="https://publications.waset.org/search?q=sub-region%20gradient" title=" sub-region gradient"> sub-region gradient</a>, <a href="https://publications.waset.org/search?q=image%20enhancement." title=" image enhancement."> image enhancement.</a> </p> <a href="https://publications.waset.org/10007889/contrast-enhancement-in-digital-images-using-an-adaptive-unsharp-masking-method" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/10007889/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/10007889/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/10007889/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/10007889/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/10007889/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/10007889/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/10007889/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/10007889/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/10007889/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/10007889/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/10007889.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1411</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">19</span> Denoising based on Wavelets and Deblurring via Self-Organizing Map for Synthetic Aperture Radar Images</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Mario%20Mastriani">Mario Mastriani</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This work deals with unsupervised image deblurring. We present a new deblurring procedure on images provided by lowresolution synthetic aperture radar (SAR) or simply by multimedia in presence of multiplicative (speckle) or additive noise, respectively. The method we propose is defined as a two-step process. First, we use an original technique for noise reduction in wavelet domain. Then, the learning of a Kohonen self-organizing map (SOM) is performed directly on the denoised image to take out it the blur. This technique has been successfully applied to real SAR images, and the simulation results are presented to demonstrate the effectiveness of the proposed algorithms. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Blur" title="Blur">Blur</a>, <a href="https://publications.waset.org/search?q=Kohonen%20self-organizing%20map" title=" Kohonen self-organizing map"> Kohonen self-organizing map</a>, <a href="https://publications.waset.org/search?q=noise" title=" noise"> noise</a>, <a href="https://publications.waset.org/search?q=speckle" title=" speckle"> speckle</a>, <a href="https://publications.waset.org/search?q=synthetic%20aperture%20radar." title=" synthetic aperture radar."> synthetic aperture radar.</a> </p> <a href="https://publications.waset.org/4726/denoising-based-on-wavelets-and-deblurring-via-self-organizing-map-for-synthetic-aperture-radar-images" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/4726/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/4726/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/4726/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/4726/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/4726/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/4726/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/4726/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/4726/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/4726/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/4726/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/4726.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1735</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">18</span> Source Optimisation of Laser-Plasma Bremmstrahlung for Applications in Engineering Imaging</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=R.J.%20Clarke">R.J. Clarke</a>, <a href="https://publications.waset.org/search?q=D.%20Neely"> D. Neely</a>, <a href="https://publications.waset.org/search?q=S.%20Blake"> S. Blake</a>, <a href="https://publications.waset.org/search?q=D.C.%20Carroll"> D.C. Carroll</a>, <a href="https://publications.waset.org/search?q=J.S.%20Green"> J.S. Green</a>, <a href="https://publications.waset.org/search?q=R.%20Heathcote"> R. Heathcote</a>, <a href="https://publications.waset.org/search?q=M.%20Notley"> M. Notley</a> </p> <p class="card-text"><strong>Abstract:</strong></p> High Power Lasers produce an intense burst of Bremmstrahlung radiation which has potential applications in broadband x-ray radiography. Since the radiation produced is through the interaction of accelerated electrons with the remaining laser target, these bursts are extremely short – in the region of a few ps. As a result, the laser-produced x-rays are capable of imaging complex dynamic objects with zero motion blur. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Bremmstrahlung" title="Bremmstrahlung">Bremmstrahlung</a>, <a href="https://publications.waset.org/search?q=Imaging" title=" Imaging"> Imaging</a>, <a href="https://publications.waset.org/search?q=Laser" title=" Laser"> Laser</a>, <a href="https://publications.waset.org/search?q=Plasma" title=" Plasma"> Plasma</a>, <a href="https://publications.waset.org/search?q=Radiography" title=" Radiography"> Radiography</a>, <a href="https://publications.waset.org/search?q=x-ray." title=" x-ray."> x-ray.</a> </p> <a href="https://publications.waset.org/1263/source-optimisation-of-laser-plasma-bremmstrahlung-for-applications-in-engineering-imaging" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/1263/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/1263/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/1263/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/1263/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/1263/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/1263/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/1263/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/1263/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/1263/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/1263/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/1263.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1705</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">17</span> e-Commerce versus m-Commerce: Where is the Dividing Line?</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Priscilla%20Omonedo">Priscilla Omonedo</a>, <a href="https://publications.waset.org/search?q=Paul%20Bocij"> Paul Bocij</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>Since the emergence of e-Commerce, the world of business has witnessed a radical shift in the way business activities are conducted. However, the emergence of m-Commerce has further pushed the boundaries of virtual commerce revolution. As a result, there seems to be a growing blur in the distinction between e- Commerce and m-Commerce. In addition, existing definitions for both forms of commerce highlight characteristics (e.g. type of device and activity conducted) that may be applicable to both concepts. The aim of this paper is to identify the characteristics that help define and delineate between e- and m- Commerce. The paper concludes that characteristics of mobility, ubiquity and immediacy provide a clearer and simpler template to distinguish between e-Commerce and m- Commerce.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=e-Commerce" title="e-Commerce">e-Commerce</a>, <a href="https://publications.waset.org/search?q=m-Commerce" title=" m-Commerce"> m-Commerce</a>, <a href="https://publications.waset.org/search?q=mobility" title=" mobility"> mobility</a>, <a href="https://publications.waset.org/search?q=ubiquity." title=" ubiquity."> ubiquity.</a> </p> <a href="https://publications.waset.org/9999848/e-commerce-versus-m-commerce-where-is-the-dividing-line" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/9999848/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/9999848/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/9999848/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/9999848/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/9999848/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/9999848/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/9999848/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/9999848/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/9999848/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/9999848/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/9999848.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">5360</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">16</span> Evaluation of Video Quality Metrics and Performance Comparison on Contents Taken from Most Commonly Used Devices</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Pratik%20Dhabal%20Deo">Pratik Dhabal Deo</a>, <a href="https://publications.waset.org/search?q=Manoj%20P."> Manoj P.</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>With the increasing number of social media users, the amount of video content available has also significantly increased. Currently, the number of smartphone users is at its peak, and many are increasingly using their smartphones as their main photography and recording devices. There have been a lot of developments in the field of video quality assessment in since the past years and more research on various other aspects of video and image are being done. Datasets that contain a huge number of videos from different high-end devices make it difficult to analyze the performance of the metrics on the content from most used devices even if they contain contents taken in poor lighting conditions using lower-end devices. These devices face a lot of distortions due to various factors since the spectrum of contents recorded on these devices is huge. In this paper, we have presented an analysis of the objective Video Quality Analysis (VQA) metrics on contents taken only from most used devices and their performance on them, focusing on full-reference metrics. To carry out this research, we created a custom dataset containing a total of 90 videos that have been taken from three most commonly used devices, and Android smartphone, an iOS smartphone and a Digital Single-Lens Reflex (DSLR) camera. On the videos taken on each of these devices, the six most common types of distortions that users face have been applied in addition to already existing H.264 compression based on four reference videos. These six applied distortions have three levels of degradation each. A total of the five most popular VQA metrics have been evaluated on this dataset and the highest values and the lowest values of each of the metrics on the distortions have been recorded. Finally, it is found that blur is the artifact on which most of the metrics did not perform well. Thus, in order to understand the results better the amount of blur in the data set has been calculated and an additional evaluation of the metrics was done using High Efficiency Video Coding (HEVC) codec, which is the next version of H.264 compression, on the camera that proved to be the sharpest among the devices. The results have shown that as the resolution increases, the performance of the metrics tends to become more accurate and the best performing metric among them is VQM with very few inconsistencies and inaccurate results when the compression applied is H.264, but when the compression is applied is HEVC, Structural Similarity (SSIM) metric and Video Multimethod Assessment Fusion (VMAF) have performed significantly better.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Distortion" title="Distortion">Distortion</a>, <a href="https://publications.waset.org/search?q=metrics" title=" metrics"> metrics</a>, <a href="https://publications.waset.org/search?q=recording" title=" recording"> recording</a>, <a href="https://publications.waset.org/search?q=frame%20rate" title=" frame rate"> frame rate</a>, <a href="https://publications.waset.org/search?q=video%20quality%20assessment." title=" video quality assessment."> video quality assessment.</a> </p> <a href="https://publications.waset.org/10012920/evaluation-of-video-quality-metrics-and-performance-comparison-on-contents-taken-from-most-commonly-used-devices" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/10012920/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/10012920/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/10012920/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/10012920/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/10012920/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/10012920/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/10012920/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/10012920/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/10012920/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/10012920/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/10012920.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">366</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">15</span> An Adaptive Model for Blind Image Restoration using Bayesian Approach</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=S.K.%20Satpathy">S.K. Satpathy</a>, <a href="https://publications.waset.org/search?q=S.K.%20Nayak"> S.K. Nayak</a>, <a href="https://publications.waset.org/search?q=K.%20K.%20Nagwanshi"> K. K. Nagwanshi</a>, <a href="https://publications.waset.org/search?q=S.%20Panda"> S. Panda</a>, <a href="https://publications.waset.org/search?q=C.%20Ardil"> C. Ardil</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Image restoration involves elimination of noise. Filtering techniques were adopted so far to restore images since last five decades. In this paper, we consider the problem of image restoration degraded by a blur function and corrupted by random noise. A method for reducing additive noise in images by explicit analysis of local image statistics is introduced and compared to other noise reduction methods. The proposed method, which makes use of an a priori noise model, has been evaluated on various types of images. Bayesian based algorithms and technique of image processing have been described and substantiated with experimentation using MATLAB. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Image%20Restoration" title="Image Restoration">Image Restoration</a>, <a href="https://publications.waset.org/search?q=Probability%20DensityFunction%20%28PDF%29" title=" Probability DensityFunction (PDF)"> Probability DensityFunction (PDF)</a>, <a href="https://publications.waset.org/search?q=Neural%20Networks" title=" Neural Networks"> Neural Networks</a>, <a href="https://publications.waset.org/search?q=Bayesian%20Classifier." title=" Bayesian Classifier."> Bayesian Classifier.</a> </p> <a href="https://publications.waset.org/14145/an-adaptive-model-for-blind-image-restoration-using-bayesian-approach" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/14145/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/14145/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/14145/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/14145/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/14145/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/14145/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/14145/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/14145/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/14145/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/14145/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/14145.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">2247</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">14</span> Exact Image Super-Resolution for Pure Translational Motion and Shift-Invariant Blur</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Fatih%20Kara">Fatih Kara</a>, <a href="https://publications.waset.org/search?q=Cabir%20Vural"> Cabir Vural</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this work, a special case of the image superresolution problem where the only type of motion is global translational motion and the blurs are shift-invariant is investigated. The necessary conditions for exact reconstruction of the original image by using finite impulse-response reconstruction filters are developed. Given that the conditions are satisfied, a method for exact super-resolution is presented and some simulation results are shown. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Image%20processing" title="Image processing">Image processing</a>, <a href="https://publications.waset.org/search?q=image%20super-resolution" title=" image super-resolution"> image super-resolution</a>, <a href="https://publications.waset.org/search?q=finite%0Aimpulse-response%20filters" title=" finite impulse-response filters"> finite impulse-response filters</a>, <a href="https://publications.waset.org/search?q=existence-uniqueness%20conditions." title=" existence-uniqueness conditions."> existence-uniqueness conditions.</a> </p> <a href="https://publications.waset.org/14908/exact-image-super-resolution-for-pure-translational-motion-and-shift-invariant-blur" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/14908/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/14908/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/14908/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/14908/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/14908/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/14908/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/14908/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/14908/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/14908/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/14908/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/14908.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1303</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">13</span> Supercompression for Full-HD and 4k-3D (8k)Digital TV Systems</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Mario%20Mastriani">Mario Mastriani</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this work, we developed the concept of supercompression, i.e., compression above the compression standard used. In this context, both compression rates are multiplied. In fact, supercompression is based on super-resolution. That is to say, supercompression is a data compression technique that superpose spatial image compression on top of bit-per-pixel compression to achieve very high compression ratios. If the compression ratio is very high, then we use a convolutive mask inside decoder that restores the edges, eliminating the blur. Finally, both, the encoder and the complete decoder are implemented on General-Purpose computation on Graphics Processing Units (GPGPU) cards. Specifically, the mentio-ned mask is coded inside texture memory of a GPGPU. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=General-Purpose%20computation%20on%20Graphics%20Processing%20Units" title="General-Purpose computation on Graphics Processing Units">General-Purpose computation on Graphics Processing Units</a>, <a href="https://publications.waset.org/search?q=Image%20Compression" title=" Image Compression"> Image Compression</a>, <a href="https://publications.waset.org/search?q=Interpolation" title=" Interpolation"> Interpolation</a>, <a href="https://publications.waset.org/search?q=Super-resolution." title=" Super-resolution."> Super-resolution.</a> </p> <a href="https://publications.waset.org/7664/supercompression-for-full-hd-and-4k-3d-8kdigital-tv-systems" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/7664/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/7664/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/7664/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/7664/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/7664/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/7664/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/7664/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/7664/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/7664/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/7664/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/7664.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1979</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">12</span> DeClEx-Processing Pipeline for Tumor Classification</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Gaurav%20Shinde">Gaurav Shinde</a>, <a href="https://publications.waset.org/search?q=Sai%20Charan%20Gongiguntla"> Sai Charan Gongiguntla</a>, <a href="https://publications.waset.org/search?q=Prajwal%20Shirur"> Prajwal Shirur</a>, <a href="https://publications.waset.org/search?q=Ahmed%20Hambaba"> Ahmed Hambaba</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>Health issues are significantly increasing, putting a substantial strain on healthcare services. This has accelerated the integration of machine learning in healthcare, particularly following the COVID-19 pandemic. The utilization of machine learning in healthcare has grown significantly. We introduce DeClEx, a pipeline which ensures that data mirrors real-world settings by incorporating gaussian noise and blur and employing autoencoders to learn intermediate feature representations. Subsequently, our convolutional neural network, paired with spatial attention, provides comparable accuracy to state-of-the-art pre-trained models while achieving a threefold improvement in training speed. Furthermore, we provide interpretable results using explainable AI techniques. We integrate denoising and deblurring, classification and explainability in a single pipeline called DeClEx. </p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Machine%20learning" title="Machine learning">Machine learning</a>, <a href="https://publications.waset.org/search?q=healthcare" title=" healthcare"> healthcare</a>, <a href="https://publications.waset.org/search?q=classification" title=" classification"> classification</a>, <a href="https://publications.waset.org/search?q=explainability." title=" explainability."> explainability.</a> </p> <a href="https://publications.waset.org/10013804/declex-processing-pipeline-for-tumor-classification" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/10013804/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/10013804/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/10013804/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/10013804/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/10013804/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/10013804/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/10013804/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/10013804/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/10013804/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/10013804/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/10013804.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">66</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">11</span> Video Quality Control Using a ROI and Two- Component Weighted Metrics</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Petra%20Heribanov%C3%A1">Petra Heribanová</a>, <a href="https://publications.waset.org/search?q=Jaroslav%20Polec"> Jaroslav Polec</a>, <a href="https://publications.waset.org/search?q=Michal%20Martinovi%C4%8D"> Michal Martinovič</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>In this paper we propose a new content-weighted method for full reference (FR) video quality control using a region of interest (ROI) and wherein two-component weighted metrics for Deaf People Video Communication. In our approach, an image is partitioned into region of interest and into region &quot;dry-as-dust&quot;, then region of interest is partitioned into two parts: edges and background (smooth regions), while the another methods (metrics) combined and weighted three or more parts as edges, edges errors, texture, smooth regions, blur, block distance etc. as we proposed. Using another idea that different image regions from deaf people video communication have different perceptual significance relative to quality. Intensity edges certainly contain considerable image information and are perceptually significant.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Video%20quality%20assessment" title="Video quality assessment">Video quality assessment</a>, <a href="https://publications.waset.org/search?q=weighted%20MSE." title=" weighted MSE."> weighted MSE.</a> </p> <a href="https://publications.waset.org/16328/video-quality-control-using-a-roi-and-two-component-weighted-metrics" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/16328/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/16328/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/16328/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/16328/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/16328/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/16328/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/16328/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/16328/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/16328/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/16328/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/16328.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1981</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">10</span> Single Frame Supercompression of Still Images,Video, High Definition TV and Digital Cinema</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Mario%20Mastriani">Mario Mastriani</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Super-resolution is nowadays used for a high-resolution image produced from several low-resolution noisy frames. In this work, we consider the problem of high-quality interpolation of a single noise-free image. Such images may come from different sources, i.e., they may be frames of videos, individual pictures, etc. On the other hand, in the encoder we apply a downsampling via bidimen-sional interpolation of each frame, and in the decoder we apply a upsampling by which we restore the original size of the image. If the compression ratio is very high, then we use a convolutive mask that restores the edges, eliminating the blur. Finally, both, the encoder and the complete decoder are implemented on General-Purpose computation on Graphics Processing Units (GPGPU) cards. In fact, the mentioned mask is coded inside texture memory of a GPGPU. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=General-Purpose%20computation%20on%20Graphics%20ProcessingUnits" title="General-Purpose computation on Graphics ProcessingUnits">General-Purpose computation on Graphics ProcessingUnits</a>, <a href="https://publications.waset.org/search?q=Image%20Compression" title=" Image Compression"> Image Compression</a>, <a href="https://publications.waset.org/search?q=Interpolation" title=" Interpolation"> Interpolation</a>, <a href="https://publications.waset.org/search?q=Super-resolution." title=" Super-resolution."> Super-resolution.</a> </p> <a href="https://publications.waset.org/15315/single-frame-supercompression-of-still-imagesvideo-high-definition-tv-and-digital-cinema" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/15315/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/15315/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/15315/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/15315/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/15315/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/15315/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/15315/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/15315/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/15315/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/15315/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/15315.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1999</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">9</span> IT/IS Organisation Design in the Digital Age – A Literature Review</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Dominik%20Krimpmann">Dominik Krimpmann</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>Information technology and information systems are currently at a tipping point. The digital age fundamentally transforms a large number of industries in the ways they work. Lines between business and technology blur. Researchers have acknowledged that this is the time in which the IT/IS organisation needs to re-strategize itself. In this paper, the author provides a structured review of the IS and organisation design literature addressing the question of how the digital age changes the design categories of an IT/IS organisation design. The findings show that most papers just analyse single aspects of either IT/IS relevant information or generic organisation design elements but miss a holistic &lsquo;big-picture&rsquo; onto an IT/IS organisation design. This paper creates a holistic IT/IS organisation design framework bringing together the IS research strand, the digital strand and the generic organisation design strand. The research identified four IT/IS organisation design categories (strategy, structure, processes and people) and discusses the importance of two additional categories (sourcing and governance). The authors findings point to a first anchor point from which further research needs to be conducted to develop a holistic IT/IS organisation design framework.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=IT%2FIS%20strategy" title="IT/IS strategy">IT/IS strategy</a>, <a href="https://publications.waset.org/search?q=IT%2FIS%20organisation%20design" title=" IT/IS organisation design"> IT/IS organisation design</a>, <a href="https://publications.waset.org/search?q=digital%20age" title=" digital age"> digital age</a>, <a href="https://publications.waset.org/search?q=organisational%20effectiveness" title=" organisational effectiveness"> organisational effectiveness</a>, <a href="https://publications.waset.org/search?q=literature%20review." title=" literature review."> literature review.</a> </p> <a href="https://publications.waset.org/10001060/itis-organisation-design-in-the-digital-age-a-literature-review" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/10001060/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/10001060/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/10001060/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/10001060/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/10001060/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/10001060/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/10001060/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/10001060/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/10001060/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/10001060/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/10001060.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">3742</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">8</span> Subjective Versus Objective Assessment for Magnetic Resonance Images</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Heshalini%20Rajagopal">Heshalini Rajagopal</a>, <a href="https://publications.waset.org/search?q=Li%20Sze%20Chow"> Li Sze Chow</a>, <a href="https://publications.waset.org/search?q=Raveendran%20Paramesran"> Raveendran Paramesran</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Magnetic Resonance Imaging (MRI) is one of the most important medical imaging modality. Subjective assessment of the image quality is regarded as the gold standard to evaluate MR images. In this study, a database of 210 MR images which contains ten reference images and 200 distorted images is presented. The reference images were distorted with four types of distortions: Rician Noise, Gaussian White Noise, Gaussian Blur and DCT compression. The 210 images were assessed by ten subjects. The subjective scores were presented in Difference Mean Opinion Score (DMOS). The DMOS values were compared with four FR-IQA metrics. We have used Pearson Linear Coefficient (PLCC) and Spearman Rank Order Correlation Coefficient (SROCC) to validate the DMOS values. The high correlation values of PLCC and SROCC shows that the DMOS values are close to the objective FR-IQA metrics. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Medical%20Resonance%20%28MR%29%20images" title="Medical Resonance (MR) images">Medical Resonance (MR) images</a>, <a href="https://publications.waset.org/search?q=Difference%20Mean%0D%0AOpinion%20Score%20%28DMOS%29" title=" Difference Mean Opinion Score (DMOS)"> Difference Mean Opinion Score (DMOS)</a>, <a href="https://publications.waset.org/search?q=Full%20Reference%20Image%20Quality%20Assessment%0D%0A%28FR-IQA%29." title=" Full Reference Image Quality Assessment (FR-IQA)."> Full Reference Image Quality Assessment (FR-IQA).</a> </p> <a href="https://publications.waset.org/10002926/subjective-versus-objective-assessment-for-magnetic-resonance-images" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/10002926/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/10002926/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/10002926/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/10002926/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/10002926/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/10002926/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/10002926/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/10002926/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/10002926/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/10002926/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/10002926.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">2218</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">7</span> Bridging Consumer-Farmer Mobile Application Divide</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=A.%20Hol">A. Hol</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Electronic mediums such as websites, feeds, blogs and social media sites are on a daily basis influencing our decision making, are improving our productivity and are shaping futures of many consumers and service/product providers. This research identifies that both customers and business providers heavily rely on smart phone applications. Based on this, mobile applications available on iTunes store were studied. It was identified that fruit and vegetable related applications used by consumers can broadly be categorized into purchase applications, diaries, tracking health applications, trip farm location and cooking applications. On the other hand, applications used by farmers can broadly be classified as: weather tracking, pests / fertilizer applications and general social media applications such as Facebook. To blur this farmer-consumer application divide, our research utilizes Context Specific eTransformation Framework and based on it identifies characteristic future consumer-farmer applications will need to have so that the current divide can be narrowed and consequently better farmerconsumer supply chain link established. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Smart%20Phone%20Applications" title="Smart Phone Applications">Smart Phone Applications</a>, <a href="https://publications.waset.org/search?q=SME" title=" SME"> SME</a>, <a href="https://publications.waset.org/search?q=Farmers" title=" Farmers"> Farmers</a>, <a href="https://publications.waset.org/search?q=Consumer" title=" Consumer"> Consumer</a>, <a href="https://publications.waset.org/search?q=Fruit%20and%20Vegetable" title=" Fruit and Vegetable"> Fruit and Vegetable</a>, <a href="https://publications.waset.org/search?q=Technology" title=" Technology"> Technology</a>, <a href="https://publications.waset.org/search?q=Business%20Innovation." title=" Business Innovation."> Business Innovation.</a> </p> <a href="https://publications.waset.org/10002766/bridging-consumer-farmer-mobile-application-divide" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/10002766/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/10002766/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/10002766/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/10002766/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/10002766/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/10002766/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/10002766/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/10002766/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/10002766/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/10002766/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/10002766.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1559</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">6</span> A Local Invariant Generalized Hough Transform Method for Integrated Circuit Visual Positioning</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Fei%20Long%20Wei">Fei Long Wei</a>, <a href="https://publications.waset.org/search?q=Hua%20Yang"> Hua Yang</a>, <a href="https://publications.waset.org/search?q=Hai%20Tao%20Zhang"> Hai Tao Zhang</a>, <a href="https://publications.waset.org/search?q=Zhou%20Ping%20Yin"> Zhou Ping Yin</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>In this study, an local invariant generalized Houghtransform (LI-GHT) method is proposed for integrated circuit (IC) visual positioning. The original generalized Hough transform (GHT) is robust to external noise; however, it is not suitable for visual positioning of IC chips due to the four-dimensionality (4D) of parameter space which leads to the substantial storage requirement and high computational complexity. The proposed LI-GHT method can reduce the dimensionality of parameter space to 2D thanks to the rotational invariance of local invariant geometric feature and it can estimate the accuracy position and rotation angle of IC chips in real-time under noise and blur influence. The experiment results show that the proposed LI-GHT can estimate position and rotation angle of IC chips with high accuracy and fast speed. The proposed LI-GHT algorithm was implemented in IC visual positioning system of radio frequency identification (RFID) packaging equipment.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Integrated%20Circuit%20Visual%20Positioning" title="Integrated Circuit Visual Positioning">Integrated Circuit Visual Positioning</a>, <a href="https://publications.waset.org/search?q=Generalized%20Hough%20Transform" title=" Generalized Hough Transform"> Generalized Hough Transform</a>, <a href="https://publications.waset.org/search?q=Local%20invariant%20Generalized%20Hough%20Transform" title=" Local invariant Generalized Hough Transform"> Local invariant Generalized Hough Transform</a>, <a href="https://publications.waset.org/search?q=ICpacking%20equipment." title=" ICpacking equipment."> ICpacking equipment.</a> </p> <a href="https://publications.waset.org/9996827/a-local-invariant-generalized-hough-transform-method-for-integrated-circuit-visual-positioning" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/9996827/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/9996827/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/9996827/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/9996827/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/9996827/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/9996827/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/9996827/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/9996827/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/9996827/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/9996827/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/9996827.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">2208</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">5</span> Performance of Compound Enhancement Algorithms on Dental Radiograph Images</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=S.A.Ahmad">S.A.Ahmad</a>, <a href="https://publications.waset.org/search?q=M.N.Taib"> M.N.Taib</a>, <a href="https://publications.waset.org/search?q=N.E.A.Khalid"> N.E.A.Khalid</a>, <a href="https://publications.waset.org/search?q=R.Ahmad"> R.Ahmad</a>, <a href="https://publications.waset.org/search?q=H.Taib"> H.Taib</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The purpose of this research is to compare the original intra-oral digital dental radiograph images with images that are enhanced using a combination of image processing algorithms. Intraoral digital dental radiograph images are often noisy, blur edges and low in contrast. A combination of sharpening and enhancement method are used to overcome these problems. Three types of proposed compound algorithms used are Sharp Adaptive Histogram Equalization (SAHE), Sharp Median Adaptive Histogram Equalization (SMAHE) and Sharp Contrast adaptive histogram equalization (SCLAHE). This paper presents an initial study of the perception of six dentists on the details of abnormal pathologies and improvement of image quality in ten intra-oral radiographs. The research focus on the detection of only three types of pathology which is periapical radiolucency, widen periodontal ligament space and loss of lamina dura. The overall result shows that SCLAHE-s slightly improve the appearance of dental abnormalities- over the original image and also outperform the other two proposed compound algorithms. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=intra-oral%20dental%20radiograph" title="intra-oral dental radiograph">intra-oral dental radiograph</a>, <a href="https://publications.waset.org/search?q=histogram%20equalization" title=" histogram equalization"> histogram equalization</a>, <a href="https://publications.waset.org/search?q=sharpening" title="sharpening">sharpening</a>, <a href="https://publications.waset.org/search?q=CLAHE." title=" CLAHE."> CLAHE.</a> </p> <a href="https://publications.waset.org/4116/performance-of-compound-enhancement-algorithms-on-dental-radiograph-images" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/4116/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/4116/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/4116/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/4116/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/4116/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/4116/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/4116/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/4116/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/4116/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/4116/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/4116.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1784</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">4</span> EEG-Based Fractal Analysis of Different Motor Imagery Tasks using Critical Exponent Method</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Montri%20Phothisonothai">Montri Phothisonothai</a>, <a href="https://publications.waset.org/search?q=Masahiro%20Nakagawa"> Masahiro Nakagawa</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>The objective of this paper is to characterize the spontaneous Electroencephalogram (EEG) signals of four different motor imagery tasks and to show hereby a possible solution for the present binary communication between the brain and a machine ora Brain-Computer Interface (BCI). The processing technique used in this paper was the fractal analysis evaluated by the Critical Exponent Method (CEM). The EEG signal was registered in 5 healthy subjects,sampling 15 measuring channels at 1024 Hz.Each channel was preprocessed by the Laplacian space ltering so as to reduce the space blur and therefore increase the spaceresolution. The EEG of each channel was segmented and its Fractaldimension (FD) calculated. The FD was evaluated in the time interval corresponding to the motor imagery and averaged out for all the subjects (each channel). In order to characterize the FD distribution,the linear regression curves of FD over the electrodes position were applied. The differences FD between the proposed mental tasks are quantied and evaluated for each experimental subject. The obtained results of the proposed method are a substantial fractal dimension in the EEG signal of motor imagery tasks and can be considerably utilized as the multiple-states BCI applications.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=electroencephalogram%20%28EEG%29" title="electroencephalogram (EEG)">electroencephalogram (EEG)</a>, <a href="https://publications.waset.org/search?q=motor%20imagery%20tasks" title=" motor imagery tasks"> motor imagery tasks</a>, <a href="https://publications.waset.org/search?q=mental%20tasks" title=" mental tasks"> mental tasks</a>, <a href="https://publications.waset.org/search?q=biomedical%20signals%20processing" title=" biomedical signals processing"> biomedical signals processing</a>, <a href="https://publications.waset.org/search?q=human-machine%20interface" title=" human-machine interface"> human-machine interface</a>, <a href="https://publications.waset.org/search?q=fractal%20analysis" title=" fractal analysis"> fractal analysis</a>, <a href="https://publications.waset.org/search?q=critical%20exponent%20method%20%28CEM%29." title=" critical exponent method (CEM)."> critical exponent method (CEM).</a> </p> <a href="https://publications.waset.org/9878/eeg-based-fractal-analysis-of-different-motor-imagery-tasks-using-critical-exponent-method" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/9878/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/9878/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/9878/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/9878/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/9878/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/9878/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/9878/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/9878/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/9878/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/9878/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/9878.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">2259</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3</span> Fused Structure and Texture (FST) Features for Improved Pedestrian Detection </h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Hussin%20K.%20Ragb">Hussin K. Ragb</a>, <a href="https://publications.waset.org/search?q=Vijayan%20K.%20Asari"> Vijayan K. Asari </a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, we present a pedestrian detection descriptor called Fused Structure and Texture (FST) features based on the combination of the local phase information with the texture features. Since the phase of the signal conveys more structural information than the magnitude, the phase congruency concept is used to capture the structural features. On the other hand, the Center-Symmetric Local Binary Pattern (CSLBP) approach is used to capture the texture information of the image. The dimension less quantity of the phase congruency and the robustness of the CSLBP operator on the flat images, as well as the blur and illumination changes, lead the proposed descriptor to be more robust and less sensitive to the light variations. The proposed descriptor can be formed by extracting the phase congruency and the CSLBP values of each pixel of the image with respect to its neighborhood. The histogram of the oriented phase and the histogram of the CSLBP values for the local regions in the image are computed and concatenated to construct the FST descriptor. Several experiments were conducted on INRIA and the low resolution DaimlerChrysler datasets to evaluate the detection performance of the pedestrian detection system that is based on the FST descriptor. A linear Support Vector Machine (SVM) is used to train the pedestrian classifier. These experiments showed that the proposed FST descriptor has better detection performance over a set of state of the art feature extraction methodologies. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Pedestrian%20detection" title="Pedestrian detection">Pedestrian detection</a>, <a href="https://publications.waset.org/search?q=phase%20congruency" title=" phase congruency"> phase congruency</a>, <a href="https://publications.waset.org/search?q=local%20phase" title=" local phase"> local phase</a>, <a href="https://publications.waset.org/search?q=LBP%20features" title=" LBP features"> LBP features</a>, <a href="https://publications.waset.org/search?q=CSLBP%20features" title=" CSLBP features"> CSLBP features</a>, <a href="https://publications.waset.org/search?q=FST%20descriptor." title=" FST descriptor."> FST descriptor.</a> </p> <a href="https://publications.waset.org/10003679/fused-structure-and-texture-fst-features-for-improved-pedestrian-detection" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/10003679/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/10003679/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/10003679/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/10003679/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/10003679/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/10003679/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/10003679/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/10003679/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/10003679/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/10003679/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/10003679.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">1489</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">2</span> Image Restoration in Non-Linear Filtering Domain using MDB approach</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=S.%20K.%20Satpathy">S. K. Satpathy</a>, <a href="https://publications.waset.org/search?q=S.%20Panda"> S. Panda</a>, <a href="https://publications.waset.org/search?q=K.%20K.%20Nagwanshi"> K. K. Nagwanshi</a>, <a href="https://publications.waset.org/search?q=C.%20Ardil"> C. Ardil</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>This paper proposes a new technique based on nonlinear Minmax Detector Based (MDB) filter for image restoration. The aim of image enhancement is to reconstruct the true image from the corrupted image. The process of image acquisition frequently leads to degradation and the quality of the digitized image becomes inferior to the original image. Image degradation can be due to the addition of different types of noise in the original image. Image noise can be modeled of many types and impulse noise is one of them. Impulse noise generates pixels with gray value not consistent with their local neighborhood. It appears as a sprinkle of both light and dark or only light spots in the image. Filtering is a technique for enhancing the image. Linear filter is the filtering in which the value of an output pixel is a linear combination of neighborhood values, which can produce blur in the image. Thus a variety of smoothing techniques have been developed that are non linear. Median filter is the one of the most popular non-linear filter. When considering a small neighborhood it is highly efficient but for large window and in case of high noise it gives rise to more blurring to image. The Centre Weighted Mean (CWM) filter has got a better average performance over the median filter. However the original pixel corrupted and noise reduction is substantial under high noise condition. Hence this technique has also blurring affect on the image. To illustrate the superiority of the proposed approach, the proposed new scheme has been simulated along with the standard ones and various restored performance measures have been compared.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Filtering" title="Filtering">Filtering</a>, <a href="https://publications.waset.org/search?q=Minmax%20Detector%20Based%20%28MDB%29" title=" Minmax Detector Based (MDB)"> Minmax Detector Based (MDB)</a>, <a href="https://publications.waset.org/search?q=noise" title=" noise"> noise</a>, <a href="https://publications.waset.org/search?q=centre%20weighted%20mean%20filter" title="centre weighted mean filter">centre weighted mean filter</a>, <a href="https://publications.waset.org/search?q=PSNR" title=" PSNR"> PSNR</a>, <a href="https://publications.waset.org/search?q=restoration." title=" restoration."> restoration.</a> </p> <a href="https://publications.waset.org/11097/image-restoration-in-non-linear-filtering-domain-using-mdb-approach" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/11097/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/11097/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/11097/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/11097/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/11097/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/11097/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/11097/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/11097/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/11097/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/11097/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/11097.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">2739</span> </span> </div> </div> <div class="card publication-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1</span> A Survey of Online User Perspectives and Age Profile in an Undergraduate Fundamental Business Technology Course</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/search?q=Danielle%20Morin">Danielle Morin</a>, <a href="https://publications.waset.org/search?q=Jennifer%20D.%20E.%20Thomas"> Jennifer D. E. Thomas</a>, <a href="https://publications.waset.org/search?q=Raafat%20G.%20Saade"> Raafat G. Saade</a>, <a href="https://publications.waset.org/search?q=Daniela%20Petrachi"> Daniela Petrachi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> <p>Over the past few decades, more and more students choose to enroll in online classes instead of attending in-class lectures. While past studies consider students&rsquo; attitudes towards online education and how their grades differed from in-class lectures, the profile of the online student remains a blur. To shed light on this, an online survey was administered to about 1,500 students enrolled in an undergraduate Fundamental Business Technology course at a Canadian University. The survey was comprised of questions on students&rsquo; demographics, their reasons for choosing online courses, their expectations towards the course, the communication channels they use for the course with fellow students and with the instructor. This paper focused on the research question: Do the perspectives of online students concerning the online experience, in general, and in the course in particular, differ according to age profile? After several statistical analyses, it was found that age does have an impact on the reasons why students select online classes instead of in-class. For example, it was found that the perception that an online course might be easier than in-class delivery was a more important reason for younger students than for older ones. Similarly, the influence of friends is much more important for younger students, than for older students. Similar results were found when analyzing students&rsquo; expectation about the online course and their use of communication tools. Overall, the age profile of online users had an impact on reasons, expectations and means of communication in an undergraduate Fundamental Business Technology course. It is left to be seen if this holds true across other courses, graduate and undergraduate.</p> <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/search?q=Communication%20channels" title="Communication channels">Communication channels</a>, <a href="https://publications.waset.org/search?q=fundamentals%20of%20business%20technology" title=" fundamentals of business technology"> fundamentals of business technology</a>, <a href="https://publications.waset.org/search?q=online%20classes" title=" online classes"> online classes</a>, <a href="https://publications.waset.org/search?q=pedagogy" title=" pedagogy"> pedagogy</a>, <a href="https://publications.waset.org/search?q=user%20age%20profile" title=" user age profile"> user age profile</a>, <a href="https://publications.waset.org/search?q=user%20perspectives." title=" user perspectives."> user perspectives.</a> </p> <a href="https://publications.waset.org/10009134/a-survey-of-online-user-perspectives-and-age-profile-in-an-undergraduate-fundamental-business-technology-course" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/10009134/apa" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">APA</a> <a href="https://publications.waset.org/10009134/bibtex" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">BibTeX</a> <a href="https://publications.waset.org/10009134/chicago" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Chicago</a> <a href="https://publications.waset.org/10009134/endnote" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">EndNote</a> <a href="https://publications.waset.org/10009134/harvard" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">Harvard</a> <a href="https://publications.waset.org/10009134/json" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">JSON</a> <a href="https://publications.waset.org/10009134/mla" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">MLA</a> <a href="https://publications.waset.org/10009134/ris" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">RIS</a> <a href="https://publications.waset.org/10009134/xml" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">XML</a> <a href="https://publications.waset.org/10009134/iso690" target="_blank" rel="nofollow" class="btn btn-primary btn-sm">ISO 690</a> <a href="https://publications.waset.org/10009134.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">768</span> </span> </div> </div> </div> </main> <footer> <div id="infolinks" class="pt-3 pb-2"> <div class="container"> <div style="background-color:#f5f5f5;" class="p-3"> <div class="row"> <div class="col-md-2"> <ul class="list-unstyled"> About <li><a href="https://waset.org/page/support">About Us</a></li> <li><a href="https://waset.org/page/support#legal-information">Legal</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/WASET-16th-foundational-anniversary.pdf">WASET celebrates its 16th foundational anniversary</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Account <li><a href="https://waset.org/profile">My Account</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Explore <li><a href="https://waset.org/disciplines">Disciplines</a></li> <li><a href="https://waset.org/conferences">Conferences</a></li> <li><a href="https://waset.org/conference-programs">Conference Program</a></li> <li><a href="https://waset.org/committees">Committees</a></li> <li><a href="https://publications.waset.org">Publications</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Research <li><a href="https://publications.waset.org/abstracts">Abstracts</a></li> <li><a href="https://publications.waset.org">Periodicals</a></li> <li><a href="https://publications.waset.org/archive">Archive</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Open Science <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Philosophy.pdf">Open Science Philosophy</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Award.pdf">Open Science Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Society-Open-Science-and-Open-Innovation.pdf">Open Innovation</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Postdoctoral-Fellowship-Award.pdf">Postdoctoral Fellowship Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Scholarly-Research-Review.pdf">Scholarly Research Review</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Support <li><a href="https://waset.org/page/support">Support</a></li> <li><a href="https://waset.org/profile/messages/create">Contact Us</a></li> <li><a href="https://waset.org/profile/messages/create">Report Abuse</a></li> </ul> </div> </div> </div> </div> </div> <div class="container text-center"> <hr style="margin-top:0;margin-bottom:.3rem;"> <a href="https://creativecommons.org/licenses/by/4.0/" target="_blank" class="text-muted small">Creative Commons Attribution 4.0 International License</a> <div id="copy" class="mt-2">&copy; 2024 World Academy of Science, Engineering and Technology</div> </div> </footer> <a href="javascript:" id="return-to-top"><i class="fas fa-arrow-up"></i></a> <div class="modal" id="modal-template"> <div class="modal-dialog"> <div class="modal-content"> <div class="row m-0 mt-1"> <div class="col-md-12"> <button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">&times;</span></button> </div> </div> <div class="modal-body"></div> </div> </div> </div> <script src="https://cdn.waset.org/static/plugins/jquery-3.3.1.min.js"></script> <script src="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/js/bootstrap.bundle.min.js"></script> <script src="https://cdn.waset.org/static/js/site.js?v=150220211556"></script> <script> jQuery(document).ready(function() { /*jQuery.get("https://publications.waset.org/xhr/user-menu", function (response) { jQuery('#mainNavMenu').append(response); });*/ jQuery.get({ url: "https://publications.waset.org/xhr/user-menu", cache: false }).then(function(response){ jQuery('#mainNavMenu').append(response); }); }); </script> </body> </html>

Pages: 1 2 3 4 5 6 7 8 9 10