CINXE.COM
Search results for: image similarity
<!DOCTYPE html> <html lang="en" dir="ltr"> <head> <!-- Google tag (gtag.js) --> <script async src="https://www.googletagmanager.com/gtag/js?id=G-P63WKM1TM1"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('js', new Date()); gtag('config', 'G-P63WKM1TM1'); </script> <!-- Yandex.Metrika counter --> <script type="text/javascript" > (function(m,e,t,r,i,k,a){m[i]=m[i]||function(){(m[i].a=m[i].a||[]).push(arguments)}; m[i].l=1*new Date(); for (var j = 0; j < document.scripts.length; j++) {if (document.scripts[j].src === r) { return; }} k=e.createElement(t),a=e.getElementsByTagName(t)[0],k.async=1,k.src=r,a.parentNode.insertBefore(k,a)}) (window, document, "script", "https://mc.yandex.ru/metrika/tag.js", "ym"); ym(55165297, "init", { clickmap:false, trackLinks:true, accurateTrackBounce:true, webvisor:false }); </script> <noscript><div><img src="https://mc.yandex.ru/watch/55165297" style="position:absolute; left:-9999px;" alt="" /></div></noscript> <!-- /Yandex.Metrika counter --> <!-- Matomo --> <!-- End Matomo Code --> <title>Search results for: image similarity</title> <meta name="description" content="Search results for: image similarity"> <meta name="keywords" content="image similarity"> <meta name="viewport" content="width=device-width, initial-scale=1, minimum-scale=1, maximum-scale=1, user-scalable=no"> <meta charset="utf-8"> <link href="https://cdn.waset.org/favicon.ico" type="image/x-icon" rel="shortcut icon"> <link href="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/css/bootstrap.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/plugins/fontawesome/css/all.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/css/site.css?v=150220211555" rel="stylesheet"> </head> <body> <header> <div class="container"> <nav class="navbar navbar-expand-lg navbar-light"> <a class="navbar-brand" href="https://waset.org"> <img src="https://cdn.waset.org/static/images/wasetc.png" alt="Open Science Research Excellence" title="Open Science Research Excellence" /> </a> <button class="d-block d-lg-none navbar-toggler ml-auto" type="button" data-toggle="collapse" data-target="#navbarMenu" aria-controls="navbarMenu" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <div class="w-100"> <div class="d-none d-lg-flex flex-row-reverse"> <form method="get" action="https://waset.org/search" class="form-inline my-2 my-lg-0"> <input class="form-control mr-sm-2" type="search" placeholder="Search Conferences" value="image similarity" name="q" aria-label="Search"> <button class="btn btn-light my-2 my-sm-0" type="submit"><i class="fas fa-search"></i></button> </form> </div> <div class="collapse navbar-collapse mt-1" id="navbarMenu"> <ul class="navbar-nav ml-auto align-items-center" id="mainNavMenu"> <li class="nav-item"> <a class="nav-link" href="https://waset.org/conferences" title="Conferences in 2024/2025/2026">Conferences</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/disciplines" title="Disciplines">Disciplines</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/committees" rel="nofollow">Committees</a> </li> <li class="nav-item dropdown"> <a class="nav-link dropdown-toggle" href="#" id="navbarDropdownPublications" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Publications </a> <div class="dropdown-menu" aria-labelledby="navbarDropdownPublications"> <a class="dropdown-item" href="https://publications.waset.org/abstracts">Abstracts</a> <a class="dropdown-item" href="https://publications.waset.org">Periodicals</a> <a class="dropdown-item" href="https://publications.waset.org/archive">Archive</a> </div> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/page/support" title="Support">Support</a> </li> </ul> </div> </div> </nav> </div> </header> <main> <div class="container mt-4"> <div class="row"> <div class="col-md-9 mx-auto"> <form method="get" action="https://publications.waset.org/abstracts/search"> <div id="custom-search-input"> <div class="input-group"> <i class="fas fa-search"></i> <input type="text" class="search-query" name="q" placeholder="Author, Title, Abstract, Keywords" value="image similarity"> <input type="submit" class="btn_search" value="Search"> </div> </div> </form> </div> </div> <div class="row mt-3"> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Commenced</strong> in January 2007</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Frequency:</strong> Monthly</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Edition:</strong> International</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Paper Count:</strong> 3358</div> </div> </div> </div> <h1 class="mt-3 mb-3 text-center" style="font-size:1.6rem;">Search results for: image similarity</h1> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3358</span> Quick Similarity Measurement of Binary Images via Probabilistic Pixel Mapping</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Adnan%20A.%20Y.%20Mustafa">Adnan A. Y. Mustafa</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper we present a quick technique to measure the similarity between binary images. The technique is based on a probabilistic mapping approach and is fast because only a minute percentage of the image pixels need to be compared to measure the similarity, and not the whole image. We exploit the power of the Probabilistic Matching Model for Binary Images (PMMBI) to arrive at an estimate of the similarity. We show that the estimate is a good approximation of the actual value, and the quality of the estimate can be improved further with increased image mappings. Furthermore, the technique is image size invariant; the similarity between big images can be measured as fast as that for small images. Examples of trials conducted on real images are presented. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=big%20images" title="big images">big images</a>, <a href="https://publications.waset.org/abstracts/search?q=binary%20images" title=" binary images"> binary images</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20matching" title=" image matching"> image matching</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20similarity" title=" image similarity"> image similarity</a> </p> <a href="https://publications.waset.org/abstracts/89963/quick-similarity-measurement-of-binary-images-via-probabilistic-pixel-mapping" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/89963.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">196</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3357</span> A Context-Sensitive Algorithm for Media Similarity Search </h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Guang-Ho%20Cha">Guang-Ho Cha</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents a context-sensitive media similarity search algorithm. One of the central problems regarding media search is the semantic gap between the low-level features computed automatically from media data and the human interpretation of them. This is because the notion of similarity is usually based on high-level abstraction but the low-level features do not sometimes reflect the human perception. Many media search algorithms have used the Minkowski metric to measure similarity between image pairs. However those functions cannot adequately capture the aspects of the characteristics of the human visual system as well as the nonlinear relationships in contextual information given by images in a collection. Our search algorithm tackles this problem by employing a similarity measure and a ranking strategy that reflect the nonlinearity of human perception and contextual information in a dataset. Similarity search in an image database based on this contextual information shows encouraging experimental results. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=context-sensitive%20search" title="context-sensitive search">context-sensitive search</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20search" title=" image search"> image search</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20ranking" title=" similarity ranking"> similarity ranking</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20search" title=" similarity search"> similarity search</a> </p> <a href="https://publications.waset.org/abstracts/65150/a-context-sensitive-algorithm-for-media-similarity-search" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/65150.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">365</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3356</span> Comparative Analysis of Dissimilarity Detection between Binary Images Based on Equivalency and Non-Equivalency of Image Inversion</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Adnan%20A.%20Y.%20Mustafa">Adnan A. Y. Mustafa</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Image matching is a fundamental problem that arises frequently in many aspects of robot and computer vision. It can become a time-consuming process when matching images to a database consisting of hundreds of images, especially if the images are big. One approach to reducing the time complexity of the matching process is to reduce the search space in a pre-matching stage, by simply removing dissimilar images quickly. The Probabilistic Matching Model for Binary Images (PMMBI) showed that dissimilarity detection between binary images can be accomplished quickly by random pixel mapping and is size invariant. The model is based on the gamma binary similarity distance that recognizes an image and its inverse as containing the same scene and hence considers them to be the same image. However, in many applications, an image and its inverse are not treated as being the same but rather dissimilar. In this paper, we present a comparative analysis of dissimilarity detection between PMMBI based on the gamma binary similarity distance and a modified PMMBI model based on a similarity distance that does distinguish between an image and its inverse as being dissimilar. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=binary%20image" title="binary image">binary image</a>, <a href="https://publications.waset.org/abstracts/search?q=dissimilarity%20detection" title=" dissimilarity detection"> dissimilarity detection</a>, <a href="https://publications.waset.org/abstracts/search?q=probabilistic%20matching%20model%20for%20binary%20images" title=" probabilistic matching model for binary images"> probabilistic matching model for binary images</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20mapping" title=" image mapping"> image mapping</a> </p> <a href="https://publications.waset.org/abstracts/113778/comparative-analysis-of-dissimilarity-detection-between-binary-images-based-on-equivalency-and-non-equivalency-of-image-inversion" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/113778.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">153</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3355</span> Generation of Photo-Mosaic Images through Block Matching and Color Adjustment</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hae-Yeoun%20Lee">Hae-Yeoun Lee</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Mosaic refers to a technique that makes image by gathering lots of small materials in various colours. This paper presents an automatic algorithm that makes the photomosaic image using photos. The algorithm is composed of four steps: Partition and feature extraction, block matching, redundancy removal and colour adjustment. The input image is partitioned in the small block to extract feature. Each block is matched to find similar photo in database by comparing similarity with Euclidean difference between blocks. The intensity of the block is adjusted to enhance the similarity of image by replacing the value of light and darkness with that of relevant block. Further, the quality of image is improved by minimizing the redundancy of tiles in the adjacent blocks. Experimental results support that the proposed algorithm is excellent in quantitative analysis and qualitative analysis. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=photomosaic" title="photomosaic">photomosaic</a>, <a href="https://publications.waset.org/abstracts/search?q=Euclidean%20distance" title=" Euclidean distance"> Euclidean distance</a>, <a href="https://publications.waset.org/abstracts/search?q=block%20matching" title=" block matching"> block matching</a>, <a href="https://publications.waset.org/abstracts/search?q=intensity%20adjustment" title=" intensity adjustment"> intensity adjustment</a> </p> <a href="https://publications.waset.org/abstracts/7022/generation-of-photo-mosaic-images-through-block-matching-and-color-adjustment" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/7022.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">278</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3354</span> GPU Accelerated Fractal Image Compression for Medical Imaging in Parallel Computing Platform</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Md.%20Enamul%20Haque">Md. Enamul Haque</a>, <a href="https://publications.waset.org/abstracts/search?q=Abdullah%20Al%20Kaisan"> Abdullah Al Kaisan</a>, <a href="https://publications.waset.org/abstracts/search?q=Mahmudur%20R.%20Saniat"> Mahmudur R. Saniat</a>, <a href="https://publications.waset.org/abstracts/search?q=Aminur%20Rahman"> Aminur Rahman</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, we have implemented both sequential and parallel version of fractal image compression algorithms using CUDA (Compute Unified Device Architecture) programming model for parallelizing the program in Graphics Processing Unit for medical images, as they are highly similar within the image itself. There is several improvements in the implementation of the algorithm as well. Fractal image compression is based on the self similarity of an image, meaning an image having similarity in majority of the regions. We take this opportunity to implement the compression algorithm and monitor the effect of it using both parallel and sequential implementation. Fractal compression has the property of high compression rate and the dimensionless scheme. Compression scheme for fractal image is of two kinds, one is encoding and another is decoding. Encoding is very much computational expensive. On the other hand decoding is less computational. The application of fractal compression to medical images would allow obtaining much higher compression ratios. While the fractal magnification an inseparable feature of the fractal compression would be very useful in presenting the reconstructed image in a highly readable form. However, like all irreversible methods, the fractal compression is connected with the problem of information loss, which is especially troublesome in the medical imaging. A very time consuming encoding process, which can last even several hours, is another bothersome drawback of the fractal compression. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=accelerated%20GPU" title="accelerated GPU">accelerated GPU</a>, <a href="https://publications.waset.org/abstracts/search?q=CUDA" title=" CUDA"> CUDA</a>, <a href="https://publications.waset.org/abstracts/search?q=parallel%20computing" title=" parallel computing"> parallel computing</a>, <a href="https://publications.waset.org/abstracts/search?q=fractal%20image%20compression" title=" fractal image compression"> fractal image compression</a> </p> <a href="https://publications.waset.org/abstracts/5645/gpu-accelerated-fractal-image-compression-for-medical-imaging-in-parallel-computing-platform" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/5645.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">335</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3353</span> Graph Cuts Segmentation Approach Using a Patch-Based Similarity Measure Applied for Interactive CT Lung Image Segmentation</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Aicha%20Majda">Aicha Majda</a>, <a href="https://publications.waset.org/abstracts/search?q=Abdelhamid%20El%20Hassani"> Abdelhamid El Hassani</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Lung CT image segmentation is a prerequisite in lung CT image analysis. Most of the conventional methods need a post-processing to deal with the abnormal lung CT scans such as lung nodules or other lesions. The simplest similarity measure in the standard Graph Cuts Algorithm consists of directly comparing the pixel values of the two neighboring regions, which is not accurate because this kind of metrics is extremely sensitive to minor transformations such as noise or other artifacts problems. In this work, we propose an improved version of the standard graph cuts algorithm based on the Patch-Based similarity metric. The boundary penalty term in the graph cut algorithm is defined Based on Patch-Based similarity measurement instead of the simple intensity measurement in the standard method. The weights between each pixel and its neighboring pixels are Based on the obtained new term. The graph is then created using theses weights between its nodes. Finally, the segmentation is completed with the minimum cut/Max-Flow algorithm. Experimental results show that the proposed method is very accurate and efficient, and can directly provide explicit lung regions without any post-processing operations compared to the standard method. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=graph%20cuts" title="graph cuts">graph cuts</a>, <a href="https://publications.waset.org/abstracts/search?q=lung%20CT%20scan" title=" lung CT scan"> lung CT scan</a>, <a href="https://publications.waset.org/abstracts/search?q=lung%20parenchyma%20segmentation" title=" lung parenchyma segmentation"> lung parenchyma segmentation</a>, <a href="https://publications.waset.org/abstracts/search?q=patch-based%20similarity%20metric" title=" patch-based similarity metric"> patch-based similarity metric</a> </p> <a href="https://publications.waset.org/abstracts/87346/graph-cuts-segmentation-approach-using-a-patch-based-similarity-measure-applied-for-interactive-ct-lung-image-segmentation" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/87346.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">169</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3352</span> A Nonlocal Means Algorithm for Poisson Denoising Based on Information Geometry</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Dongxu%20Chen">Dongxu Chen</a>, <a href="https://publications.waset.org/abstracts/search?q=Yipeng%20Li"> Yipeng Li</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents an information geometry NonlocalMeans(NLM) algorithm for Poisson denoising. NLM estimates a noise-free pixel as a weighted average of image pixels, where each pixel is weighted according to the similarity between image patches in Euclidean space. In this work, every pixel is a Poisson distribution locally estimated by Maximum Likelihood (ML), all distributions consist of a statistical manifold. A NLM denoising algorithm is conducted on the statistical manifold where Fisher information matrix can be used for computing distribution geodesics referenced as the similarity between patches. This approach was demonstrated to be competitive with related state-of-the-art methods. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=image%20denoising" title="image denoising">image denoising</a>, <a href="https://publications.waset.org/abstracts/search?q=Poisson%20noise" title=" Poisson noise"> Poisson noise</a>, <a href="https://publications.waset.org/abstracts/search?q=information%20geometry" title=" information geometry"> information geometry</a>, <a href="https://publications.waset.org/abstracts/search?q=nonlocal-means" title=" nonlocal-means"> nonlocal-means</a> </p> <a href="https://publications.waset.org/abstracts/51221/a-nonlocal-means-algorithm-for-poisson-denoising-based-on-information-geometry" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/51221.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">285</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3351</span> Saliency Detection Using a Background Probability Model</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Junling%20Li">Junling Li</a>, <a href="https://publications.waset.org/abstracts/search?q=Fang%20Meng"> Fang Meng</a>, <a href="https://publications.waset.org/abstracts/search?q=Yichun%20Zhang"> Yichun Zhang</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Image saliency detection has been long studied, while several challenging problems are still unsolved, such as detecting saliency inaccurately in complex scenes or suppressing salient objects in the image borders. In this paper, we propose a new saliency detection algorithm in order to solving these problems. We represent the image as a graph with superixels as nodes. By considering appearance similarity between the boundary and the background, the proposed method chooses non-saliency boundary nodes as background priors to construct the background probability model. The probability that each node belongs to the model is computed, which measures its similarity with backgrounds. Thus we can calculate saliency by the transformed probability as a metric. We compare our algorithm with ten-state-of-the-art salient detection methods on the public database. Experimental results show that our simple and effective approach can attack those challenging problems that had been baffling in image saliency detection. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=visual%20saliency" title="visual saliency">visual saliency</a>, <a href="https://publications.waset.org/abstracts/search?q=background%20probability" title=" background probability"> background probability</a>, <a href="https://publications.waset.org/abstracts/search?q=boundary%20knowledge" title=" boundary knowledge"> boundary knowledge</a>, <a href="https://publications.waset.org/abstracts/search?q=background%20priors" title=" background priors"> background priors</a> </p> <a href="https://publications.waset.org/abstracts/13729/saliency-detection-using-a-background-probability-model" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/13729.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">429</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3350</span> A Comparison between Different Segmentation Techniques Used in Medical Imaging </h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ibtihal%20D.%20Mustafa">Ibtihal D. Mustafa</a>, <a href="https://publications.waset.org/abstracts/search?q=Mawia%20A.%20Hassan"> Mawia A. Hassan</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Tumor segmentation from MRI image is important part of medical images experts. This is particularly a challenging task because of the high assorting appearance of tumor tissue among different patients. MRI images are advance of medical imaging because it is give richer information about human soft tissue. There are different segmentation techniques to detect MRI brain tumor. In this paper, different procedure segmentation methods are used to segment brain tumors and compare the result of segmentations by using correlation and structural similarity index (SSIM) to analysis and see the best technique that could be applied to MRI image. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=MRI" title="MRI">MRI</a>, <a href="https://publications.waset.org/abstracts/search?q=segmentation" title=" segmentation"> segmentation</a>, <a href="https://publications.waset.org/abstracts/search?q=correlation" title=" correlation"> correlation</a>, <a href="https://publications.waset.org/abstracts/search?q=structural%20similarity" title=" structural similarity"> structural similarity</a> </p> <a href="https://publications.waset.org/abstracts/51091/a-comparison-between-different-segmentation-techniques-used-in-medical-imaging" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/51091.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">410</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3349</span> A Deep Learning Based Approach for Dynamically Selecting Pre-processing Technique for Images</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Revoti%20Prasad%20Bora">Revoti Prasad Bora</a>, <a href="https://publications.waset.org/abstracts/search?q=Nikita%20Katyal"> Nikita Katyal</a>, <a href="https://publications.waset.org/abstracts/search?q=Saurabh%20Yadav"> Saurabh Yadav</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Pre-processing plays an important role in various image processing applications. Most of the time due to the similar nature of images, a particular pre-processing or a set of pre-processing steps are sufficient to produce the desired results. However, in the education domain, there is a wide variety of images in various aspects like images with line-based diagrams, chemical formulas, mathematical equations, etc. Hence a single pre-processing or a set of pre-processing steps may not yield good results. Therefore, a Deep Learning based approach for dynamically selecting a relevant pre-processing technique for each image is proposed. The proposed method works as a classifier to detect hidden patterns in the images and predicts the relevant pre-processing technique needed for the image. This approach experimented for an image similarity matching problem but it can be adapted to other use cases too. Experimental results showed significant improvement in average similarity ranking with the proposed method as opposed to static pre-processing techniques. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=deep-learning" title="deep-learning">deep-learning</a>, <a href="https://publications.waset.org/abstracts/search?q=classification" title=" classification"> classification</a>, <a href="https://publications.waset.org/abstracts/search?q=pre-processing" title=" pre-processing"> pre-processing</a>, <a href="https://publications.waset.org/abstracts/search?q=computer%20vision" title=" computer vision"> computer vision</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20processing" title=" image processing"> image processing</a>, <a href="https://publications.waset.org/abstracts/search?q=educational%20data%20mining" title=" educational data mining"> educational data mining</a> </p> <a href="https://publications.waset.org/abstracts/148397/a-deep-learning-based-approach-for-dynamically-selecting-pre-processing-technique-for-images" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/148397.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">163</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3348</span> A Similarity Measure for Classification and Clustering in Image Based Medical and Text Based Banking Applications</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=K.%20P.%20Sandesh">K. P. Sandesh</a>, <a href="https://publications.waset.org/abstracts/search?q=M.%20H.%20Suman"> M. H. Suman</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Text processing plays an important role in information retrieval, data-mining, and web search. Measuring the similarity between the documents is an important operation in the text processing field. In this project, a new similarity measure is proposed. To compute the similarity between two documents with respect to a feature the proposed measure takes the following three cases into account: (1) The feature appears in both documents; (2) The feature appears in only one document and; (3) The feature appears in none of the documents. The proposed measure is extended to gauge the similarity between two sets of documents. The effectiveness of our measure is evaluated on several real-world data sets for text classification and clustering problems, especially in banking and health sectors. The results show that the performance obtained by the proposed measure is better than that achieved by the other measures. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=document%20classification" title="document classification">document classification</a>, <a href="https://publications.waset.org/abstracts/search?q=document%20clustering" title=" document clustering"> document clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=entropy" title=" entropy"> entropy</a>, <a href="https://publications.waset.org/abstracts/search?q=accuracy" title=" accuracy"> accuracy</a>, <a href="https://publications.waset.org/abstracts/search?q=classifiers" title=" classifiers"> classifiers</a>, <a href="https://publications.waset.org/abstracts/search?q=clustering%20algorithms" title=" clustering algorithms"> clustering algorithms</a> </p> <a href="https://publications.waset.org/abstracts/22708/a-similarity-measure-for-classification-and-clustering-in-image-based-medical-and-text-based-banking-applications" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/22708.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">518</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3347</span> Multi-Objective Optimal Threshold Selection for Similarity Functions in Siamese Networks for Semantic Textual Similarity Tasks</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Kriuk%20Boris">Kriuk Boris</a>, <a href="https://publications.waset.org/abstracts/search?q=Kriuk%20Fedor"> Kriuk Fedor</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents a comparative study of fundamental similarity functions for Siamese networks in semantic textual similarity (STS) tasks. We evaluate various similarity functions using the STS Benchmark dataset, analyzing their performance and stability. Additionally, we introduce a multi-objective approach for optimal threshold selection. Our findings provide insights into the effectiveness of different similarity functions and offer a straightforward method for threshold selection optimization, contributing to the advancement of Siamese network architectures in STS applications. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=siamese%20networks" title="siamese networks">siamese networks</a>, <a href="https://publications.waset.org/abstracts/search?q=semantic%20textual%20similarity" title=" semantic textual similarity"> semantic textual similarity</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20functions" title=" similarity functions"> similarity functions</a>, <a href="https://publications.waset.org/abstracts/search?q=STS%20benchmark%20dataset" title=" STS benchmark dataset"> STS benchmark dataset</a>, <a href="https://publications.waset.org/abstracts/search?q=threshold%20selection" title=" threshold selection"> threshold selection</a> </p> <a href="https://publications.waset.org/abstracts/187407/multi-objective-optimal-threshold-selection-for-similarity-functions-in-siamese-networks-for-semantic-textual-similarity-tasks" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/187407.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">37</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3346</span> Similarity Based Retrieval in Case Based Reasoning for Analysis of Medical Images</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=M.%20Dasgupta">M. Dasgupta</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20Banerjee"> S. Banerjee</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Content Based Image Retrieval (CBIR) coupled with Case Based Reasoning (CBR) is a paradigm that is becoming increasingly popular in the diagnosis and therapy planning of medical ailments utilizing the digital content of medical images. This paper presents a survey of some of the promising approaches used in the detection of abnormalities in retina images as well in mammographic screening and detection of regions of interest in MRI scans of the brain. We also describe our proposed algorithm to detect hard exudates in fundus images of the retina of Diabetic Retinopathy patients. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=case%20based%20reasoning" title="case based reasoning">case based reasoning</a>, <a href="https://publications.waset.org/abstracts/search?q=exudates" title=" exudates"> exudates</a>, <a href="https://publications.waset.org/abstracts/search?q=retina%20image" title=" retina image"> retina image</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20based%20retrieval" title=" similarity based retrieval"> similarity based retrieval</a> </p> <a href="https://publications.waset.org/abstracts/2992/similarity-based-retrieval-in-case-based-reasoning-for-analysis-of-medical-images" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/2992.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">348</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3345</span> High-Capacity Image Steganography using Wavelet-based Fusion on Deep Convolutional Neural Networks</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Amal%20Khalifa">Amal Khalifa</a>, <a href="https://publications.waset.org/abstracts/search?q=Nicolas%20Vana%20Santos"> Nicolas Vana Santos</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Steganography has been known for centuries as an efficient approach for covert communication. Due to its popularity and ease of access, image steganography has attracted researchers to find secure techniques for hiding information within an innocent looking cover image. In this research, we propose a novel deep-learning approach to digital image steganography. The proposed method, DeepWaveletFusion, uses convolutional neural networks (CNN) to hide a secret image into a cover image of the same size. Two CNNs are trained back-to-back to merge the Discrete Wavelet Transform (DWT) of both colored images and eventually be able to blindly extract the hidden image. Based on two different image similarity metrics, a weighted gain function is used to guide the learning process and maximize the quality of the retrieved secret image and yet maintaining acceptable imperceptibility. Experimental results verified the high recoverability of DeepWaveletFusion which outperformed similar deep-learning-based methods. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=deep%20learning" title="deep learning">deep learning</a>, <a href="https://publications.waset.org/abstracts/search?q=steganography" title=" steganography"> steganography</a>, <a href="https://publications.waset.org/abstracts/search?q=image" title=" image"> image</a>, <a href="https://publications.waset.org/abstracts/search?q=discrete%20wavelet%20transform" title=" discrete wavelet transform"> discrete wavelet transform</a>, <a href="https://publications.waset.org/abstracts/search?q=fusion" title=" fusion"> fusion</a> </p> <a href="https://publications.waset.org/abstracts/170293/high-capacity-image-steganography-using-wavelet-based-fusion-on-deep-convolutional-neural-networks" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/170293.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">90</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3344</span> Robust Image Registration Based on an Adaptive Normalized Mutual Information Metric</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Huda%20Algharib">Huda Algharib</a>, <a href="https://publications.waset.org/abstracts/search?q=Amal%20Algharib"> Amal Algharib</a>, <a href="https://publications.waset.org/abstracts/search?q=Hanan%20Algharib"> Hanan Algharib</a>, <a href="https://publications.waset.org/abstracts/search?q=Ali%20Mohammad%20Alqudah"> Ali Mohammad Alqudah</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Image registration is an important topic for many imaging systems and computer vision applications. The standard image registration techniques such as Mutual information/ Normalized mutual information -based methods have a limited performance because they do not consider the spatial information or the relationships between the neighbouring pixels or voxels. In addition, the amount of image noise may significantly affect the registration accuracy. Therefore, this paper proposes an efficient method that explicitly considers the relationships between the adjacent pixels, where the gradient information of the reference and scene images is extracted first, and then the cosine similarity of the extracted gradient information is computed and used to improve the accuracy of the standard normalized mutual information measure. Our experimental results on different data types (i.e. CT, MRI and thermal images) show that the proposed method outperforms a number of image registration techniques in terms of the accuracy. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=image%20registration" title="image registration">image registration</a>, <a href="https://publications.waset.org/abstracts/search?q=mutual%20information" title=" mutual information"> mutual information</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20gradients" title=" image gradients"> image gradients</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20transformations" title=" image transformations"> image transformations</a> </p> <a href="https://publications.waset.org/abstracts/82815/robust-image-registration-based-on-an-adaptive-normalized-mutual-information-metric" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/82815.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">248</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3343</span> Lifting Wavelet Transform and Singular Values Decomposition for Secure Image Watermarking</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Siraa%20Ben%20Ftima">Siraa Ben Ftima</a>, <a href="https://publications.waset.org/abstracts/search?q=Mourad%20Talbi"> Mourad Talbi</a>, <a href="https://publications.waset.org/abstracts/search?q=Tahar%20Ezzedine"> Tahar Ezzedine</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, we present a technique of secure watermarking of grayscale and color images. This technique consists in applying the Singular Value Decomposition (SVD) in LWT (Lifting Wavelet Transform) domain in order to insert the watermark image (grayscale) in the host image (grayscale or color image). It also uses signature in the embedding and extraction steps. The technique is applied on a number of grayscale and color images. The performance of this technique is proved by the PSNR (Pick Signal to Noise Ratio), the MSE (Mean Square Error) and the SSIM (structural similarity) computations. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=lifting%20wavelet%20transform%20%28LWT%29" title="lifting wavelet transform (LWT)">lifting wavelet transform (LWT)</a>, <a href="https://publications.waset.org/abstracts/search?q=sub-space%20vectorial%20decomposition" title=" sub-space vectorial decomposition"> sub-space vectorial decomposition</a>, <a href="https://publications.waset.org/abstracts/search?q=secure" title=" secure"> secure</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20watermarking" title=" image watermarking"> image watermarking</a>, <a href="https://publications.waset.org/abstracts/search?q=watermark" title=" watermark"> watermark</a> </p> <a href="https://publications.waset.org/abstracts/70998/lifting-wavelet-transform-and-singular-values-decomposition-for-secure-image-watermarking" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/70998.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">276</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3342</span> Approximately Similarity Measurement of Web Sites Using Genetic Algorithms and Binary Trees</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Doru%20Anastasiu%20Popescu">Doru Anastasiu Popescu</a>, <a href="https://publications.waset.org/abstracts/search?q=Dan%20R%C4%83dulescu"> Dan R膬dulescu</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, we determine the similarity of two HTML web applications. We are going to use a genetic algorithm in order to determine the most significant web pages of each application (we are not going to use every web page of a site). Using these significant web pages, we will find the similarity value between the two applications. The algorithm is going to be efficient because we are going to use a reduced number of web pages for comparisons but it will return an approximate value of the similarity. The binary trees are used to keep the tags from the significant pages. The algorithm was implemented in Java language. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Tag" title="Tag">Tag</a>, <a href="https://publications.waset.org/abstracts/search?q=HTML" title=" HTML"> HTML</a>, <a href="https://publications.waset.org/abstracts/search?q=web%20page" title=" web page"> web page</a>, <a href="https://publications.waset.org/abstracts/search?q=genetic%20algorithm" title=" genetic algorithm"> genetic algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20value" title=" similarity value"> similarity value</a>, <a href="https://publications.waset.org/abstracts/search?q=binary%20tree" title=" binary tree"> binary tree</a> </p> <a href="https://publications.waset.org/abstracts/50460/approximately-similarity-measurement-of-web-sites-using-genetic-algorithms-and-binary-trees" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/50460.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">355</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3341</span> Measuring Text-Based Semantics Relatedness Using WordNet</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Madiha%20Khan">Madiha Khan</a>, <a href="https://publications.waset.org/abstracts/search?q=Sidrah%20Ramzan"> Sidrah Ramzan</a>, <a href="https://publications.waset.org/abstracts/search?q=Seemab%20Khan"> Seemab Khan</a>, <a href="https://publications.waset.org/abstracts/search?q=Shahzad%20Hassan"> Shahzad Hassan</a>, <a href="https://publications.waset.org/abstracts/search?q=Kamran%20Saeed"> Kamran Saeed</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Measuring semantic similarity between texts is calculating semantic relatedness between texts using various techniques. Our web application (Measuring Relatedness of Concepts-MRC) allows user to input two text corpuses and get semantic similarity percentage between both using WordNet. Our application goes through five stages for the computation of semantic relatedness. Those stages are: Preprocessing (extracts keywords from content), Feature Extraction (classification of words into Parts-of-Speech), Synonyms Extraction (retrieves synonyms against each keyword), Measuring Similarity (using keywords and synonyms, similarity is measured) and Visualization (graphical representation of similarity measure). Hence the user can measure similarity on basis of features as well. The end result is a percentage score and the word(s) which form the basis of similarity between both texts with use of different tools on same platform. In future work we look forward for a Web as a live corpus application that provides a simpler and user friendly tool to compare documents and extract useful information. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Graphviz%20representation" title="Graphviz representation">Graphviz representation</a>, <a href="https://publications.waset.org/abstracts/search?q=semantic%20relatedness" title=" semantic relatedness"> semantic relatedness</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20measurement" title=" similarity measurement"> similarity measurement</a>, <a href="https://publications.waset.org/abstracts/search?q=WordNet%20similarity" title=" WordNet similarity"> WordNet similarity</a> </p> <a href="https://publications.waset.org/abstracts/95106/measuring-text-based-semantics-relatedness-using-wordnet" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/95106.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">237</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3340</span> A Hybrid Watermarking Scheme Using Discrete and Discrete Stationary Wavelet Transformation For Color Images</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=B%C3%BClent%20Kantar">B眉lent Kantar</a>, <a href="https://publications.waset.org/abstracts/search?q=Numan%20%C3%9Cnald%C4%B1"> Numan 脺nald谋</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents a new method which includes robust and invisible digital watermarking on images that is colored. Colored images are used as watermark. Frequency region is used for digital watermarking. Discrete wavelet transform and discrete stationary wavelet transform are used for frequency region transformation. Low, medium and high frequency coefficients are obtained by applying the two-level discrete wavelet transform to the original image. Low frequency coefficients are obtained by applying one level discrete stationary wavelet transform separately to all frequency coefficient of the two-level discrete wavelet transformation of the original image. For every low frequency coefficient obtained from one level discrete stationary wavelet transformation, watermarks are added. Watermarks are added to all frequency coefficients of two-level discrete wavelet transform. Totally, four watermarks are added to original image. In order to get back the watermark, the original and watermarked images are applied with two-level discrete wavelet transform and one level discrete stationary wavelet transform. The watermark is obtained from difference of the discrete stationary wavelet transform of the low frequency coefficients. A total of four watermarks are obtained from all frequency of two-level discrete wavelet transform. Obtained watermark results are compared with real watermark results, and a similarity result is obtained. A watermark is obtained from the highest similarity values. Proposed methods of watermarking are tested against attacks of the geometric and image processing. The results show that proposed watermarking method is robust and invisible. All features of frequencies of two level discrete wavelet transform watermarking are combined to get back the watermark from the watermarked image. Watermarks have been added to the image by converting the binary image. These operations provide us with better results in getting back the watermark from watermarked image by attacking of the geometric and image processing. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=watermarking" title="watermarking">watermarking</a>, <a href="https://publications.waset.org/abstracts/search?q=DWT" title=" DWT"> DWT</a>, <a href="https://publications.waset.org/abstracts/search?q=DSWT" title=" DSWT"> DSWT</a>, <a href="https://publications.waset.org/abstracts/search?q=copy%20right%20protection" title=" copy right protection"> copy right protection</a>, <a href="https://publications.waset.org/abstracts/search?q=RGB" title=" RGB "> RGB </a> </p> <a href="https://publications.waset.org/abstracts/16927/a-hybrid-watermarking-scheme-using-discrete-and-discrete-stationary-wavelet-transformation-for-color-images" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/16927.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">535</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3339</span> Plagiarism Detection for Flowchart and Figures in Texts</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ahmadu%20Maidorawa">Ahmadu Maidorawa</a>, <a href="https://publications.waset.org/abstracts/search?q=Idrissa%20Djibo"> Idrissa Djibo</a>, <a href="https://publications.waset.org/abstracts/search?q=Muhammad%20Tella"> Muhammad Tella </a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents a method for detecting flow chart and figure plagiarism based on shape of image processing and multimedia retrieval. The method managed to retrieve flowcharts with ranked similarity according to different matching sets. Plagiarism detection is well known phenomenon in the academic arena. Copying other people is considered as serious offense that needs to be checked. There are many plagiarism detection systems such as turn-it-in that has been developed to provide these checks. Most, if not all, discard the figures and charts before checking for plagiarism. Discarding the figures and charts result in look holes that people can take advantage. That means people can plagiarize figures and charts easily without the current plagiarism systems detecting it. There are very few papers which talks about flowcharts plagiarism detection. Therefore, there is a need to develop a system that will detect plagiarism in figures and charts. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=flowchart" title="flowchart">flowchart</a>, <a href="https://publications.waset.org/abstracts/search?q=multimedia%20retrieval" title=" multimedia retrieval"> multimedia retrieval</a>, <a href="https://publications.waset.org/abstracts/search?q=figures%20similarity" title=" figures similarity"> figures similarity</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20comparison" title=" image comparison"> image comparison</a>, <a href="https://publications.waset.org/abstracts/search?q=figure%20retrieval" title=" figure retrieval"> figure retrieval</a> </p> <a href="https://publications.waset.org/abstracts/32548/plagiarism-detection-for-flowchart-and-figures-in-texts" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/32548.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">464</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3338</span> Review and Suggestions of the Similarity between Employee and Its Workplace</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Gi%20Ryung%20Song">Gi Ryung Song</a>, <a href="https://publications.waset.org/abstracts/search?q=Kyoung%20Seok%20Kim"> Kyoung Seok Kim</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This study reviewed the literature that focused on similarity of various characteristics such as values, personality, or demographics between employee and other elements in its organization for example employee with leader, job, and organization. We divided a body of this study into two parts and organized and demonstrated recent studies in first part. Three issues appeared in this part, which are statistical ways of measuring similarity, supervisor-subordinate similarity, and person-organization fit with person-job fit. In the latter part, based on the three issues of recent studies, we suggested three propositions about points that the recent studies missed or the studies did not orient. First proposition argued about the direction of similarity, which could also be interpreted as there is causal relation between employee and its workplace environments. Second, we suggested a consideration of eliminating common variance buried in one鈥檚 characteristics or its profiles. Third proposition was about the similarity of extra role behavior between individual and organization, and we treated this organization鈥檚 level of extra role behavior as a kind of its culture. In doing so, similarity of individual鈥檚 extra role behavior and organization鈥檚 has the meaning that individual鈥檚 congruence against their organization culture. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=similarity" title="similarity">similarity</a>, <a href="https://publications.waset.org/abstracts/search?q=person-organization%20fit" title=" person-organization fit"> person-organization fit</a>, <a href="https://publications.waset.org/abstracts/search?q=supervisor-subordinate%20similarity" title=" supervisor-subordinate similarity"> supervisor-subordinate similarity</a>, <a href="https://publications.waset.org/abstracts/search?q=literature%20review" title=" literature review"> literature review</a> </p> <a href="https://publications.waset.org/abstracts/54492/review-and-suggestions-of-the-similarity-between-employee-and-its-workplace" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/54492.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">283</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3337</span> Medical Image Classification Using Legendre Multifractal Spectrum Features</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=R.%20Korchiyne">R. Korchiyne</a>, <a href="https://publications.waset.org/abstracts/search?q=A.%20Sbihi"> A. Sbihi</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20M.%20Farssi"> S. M. Farssi</a>, <a href="https://publications.waset.org/abstracts/search?q=R.%20Touahni"> R. Touahni</a>, <a href="https://publications.waset.org/abstracts/search?q=M.%20Tahiri%20Alaoui"> M. Tahiri Alaoui</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Trabecular bone structure is important texture in the study of osteoporosis. Legendre multifractal spectrum can reflect the complex and self-similarity characteristic of structures. The main objective of this paper is to develop a new technique of medical image classification based on Legendre multifractal spectrum. Novel features have been developed from basic geometrical properties of this spectrum in a supervised image classification. The proposed method has been successfully used to classify medical images of bone trabeculations, and could be a useful supplement to the clinical observations for osteoporosis diagnosis. A comparative study with existing data reveals that the results of this approach are concordant. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=multifractal%20analysis" title="multifractal analysis">multifractal analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=medical%20image" title=" medical image"> medical image</a>, <a href="https://publications.waset.org/abstracts/search?q=osteoporosis" title=" osteoporosis"> osteoporosis</a>, <a href="https://publications.waset.org/abstracts/search?q=fractal%20dimension" title=" fractal dimension"> fractal dimension</a>, <a href="https://publications.waset.org/abstracts/search?q=Legendre%20spectrum" title=" Legendre spectrum"> Legendre spectrum</a>, <a href="https://publications.waset.org/abstracts/search?q=supervised%20classification" title=" supervised classification"> supervised classification</a> </p> <a href="https://publications.waset.org/abstracts/15795/medical-image-classification-using-legendre-multifractal-spectrum-features" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/15795.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">514</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3336</span> 2D Fingerprint Performance for PubChem Chemical Database</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Fatimah%20Zawani%20Abdullah">Fatimah Zawani Abdullah</a>, <a href="https://publications.waset.org/abstracts/search?q=Shereena%20Mohd%20Arif"> Shereena Mohd Arif</a>, <a href="https://publications.waset.org/abstracts/search?q=Nurul%20Malim"> Nurul Malim</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The study of molecular similarity search in chemical database is increasingly widespread, especially in the area of drug discovery. Similarity search is an application in the field of Chemoinformatics to measure the similarity between the molecular structure which is known as the query and the structure of chemical compounds in the database. Similarity search is also one of the approaches in virtual screening which involves computational techniques and scoring the probabilities of activity. The main objective of this work is to determine the best fingerprint when compared to the other five fingerprints selected in this study using PubChem chemical dataset. This paper will discuss the similarity searching process conducted using 6 types of descriptors, which are ECFP4, ECFC4, FCFP4, FCFC4, SRECFC4 and SRFCFC4 on 15 activity classes of PubChem dataset using Tanimoto coefficient to calculate the similarity between the query structures and each of the database structure. The results suggest that ECFP4 performs the best to be used with Tanimoto coefficient in the PubChem dataset. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=2D%20fingerprints" title="2D fingerprints">2D fingerprints</a>, <a href="https://publications.waset.org/abstracts/search?q=Tanimoto" title=" Tanimoto"> Tanimoto</a>, <a href="https://publications.waset.org/abstracts/search?q=PubChem" title=" PubChem"> PubChem</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20searching" title=" similarity searching"> similarity searching</a>, <a href="https://publications.waset.org/abstracts/search?q=chemoinformatics" title=" chemoinformatics"> chemoinformatics</a> </p> <a href="https://publications.waset.org/abstracts/15097/2d-fingerprint-performance-for-pubchem-chemical-database" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/15097.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">293</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3335</span> Study on the Self-Location Estimate by the Evolutional Triangle Similarity Matching Using Artificial Bee Colony Algorithm</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Yuji%20Kageyama">Yuji Kageyama</a>, <a href="https://publications.waset.org/abstracts/search?q=Shin%20Nagata"> Shin Nagata</a>, <a href="https://publications.waset.org/abstracts/search?q=Tatsuya%20Takino"> Tatsuya Takino</a>, <a href="https://publications.waset.org/abstracts/search?q=Izuru%20Nomura"> Izuru Nomura</a>, <a href="https://publications.waset.org/abstracts/search?q=Hiroyuki%20Kamata"> Hiroyuki Kamata</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In previous study, technique to estimate a self-location by using a lunar image is proposed. We consider the improvement of the conventional method in consideration of FPGA implementation in this paper. Specifically, we introduce Artificial Bee Colony algorithm for reduction of search time. In addition, we use fixed point arithmetic to enable high-speed operation on FPGA. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=SLIM" title="SLIM">SLIM</a>, <a href="https://publications.waset.org/abstracts/search?q=Artificial%20Bee%20Colony%20Algorithm" title=" Artificial Bee Colony Algorithm"> Artificial Bee Colony Algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=location%20estimate" title=" location estimate"> location estimate</a>, <a href="https://publications.waset.org/abstracts/search?q=evolutional%20triangle%20similarity" title=" evolutional triangle similarity"> evolutional triangle similarity</a> </p> <a href="https://publications.waset.org/abstracts/19303/study-on-the-self-location-estimate-by-the-evolutional-triangle-similarity-matching-using-artificial-bee-colony-algorithm" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/19303.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">518</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3334</span> Image Features Comparison-Based Position Estimation Method Using a Camera Sensor</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Jinseon%20Song">Jinseon Song</a>, <a href="https://publications.waset.org/abstracts/search?q=Yongwan%20Park"> Yongwan Park</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, propose method that can user’s position that based on database is built from single camera. Previous positioning calculate distance by arrival-time of signal like GPS (Global Positioning System), RF(Radio Frequency). However, these previous method have weakness because these have large error range according to signal interference. Method for solution estimate position by camera sensor. But, signal camera is difficult to obtain relative position data and stereo camera is difficult to provide real-time position data because of a lot of image data, too. First of all, in this research we build image database at space that able to provide positioning service with single camera. Next, we judge similarity through image matching of database image and transmission image from user. Finally, we decide position of user through position of most similar database image. For verification of propose method, we experiment at real-environment like indoor and outdoor. Propose method is wide positioning range and this method can verify not only position of user but also direction. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=positioning" title="positioning">positioning</a>, <a href="https://publications.waset.org/abstracts/search?q=distance" title=" distance"> distance</a>, <a href="https://publications.waset.org/abstracts/search?q=camera" title=" camera"> camera</a>, <a href="https://publications.waset.org/abstracts/search?q=features" title=" features"> features</a>, <a href="https://publications.waset.org/abstracts/search?q=SURF%28Speed-Up%20Robust%20Features%29" title=" SURF(Speed-Up Robust Features)"> SURF(Speed-Up Robust Features)</a>, <a href="https://publications.waset.org/abstracts/search?q=database" title=" database"> database</a>, <a href="https://publications.waset.org/abstracts/search?q=estimation" title=" estimation"> estimation</a> </p> <a href="https://publications.waset.org/abstracts/11844/image-features-comparison-based-position-estimation-method-using-a-camera-sensor" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/11844.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">349</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3333</span> A Robust Spatial Feature Extraction Method for Facial Expression Recognition</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=H.%20G.%20C.%20P.%20Dinesh">H. G. C. P. Dinesh</a>, <a href="https://publications.waset.org/abstracts/search?q=G.%20Tharshini"> G. Tharshini</a>, <a href="https://publications.waset.org/abstracts/search?q=M.%20P.%20B.%20Ekanayake"> M. P. B. Ekanayake</a>, <a href="https://publications.waset.org/abstracts/search?q=G.%20M.%20R.%20I.%20Godaliyadda"> G. M. R. I. Godaliyadda</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents a new spatial feature extraction method based on principle component analysis (PCA) and Fisher Discernment Analysis (FDA) for facial expression recognition. It not only extracts reliable features for classification, but also reduces the feature space dimensions of pattern samples. In this method, first each gray scale image is considered in its entirety as the measurement matrix. Then, principle components (PCs) of row vectors of this matrix and variance of these row vectors along PCs are estimated. Therefore, this method would ensure the preservation of spatial information of the facial image. Afterwards, by incorporating the spectral information of the eigen-filters derived from the PCs, a feature vector was constructed, for a given image. Finally, FDA was used to define a set of basis in a reduced dimension subspace such that the optimal clustering is achieved. The method of FDA defines an inter-class scatter matrix and intra-class scatter matrix to enhance the compactness of each cluster while maximizing the distance between cluster marginal points. In order to matching the test image with the training set, a cosine similarity based Bayesian classification was used. The proposed method was tested on the Cohn-Kanade database and JAFFE database. It was observed that the proposed method which incorporates spatial information to construct an optimal feature space outperforms the standard PCA and FDA based methods. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=facial%20expression%20recognition" title="facial expression recognition">facial expression recognition</a>, <a href="https://publications.waset.org/abstracts/search?q=principle%20component%20analysis%20%28PCA%29" title=" principle component analysis (PCA)"> principle component analysis (PCA)</a>, <a href="https://publications.waset.org/abstracts/search?q=fisher%20discernment%20analysis%20%28FDA%29" title=" fisher discernment analysis (FDA)"> fisher discernment analysis (FDA)</a>, <a href="https://publications.waset.org/abstracts/search?q=eigen-filter" title=" eigen-filter"> eigen-filter</a>, <a href="https://publications.waset.org/abstracts/search?q=cosine%20similarity" title=" cosine similarity"> cosine similarity</a>, <a href="https://publications.waset.org/abstracts/search?q=bayesian%20classifier" title=" bayesian classifier"> bayesian classifier</a>, <a href="https://publications.waset.org/abstracts/search?q=f-measure" title=" f-measure"> f-measure</a> </p> <a href="https://publications.waset.org/abstracts/36459/a-robust-spatial-feature-extraction-method-for-facial-expression-recognition" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/36459.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">425</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3332</span> Hit-Or-Miss Transform as a Tool for Similar Shape Detection</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Osama%20Mohamed%20Elrajubi">Osama Mohamed Elrajubi</a>, <a href="https://publications.waset.org/abstracts/search?q=Idris%20El-Feghi"> Idris El-Feghi</a>, <a href="https://publications.waset.org/abstracts/search?q=Mohamed%20Abu%20Baker%20Saghayer"> Mohamed Abu Baker Saghayer</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper describes an identification of specific shapes within binary images using the morphological Hit-or-Miss Transform (HMT). Hit-or-Miss transform is a general binary morphological operation that can be used in searching of particular patterns of foreground and background pixels in an image. It is actually a basic operation of binary morphology since almost all other binary morphological operators are derived from it. The input of this method is a binary image and a structuring element (a template which will be searched in a binary image) while the output is another binary image. In this paper a modification of Hit-or-Miss transform has been proposed. The accuracy of algorithm is adjusted according to the similarity of the template and the sought template. The implementation of this method has been done by C language. The algorithm has been tested on several images and the results have shown that this new method can be used for similar shape detection. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=hit-or-miss%20operator%20transform" title="hit-or-miss operator transform">hit-or-miss operator transform</a>, <a href="https://publications.waset.org/abstracts/search?q=HMT" title=" HMT"> HMT</a>, <a href="https://publications.waset.org/abstracts/search?q=binary%20morphological%20operation" title=" binary morphological operation"> binary morphological operation</a>, <a href="https://publications.waset.org/abstracts/search?q=shape%20detection" title=" shape detection"> shape detection</a>, <a href="https://publications.waset.org/abstracts/search?q=binary%20images%20processing" title=" binary images processing"> binary images processing</a> </p> <a href="https://publications.waset.org/abstracts/11881/hit-or-miss-transform-as-a-tool-for-similar-shape-detection" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/11881.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">331</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3331</span> Similarity Based Membership of Elements to Uncertain Concept in Information System</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=M.%20Kamel%20El-Sayed">M. Kamel El-Sayed</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The process of determining the degree of membership for an element to an uncertain concept has been found in many ways, using equivalence and symmetry relations in information systems. In the case of similarity, these methods did not take into account the degree of symmetry between elements. In this paper, we use a new definition for finding the membership based on the degree of symmetry. We provide an example to clarify the suggested methods and compare it with previous methods. This method opens the door to more accurate decisions in information systems. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=information%20system" title="information system">information system</a>, <a href="https://publications.waset.org/abstracts/search?q=uncertain%20concept" title=" uncertain concept"> uncertain concept</a>, <a href="https://publications.waset.org/abstracts/search?q=membership%20function" title=" membership function"> membership function</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20relation" title=" similarity relation"> similarity relation</a>, <a href="https://publications.waset.org/abstracts/search?q=degree%20of%20similarity" title=" degree of similarity"> degree of similarity</a> </p> <a href="https://publications.waset.org/abstracts/88086/similarity-based-membership-of-elements-to-uncertain-concept-in-information-system" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/88086.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">223</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3330</span> Design and Implementation of Image Super-Resolution for Myocardial Image</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=M.%20V.%20Chidananda%20Murthy">M. V. Chidananda Murthy</a>, <a href="https://publications.waset.org/abstracts/search?q=M.%20Z.%20Kurian"> M. Z. Kurian</a>, <a href="https://publications.waset.org/abstracts/search?q=H.%20S.%20Guruprasad"> H. S. Guruprasad</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Super-resolution is the technique of intelligently upscaling images, avoiding artifacts or blurring, and deals with the recovery of a high-resolution image from one or more low-resolution images. Single-image super-resolution is a process of obtaining a high-resolution image from a set of low-resolution observations by signal processing. While super-resolution has been demonstrated to improve image quality in scaled down images in the image domain, its effects on the Fourier-based technique remains unknown. Super-resolution substantially improved the spatial resolution of the patient LGE images by sharpening the edges of the heart and the scar. This paper aims at investigating the effects of single image super-resolution on Fourier-based and image based methods of scale-up. In this paper, first, generate a training phase of the low-resolution image and high-resolution image to obtain dictionary. In the test phase, first, generate a patch and then difference of high-resolution image and interpolation image from the low-resolution image. Next simulation of the image is obtained by applying convolution method to the dictionary creation image and patch extracted the image. Finally, super-resolution image is obtained by combining the fused image and difference of high-resolution and interpolated image. Super-resolution reduces image errors and improves the image quality. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=image%20dictionary%20creation" title="image dictionary creation">image dictionary creation</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20super-resolution" title=" image super-resolution"> image super-resolution</a>, <a href="https://publications.waset.org/abstracts/search?q=LGE%20images" title=" LGE images"> LGE images</a>, <a href="https://publications.waset.org/abstracts/search?q=patch%20extraction" title=" patch extraction"> patch extraction</a> </p> <a href="https://publications.waset.org/abstracts/59494/design-and-implementation-of-image-super-resolution-for-myocardial-image" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/59494.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">375</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3329</span> Agglomerative Hierarchical Clustering Using the T胃 Family of Similarity Measures</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Salima%20Kouici">Salima Kouici</a>, <a href="https://publications.waset.org/abstracts/search?q=Abdelkader%20Khelladi"> Abdelkader Khelladi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this work, we begin with the presentation of the T胃 family of usual similarity measures concerning multidimensional binary data. Subsequently, some properties of these measures are proposed. Finally, the impact of the use of different inter-elements measures on the results of the Agglomerative Hierarchical Clustering Methods is studied. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=binary%20data" title="binary data">binary data</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20measure" title=" similarity measure"> similarity measure</a>, <a href="https://publications.waset.org/abstracts/search?q=T%CE%B8%20measures" title=" T胃 measures"> T胃 measures</a>, <a href="https://publications.waset.org/abstracts/search?q=agglomerative%20hierarchical%20clustering" title=" agglomerative hierarchical clustering"> agglomerative hierarchical clustering</a> </p> <a href="https://publications.waset.org/abstracts/13108/agglomerative-hierarchical-clustering-using-the-tth-family-of-similarity-measures" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/13108.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">481</span> </span> </div> </div> <ul class="pagination"> <li class="page-item disabled"><span class="page-link">‹</span></li> <li class="page-item active"><span class="page-link">1</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=2">2</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=3">3</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=4">4</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=5">5</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=6">6</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=7">7</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=8">8</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=9">9</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=10">10</a></li> <li class="page-item disabled"><span class="page-link">...</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=111">111</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=112">112</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=image%20similarity&page=2" rel="next">›</a></li> </ul> </div> </main> <footer> <div id="infolinks" class="pt-3 pb-2"> <div class="container"> <div style="background-color:#f5f5f5;" class="p-3"> <div class="row"> <div class="col-md-2"> <ul class="list-unstyled"> About <li><a href="https://waset.org/page/support">About Us</a></li> <li><a href="https://waset.org/page/support#legal-information">Legal</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/WASET-16th-foundational-anniversary.pdf">WASET celebrates its 16th foundational anniversary</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Account <li><a href="https://waset.org/profile">My Account</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Explore <li><a href="https://waset.org/disciplines">Disciplines</a></li> <li><a href="https://waset.org/conferences">Conferences</a></li> <li><a href="https://waset.org/conference-programs">Conference Program</a></li> <li><a href="https://waset.org/committees">Committees</a></li> <li><a href="https://publications.waset.org">Publications</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Research <li><a href="https://publications.waset.org/abstracts">Abstracts</a></li> <li><a href="https://publications.waset.org">Periodicals</a></li> <li><a href="https://publications.waset.org/archive">Archive</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Open Science <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Philosophy.pdf">Open Science Philosophy</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Award.pdf">Open Science Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Society-Open-Science-and-Open-Innovation.pdf">Open Innovation</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Postdoctoral-Fellowship-Award.pdf">Postdoctoral Fellowship Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Scholarly-Research-Review.pdf">Scholarly Research Review</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Support <li><a href="https://waset.org/page/support">Support</a></li> <li><a href="https://waset.org/profile/messages/create">Contact Us</a></li> <li><a href="https://waset.org/profile/messages/create">Report Abuse</a></li> </ul> </div> </div> </div> </div> </div> <div class="container text-center"> <hr style="margin-top:0;margin-bottom:.3rem;"> <a href="https://creativecommons.org/licenses/by/4.0/" target="_blank" class="text-muted small">Creative Commons Attribution 4.0 International License</a> <div id="copy" class="mt-2">© 2024 World Academy of Science, Engineering and Technology</div> </div> </footer> <a href="javascript:" id="return-to-top"><i class="fas fa-arrow-up"></i></a> <div class="modal" id="modal-template"> <div class="modal-dialog"> <div class="modal-content"> <div class="row m-0 mt-1"> <div class="col-md-12"> <button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">×</span></button> </div> </div> <div class="modal-body"></div> </div> </div> </div> <script src="https://cdn.waset.org/static/plugins/jquery-3.3.1.min.js"></script> <script src="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/js/bootstrap.bundle.min.js"></script> <script src="https://cdn.waset.org/static/js/site.js?v=150220211556"></script> <script> jQuery(document).ready(function() { /*jQuery.get("https://publications.waset.org/xhr/user-menu", function (response) { jQuery('#mainNavMenu').append(response); });*/ jQuery.get({ url: "https://publications.waset.org/xhr/user-menu", cache: false }).then(function(response){ jQuery('#mainNavMenu').append(response); }); }); </script> </body> </html>