CINXE.COM

Search results for: density-based clustering

<!DOCTYPE html> <html lang="en" dir="ltr"> <head> <!-- Google tag (gtag.js) --> <script async src="https://www.googletagmanager.com/gtag/js?id=G-P63WKM1TM1"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('js', new Date()); gtag('config', 'G-P63WKM1TM1'); </script> <!-- Yandex.Metrika counter --> <script type="text/javascript" > (function(m,e,t,r,i,k,a){m[i]=m[i]||function(){(m[i].a=m[i].a||[]).push(arguments)}; m[i].l=1*new Date(); for (var j = 0; j < document.scripts.length; j++) {if (document.scripts[j].src === r) { return; }} k=e.createElement(t),a=e.getElementsByTagName(t)[0],k.async=1,k.src=r,a.parentNode.insertBefore(k,a)}) (window, document, "script", "https://mc.yandex.ru/metrika/tag.js", "ym"); ym(55165297, "init", { clickmap:false, trackLinks:true, accurateTrackBounce:true, webvisor:false }); </script> <noscript><div><img src="https://mc.yandex.ru/watch/55165297" style="position:absolute; left:-9999px;" alt="" /></div></noscript> <!-- /Yandex.Metrika counter --> <!-- Matomo --> <!-- End Matomo Code --> <title>Search results for: density-based clustering</title> <meta name="description" content="Search results for: density-based clustering"> <meta name="keywords" content="density-based clustering"> <meta name="viewport" content="width=device-width, initial-scale=1, minimum-scale=1, maximum-scale=1, user-scalable=no"> <meta charset="utf-8"> <link href="https://cdn.waset.org/favicon.ico" type="image/x-icon" rel="shortcut icon"> <link href="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/css/bootstrap.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/plugins/fontawesome/css/all.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/css/site.css?v=150220211555" rel="stylesheet"> </head> <body> <header> <div class="container"> <nav class="navbar navbar-expand-lg navbar-light"> <a class="navbar-brand" href="https://waset.org"> <img src="https://cdn.waset.org/static/images/wasetc.png" alt="Open Science Research Excellence" title="Open Science Research Excellence" /> </a> <button class="d-block d-lg-none navbar-toggler ml-auto" type="button" data-toggle="collapse" data-target="#navbarMenu" aria-controls="navbarMenu" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <div class="w-100"> <div class="d-none d-lg-flex flex-row-reverse"> <form method="get" action="https://waset.org/search" class="form-inline my-2 my-lg-0"> <input class="form-control mr-sm-2" type="search" placeholder="Search Conferences" value="density-based clustering" name="q" aria-label="Search"> <button class="btn btn-light my-2 my-sm-0" type="submit"><i class="fas fa-search"></i></button> </form> </div> <div class="collapse navbar-collapse mt-1" id="navbarMenu"> <ul class="navbar-nav ml-auto align-items-center" id="mainNavMenu"> <li class="nav-item"> <a class="nav-link" href="https://waset.org/conferences" title="Conferences in 2024/2025/2026">Conferences</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/disciplines" title="Disciplines">Disciplines</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/committees" rel="nofollow">Committees</a> </li> <li class="nav-item dropdown"> <a class="nav-link dropdown-toggle" href="#" id="navbarDropdownPublications" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Publications </a> <div class="dropdown-menu" aria-labelledby="navbarDropdownPublications"> <a class="dropdown-item" href="https://publications.waset.org/abstracts">Abstracts</a> <a class="dropdown-item" href="https://publications.waset.org">Periodicals</a> <a class="dropdown-item" href="https://publications.waset.org/archive">Archive</a> </div> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/page/support" title="Support">Support</a> </li> </ul> </div> </div> </nav> </div> </header> <main> <div class="container mt-4"> <div class="row"> <div class="col-md-9 mx-auto"> <form method="get" action="https://publications.waset.org/abstracts/search"> <div id="custom-search-input"> <div class="input-group"> <i class="fas fa-search"></i> <input type="text" class="search-query" name="q" placeholder="Author, Title, Abstract, Keywords" value="density-based clustering"> <input type="submit" class="btn_search" value="Search"> </div> </div> </form> </div> </div> <div class="row mt-3"> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Commenced</strong> in January 2007</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Frequency:</strong> Monthly</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Edition:</strong> International</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Paper Count:</strong> 609</div> </div> </div> </div> <h1 class="mt-3 mb-3 text-center" style="font-size:1.6rem;">Search results for: density-based clustering</h1> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">609</span> Mitigating the Negative Effect of Intrabrand Clustering: The Role of Interbrand Clustering and Firm Size</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Moeen%20Naseer%20Butt">Moeen Naseer Butt</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Clustering –geographic concentrations of entities– has recently received more attention in marketing research and has been shown to affect multiple outcomes. This study investigates the impact of intrabrand clustering (clustering of same-brand outlets) on an outlet’s quality performance. Further, it assesses the moderating effects of interbrand clustering (clustering of other-brand outlets) and firm size. An examination of approximately 21,000 food service establishments in New York State in 2019 finds that the impact of intrabrand clustering on an outlet’s quality performance is context-dependent. Specifically, intrabrand clustering decreases, whereas interbrand clustering and firm size help increase the outlet’s performance. Additionally, this study finds that the role of firm size is more substantial than interbrand clustering in mitigating the adverse effects of intrabrand clustering on outlet quality performance. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=intraband%20clustering" title="intraband clustering">intraband clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=interbrand%20clustering" title=" interbrand clustering"> interbrand clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=firm%20size" title=" firm size"> firm size</a>, <a href="https://publications.waset.org/abstracts/search?q=brand%20competition" title=" brand competition"> brand competition</a>, <a href="https://publications.waset.org/abstracts/search?q=outlet%20performance" title=" outlet performance"> outlet performance</a>, <a href="https://publications.waset.org/abstracts/search?q=quality%20violations" title=" quality violations"> quality violations</a> </p> <a href="https://publications.waset.org/abstracts/155035/mitigating-the-negative-effect-of-intrabrand-clustering-the-role-of-interbrand-clustering-and-firm-size" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/155035.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">188</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">608</span> Hierarchical Clustering Algorithms in Data Mining</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Z.%20Abdullah">Z. Abdullah</a>, <a href="https://publications.waset.org/abstracts/search?q=A.%20R.%20Hamdan"> A. R. Hamdan</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Clustering is a process of grouping objects and data into groups of clusters to ensure that data objects from the same cluster are identical to each other. Clustering algorithms in one of the areas in data mining and it can be classified into partition, hierarchical, density based, and grid-based. Therefore, in this paper, we do a survey and review for four major hierarchical clustering algorithms called CURE, ROCK, CHAMELEON, and BIRCH. The obtained state of the art of these algorithms will help in eliminating the current problems, as well as deriving more robust and scalable algorithms for clustering. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=clustering" title="clustering">clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=unsupervised%20learning" title=" unsupervised learning"> unsupervised learning</a>, <a href="https://publications.waset.org/abstracts/search?q=algorithms" title=" algorithms"> algorithms</a>, <a href="https://publications.waset.org/abstracts/search?q=hierarchical" title=" hierarchical"> hierarchical</a> </p> <a href="https://publications.waset.org/abstracts/31217/hierarchical-clustering-algorithms-in-data-mining" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/31217.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">885</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">607</span> Hybrid Hierarchical Clustering Approach for Community Detection in Social Network</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Radhia%20Toujani">Radhia Toujani</a>, <a href="https://publications.waset.org/abstracts/search?q=Jalel%20Akaichi"> Jalel Akaichi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Social Networks generally present a hierarchy of communities. To determine these communities and the relationship between them, detection algorithms should be applied. Most of the existing algorithms, proposed for hierarchical communities identification, are based on either agglomerative clustering or divisive clustering. In this paper, we present a hybrid hierarchical clustering approach for community detection based on both bottom-up and bottom-down clustering. Obviously, our approach provides more relevant community structure than hierarchical method which considers only divisive or agglomerative clustering to identify communities. Moreover, we performed some comparative experiments to enhance the quality of the clustering results and to show the effectiveness of our algorithm. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=agglomerative%20hierarchical%20clustering" title="agglomerative hierarchical clustering">agglomerative hierarchical clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=community%20structure" title=" community structure"> community structure</a>, <a href="https://publications.waset.org/abstracts/search?q=divisive%20hierarchical%20clustering" title=" divisive hierarchical clustering"> divisive hierarchical clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=hybrid%20hierarchical%20clustering" title=" hybrid hierarchical clustering"> hybrid hierarchical clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=opinion%20mining" title=" opinion mining"> opinion mining</a>, <a href="https://publications.waset.org/abstracts/search?q=social%20network" title=" social network"> social network</a>, <a href="https://publications.waset.org/abstracts/search?q=social%20network%20analysis" title=" social network analysis"> social network analysis</a> </p> <a href="https://publications.waset.org/abstracts/63702/hybrid-hierarchical-clustering-approach-for-community-detection-in-social-network" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/63702.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">365</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">606</span> Performance Analysis of Hierarchical Agglomerative Clustering in a Wireless Sensor Network Using Quantitative Data</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Tapan%20Jain">Tapan Jain</a>, <a href="https://publications.waset.org/abstracts/search?q=Davender%20Singh%20Saini"> Davender Singh Saini</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Clustering is a useful mechanism in wireless sensor networks which helps to cope with scalability and data transmission problems. The basic aim of our research work is to provide efficient clustering using Hierarchical agglomerative clustering (HAC). If the distance between the sensing nodes is calculated using their location then it’s quantitative HAC. This paper compares the various agglomerative clustering techniques applied in a wireless sensor network using the quantitative data. The simulations are done in MATLAB and the comparisons are made between the different protocols using dendrograms. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=routing" title="routing">routing</a>, <a href="https://publications.waset.org/abstracts/search?q=hierarchical%20clustering" title=" hierarchical clustering"> hierarchical clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=agglomerative" title=" agglomerative"> agglomerative</a>, <a href="https://publications.waset.org/abstracts/search?q=quantitative" title=" quantitative"> quantitative</a>, <a href="https://publications.waset.org/abstracts/search?q=wireless%20sensor%20network" title=" wireless sensor network"> wireless sensor network</a> </p> <a href="https://publications.waset.org/abstracts/3593/performance-analysis-of-hierarchical-agglomerative-clustering-in-a-wireless-sensor-network-using-quantitative-data" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/3593.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">615</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">605</span> Flowing Online Vehicle GPS Data Clustering Using a New Parallel K-Means Algorithm</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Orhun%20Vural">Orhun Vural</a>, <a href="https://publications.waset.org/abstracts/search?q=Oguz%20%20Bayat"> Oguz Bayat</a>, <a href="https://publications.waset.org/abstracts/search?q=Rustu%20Akay"> Rustu Akay</a>, <a href="https://publications.waset.org/abstracts/search?q=Osman%20N.%20Ucan"> Osman N. Ucan</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This study presents a new parallel approach clustering of GPS data. Evaluation has been made by comparing execution time of various clustering algorithms on GPS data. This paper aims to propose a parallel based on neighborhood K-means algorithm to make it faster. The proposed parallelization approach assumes that each GPS data represents a vehicle and to communicate between vehicles close to each other after vehicles are clustered. This parallelization approach has been examined on different sized continuously changing GPS data and compared with serial K-means algorithm and other serial clustering algorithms. The results demonstrated that proposed parallel K-means algorithm has been shown to work much faster than other clustering algorithms. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=parallel%20k-means%20algorithm" title="parallel k-means algorithm">parallel k-means algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=parallel%20clustering" title=" parallel clustering"> parallel clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=clustering%20algorithms" title=" clustering algorithms"> clustering algorithms</a>, <a href="https://publications.waset.org/abstracts/search?q=clustering%20on%20flowing%20data" title=" clustering on flowing data"> clustering on flowing data</a> </p> <a href="https://publications.waset.org/abstracts/86622/flowing-online-vehicle-gps-data-clustering-using-a-new-parallel-k-means-algorithm" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/86622.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">222</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">604</span> Semi-Supervised Hierarchical Clustering Given a Reference Tree of Labeled Documents</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ying%20Zhao">Ying Zhao</a>, <a href="https://publications.waset.org/abstracts/search?q=Xingyan%20Bin"> Xingyan Bin</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Semi-supervised clustering algorithms have been shown effective to improve clustering process with even limited supervision. However, semi-supervised hierarchical clustering remains challenging due to the complexities of expressing constraints for agglomerative clustering algorithms. This paper proposes novel semi-supervised agglomerative clustering algorithms to build a hierarchy based on a known reference tree. We prove that by enforcing distance constraints defined by a reference tree during the process of hierarchical clustering, the resultant tree is guaranteed to be consistent with the reference tree. We also propose a framework that allows the hierarchical tree generation be aware of levels of levels of the agglomerative tree under creation, so that metric weights can be learned and adopted at each level in a recursive fashion. The experimental evaluation shows that the additional cost of our contraint-based semi-supervised hierarchical clustering algorithm (HAC) is negligible, and our combined semi-supervised HAC algorithm outperforms the state-of-the-art algorithms on real-world datasets. The experiments also show that our proposed methods can improve clustering performance even with a small number of unevenly distributed labeled data. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=semi-supervised%20clustering" title="semi-supervised clustering">semi-supervised clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=hierarchical%0D%0Aagglomerative%20clustering" title=" hierarchical agglomerative clustering"> hierarchical agglomerative clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=reference%20trees" title=" reference trees"> reference trees</a>, <a href="https://publications.waset.org/abstracts/search?q=distance%20constraints" title=" distance constraints "> distance constraints </a> </p> <a href="https://publications.waset.org/abstracts/19478/semi-supervised-hierarchical-clustering-given-a-reference-tree-of-labeled-documents" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/19478.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">547</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">603</span> Fuzzy Optimization Multi-Objective Clustering Ensemble Model for Multi-Source Data Analysis</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=C.%20B.%20Le">C. B. Le</a>, <a href="https://publications.waset.org/abstracts/search?q=V.%20N.%20Pham"> V. N. Pham</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In modern data analysis, multi-source data appears more and more in real applications. Multi-source data clustering has emerged as a important issue in the data mining and machine learning community. Different data sources provide information about different data. Therefore, multi-source data linking is essential to improve clustering performance. However, in practice multi-source data is often heterogeneous, uncertain, and large. This issue is considered a major challenge from multi-source data. Ensemble is a versatile machine learning model in which learning techniques can work in parallel, with big data. Clustering ensemble has been shown to outperform any standard clustering algorithm in terms of accuracy and robustness. However, most of the traditional clustering ensemble approaches are based on single-objective function and single-source data. This paper proposes a new clustering ensemble method for multi-source data analysis. The fuzzy optimized multi-objective clustering ensemble method is called FOMOCE. Firstly, a clustering ensemble mathematical model based on the structure of multi-objective clustering function, multi-source data, and dark knowledge is introduced. Then, rules for extracting dark knowledge from the input data, clustering algorithms, and base clusterings are designed and applied. Finally, a clustering ensemble algorithm is proposed for multi-source data analysis. The experiments were performed on the standard sample data set. The experimental results demonstrate the superior performance of the FOMOCE method compared to the existing clustering ensemble methods and multi-source clustering methods. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=clustering%20ensemble" title="clustering ensemble">clustering ensemble</a>, <a href="https://publications.waset.org/abstracts/search?q=multi-source" title=" multi-source"> multi-source</a>, <a href="https://publications.waset.org/abstracts/search?q=multi-objective" title=" multi-objective"> multi-objective</a>, <a href="https://publications.waset.org/abstracts/search?q=fuzzy%20clustering" title=" fuzzy clustering"> fuzzy clustering</a> </p> <a href="https://publications.waset.org/abstracts/136598/fuzzy-optimization-multi-objective-clustering-ensemble-model-for-multi-source-data-analysis" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/136598.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">189</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">602</span> ACOPIN: An ACO Algorithm with TSP Approach for Clustering Proteins in Protein Interaction Networks</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Jamaludin%20Sallim">Jamaludin Sallim</a>, <a href="https://publications.waset.org/abstracts/search?q=Rozlina%20Mohamed"> Rozlina Mohamed</a>, <a href="https://publications.waset.org/abstracts/search?q=Roslina%20Abdul%20Hamid"> Roslina Abdul Hamid</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, we proposed an Ant Colony Optimization (ACO) algorithm together with Traveling Salesman Problem (TSP) approach to investigate the clustering problem in Protein Interaction Networks (PIN). We named this combination as ACOPIN. The purpose of this work is two-fold. First, to test the efficacy of ACO in clustering PIN and second, to propose the simple generalization of the ACO algorithm that might allow its application in clustering proteins in PIN. We split this paper to three main sections. First, we describe the PIN and clustering proteins in PIN. Second, we discuss the steps involved in each phase of ACO algorithm. Finally, we present some results of the investigation with the clustering patterns. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=ant%20colony%20optimization%20algorithm" title="ant colony optimization algorithm">ant colony optimization algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=searching%20algorithm" title=" searching algorithm"> searching algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=protein%20functional%20module" title=" protein functional module"> protein functional module</a>, <a href="https://publications.waset.org/abstracts/search?q=protein%20interaction%20network" title=" protein interaction network "> protein interaction network </a> </p> <a href="https://publications.waset.org/abstracts/22367/acopin-an-aco-algorithm-with-tsp-approach-for-clustering-proteins-in-protein-interaction-networks" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/22367.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">612</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">601</span> Spectral Clustering for Manufacturing Cell Formation</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Yessica%20Nataliani">Yessica Nataliani</a>, <a href="https://publications.waset.org/abstracts/search?q=Miin-Shen%20Yang"> Miin-Shen Yang</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Cell formation (CF) is an important step in group technology. It is used in designing cellular manufacturing systems using similarities between parts in relation to machines so that it can identify part families and machine groups. There are many CF methods in the literature, but there is less spectral clustering used in CF. In this paper, we propose a spectral clustering algorithm for machine-part CF. Some experimental examples are used to illustrate its efficiency. Overall, the spectral clustering algorithm can be used in CF with a wide variety of machine/part matrices. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=group%20technology" title="group technology">group technology</a>, <a href="https://publications.waset.org/abstracts/search?q=cell%20formation" title=" cell formation"> cell formation</a>, <a href="https://publications.waset.org/abstracts/search?q=spectral%20clustering" title=" spectral clustering"> spectral clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=grouping%20efficiency" title=" grouping efficiency"> grouping efficiency</a> </p> <a href="https://publications.waset.org/abstracts/72294/spectral-clustering-for-manufacturing-cell-formation" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/72294.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">408</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">600</span> Investigation of Clustering Algorithms Used in Wireless Sensor Networks</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Naim%20Karasekreter">Naim Karasekreter</a>, <a href="https://publications.waset.org/abstracts/search?q=Ugur%20Fidan"> Ugur Fidan</a>, <a href="https://publications.waset.org/abstracts/search?q=Fatih%20Basciftci"> Fatih Basciftci</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Wireless sensor networks are networks in which more than one sensor node is organized among themselves. The working principle is based on the transfer of the sensed data over the other nodes in the network to the central station. Wireless sensor networks concentrate on routing algorithms, energy efficiency and clustering algorithms. In the clustering method, the nodes in the network are divided into clusters using different parameters and the most suitable cluster head is selected from among them. The data to be sent to the center is sent per cluster, and the cluster head is transmitted to the center. With this method, the network traffic is reduced and the energy efficiency of the nodes is increased. In this study, clustering algorithms were examined in terms of clustering performances and cluster head selection characteristics to try to identify weak and strong sides. This work is supported by the Project 17.Kariyer.123 of Afyon Kocatepe University BAP Commission. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=wireless%20sensor%20networks%20%28WSN%29" title="wireless sensor networks (WSN)">wireless sensor networks (WSN)</a>, <a href="https://publications.waset.org/abstracts/search?q=clustering%20algorithm" title=" clustering algorithm"> clustering algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=cluster%20head" title=" cluster head"> cluster head</a>, <a href="https://publications.waset.org/abstracts/search?q=clustering" title=" clustering"> clustering</a> </p> <a href="https://publications.waset.org/abstracts/78846/investigation-of-clustering-algorithms-used-in-wireless-sensor-networks" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/78846.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">513</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">599</span> A Comparative Study of Multi-SOM Algorithms for Determining the Optimal Number of Clusters</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Im%C3%A8n%20Khanchouch">Imèn Khanchouch</a>, <a href="https://publications.waset.org/abstracts/search?q=Malika%20Charrad"> Malika Charrad</a>, <a href="https://publications.waset.org/abstracts/search?q=Mohamed%20Limam"> Mohamed Limam</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The interpretation of the quality of clusters and the determination of the optimal number of clusters is still a crucial problem in clustering. We focus in this paper on multi-SOM clustering method which overcomes the problem of extracting the number of clusters from the SOM map through the use of a clustering validity index. We then tested multi-SOM using real and artificial data sets with different evaluation criteria not used previously such as Davies Bouldin index, Dunn index and silhouette index. The developed multi-SOM algorithm is compared to k-means and Birch methods. Results show that it is more efficient than classical clustering methods. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=clustering" title="clustering">clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=SOM" title=" SOM"> SOM</a>, <a href="https://publications.waset.org/abstracts/search?q=multi-SOM" title=" multi-SOM"> multi-SOM</a>, <a href="https://publications.waset.org/abstracts/search?q=DB%20index" title=" DB index"> DB index</a>, <a href="https://publications.waset.org/abstracts/search?q=Dunn%20index" title=" Dunn index"> Dunn index</a>, <a href="https://publications.waset.org/abstracts/search?q=silhouette%20index" title=" silhouette index"> silhouette index</a> </p> <a href="https://publications.waset.org/abstracts/17422/a-comparative-study-of-multi-som-algorithms-for-determining-the-optimal-number-of-clusters" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/17422.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">599</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">598</span> A Fuzzy Kernel K-Medoids Algorithm for Clustering Uncertain Data Objects</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Behnam%20Tavakkol">Behnam Tavakkol</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Uncertain data mining algorithms use different ways to consider uncertainty in data such as by representing a data object as a sample of points or a probability distribution. Fuzzy methods have long been used for clustering traditional (certain) data objects. They are used to produce non-crisp cluster labels. For uncertain data, however, besides some uncertain fuzzy k-medoids algorithms, not many other fuzzy clustering methods have been developed. In this work, we develop a fuzzy kernel k-medoids algorithm for clustering uncertain data objects. The developed fuzzy kernel k-medoids algorithm is superior to existing fuzzy k-medoids algorithms in clustering data sets with non-linearly separable clusters. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=clustering%20algorithm" title="clustering algorithm">clustering algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=fuzzy%20methods" title=" fuzzy methods"> fuzzy methods</a>, <a href="https://publications.waset.org/abstracts/search?q=kernel%20k-medoids" title=" kernel k-medoids"> kernel k-medoids</a>, <a href="https://publications.waset.org/abstracts/search?q=uncertain%20data" title=" uncertain data"> uncertain data</a> </p> <a href="https://publications.waset.org/abstracts/123501/a-fuzzy-kernel-k-medoids-algorithm-for-clustering-uncertain-data-objects" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/123501.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">215</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">597</span> An Experimental Study on Some Conventional and Hybrid Models of Fuzzy Clustering</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Jeugert%20Kujtila">Jeugert Kujtila</a>, <a href="https://publications.waset.org/abstracts/search?q=Kristi%20Hoxhalli"> Kristi Hoxhalli</a>, <a href="https://publications.waset.org/abstracts/search?q=Ramazan%20Dalipi"> Ramazan Dalipi</a>, <a href="https://publications.waset.org/abstracts/search?q=Erjon%20Cota"> Erjon Cota</a>, <a href="https://publications.waset.org/abstracts/search?q=Ardit%20Murati"> Ardit Murati</a>, <a href="https://publications.waset.org/abstracts/search?q=Erind%20Bedalli"> Erind Bedalli</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Clustering is a versatile instrument in the analysis of collections of data providing insights of the underlying structures of the dataset and enhancing the modeling capabilities. The fuzzy approach to the clustering problem increases the flexibility involving the concept of partial memberships (some value in the continuous interval [0, 1]) of the instances in the clusters. Several fuzzy clustering algorithms have been devised like FCM, Gustafson-Kessel, Gath-Geva, kernel-based FCM, PCM etc. Each of these algorithms has its own advantages and drawbacks, so none of these algorithms would be able to perform superiorly in all datasets. In this paper we will experimentally compare FCM, GK, GG algorithm and a hybrid two-stage fuzzy clustering model combining the FCM and Gath-Geva algorithms. Firstly we will theoretically dis-cuss the advantages and drawbacks for each of these algorithms and we will describe the hybrid clustering model exploiting the advantages and diminishing the drawbacks of each algorithm. Secondly we will experimentally compare the accuracy of the hybrid model by applying it on several benchmark and synthetic datasets. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=fuzzy%20clustering" title="fuzzy clustering">fuzzy clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=fuzzy%20c-means%20algorithm%20%28FCM%29" title=" fuzzy c-means algorithm (FCM)"> fuzzy c-means algorithm (FCM)</a>, <a href="https://publications.waset.org/abstracts/search?q=Gustafson-Kessel%20algorithm" title=" Gustafson-Kessel algorithm"> Gustafson-Kessel algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=hybrid%20clustering%20model" title=" hybrid clustering model"> hybrid clustering model</a> </p> <a href="https://publications.waset.org/abstracts/67863/an-experimental-study-on-some-conventional-and-hybrid-models-of-fuzzy-clustering" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/67863.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">514</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">596</span> Using Closed Frequent Itemsets for Hierarchical Document Clustering</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Cheng-Jhe%20Lee">Cheng-Jhe Lee</a>, <a href="https://publications.waset.org/abstracts/search?q=Chiun-Chieh%20Hsu"> Chiun-Chieh Hsu</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Due to the rapid development of the Internet and the increased availability of digital documents, the excessive information on the Internet has led to information overflow problem. In order to solve these problems for effective information retrieval, document clustering in text mining becomes a popular research topic. Clustering is the unsupervised classification of data items into groups without the need of training data. Many conventional document clustering methods perform inefficiently for large document collections because they were originally designed for relational database. Therefore they are impractical in real-world document clustering and require special handling for high dimensionality and high volume. We propose the FIHC (Frequent Itemset-based Hierarchical Clustering) method, which is a hierarchical clustering method developed for document clustering, where the intuition of FIHC is that there exist some common words for each cluster. FIHC uses such words to cluster documents and builds hierarchical topic tree. In this paper, we combine FIHC algorithm with ontology to solve the semantic problem and mine the meaning behind the words in documents. Furthermore, we use the closed frequent itemsets instead of only use frequent itemsets, which increases efficiency and scalability. The experimental results show that our method is more accurate than those of well-known document clustering algorithms. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=FIHC" title="FIHC">FIHC</a>, <a href="https://publications.waset.org/abstracts/search?q=documents%20clustering" title=" documents clustering"> documents clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=ontology" title=" ontology"> ontology</a>, <a href="https://publications.waset.org/abstracts/search?q=closed%20frequent%20itemset" title=" closed frequent itemset"> closed frequent itemset</a> </p> <a href="https://publications.waset.org/abstracts/41381/using-closed-frequent-itemsets-for-hierarchical-document-clustering" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/41381.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">399</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">595</span> Improved K-Means Clustering Algorithm Using RHadoop with Combiner</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ji%20Eun%20Shin">Ji Eun Shin</a>, <a href="https://publications.waset.org/abstracts/search?q=Dong%20Hoon%20Lim"> Dong Hoon Lim</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Data clustering is a common technique used in data analysis and is used in many applications, such as artificial intelligence, pattern recognition, economics, ecology, psychiatry and marketing. K-means clustering is a well-known clustering algorithm aiming to cluster a set of data points to a predefined number of clusters. In this paper, we implement K-means algorithm based on MapReduce framework with RHadoop to make the clustering method applicable to large scale data. RHadoop is a collection of R packages that allow users to manage and analyze data with Hadoop. The main idea is to introduce a combiner as a function of our map output to decrease the amount of data needed to be processed by reducers. The experimental results demonstrated that K-means algorithm using RHadoop can scale well and efficiently process large data sets on commodity hardware. We also showed that our K-means algorithm using RHadoop with combiner was faster than regular algorithm without combiner as the size of data set increases. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=big%20data" title="big data">big data</a>, <a href="https://publications.waset.org/abstracts/search?q=combiner" title=" combiner"> combiner</a>, <a href="https://publications.waset.org/abstracts/search?q=K-means%20clustering" title=" K-means clustering"> K-means clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=RHadoop" title=" RHadoop"> RHadoop</a> </p> <a href="https://publications.waset.org/abstracts/41570/improved-k-means-clustering-algorithm-using-rhadoop-with-combiner" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/41570.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">439</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">594</span> Application of Data Mining for Aquifer Environmental Assessment</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Saman%20Javadi">Saman Javadi</a>, <a href="https://publications.waset.org/abstracts/search?q=Mehdi%20Hashemy"> Mehdi Hashemy</a>, <a href="https://publications.waset.org/abstracts/search?q=Mohahammad%20Mahmoodi"> Mohahammad Mahmoodi </a> </p> <p class="card-text"><strong>Abstract:</strong></p> Vulnerability maps are employed as an important solution in order to handle entrance of pollution into the aquifers. The common way to provide vulnerability map is DRASTIC. Meanwhile, application of the method is not easy to apply for any aquifer due to choosing appropriate constant values of weights and ranks. In this study, a new approach using k-means clustering is applied to make vulnerability maps. Four features of depth to groundwater, hydraulic conductivity, recharge value and vadose zone were considered at the same time as features of clustering. Five regions are recognized out of the case study represent zones with different level of vulnerability. The finding results show that clustering provides a realistic vulnerability map so that, Pearson’s correlation coefficients between nitrate concentrations and clustering vulnerability is obtained 61%. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=clustering" title="clustering">clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=data%20mining" title=" data mining"> data mining</a>, <a href="https://publications.waset.org/abstracts/search?q=groundwater" title=" groundwater"> groundwater</a>, <a href="https://publications.waset.org/abstracts/search?q=vulnerability%20assessment" title=" vulnerability assessment "> vulnerability assessment </a> </p> <a href="https://publications.waset.org/abstracts/20164/application-of-data-mining-for-aquifer-environmental-assessment" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/20164.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">603</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">593</span> 3D Mesh Coarsening via Uniform Clustering</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Shuhua%20Lai">Shuhua Lai</a>, <a href="https://publications.waset.org/abstracts/search?q=Kairui%20Chen"> Kairui Chen</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, we present a fast and efficient mesh coarsening algorithm for 3D triangular meshes. Theis approach can be applied to very complex 3D meshes of arbitrary topology and with millions of vertices. The algorithm is based on the clustering of the input mesh elements, which divides the faces of an input mesh into a given number of clusters for clustering purpose by approximating the Centroidal Voronoi Tessellation of the input mesh. Once a clustering is achieved, it provides us an efficient way to construct uniform tessellations, and therefore leads to good coarsening of polygonal meshes. With proliferation of 3D scanners, this coarsening algorithm is particularly useful for reverse engineering applications of 3D models, which in many cases are dense, non-uniform, irregular and arbitrary topology. Examples demonstrating effectiveness of the new algorithm are also included in the paper. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=coarsening" title="coarsening">coarsening</a>, <a href="https://publications.waset.org/abstracts/search?q=mesh%20clustering" title=" mesh clustering"> mesh clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=shape%20approximation" title=" shape approximation"> shape approximation</a>, <a href="https://publications.waset.org/abstracts/search?q=mesh%20simplification" title=" mesh simplification"> mesh simplification</a> </p> <a href="https://publications.waset.org/abstracts/48919/3d-mesh-coarsening-via-uniform-clustering" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/48919.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">381</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">592</span> Multimodal Optimization of Density-Based Clustering Using Collective Animal Behavior Algorithm</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Kristian%20Bautista">Kristian Bautista</a>, <a href="https://publications.waset.org/abstracts/search?q=Ruben%20A.%20Idoy"> Ruben A. Idoy</a> </p> <p class="card-text"><strong>Abstract:</strong></p> A bio-inspired metaheuristic algorithm inspired by the theory of collective animal behavior (CAB) was integrated to density-based clustering modeled as multimodal optimization problem. The algorithm was tested on synthetic, Iris, Glass, Pima and Thyroid data sets in order to measure its effectiveness relative to CDE-based Clustering algorithm. Upon preliminary testing, it was found out that one of the parameter settings used was ineffective in performing clustering when applied to the algorithm prompting the researcher to do an investigation. It was revealed that fine tuning distance δ3 that determines the extent to which a given data point will be clustered helped improve the quality of cluster output. Even though the modification of distance δ3 significantly improved the solution quality and cluster output of the algorithm, results suggest that there is no difference between the population mean of the solutions obtained using the original and modified parameter setting for all data sets. This implies that using either the original or modified parameter setting will not have any effect towards obtaining the best global and local animal positions. Results also suggest that CDE-based clustering algorithm is better than CAB-density clustering algorithm for all data sets. Nevertheless, CAB-density clustering algorithm is still a good clustering algorithm because it has correctly identified the number of classes of some data sets more frequently in a thirty trial run with a much smaller standard deviation, a potential in clustering high dimensional data sets. Thus, the researcher recommends further investigation in the post-processing stage of the algorithm. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=clustering" title="clustering">clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=metaheuristics" title=" metaheuristics"> metaheuristics</a>, <a href="https://publications.waset.org/abstracts/search?q=collective%20animal%20behavior%20algorithm" title=" collective animal behavior algorithm"> collective animal behavior algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering" title=" density-based clustering"> density-based clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=multimodal%20optimization" title=" multimodal optimization"> multimodal optimization</a> </p> <a href="https://publications.waset.org/abstracts/94254/multimodal-optimization-of-density-based-clustering-using-collective-animal-behavior-algorithm" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/94254.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">231</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">591</span> Anomaly Detection Based Fuzzy K-Mode Clustering for Categorical Data</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Murat%20Yazici">Murat Yazici</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Anomalies are irregularities found in data that do not adhere to a well-defined standard of normal behavior. The identification of outliers or anomalies in data has been a subject of study within the statistics field since the 1800s. Over time, a variety of anomaly detection techniques have been developed in several research communities. The cluster analysis can be used to detect anomalies. It is the process of associating data with clusters that are as similar as possible while dissimilar clusters are associated with each other. Many of the traditional cluster algorithms have limitations in dealing with data sets containing categorical properties. To detect anomalies in categorical data, fuzzy clustering approach can be used with its advantages. The fuzzy k-Mode (FKM) clustering algorithm, which is one of the fuzzy clustering approaches, by extension to the k-means algorithm, is reported for clustering datasets with categorical values. It is a form of clustering: each point can be associated with more than one cluster. In this paper, anomaly detection is performed on two simulated data by using the FKM cluster algorithm. As a significance of the study, the FKM cluster algorithm allows to determine anomalies with their abnormality degree in contrast to numerous anomaly detection algorithms. According to the results, the FKM cluster algorithm illustrated good performance in the anomaly detection of data, including both one anomaly and more than one anomaly. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=fuzzy%20k-mode%20clustering" title="fuzzy k-mode clustering">fuzzy k-mode clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=anomaly%20detection" title=" anomaly detection"> anomaly detection</a>, <a href="https://publications.waset.org/abstracts/search?q=noise" title=" noise"> noise</a>, <a href="https://publications.waset.org/abstracts/search?q=categorical%20data" title=" categorical data"> categorical data</a> </p> <a href="https://publications.waset.org/abstracts/185305/anomaly-detection-based-fuzzy-k-mode-clustering-for-categorical-data" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/185305.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">54</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">590</span> Chemical Reaction Algorithm for Expectation Maximization Clustering</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Li%20Ni">Li Ni</a>, <a href="https://publications.waset.org/abstracts/search?q=Pen%20ManMan"> Pen ManMan</a>, <a href="https://publications.waset.org/abstracts/search?q=Li%20KenLi"> Li KenLi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Clustering is an intensive research for some years because of its multifaceted applications, such as biology, information retrieval, medicine, business and so on. The expectation maximization (EM) is a kind of algorithm framework in clustering methods, one of the ten algorithms of machine learning. Traditionally, optimization of objective function has been the standard approach in EM. Hence, research has investigated the utility of evolutionary computing and related techniques in the regard. Chemical Reaction Optimization (CRO) is a recently established method. So the property embedded in CRO is used to solve optimization problems. This paper presents an algorithm framework (EM-CRO) with modified CRO operators based on EM cluster problems. The hybrid algorithm is mainly to solve the problem of initial value sensitivity of the objective function optimization clustering algorithm. Our experiments mainly take the EM classic algorithm:k-means and fuzzy k-means as an example, through the CRO algorithm to optimize its initial value, get K-means-CRO and FKM-CRO algorithm. The experimental results of them show that there is improved efficiency for solving objective function optimization clustering problems. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=chemical%20reaction%20optimization" title="chemical reaction optimization">chemical reaction optimization</a>, <a href="https://publications.waset.org/abstracts/search?q=expection%20maimization" title=" expection maimization"> expection maimization</a>, <a href="https://publications.waset.org/abstracts/search?q=initia" title=" initia"> initia</a>, <a href="https://publications.waset.org/abstracts/search?q=objective%20function%20clustering" title=" objective function clustering"> objective function clustering</a> </p> <a href="https://publications.waset.org/abstracts/54706/chemical-reaction-algorithm-for-expectation-maximization-clustering" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/54706.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">715</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">589</span> Self-Supervised Attributed Graph Clustering with Dual Contrastive Loss Constraints</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Lijuan%20Zhou">Lijuan Zhou</a>, <a href="https://publications.waset.org/abstracts/search?q=Mengqi%20Wu"> Mengqi Wu</a>, <a href="https://publications.waset.org/abstracts/search?q=Changyong%20Niu"> Changyong Niu</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Attributed graph clustering can utilize the graph topology and node attributes to uncover hidden community structures and patterns in complex networks, aiding in the understanding and analysis of complex systems. Utilizing contrastive learning for attributed graph clustering can effectively exploit meaningful implicit relationships between data. However, existing attributed graph clustering methods based on contrastive learning suffer from the following drawbacks: 1) Complex data augmentation increases computational cost, and inappropriate data augmentation may lead to semantic drift. 2) The selection of positive and negative samples neglects the intrinsic cluster structure learned from graph topology and node attributes. Therefore, this paper proposes a method called self-supervised Attributed Graph Clustering with Dual Contrastive Loss constraints (AGC-DCL). Firstly, Siamese Multilayer Perceptron (MLP) encoders are employed to generate two views separately to avoid complex data augmentation. Secondly, the neighborhood contrastive loss is introduced to constrain node representation using local topological structure while effectively embedding attribute information through attribute reconstruction. Additionally, clustering-oriented contrastive loss is applied to fully utilize clustering information in global semantics for discriminative node representations, regarding the cluster centers from two views as negative samples to fully leverage effective clustering information from different views. Comparative clustering results with existing attributed graph clustering algorithms on six datasets demonstrate the superiority of the proposed method. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=attributed%20graph%20clustering" title="attributed graph clustering">attributed graph clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=contrastive%20learning" title=" contrastive learning"> contrastive learning</a>, <a href="https://publications.waset.org/abstracts/search?q=clustering-oriented" title=" clustering-oriented"> clustering-oriented</a>, <a href="https://publications.waset.org/abstracts/search?q=self-supervised%20learning" title=" self-supervised learning"> self-supervised learning</a> </p> <a href="https://publications.waset.org/abstracts/185262/self-supervised-attributed-graph-clustering-with-dual-contrastive-loss-constraints" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/185262.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">54</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">588</span> Decision Trees Constructing Based on K-Means Clustering Algorithm</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Loai%20Abdallah">Loai Abdallah</a>, <a href="https://publications.waset.org/abstracts/search?q=Malik%20Yousef"> Malik Yousef</a> </p> <p class="card-text"><strong>Abstract:</strong></p> A domain space for the data should reflect the actual similarity between objects. Since objects belonging to the same cluster usually share some common traits even though their geometric distance might be relatively large. In general, the Euclidean distance of data points that represented by large number of features is not capturing the actual relation between those points. In this study, we propose a new method to construct a different space that is based on clustering to form a new distance metric. The new distance space is based on ensemble clustering (EC). The EC distance space is defined by tracking the membership of the points over multiple runs of clustering algorithm metric. Over this distance, we train the decision trees classifier (DT-EC). The results obtained by applying DT-EC on 10 datasets confirm our hypotheses that embedding the EC space as a distance metric would improve the performance. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=ensemble%20clustering" title="ensemble clustering">ensemble clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=decision%20trees" title=" decision trees"> decision trees</a>, <a href="https://publications.waset.org/abstracts/search?q=classification" title=" classification"> classification</a>, <a href="https://publications.waset.org/abstracts/search?q=K%20nearest%20neighbors" title=" K nearest neighbors"> K nearest neighbors</a> </p> <a href="https://publications.waset.org/abstracts/89656/decision-trees-constructing-based-on-k-means-clustering-algorithm" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/89656.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">191</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">587</span> A Non-parametric Clustering Approach for Multivariate Geostatistical Data</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Francky%20Fouedjio">Francky Fouedjio</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Multivariate geostatistical data have become omnipresent in the geosciences and pose substantial analysis challenges. One of them is the grouping of data locations into spatially contiguous clusters so that data locations within the same cluster are more similar while clusters are different from each other, in some sense. Spatially contiguous clusters can significantly improve the interpretation that turns the resulting clusters into meaningful geographical subregions. In this paper, we develop an agglomerative hierarchical clustering approach that takes into account the spatial dependency between observations. It relies on a dissimilarity matrix built from a non-parametric kernel estimator of the spatial dependence structure of data. It integrates existing methods to find the optimal cluster number and to evaluate the contribution of variables to the clustering. The capability of the proposed approach to provide spatially compact, connected and meaningful clusters is assessed using bivariate synthetic dataset and multivariate geochemical dataset. The proposed clustering method gives satisfactory results compared to other similar geostatistical clustering methods. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=clustering" title="clustering">clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=geostatistics" title=" geostatistics"> geostatistics</a>, <a href="https://publications.waset.org/abstracts/search?q=multivariate%20data" title=" multivariate data"> multivariate data</a>, <a href="https://publications.waset.org/abstracts/search?q=non-parametric" title=" non-parametric"> non-parametric</a> </p> <a href="https://publications.waset.org/abstracts/46870/a-non-parametric-clustering-approach-for-multivariate-geostatistical-data" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/46870.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">477</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">586</span> Power Iteration Clustering Based on Deflation Technique on Large Scale Graphs </h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Taysir%20Soliman">Taysir Soliman</a> </p> <p class="card-text"><strong>Abstract:</strong></p> One of the current popular clustering techniques is Spectral Clustering (SC) because of its advantages over conventional approaches such as hierarchical clustering, k-means, etc. and other techniques as well. However, one of the disadvantages of SC is the time consuming process because it requires computing the eigenvectors. In the past to overcome this disadvantage, a number of attempts have been proposed such as the Power Iteration Clustering (PIC) technique, which is one of versions from SC; some of PIC advantages are: 1) its scalability and efficiency, 2) finding one pseudo-eigenvectors instead of computing eigenvectors, and 3) linear combination of the eigenvectors in linear time. However, its worst disadvantage is an inter-class collision problem because it used only one pseudo-eigenvectors which is not enough. Previous researchers developed Deflation-based Power Iteration Clustering (DPIC) to overcome problems of PIC technique on inter-class collision with the same efficiency of PIC. In this paper, we developed Parallel DPIC (PDPIC) to improve the time and memory complexity which is run on apache spark framework using sparse matrix. To test the performance of PDPIC, we compared it to SC, ESCG, ESCALG algorithms on four small graph benchmark datasets and nine large graph benchmark datasets, where PDPIC proved higher accuracy and better time consuming than other compared algorithms. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=spectral%20clustering" title="spectral clustering">spectral clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=power%20iteration%20clustering" title=" power iteration clustering"> power iteration clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=deflation-based%20power%20iteration%20clustering" title=" deflation-based power iteration clustering"> deflation-based power iteration clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=Apache%20spark" title=" Apache spark"> Apache spark</a>, <a href="https://publications.waset.org/abstracts/search?q=large%20graph" title=" large graph"> large graph</a> </p> <a href="https://publications.waset.org/abstracts/120360/power-iteration-clustering-based-on-deflation-technique-on-large-scale-graphs" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/120360.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">189</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">585</span> Agglomerative Hierarchical Clustering Using the Tθ Family of Similarity Measures</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Salima%20Kouici">Salima Kouici</a>, <a href="https://publications.waset.org/abstracts/search?q=Abdelkader%20Khelladi"> Abdelkader Khelladi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this work, we begin with the presentation of the Tθ family of usual similarity measures concerning multidimensional binary data. Subsequently, some properties of these measures are proposed. Finally, the impact of the use of different inter-elements measures on the results of the Agglomerative Hierarchical Clustering Methods is studied. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=binary%20data" title="binary data">binary data</a>, <a href="https://publications.waset.org/abstracts/search?q=similarity%20measure" title=" similarity measure"> similarity measure</a>, <a href="https://publications.waset.org/abstracts/search?q=T%CE%B8%20measures" title=" Tθ measures"> Tθ measures</a>, <a href="https://publications.waset.org/abstracts/search?q=agglomerative%20hierarchical%20clustering" title=" agglomerative hierarchical clustering"> agglomerative hierarchical clustering</a> </p> <a href="https://publications.waset.org/abstracts/13108/agglomerative-hierarchical-clustering-using-the-tth-family-of-similarity-measures" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/13108.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">482</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">584</span> Finding Bicluster on Gene Expression Data of Lymphoma Based on Singular Value Decomposition and Hierarchical Clustering</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Alhadi%20Bustaman">Alhadi Bustaman</a>, <a href="https://publications.waset.org/abstracts/search?q=Soeganda%20Formalidin"> Soeganda Formalidin</a>, <a href="https://publications.waset.org/abstracts/search?q=Titin%20Siswantining"> Titin Siswantining</a> </p> <p class="card-text"><strong>Abstract:</strong></p> DNA microarray technology is used to analyze thousand gene expression data simultaneously and a very important task for drug development and test, function annotation, and cancer diagnosis. Various clustering methods have been used for analyzing gene expression data. However, when analyzing very large and heterogeneous collections of gene expression data, conventional clustering methods often cannot produce a satisfactory solution. Biclustering algorithm has been used as an alternative approach to identifying structures from gene expression data. In this paper, we introduce a transform technique based on singular value decomposition to identify normalized matrix of gene expression data followed by Mixed-Clustering algorithm and the Lift algorithm, inspired in the node-deletion and node-addition phases proposed by Cheng and Church based on Agglomerative Hierarchical Clustering (AHC). Experimental study on standard datasets demonstrated the effectiveness of the algorithm in gene expression data. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=agglomerative%20hierarchical%20clustering%20%28AHC%29" title="agglomerative hierarchical clustering (AHC)">agglomerative hierarchical clustering (AHC)</a>, <a href="https://publications.waset.org/abstracts/search?q=biclustering" title=" biclustering"> biclustering</a>, <a href="https://publications.waset.org/abstracts/search?q=gene%20expression%20data" title=" gene expression data"> gene expression data</a>, <a href="https://publications.waset.org/abstracts/search?q=lymphoma" title=" lymphoma"> lymphoma</a>, <a href="https://publications.waset.org/abstracts/search?q=singular%20value%20decomposition%20%28SVD%29" title=" singular value decomposition (SVD)"> singular value decomposition (SVD)</a> </p> <a href="https://publications.waset.org/abstracts/72889/finding-bicluster-on-gene-expression-data-of-lymphoma-based-on-singular-value-decomposition-and-hierarchical-clustering" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/72889.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">279</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">583</span> K-Means Clustering-Based Infinite Feature Selection Method</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Seyyedeh%20Faezeh%20Hassani%20Ziabari">Seyyedeh Faezeh Hassani Ziabari</a>, <a href="https://publications.waset.org/abstracts/search?q=Sadegh%20Eskandari"> Sadegh Eskandari</a>, <a href="https://publications.waset.org/abstracts/search?q=Maziar%20Salahi"> Maziar Salahi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Infinite Feature Selection (IFS) algorithm is an efficient feature selection algorithm that selects a subset of features of all sizes (including infinity). In this paper, we present an improved version of it, called clustering IFS (CIFS), by clustering the dataset in advance. To do so, first, we apply the K-means algorithm to cluster the dataset, then we apply IFS. In the CIFS method, the spatial and temporal complexities are reduced compared to the IFS method. Experimental results on 6 datasets show the superiority of CIFS compared to IFS in terms of accuracy, running time, and memory consumption. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=feature%20selection" title="feature selection">feature selection</a>, <a href="https://publications.waset.org/abstracts/search?q=infinite%20feature%20selection" title=" infinite feature selection"> infinite feature selection</a>, <a href="https://publications.waset.org/abstracts/search?q=clustering" title=" clustering"> clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=graph" title=" graph"> graph</a> </p> <a href="https://publications.waset.org/abstracts/155406/k-means-clustering-based-infinite-feature-selection-method" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/155406.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">128</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">582</span> An Improved K-Means Algorithm for Gene Expression Data Clustering</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Billel%20Kenidra">Billel Kenidra</a>, <a href="https://publications.waset.org/abstracts/search?q=Mohamed%20Benmohammed"> Mohamed Benmohammed</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Data mining technique used in the field of clustering is a subject of active research and assists in biological pattern recognition and extraction of new knowledge from raw data. Clustering means the act of partitioning an unlabeled dataset into groups of similar objects. Each group, called a cluster, consists of objects that are similar between themselves and dissimilar to objects of other groups. Several clustering methods are based on partitional clustering. This category attempts to directly decompose the dataset into a set of disjoint clusters leading to an integer number of clusters that optimizes a given criterion function. The criterion function may emphasize a local or a global structure of the data, and its optimization is an iterative relocation procedure. The K-Means algorithm is one of the most widely used partitional clustering techniques. Since K-Means is extremely sensitive to the initial choice of centers and a poor choice of centers may lead to a local optimum that is quite inferior to the global optimum, we propose a strategy to initiate K-Means centers. The improved K-Means algorithm is compared with the original K-Means, and the results prove how the efficiency has been significantly improved. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=microarray%20data%20mining" title="microarray data mining">microarray data mining</a>, <a href="https://publications.waset.org/abstracts/search?q=biological%20pattern%20recognition" title=" biological pattern recognition"> biological pattern recognition</a>, <a href="https://publications.waset.org/abstracts/search?q=partitional%20clustering" title=" partitional clustering"> partitional clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=k-means%20algorithm" title=" k-means algorithm"> k-means algorithm</a>, <a href="https://publications.waset.org/abstracts/search?q=centroid%20initialization" title=" centroid initialization"> centroid initialization</a> </p> <a href="https://publications.waset.org/abstracts/83541/an-improved-k-means-algorithm-for-gene-expression-data-clustering" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/83541.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">190</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">581</span> Clustering Categorical Data Using the K-Means Algorithm and the Attribute’s Relative Frequency</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Semeh%20Ben%20Salem">Semeh Ben Salem</a>, <a href="https://publications.waset.org/abstracts/search?q=Sami%20Naouali"> Sami Naouali</a>, <a href="https://publications.waset.org/abstracts/search?q=Moetez%20Sallami"> Moetez Sallami</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Clustering is a well known data mining technique used in pattern recognition and information retrieval. The initial dataset to be clustered can either contain categorical or numeric data. Each type of data has its own specific clustering algorithm. In this context, two algorithms are proposed: the <em>k</em>-means for clustering numeric datasets and the <em>k</em>-modes for categorical datasets. The main encountered problem in data mining applications is clustering categorical dataset so relevant in the datasets. One main issue to achieve the clustering process on categorical values is to transform the categorical attributes into numeric measures and directly apply the <em>k</em>-means algorithm instead the <em>k</em>-modes. In this paper, it is proposed to experiment an approach based on the previous issue by transforming the categorical values into numeric ones using the relative frequency of each modality in the attributes. The proposed approach is compared with a previously method based on transforming the categorical datasets into binary values. The scalability and accuracy of the two methods are experimented. The obtained results show that our proposed method outperforms the binary method in all cases. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=clustering" title="clustering">clustering</a>, <a href="https://publications.waset.org/abstracts/search?q=unsupervised%20learning" title=" unsupervised learning"> unsupervised learning</a>, <a href="https://publications.waset.org/abstracts/search?q=pattern%20recognition" title=" pattern recognition"> pattern recognition</a>, <a href="https://publications.waset.org/abstracts/search?q=categorical%20datasets" title=" categorical datasets"> categorical datasets</a>, <a href="https://publications.waset.org/abstracts/search?q=knowledge%20discovery" title=" knowledge discovery"> knowledge discovery</a>, <a href="https://publications.waset.org/abstracts/search?q=k-means" title=" k-means"> k-means</a> </p> <a href="https://publications.waset.org/abstracts/73588/clustering-categorical-data-using-the-k-means-algorithm-and-the-attributes-relative-frequency" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/73588.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">260</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">580</span> Generalization of Clustering Coefficient on Lattice Networks Applied to Criminal Networks</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Christian%20H.%20Sanabria-Monta%C3%B1a">Christian H. Sanabria-Montaña</a>, <a href="https://publications.waset.org/abstracts/search?q=Rodrigo%20Huerta-Quintanilla"> Rodrigo Huerta-Quintanilla</a> </p> <p class="card-text"><strong>Abstract:</strong></p> A lattice network is a special type of network in which all nodes have the same number of links, and its boundary conditions are periodic. The most basic lattice network is the ring, a one-dimensional network with periodic border conditions. In contrast, the Cartesian product of d rings forms a d-dimensional lattice network. An analytical expression currently exists for the clustering coefficient in this type of network, but the theoretical value is valid only up to certain connectivity value; in other words, the analytical expression is incomplete. Here we obtain analytically the clustering coefficient expression in d-dimensional lattice networks for any link density. Our analytical results show that the clustering coefficient for a lattice network with density of links that tend to 1, leads to the value of the clustering coefficient of a fully connected network. We developed a model on criminology in which the generalized clustering coefficient expression is applied. The model states that delinquents learn the know-how of crime business by sharing knowledge, directly or indirectly, with their friends of the gang. This generalization shed light on the network properties, which is important to develop new models in different fields where network structure plays an important role in the system dynamic, such as criminology, evolutionary game theory, econophysics, among others. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=clustering%20coefficient" title="clustering coefficient">clustering coefficient</a>, <a href="https://publications.waset.org/abstracts/search?q=criminology" title=" criminology"> criminology</a>, <a href="https://publications.waset.org/abstracts/search?q=generalized" title=" generalized"> generalized</a>, <a href="https://publications.waset.org/abstracts/search?q=regular%20network%20d-dimensional" title=" regular network d-dimensional"> regular network d-dimensional</a> </p> <a href="https://publications.waset.org/abstracts/71972/generalization-of-clustering-coefficient-on-lattice-networks-applied-to-criminal-networks" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/71972.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">411</span> </span> </div> </div> <ul class="pagination"> <li class="page-item disabled"><span class="page-link">&lsaquo;</span></li> <li class="page-item active"><span class="page-link">1</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=2">2</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=3">3</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=4">4</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=5">5</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=6">6</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=7">7</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=8">8</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=9">9</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=10">10</a></li> <li class="page-item disabled"><span class="page-link">...</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=20">20</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=21">21</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=density-based%20%20clustering&amp;page=2" rel="next">&rsaquo;</a></li> </ul> </div> </main> <footer> <div id="infolinks" class="pt-3 pb-2"> <div class="container"> <div style="background-color:#f5f5f5;" class="p-3"> <div class="row"> <div class="col-md-2"> <ul class="list-unstyled"> About <li><a href="https://waset.org/page/support">About Us</a></li> <li><a href="https://waset.org/page/support#legal-information">Legal</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/WASET-16th-foundational-anniversary.pdf">WASET celebrates its 16th foundational anniversary</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Account <li><a href="https://waset.org/profile">My Account</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Explore <li><a href="https://waset.org/disciplines">Disciplines</a></li> <li><a href="https://waset.org/conferences">Conferences</a></li> <li><a href="https://waset.org/conference-programs">Conference Program</a></li> <li><a href="https://waset.org/committees">Committees</a></li> <li><a href="https://publications.waset.org">Publications</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Research <li><a href="https://publications.waset.org/abstracts">Abstracts</a></li> <li><a href="https://publications.waset.org">Periodicals</a></li> <li><a href="https://publications.waset.org/archive">Archive</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Open Science <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Philosophy.pdf">Open Science Philosophy</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Award.pdf">Open Science Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Society-Open-Science-and-Open-Innovation.pdf">Open Innovation</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Postdoctoral-Fellowship-Award.pdf">Postdoctoral Fellowship Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Scholarly-Research-Review.pdf">Scholarly Research Review</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Support <li><a href="https://waset.org/page/support">Support</a></li> <li><a href="https://waset.org/profile/messages/create">Contact Us</a></li> <li><a href="https://waset.org/profile/messages/create">Report Abuse</a></li> </ul> </div> </div> </div> </div> </div> <div class="container text-center"> <hr style="margin-top:0;margin-bottom:.3rem;"> <a href="https://creativecommons.org/licenses/by/4.0/" target="_blank" class="text-muted small">Creative Commons Attribution 4.0 International License</a> <div id="copy" class="mt-2">&copy; 2024 World Academy of Science, Engineering and Technology</div> </div> </footer> <a href="javascript:" id="return-to-top"><i class="fas fa-arrow-up"></i></a> <div class="modal" id="modal-template"> <div class="modal-dialog"> <div class="modal-content"> <div class="row m-0 mt-1"> <div class="col-md-12"> <button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">&times;</span></button> </div> </div> <div class="modal-body"></div> </div> </div> </div> <script src="https://cdn.waset.org/static/plugins/jquery-3.3.1.min.js"></script> <script src="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/js/bootstrap.bundle.min.js"></script> <script src="https://cdn.waset.org/static/js/site.js?v=150220211556"></script> <script> jQuery(document).ready(function() { /*jQuery.get("https://publications.waset.org/xhr/user-menu", function (response) { jQuery('#mainNavMenu').append(response); });*/ jQuery.get({ url: "https://publications.waset.org/xhr/user-menu", cache: false }).then(function(response){ jQuery('#mainNavMenu').append(response); }); }); </script> </body> </html>

Pages: 1 2 3 4 5 6 7 8 9 10