CINXE.COM
Search results for: Bert Moolman
<!DOCTYPE html> <html lang="en" dir="ltr"> <head> <!-- Google tag (gtag.js) --> <script async src="https://www.googletagmanager.com/gtag/js?id=G-P63WKM1TM1"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('js', new Date()); gtag('config', 'G-P63WKM1TM1'); </script> <!-- Yandex.Metrika counter --> <script type="text/javascript" > (function(m,e,t,r,i,k,a){m[i]=m[i]||function(){(m[i].a=m[i].a||[]).push(arguments)}; m[i].l=1*new Date(); for (var j = 0; j < document.scripts.length; j++) {if (document.scripts[j].src === r) { return; }} k=e.createElement(t),a=e.getElementsByTagName(t)[0],k.async=1,k.src=r,a.parentNode.insertBefore(k,a)}) (window, document, "script", "https://mc.yandex.ru/metrika/tag.js", "ym"); ym(55165297, "init", { clickmap:false, trackLinks:true, accurateTrackBounce:true, webvisor:false }); </script> <noscript><div><img src="https://mc.yandex.ru/watch/55165297" style="position:absolute; left:-9999px;" alt="" /></div></noscript> <!-- /Yandex.Metrika counter --> <!-- Matomo --> <!-- End Matomo Code --> <title>Search results for: Bert Moolman</title> <meta name="description" content="Search results for: Bert Moolman"> <meta name="keywords" content="Bert Moolman"> <meta name="viewport" content="width=device-width, initial-scale=1, minimum-scale=1, maximum-scale=1, user-scalable=no"> <meta charset="utf-8"> <link href="https://cdn.waset.org/favicon.ico" type="image/x-icon" rel="shortcut icon"> <link href="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/css/bootstrap.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/plugins/fontawesome/css/all.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/css/site.css?v=150220211555" rel="stylesheet"> </head> <body> <header> <div class="container"> <nav class="navbar navbar-expand-lg navbar-light"> <a class="navbar-brand" href="https://waset.org"> <img src="https://cdn.waset.org/static/images/wasetc.png" alt="Open Science Research Excellence" title="Open Science Research Excellence" /> </a> <button class="d-block d-lg-none navbar-toggler ml-auto" type="button" data-toggle="collapse" data-target="#navbarMenu" aria-controls="navbarMenu" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <div class="w-100"> <div class="d-none d-lg-flex flex-row-reverse"> <form method="get" action="https://waset.org/search" class="form-inline my-2 my-lg-0"> <input class="form-control mr-sm-2" type="search" placeholder="Search Conferences" value="Bert Moolman" name="q" aria-label="Search"> <button class="btn btn-light my-2 my-sm-0" type="submit"><i class="fas fa-search"></i></button> </form> </div> <div class="collapse navbar-collapse mt-1" id="navbarMenu"> <ul class="navbar-nav ml-auto align-items-center" id="mainNavMenu"> <li class="nav-item"> <a class="nav-link" href="https://waset.org/conferences" title="Conferences in 2024/2025/2026">Conferences</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/disciplines" title="Disciplines">Disciplines</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/committees" rel="nofollow">Committees</a> </li> <li class="nav-item dropdown"> <a class="nav-link dropdown-toggle" href="#" id="navbarDropdownPublications" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Publications </a> <div class="dropdown-menu" aria-labelledby="navbarDropdownPublications"> <a class="dropdown-item" href="https://publications.waset.org/abstracts">Abstracts</a> <a class="dropdown-item" href="https://publications.waset.org">Periodicals</a> <a class="dropdown-item" href="https://publications.waset.org/archive">Archive</a> </div> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/page/support" title="Support">Support</a> </li> </ul> </div> </div> </nav> </div> </header> <main> <div class="container mt-4"> <div class="row"> <div class="col-md-9 mx-auto"> <form method="get" action="https://publications.waset.org/abstracts/search"> <div id="custom-search-input"> <div class="input-group"> <i class="fas fa-search"></i> <input type="text" class="search-query" name="q" placeholder="Author, Title, Abstract, Keywords" value="Bert Moolman"> <input type="submit" class="btn_search" value="Search"> </div> </div> </form> </div> </div> <div class="row mt-3"> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Commenced</strong> in January 2007</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Frequency:</strong> Monthly</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Edition:</strong> International</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Paper Count:</strong> 46</div> </div> </div> </div> <h1 class="mt-3 mb-3 text-center" style="font-size:1.6rem;">Search results for: Bert Moolman</h1> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">46</span> BERT-Based Chinese Coreference Resolution</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Li%20Xiaoge">Li Xiaoge</a>, <a href="https://publications.waset.org/abstracts/search?q=Wang%20Chaodong"> Wang Chaodong</a> </p> <p class="card-text"><strong>Abstract:</strong></p> We introduce the first Chinese Coreference Resolution Model based on BERT (CCRM-BERT) and show that it significantly outperforms all previous work. The key idea is to consider the features of the mention, such as part of speech, width of spans, distance between spans, etc. And the influence of each features on the model is analyzed. The model computes mention embeddings that combine BERT with features. Compared to the existing state-of-the-art span-ranking approach, our model significantly improves accuracy on the Chinese OntoNotes benchmark. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=BERT" title="BERT">BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=coreference%20resolution" title=" coreference resolution"> coreference resolution</a>, <a href="https://publications.waset.org/abstracts/search?q=deep%20learning" title=" deep learning"> deep learning</a>, <a href="https://publications.waset.org/abstracts/search?q=nature%20language%20processing" title=" nature language processing"> nature language processing</a> </p> <a href="https://publications.waset.org/abstracts/145875/bert-based-chinese-coreference-resolution" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/145875.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">216</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">45</span> A Context-Centric Chatbot for Cryptocurrency Using the Bidirectional Encoder Representations from Transformers Neural Networks</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Qitao%20Xie">Qitao Xie</a>, <a href="https://publications.waset.org/abstracts/search?q=Qingquan%20Zhang"> Qingquan Zhang</a>, <a href="https://publications.waset.org/abstracts/search?q=Xiaofei%20Zhang"> Xiaofei Zhang</a>, <a href="https://publications.waset.org/abstracts/search?q=Di%20Tian"> Di Tian</a>, <a href="https://publications.waset.org/abstracts/search?q=Ruixuan%20Wen"> Ruixuan Wen</a>, <a href="https://publications.waset.org/abstracts/search?q=Ting%20Zhu"> Ting Zhu</a>, <a href="https://publications.waset.org/abstracts/search?q=Ping%20Yi"> Ping Yi</a>, <a href="https://publications.waset.org/abstracts/search?q=Xin%20Li"> Xin Li</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Inspired by the recent movement of digital currency, we are building a question answering system concerning the subject of cryptocurrency using Bidirectional Encoder Representations from Transformers (BERT). The motivation behind this work is to properly assist digital currency investors by directing them to the corresponding knowledge bases that can offer them help and increase the querying speed. BERT, one of newest language models in natural language processing, was investigated to improve the quality of generated responses. We studied different combinations of hyperparameters of the BERT model to obtain the best fit responses. Further, we created an intelligent chatbot for cryptocurrency using BERT. A chatbot using BERT shows great potential for the further advancement of a cryptocurrency market tool. We show that the BERT neural networks generalize well to other tasks by applying it successfully to cryptocurrency. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=bidirectional%20encoder%20representations%20from%20transformers" title="bidirectional encoder representations from transformers">bidirectional encoder representations from transformers</a>, <a href="https://publications.waset.org/abstracts/search?q=BERT" title=" BERT"> BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=chatbot" title=" chatbot"> chatbot</a>, <a href="https://publications.waset.org/abstracts/search?q=cryptocurrency" title=" cryptocurrency"> cryptocurrency</a>, <a href="https://publications.waset.org/abstracts/search?q=deep%20learning" title=" deep learning"> deep learning</a> </p> <a href="https://publications.waset.org/abstracts/129261/a-context-centric-chatbot-for-cryptocurrency-using-the-bidirectional-encoder-representations-from-transformers-neural-networks" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/129261.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">147</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">44</span> Exploring Bidirectional Encoder Representations from the Transformers’ Capabilities to Detect English Preposition Errors</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Dylan%20Elliott">Dylan Elliott</a>, <a href="https://publications.waset.org/abstracts/search?q=Katya%20Pertsova"> Katya Pertsova</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Preposition errors are some of the most common errors created by L2 speakers. In addition, improving error correction and detection methods remains an open issue in the realm of Natural Language Processing (NLP). This research investigates whether the bidirectional encoder representations from the transformers model (BERT) have the potential to correct preposition errors accurately enough to be useful in error correction software. This research finds that BERT performs strongly when the scope of its error correction is limited to preposition choice. The researchers used an open-source BERT model and over three hundred thousand edited sentences from Wikipedia, tagged for part of speech, where only a preposition edit had occurred. To test BERT’s ability to detect errors, a technique known as multi-level masking was used to generate suggestions based on sentence context for every prepositional environment in the test data. These suggestions were compared with the original errors in the data and their known corrections to evaluate BERT’s performance. The suggestions were further analyzed to determine if BERT more often agreed with the judgements of the Wikipedia editors. Both the untrained and fined-tuned models were compared. Finetuning led to a greater rate of error-detection which significantly improved recall, but lowered precision due to an increase in false positives or falsely flagged errors. However, in most cases, these false positives were not errors in preposition usage but merely cases where more than one preposition was possible. Furthermore, when BERT correctly identified an error, the model largely agreed with the Wikipedia editors, suggesting that BERT’s ability to detect misused prepositions is better than previously believed. To evaluate to what extent BERT’s false positives were grammatical suggestions, we plan to do a further crowd-sourcing study to test the grammaticality of BERT’s suggested sentence corrections against native speakers’ judgments. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=BERT" title="BERT">BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=grammatical%20error%20correction" title=" grammatical error correction"> grammatical error correction</a>, <a href="https://publications.waset.org/abstracts/search?q=preposition%20error%20detection" title=" preposition error detection"> preposition error detection</a>, <a href="https://publications.waset.org/abstracts/search?q=prepositions" title=" prepositions"> prepositions</a> </p> <a href="https://publications.waset.org/abstracts/152314/exploring-bidirectional-encoder-representations-from-the-transformers-capabilities-to-detect-english-preposition-errors" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/152314.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">147</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">43</span> A Comparison of Performance Indicators Between University-Level Rugby Union and Rugby Union Sevens Matches</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Pieter%20van%20den%20Berg">Pieter van den Berg</a>, <a href="https://publications.waset.org/abstracts/search?q=Retief%20Broodryk"> Retief Broodryk</a>, <a href="https://publications.waset.org/abstracts/search?q=Bert%20Moolman"> Bert Moolman</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Firstly, this study aimed to identify which performance indicators (PIs) discriminate between winning and losing university-level Rugby Union (RU) teams and, secondly, to compare the significant PIs in RU and Rugby Union Sevens (RS) at university level. Understanding the importance of PIs and their effect on match outcomes could assist coaching staff to prioritise specific game aspects during training to increase performance. Twenty randomly selected round-robin matches of the 2018 Varsity Cup (n=20), and Varsity Sports sevens (n=20) tournaments were analysed. A linear mixed model was used to determine statistical significant differences set at p≤0.05 while effect size was reported according to Cohen's d value. Results revealed that various PIs discriminated between winning and losing RU teams and that specific PIs could be observed as significant in both RU and RS. Therefore, specific identified tactical aspects of RU and RS should be prioritised to optimise performance <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=match%20success" title="match success">match success</a>, <a href="https://publications.waset.org/abstracts/search?q=notational%20analysis" title=" notational analysis"> notational analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=performance%20analysis" title=" performance analysis"> performance analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=rugby" title=" rugby"> rugby</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20analysis" title=" video analysis"> video analysis</a> </p> <a href="https://publications.waset.org/abstracts/171098/a-comparison-of-performance-indicators-between-university-level-rugby-union-and-rugby-union-sevens-matches" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/171098.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">71</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">42</span> Bridging the Data Gap for Sexism Detection in Twitter: A Semi-Supervised Approach</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Adeep%20Hande">Adeep Hande</a>, <a href="https://publications.waset.org/abstracts/search?q=Shubham%20Agarwal"> Shubham Agarwal</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This paper presents a study on identifying sexism in online texts using various state-of-the-art deep learning models based on BERT. We experimented with different feature sets and model architectures and evaluated their performance using precision, recall, F1 score, and accuracy metrics. We also explored the use of pseudolabeling technique to improve model performance. Our experiments show that the best-performing models were based on BERT, and their multilingual model achieved an F1 score of 0.83. Furthermore, the use of pseudolabeling significantly improved the performance of the BERT-based models, with the best results achieved using the pseudolabeling technique. Our findings suggest that BERT-based models with pseudolabeling hold great promise for identifying sexism in online texts with high accuracy. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=large%20language%20models" title="large language models">large language models</a>, <a href="https://publications.waset.org/abstracts/search?q=semi-supervised%20learning" title=" semi-supervised learning"> semi-supervised learning</a>, <a href="https://publications.waset.org/abstracts/search?q=sexism%20detection" title=" sexism detection"> sexism detection</a>, <a href="https://publications.waset.org/abstracts/search?q=data%20sparsity" title=" data sparsity"> data sparsity</a> </p> <a href="https://publications.waset.org/abstracts/171717/bridging-the-data-gap-for-sexism-detection-in-twitter-a-semi-supervised-approach" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/171717.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">70</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">41</span> Relation between Low Thermal Stress and Antioxidant Enzymes Activity in a Sweetening Plant: Stevia Rebaudiana Bert</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=T.%20Bettaieb">T. Bettaieb</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20Soufi"> S. Soufi</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20Arbaoui"> S. Arbaoui</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Stevia rebaudiana Bert. is a natural sweet plant. The leaves contain diterpene glycosides stevioside, rebaudiosides A-F, steviolbioside and dulcoside, which are responsible for its sweet taste and have commercial value all over the world as sugar substitute in foods and medicines. Stevia rebaudiana Bert. is sensitive temperature lower than 9°C. The possibility of its outdoor culture in Tunisian conditions demand genotypes tolerant to low temperatures. In order to evaluate the low temperature tolerance of eight genotypes of Stevia rebaudiana, the activities of superoxide dismutase (SOD), ascorbate peroxidase (APX) and catalases (CAT) were measured. Before carrying out the analyses, three genotypes of Stevia were exposed for 1 month at a temperature regime of 18°C during the day and 7°C at night similar to winter conditions in Tunisia. In response to the stress generated by low temperature, antioxidant enzymes activity revealed on native gel and quantified by spectrophotometry showed variable levels according to their degree of tolerance to low temperatures. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=chilling%20tolerance" title="chilling tolerance">chilling tolerance</a>, <a href="https://publications.waset.org/abstracts/search?q=enzymatic%20activity" title=" enzymatic activity"> enzymatic activity</a>, <a href="https://publications.waset.org/abstracts/search?q=stevia%20rebaudiana%20bert" title=" stevia rebaudiana bert"> stevia rebaudiana bert</a>, <a href="https://publications.waset.org/abstracts/search?q=low%20thermal%20stress" title=" low thermal stress"> low thermal stress</a> </p> <a href="https://publications.waset.org/abstracts/16932/relation-between-low-thermal-stress-and-antioxidant-enzymes-activity-in-a-sweetening-plant-stevia-rebaudiana-bert" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/16932.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">442</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">40</span> A Grey-Box Text Attack Framework Using Explainable AI</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Esther%20Chiramal">Esther Chiramal</a>, <a href="https://publications.waset.org/abstracts/search?q=Kelvin%20Soh%20Boon%20Kai"> Kelvin Soh Boon Kai</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Explainable AI is a strong strategy implemented to understand complex black-box model predictions in a human-interpretable language. It provides the evidence required to execute the use of trustworthy and reliable AI systems. On the other hand, however, it also opens the door to locating possible vulnerabilities in an AI model. Traditional adversarial text attack uses word substitution, data augmentation techniques, and gradient-based attacks on powerful pre-trained Bidirectional Encoder Representations from Transformers (BERT) variants to generate adversarial sentences. These attacks are generally white-box in nature and not practical as they can be easily detected by humans e.g., Changing the word from “Poor” to “Rich”. We proposed a simple yet effective Grey-box cum Black-box approach that does not require the knowledge of the model while using a set of surrogate Transformer/BERT models to perform the attack using Explainable AI techniques. As Transformers are the current state-of-the-art models for almost all Natural Language Processing (NLP) tasks, an attack generated from BERT1 is transferable to BERT2. This transferability is made possible due to the attention mechanism in the transformer that allows the model to capture long-range dependencies in a sequence. Using the power of BERT generalisation via attention, we attempt to exploit how transformers learn by attacking a few surrogate transformer variants which are all based on a different architecture. We demonstrate that this approach is highly effective to generate semantically good sentences by changing as little as one word that is not detectable by humans while still fooling other BERT models. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=BERT" title="BERT">BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=explainable%20AI" title=" explainable AI"> explainable AI</a>, <a href="https://publications.waset.org/abstracts/search?q=Grey-box%20text%20attack" title=" Grey-box text attack"> Grey-box text attack</a>, <a href="https://publications.waset.org/abstracts/search?q=transformer" title=" transformer"> transformer</a> </p> <a href="https://publications.waset.org/abstracts/156518/a-grey-box-text-attack-framework-using-explainable-ai" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/156518.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">137</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">39</span> One-Shot Text Classification with Multilingual-BERT</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hsin-Yang%20Wang">Hsin-Yang Wang</a>, <a href="https://publications.waset.org/abstracts/search?q=K.%20M.%20A.%20Salam"> K. M. A. Salam</a>, <a href="https://publications.waset.org/abstracts/search?q=Ying-Jia%20Lin"> Ying-Jia Lin</a>, <a href="https://publications.waset.org/abstracts/search?q=Daniel%20Tan"> Daniel Tan</a>, <a href="https://publications.waset.org/abstracts/search?q=Tzu-Hsuan%20Chou"> Tzu-Hsuan Chou</a>, <a href="https://publications.waset.org/abstracts/search?q=Hung-Yu%20Kao"> Hung-Yu Kao</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Detecting user intent from natural language expression has a wide variety of use cases in different natural language processing applications. Recently few-shot training has a spike of usage on commercial domains. Due to the lack of significant sample features, the downstream task performance has been limited or leads to an unstable result across different domains. As a state-of-the-art method, the pre-trained BERT model gathering the sentence-level information from a large text corpus shows improvement on several NLP benchmarks. In this research, we are proposing a method to change multi-class classification tasks into binary classification tasks, then use the confidence score to rank the results. As a language model, BERT performs well on sequence data. In our experiment, we change the objective from predicting labels into finding the relations between words in sequence data. Our proposed method achieved 71.0% accuracy in the internal intent detection dataset and 63.9% accuracy in the HuffPost dataset. Acknowledgment: This work was supported by NCKU-B109-K003, which is the collaboration between National Cheng Kung University, Taiwan, and SoftBank Corp., Tokyo. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=OSML" title="OSML">OSML</a>, <a href="https://publications.waset.org/abstracts/search?q=BERT" title=" BERT"> BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=text%20classification" title=" text classification"> text classification</a>, <a href="https://publications.waset.org/abstracts/search?q=one%20shot" title=" one shot"> one shot</a> </p> <a href="https://publications.waset.org/abstracts/135007/one-shot-text-classification-with-multilingual-bert" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/135007.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">101</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">38</span> Detecting Covid-19 Fake News Using Deep Learning Technique</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=AnjalI%20A.%20Prasad">AnjalI A. Prasad</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Nowadays, social media played an important role in spreading misinformation or fake news. This study analyzes the fake news related to the COVID-19 pandemic spread in social media. This paper aims at evaluating and comparing different approaches that are used to mitigate this issue, including popular deep learning approaches, such as CNN, RNN, LSTM, and BERT algorithm for classification. To evaluate models’ performance, we used accuracy, precision, recall, and F1-score as the evaluation metrics. And finally, compare which algorithm shows better result among the four algorithms. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=BERT" title="BERT">BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=CNN" title=" CNN"> CNN</a>, <a href="https://publications.waset.org/abstracts/search?q=LSTM" title=" LSTM"> LSTM</a>, <a href="https://publications.waset.org/abstracts/search?q=RNN" title=" RNN"> RNN</a> </p> <a href="https://publications.waset.org/abstracts/135210/detecting-covid-19-fake-news-using-deep-learning-technique" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/135210.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">205</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">37</span> A Review of Research on Pre-training Technology for Natural Language Processing</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Moquan%20Gong">Moquan Gong</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In recent years, with the rapid development of deep learning, pre-training technology for natural language processing has made great progress. The early field of natural language processing has long used word vector methods such as Word2Vec to encode text. These word vector methods can also be regarded as static pre-training techniques. However, this context-free text representation brings very limited improvement to subsequent natural language processing tasks and cannot solve the problem of word polysemy. ELMo proposes a context-sensitive text representation method that can effectively handle polysemy problems. Since then, pre-training language models such as GPT and BERT have been proposed one after another. Among them, the BERT model has significantly improved its performance on many typical downstream tasks, greatly promoting the technological development in the field of natural language processing, and has since entered the field of natural language processing. The era of dynamic pre-training technology. Since then, a large number of pre-trained language models based on BERT and XLNet have continued to emerge, and pre-training technology has become an indispensable mainstream technology in the field of natural language processing. This article first gives an overview of pre-training technology and its development history, and introduces in detail the classic pre-training technology in the field of natural language processing, including early static pre-training technology and classic dynamic pre-training technology; and then briefly sorts out a series of enlightening technologies. Pre-training technology, including improved models based on BERT and XLNet; on this basis, analyze the problems faced by current pre-training technology research; finally, look forward to the future development trend of pre-training technology. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=natural%20language%20processing" title="natural language processing">natural language processing</a>, <a href="https://publications.waset.org/abstracts/search?q=pre-training" title=" pre-training"> pre-training</a>, <a href="https://publications.waset.org/abstracts/search?q=language%20model" title=" language model"> language model</a>, <a href="https://publications.waset.org/abstracts/search?q=word%20vectors" title=" word vectors"> word vectors</a> </p> <a href="https://publications.waset.org/abstracts/183121/a-review-of-research-on-pre-training-technology-for-natural-language-processing" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/183121.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">57</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">36</span> Bidirectional Encoder Representations from Transformers Sentiment Analysis Applied to Three Presidential Pre-Candidates in Costa Rica</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=F%C3%A9lix%20David%20Su%C3%A1rez%20Bonilla">Félix David Suárez Bonilla</a> </p> <p class="card-text"><strong>Abstract:</strong></p> A sentiment analysis service to detect polarity (positive, neural, and negative), based on transfer learning, was built using a Spanish version of BERT and applied to tweets written in Spanish. The dataset that was used consisted of 11975 reviews, which were extracted from Google Play using the google-play-scrapper package. The BETO trained model used: the AdamW optimizer, a batch size of 16, a learning rate of 2x10⁻⁵ and 10 epochs. The system was tested using tweets of three presidential pre-candidates from Costa Rica. The system was finally validated using human labeled examples, achieving an accuracy of 83.3%. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=NLP" title="NLP">NLP</a>, <a href="https://publications.waset.org/abstracts/search?q=transfer%20learning" title=" transfer learning"> transfer learning</a>, <a href="https://publications.waset.org/abstracts/search?q=BERT" title=" BERT"> BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=sentiment%20analysis" title=" sentiment analysis"> sentiment analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=social%20media" title=" social media"> social media</a>, <a href="https://publications.waset.org/abstracts/search?q=opinion%20mining" title=" opinion mining"> opinion mining</a> </p> <a href="https://publications.waset.org/abstracts/144475/bidirectional-encoder-representations-from-transformers-sentiment-analysis-applied-to-three-presidential-pre-candidates-in-costa-rica" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/144475.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">174</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">35</span> Benchmarking Bert-Based Low-Resource Language: Case Uzbek NLP Models</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Jamshid%20Qodirov">Jamshid Qodirov</a>, <a href="https://publications.waset.org/abstracts/search?q=Sirojiddin%20Komolov"> Sirojiddin Komolov</a>, <a href="https://publications.waset.org/abstracts/search?q=Ravilov%20Mirahmad"> Ravilov Mirahmad</a>, <a href="https://publications.waset.org/abstracts/search?q=Olimjon%20Mirzayev"> Olimjon Mirzayev</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Nowadays, natural language processing tools play a crucial role in our daily lives, including various techniques with text processing. There are very advanced models in modern languages, such as English, Russian etc. But, in some languages, such as Uzbek, the NLP models have been developed recently. Thus, there are only a few NLP models in Uzbek language. Moreover, there is no such work that could show which Uzbek NLP model behaves in different situations and when to use them. This work tries to close this gap and compares the Uzbek NLP models existing as of the time this article was written. The authors try to compare the NLP models in two different scenarios: sentiment analysis and sentence similarity, which are the implementations of the two most common problems in the industry: classification and similarity. Another outcome from this work is two datasets for classification and sentence similarity in Uzbek language that we generated ourselves and can be useful in both industry and academia as well. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=NLP" title="NLP">NLP</a>, <a href="https://publications.waset.org/abstracts/search?q=benchmak" title=" benchmak"> benchmak</a>, <a href="https://publications.waset.org/abstracts/search?q=bert" title=" bert"> bert</a>, <a href="https://publications.waset.org/abstracts/search?q=vectorization" title=" vectorization"> vectorization</a> </p> <a href="https://publications.waset.org/abstracts/182098/benchmarking-bert-based-low-resource-language-case-uzbek-nlp-models" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/182098.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">54</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">34</span> A BERT-Based Model for Financial Social Media Sentiment Analysis</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Josiel%20Delgadillo">Josiel Delgadillo</a>, <a href="https://publications.waset.org/abstracts/search?q=Johnson%20Kinyua"> Johnson Kinyua</a>, <a href="https://publications.waset.org/abstracts/search?q=Charles%20Mutigwe"> Charles Mutigwe</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The purpose of sentiment analysis is to determine the sentiment strength (e.g., positive, negative, neutral) from a textual source for good decision-making. Natural language processing in domains such as financial markets requires knowledge of domain ontology, and pre-trained language models, such as BERT, have made significant breakthroughs in various NLP tasks by training on large-scale un-labeled generic corpora such as Wikipedia. However, sentiment analysis is a strong domain-dependent task. The rapid growth of social media has given users a platform to share their experiences and views about products, services, and processes, including financial markets. StockTwits and Twitter are social networks that allow the public to express their sentiments in real time. Hence, leveraging the success of unsupervised pre-training and a large amount of financial text available on social media platforms could potentially benefit a wide range of financial applications. This work is focused on sentiment analysis using social media text on platforms such as StockTwits and Twitter. To meet this need, SkyBERT, a domain-specific language model pre-trained and fine-tuned on financial corpora, has been developed. The results show that SkyBERT outperforms current state-of-the-art models in financial sentiment analysis. Extensive experimental results demonstrate the effectiveness and robustness of SkyBERT. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=BERT" title="BERT">BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=financial%20markets" title=" financial markets"> financial markets</a>, <a href="https://publications.waset.org/abstracts/search?q=Twitter" title=" Twitter"> Twitter</a>, <a href="https://publications.waset.org/abstracts/search?q=sentiment%20analysis" title=" sentiment analysis"> sentiment analysis</a> </p> <a href="https://publications.waset.org/abstracts/156566/a-bert-based-model-for-financial-social-media-sentiment-analysis" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/156566.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">152</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">33</span> Improving Subjective Bias Detection Using Bidirectional Encoder Representations from Transformers and Bidirectional Long Short-Term Memory</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ebipatei%20Victoria%20Tunyan">Ebipatei Victoria Tunyan</a>, <a href="https://publications.waset.org/abstracts/search?q=T.%20A.%20Cao"> T. A. Cao</a>, <a href="https://publications.waset.org/abstracts/search?q=Cheol%20Young%20Ock"> Cheol Young Ock</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Detecting subjectively biased statements is a vital task. This is because this kind of bias, when present in the text or other forms of information dissemination media such as news, social media, scientific texts, and encyclopedias, can weaken trust in the information and stir conflicts amongst consumers. Subjective bias detection is also critical for many Natural Language Processing (NLP) tasks like sentiment analysis, opinion identification, and bias neutralization. Having a system that can adequately detect subjectivity in text will boost research in the above-mentioned areas significantly. It can also come in handy for platforms like Wikipedia, where the use of neutral language is of importance. The goal of this work is to identify the subjectively biased language in text on a sentence level. With machine learning, we can solve complex AI problems, making it a good fit for the problem of subjective bias detection. A key step in this approach is to train a classifier based on BERT (Bidirectional Encoder Representations from Transformers) as upstream model. BERT by itself can be used as a classifier; however, in this study, we use BERT as data preprocessor as well as an embedding generator for a Bi-LSTM (Bidirectional Long Short-Term Memory) network incorporated with attention mechanism. This approach produces a deeper and better classifier. We evaluate the effectiveness of our model using the Wiki Neutrality Corpus (WNC), which was compiled from Wikipedia edits that removed various biased instances from sentences as a benchmark dataset, with which we also compare our model to existing approaches. Experimental analysis indicates an improved performance, as our model achieved state-of-the-art accuracy in detecting subjective bias. This study focuses on the English language, but the model can be fine-tuned to accommodate other languages. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=subjective%20bias%20detection" title="subjective bias detection">subjective bias detection</a>, <a href="https://publications.waset.org/abstracts/search?q=machine%20learning" title=" machine learning"> machine learning</a>, <a href="https://publications.waset.org/abstracts/search?q=BERT%E2%80%93BiLSTM%E2%80%93Attention" title=" BERT–BiLSTM–Attention"> BERT–BiLSTM–Attention</a>, <a href="https://publications.waset.org/abstracts/search?q=text%20classification" title=" text classification"> text classification</a>, <a href="https://publications.waset.org/abstracts/search?q=natural%20language%20processing" title=" natural language processing"> natural language processing</a> </p> <a href="https://publications.waset.org/abstracts/133543/improving-subjective-bias-detection-using-bidirectional-encoder-representations-from-transformers-and-bidirectional-long-short-term-memory" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/133543.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">130</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">32</span> Topic Sentiments toward the COVID-19 Vaccine on Twitter</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Melissa%20Vang">Melissa Vang</a>, <a href="https://publications.waset.org/abstracts/search?q=Raheyma%20Khan"> Raheyma Khan</a>, <a href="https://publications.waset.org/abstracts/search?q=Haihua%20Chen"> Haihua Chen</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The coronavirus disease 2019 (COVID‐19) pandemic has changed people's lives from all over the world. More people have turned to Twitter to engage online and discuss the COVID-19 vaccine. This study aims to present a text mining approach to identify people's attitudes towards the COVID-19 vaccine on Twitter. To achieve this purpose, we collected 54,268 COVID-19 vaccine tweets from September 01, 2020, to November 01, 2020, then the BERT model is used for the sentiment and topic analysis. The results show that people had more negative than positive attitudes about the vaccine, and countries with an increasing number of confirmed cases had a higher percentage of negative attitudes. Additionally, the topics discussed in positive and negative tweets are different. The tweet datasets can be helpful to information professionals to inform the public about vaccine-related informational resources. Our findings may have implications for understanding people's cognitions and feelings about the vaccine. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=BERT" title="BERT">BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=COVID-19%20vaccine" title=" COVID-19 vaccine"> COVID-19 vaccine</a>, <a href="https://publications.waset.org/abstracts/search?q=sentiment%20analysis" title=" sentiment analysis"> sentiment analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=topic%20modeling" title=" topic modeling"> topic modeling</a> </p> <a href="https://publications.waset.org/abstracts/138813/topic-sentiments-toward-the-covid-19-vaccine-on-twitter" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/138813.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">150</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">31</span> The Impact of Recurring Events in Fake News Detection</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ali%20Raza">Ali Raza</a>, <a href="https://publications.waset.org/abstracts/search?q=Shafiq%20Ur%20Rehman%20Khan"> Shafiq Ur Rehman Khan</a>, <a href="https://publications.waset.org/abstracts/search?q=Raja%20Sher%20Afgun%20Usmani"> Raja Sher Afgun Usmani</a>, <a href="https://publications.waset.org/abstracts/search?q=Asif%20Raza"> Asif Raza</a>, <a href="https://publications.waset.org/abstracts/search?q=Basit%20Umair"> Basit Umair</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Detection of Fake news and missing information is gaining popularity, especially after the advancement in social media and online news platforms. Social media platforms are the main and speediest source of fake news propagation, whereas online news websites contribute to fake news dissipation. In this study, we propose a framework to detect fake news using the temporal features of text and consider user feedback to identify whether the news is fake or not. In recent studies, the temporal features in text documents gain valuable consideration from Natural Language Processing and user feedback and only try to classify the textual data as fake or true. This research article indicates the impact of recurring and non-recurring events on fake and true news. We use two models BERT and Bi-LSTM to investigate, and it is concluded from BERT we get better results and 70% of true news are recurring and rest of 30% are non-recurring. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=natural%20language%20processing" title="natural language processing">natural language processing</a>, <a href="https://publications.waset.org/abstracts/search?q=fake%20news%20detection" title=" fake news detection"> fake news detection</a>, <a href="https://publications.waset.org/abstracts/search?q=machine%20learning" title=" machine learning"> machine learning</a>, <a href="https://publications.waset.org/abstracts/search?q=Bi-LSTM" title=" Bi-LSTM"> Bi-LSTM</a> </p> <a href="https://publications.waset.org/abstracts/190551/the-impact-of-recurring-events-in-fake-news-detection" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/190551.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">22</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">30</span> Evaluation of Modern Natural Language Processing Techniques via Measuring a Company's Public Perception</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Burak%20Oksuzoglu">Burak Oksuzoglu</a>, <a href="https://publications.waset.org/abstracts/search?q=Savas%20Yildirim"> Savas Yildirim</a>, <a href="https://publications.waset.org/abstracts/search?q=Ferhat%20Kutlu"> Ferhat Kutlu</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Opinion mining (OM) is one of the natural language processing (NLP) problems to determine the polarity of opinions, mostly represented on a positive-neutral-negative axis. The data for OM is usually collected from various social media platforms. In an era where social media has considerable control over companies’ futures, it’s worth understanding social media and taking actions accordingly. OM comes to the fore here as the scale of the discussion about companies increases, and it becomes unfeasible to gauge opinion on individual levels. Thus, the companies opt to automize this process by applying machine learning (ML) approaches to their data. For the last two decades, OM or sentiment analysis (SA) has been mainly performed by applying ML classification algorithms such as support vector machines (SVM) and Naïve Bayes to a bag of n-gram representations of textual data. With the advent of deep learning and its apparent success in NLP, traditional methods have become obsolete. Transfer learning paradigm that has been commonly used in computer vision (CV) problems started to shape NLP approaches and language models (LM) lately. This gave a sudden rise to the usage of the pretrained language model (PTM), which contains language representations that are obtained by training it on the large datasets using self-supervised learning objectives. The PTMs are further fine-tuned by a specialized downstream task dataset to produce efficient models for various NLP tasks such as OM, NER (Named-Entity Recognition), Question Answering (QA), and so forth. In this study, the traditional and modern NLP approaches have been evaluated for OM by using a sizable corpus belonging to a large private company containing about 76,000 comments in Turkish: SVM with a bag of n-grams, and two chosen pre-trained models, multilingual universal sentence encoder (MUSE) and bidirectional encoder representations from transformers (BERT). The MUSE model is a multilingual model that supports 16 languages, including Turkish, and it is based on convolutional neural networks. The BERT is a monolingual model in our case and transformers-based neural networks. It uses a masked language model and next sentence prediction tasks that allow the bidirectional training of the transformers. During the training phase of the architecture, pre-processing operations such as morphological parsing, stemming, and spelling correction was not used since the experiments showed that their contribution to the model performance was found insignificant even though Turkish is a highly agglutinative and inflective language. The results show that usage of deep learning methods with pre-trained models and fine-tuning achieve about 11% improvement over SVM for OM. The BERT model achieved around 94% prediction accuracy while the MUSE model achieved around 88% and SVM did around 83%. The MUSE multilingual model shows better results than SVM, but it still performs worse than the monolingual BERT model. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=BERT" title="BERT">BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=MUSE" title=" MUSE"> MUSE</a>, <a href="https://publications.waset.org/abstracts/search?q=opinion%20mining" title=" opinion mining"> opinion mining</a>, <a href="https://publications.waset.org/abstracts/search?q=pretrained%20language%20model" title=" pretrained language model"> pretrained language model</a>, <a href="https://publications.waset.org/abstracts/search?q=SVM" title=" SVM"> SVM</a>, <a href="https://publications.waset.org/abstracts/search?q=Turkish" title=" Turkish"> Turkish</a> </p> <a href="https://publications.waset.org/abstracts/131651/evaluation-of-modern-natural-language-processing-techniques-via-measuring-a-companys-public-perception" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/131651.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">146</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">29</span> Legal Judgment Prediction through Indictments via Data Visualization in Chinese</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Kuo-Chun%20Chien">Kuo-Chun Chien</a>, <a href="https://publications.waset.org/abstracts/search?q=Chia-Hui%20Chang"> Chia-Hui Chang</a>, <a href="https://publications.waset.org/abstracts/search?q=Ren-Der%20Sun"> Ren-Der Sun</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Legal Judgment Prediction (LJP) is a subtask for legal AI. Its main purpose is to use the facts of a case to predict the judgment result. In Taiwan's criminal procedure, when prosecutors complete the investigation of the case, they will decide whether to prosecute the suspect and which article of criminal law should be used based on the facts and evidence of the case. In this study, we collected 305,240 indictments from the public inquiry system of the procuratorate of the Ministry of Justice, which included 169 charges and 317 articles from 21 laws. We take the crime facts in the indictments as the main input to jointly learn the prediction model for law source, article, and charge simultaneously based on the pre-trained Bert model. For single article cases where the frequency of the charge and article are greater than 50, the prediction performance of law sources, articles, and charges reach 97.66, 92.22, and 60.52 macro-f1, respectively. To understand the big performance gap between articles and charges, we used a bipartite graph to visualize the relationship between the articles and charges, and found that the reason for the poor prediction performance was actually due to the wording precision. Some charges use the simplest words, while others may include the perpetrator or the result to make the charges more specific. For example, Article 284 of the Criminal Law may be indicted as “negligent injury”, "negligent death”, "business injury", "driving business injury", or "non-driving business injury". As another example, Article 10 of the Drug Hazard Control Regulations can be charged as “Drug Control Regulations” or “Drug Hazard Control Regulations”. In order to solve the above problems and more accurately predict the article and charge, we plan to include the article content or charge names in the input, and use the sentence-pair classification method for question-answer problems in the BERT model to improve the performance. We will also consider a sequence-to-sequence approach to charge prediction. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=legal%20judgment%20prediction" title="legal judgment prediction">legal judgment prediction</a>, <a href="https://publications.waset.org/abstracts/search?q=deep%20learning" title=" deep learning"> deep learning</a>, <a href="https://publications.waset.org/abstracts/search?q=natural%20language%20processing" title=" natural language processing"> natural language processing</a>, <a href="https://publications.waset.org/abstracts/search?q=BERT" title=" BERT"> BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=data%20visualization" title=" data visualization"> data visualization</a> </p> <a href="https://publications.waset.org/abstracts/147895/legal-judgment-prediction-through-indictments-via-data-visualization-in-chinese" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/147895.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">121</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">28</span> Ontology Expansion via Synthetic Dataset Generation and Transformer-Based Concept Extraction</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Andrey%20Khalov">Andrey Khalov</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The rapid proliferation of unstructured data in IT infrastructure management demands innovative approaches for extracting actionable knowledge. This paper presents a framework for ontology-based knowledge extraction that combines relational graph neural networks (R-GNN) with large language models (LLMs). The proposed method leverages the DOLCE framework as the foundational ontology, extending it with concepts from ITSMO for domain-specific applications in IT service management and outsourcing. A key component of this research is the use of transformer-based models, such as DeBERTa-v3-large, for automatic entity and relationship extraction from unstructured texts. Furthermore, the paper explores how transfer learning techniques can be applied to fine-tune large language models (LLaMA) for using to generate synthetic datasets to improve precision in BERT-based entity recognition and ontology alignment. The resulting IT Ontology (ITO) serves as a comprehensive knowledge base that integrates domain-specific insights from ITIL processes, enabling more efficient decision-making. Experimental results demonstrate significant improvements in knowledge extraction and relationship mapping, offering a cutting-edge solution for enhancing cognitive computing in IT service environments. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=ontology%20expansion" title="ontology expansion">ontology expansion</a>, <a href="https://publications.waset.org/abstracts/search?q=synthetic%20dataset" title=" synthetic dataset"> synthetic dataset</a>, <a href="https://publications.waset.org/abstracts/search?q=transformer%20fine-tuning" title=" transformer fine-tuning"> transformer fine-tuning</a>, <a href="https://publications.waset.org/abstracts/search?q=concept%20extraction" title=" concept extraction"> concept extraction</a>, <a href="https://publications.waset.org/abstracts/search?q=DOLCE" title=" DOLCE"> DOLCE</a>, <a href="https://publications.waset.org/abstracts/search?q=BERT" title=" BERT"> BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=taxonomy" title=" taxonomy"> taxonomy</a>, <a href="https://publications.waset.org/abstracts/search?q=LLM" title=" LLM"> LLM</a>, <a href="https://publications.waset.org/abstracts/search?q=NER" title=" NER"> NER</a> </p> <a href="https://publications.waset.org/abstracts/192579/ontology-expansion-via-synthetic-dataset-generation-and-transformer-based-concept-extraction" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/192579.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">14</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">27</span> Transformer-Driven Multi-Category Classification for an Automated Academic Strand Recommendation Framework</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ma%20Cecilia%20Siva">Ma Cecilia Siva</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This study introduces a Bidirectional Encoder Representations from Transformers (BERT)-based machine learning model aimed at improving educational counseling by automating the process of recommending academic strands for students. The framework is designed to streamline and enhance the strand selection process by analyzing students' profiles and suggesting suitable academic paths based on their interests, strengths, and goals. Data was gathered from a sample of 200 grade 10 students, which included personal essays and survey responses relevant to strand alignment. After thorough preprocessing, the text data was tokenized, label-encoded, and input into a fine-tuned BERT model set up for multi-label classification. The model was optimized for balanced accuracy and computational efficiency, featuring a multi-category classification layer with sigmoid activation for independent strand predictions. Performance metrics showed an F1 score of 88%, indicating a well-balanced model with precision at 80% and recall at 100%, demonstrating its effectiveness in providing reliable recommendations while reducing irrelevant strand suggestions. To facilitate practical use, the final deployment phase created a recommendation framework that processes new student data through the trained model and generates personalized academic strand suggestions. This automated recommendation system presents a scalable solution for academic guidance, potentially enhancing student satisfaction and alignment with educational objectives. The study's findings indicate that expanding the data set, integrating additional features, and refining the model iteratively could improve the framework's accuracy and broaden its applicability in various educational contexts. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=tokenized" title="tokenized">tokenized</a>, <a href="https://publications.waset.org/abstracts/search?q=sigmoid%20activation" title=" sigmoid activation"> sigmoid activation</a>, <a href="https://publications.waset.org/abstracts/search?q=transformer" title=" transformer"> transformer</a>, <a href="https://publications.waset.org/abstracts/search?q=multi%20category%20classification" title=" multi category classification"> multi category classification</a> </p> <a href="https://publications.waset.org/abstracts/193892/transformer-driven-multi-category-classification-for-an-automated-academic-strand-recommendation-framework" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/193892.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">8</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">26</span> AI-Based Techniques for Online Social Media Network Sentiment Analysis: A Methodical Review</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=A.%20M.%20John-Otumu">A. M. John-Otumu</a>, <a href="https://publications.waset.org/abstracts/search?q=M.%20M.%20Rahman"> M. M. Rahman</a>, <a href="https://publications.waset.org/abstracts/search?q=O.%20C.%20Nwokonkwo"> O. C. Nwokonkwo</a>, <a href="https://publications.waset.org/abstracts/search?q=M.%20C.%20Onuoha"> M. C. Onuoha</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Online social media networks have long served as a primary arena for group conversations, gossip, text-based information sharing and distribution. The use of natural language processing techniques for text classification and unbiased decision-making has not been far-fetched. Proper classification of this textual information in a given context has also been very difficult. As a result, we decided to conduct a systematic review of previous literature on sentiment classification and AI-based techniques that have been used in order to gain a better understanding of the process of designing and developing a robust and more accurate sentiment classifier that can correctly classify social media textual information of a given context between hate speech and inverted compliments with a high level of accuracy by assessing different artificial intelligence techniques. We evaluated over 250 articles from digital sources like ScienceDirect, ACM, Google Scholar, and IEEE Xplore and whittled down the number of research to 31. Findings revealed that Deep learning approaches such as CNN, RNN, BERT, and LSTM outperformed various machine learning techniques in terms of performance accuracy. A large dataset is also necessary for developing a robust sentiment classifier and can be obtained from places like Twitter, movie reviews, Kaggle, SST, and SemEval Task4. Hybrid Deep Learning techniques like CNN+LSTM, CNN+GRU, CNN+BERT outperformed single Deep Learning techniques and machine learning techniques. Python programming language outperformed Java programming language in terms of sentiment analyzer development due to its simplicity and AI-based library functionalities. Based on some of the important findings from this study, we made a recommendation for future research. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=artificial%20intelligence" title="artificial intelligence">artificial intelligence</a>, <a href="https://publications.waset.org/abstracts/search?q=natural%20language%20processing" title=" natural language processing"> natural language processing</a>, <a href="https://publications.waset.org/abstracts/search?q=sentiment%20analysis" title=" sentiment analysis"> sentiment analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=social%20network" title=" social network"> social network</a>, <a href="https://publications.waset.org/abstracts/search?q=text" title=" text"> text</a> </p> <a href="https://publications.waset.org/abstracts/146373/ai-based-techniques-for-online-social-media-network-sentiment-analysis-a-methodical-review" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/146373.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">115</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">25</span> Navigating Creditors' Interests in the Context of Business Rescue</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hermanus%20J.%20Moolman">Hermanus J. Moolman</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The COVID-19 pandemic had a severe impact on the society and companies in South Africa. This raises questions about the position of creditors of companies facing financial distress and the actions that directors should take to cater to the interests of creditors. The extent to which directors owe their duties and consideration to creditors has been the subject of debate. The directors of a solvent company owe their duties to the company in favour of its shareholders. When the company becomes insolvent, creditors are the beneficiaries of the directors’ duties. However, the intermittent phase between solvency and insolvency, otherwise referred to as the realm of insolvency, is not accounted for. The purpose of this paper is to determine whether South African company law appropriately addresses the duties that directors owe to creditors and the extent of consideration given to creditors’ interests when the company is in the realm of insolvency and has started business rescue proceedings. A comparative study on South Africa, the United States of America, the United Kingdom and international instruments was employed to achieve the purpose statement. In the United States of America and the United Kingdom, the focus shifts from shareholders to the best interests of creditors when business recue proceedings commence. Such an approach is not aligned with the purpose of the Companies Act of 2008 that calls for a balance of interests of all persons affected by a company’s financial distress and will not be suitable for the South African context. Business rescue in South Africa is relatively new when compared to the practices of the United States of America and the United Kingdom, and the entrepreneurial landscape in South Africa is still evolving. The interests of creditors are not the only interests at risk when a company is financially distressed. It is recommended that an enlightened creditor value approach is adopted for South Africa, where the interests of creditors, albeit paramount, are balanced with those of other stakeholders. This approach optimises a gradual shift in the duties of directors from shareholders to creditors, without disregarding the interests of shareholders. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=business%20rescue" title="business rescue">business rescue</a>, <a href="https://publications.waset.org/abstracts/search?q=shareholders" title=" shareholders"> shareholders</a>, <a href="https://publications.waset.org/abstracts/search?q=creditors" title=" creditors"> creditors</a>, <a href="https://publications.waset.org/abstracts/search?q=financial%20distress" title=" financial distress"> financial distress</a>, <a href="https://publications.waset.org/abstracts/search?q=balance%20of%20interests" title=" balance of interests"> balance of interests</a>, <a href="https://publications.waset.org/abstracts/search?q=alternative%20remedies" title=" alternative remedies"> alternative remedies</a>, <a href="https://publications.waset.org/abstracts/search?q=company%20law" title=" company law"> company law</a> </p> <a href="https://publications.waset.org/abstracts/179306/navigating-creditors-interests-in-the-context-of-business-rescue" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/179306.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">44</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">24</span> The Platform for Digitization of Georgian Documents</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Erekle%20Magradze">Erekle Magradze</a>, <a href="https://publications.waset.org/abstracts/search?q=Davit%20Soselia"> Davit Soselia</a>, <a href="https://publications.waset.org/abstracts/search?q=Levan%20Shughliashvili"> Levan Shughliashvili</a>, <a href="https://publications.waset.org/abstracts/search?q=Irakli%20Koberidze"> Irakli Koberidze</a>, <a href="https://publications.waset.org/abstracts/search?q=Shota%20Tsiskaridze"> Shota Tsiskaridze</a>, <a href="https://publications.waset.org/abstracts/search?q=Victor%20Kakhniashvili"> Victor Kakhniashvili</a>, <a href="https://publications.waset.org/abstracts/search?q=Tamar%20Chaghiashvili"> Tamar Chaghiashvili</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Since the beginning of active publishing activity in Georgia, voluminous printed material has been accumulated, the digitization of which is an important task. Digitized materials will be available to the audience, and it will be possible to find text in them and conduct various factual research. Digitizing scanned documents means scanning documents, extracting text from the scanned documents, and processing the text into a corresponding language model to detect inaccuracies and grammatical errors. Implementing these stages requires a unified, scalable, and automated platform, where the digital service developed for each stage will perform the task assigned to it; at the same time, it will be possible to develop these services dynamically so that there is no interruption in the work of the platform. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=NLP" title="NLP">NLP</a>, <a href="https://publications.waset.org/abstracts/search?q=OCR" title=" OCR"> OCR</a>, <a href="https://publications.waset.org/abstracts/search?q=BERT" title=" BERT"> BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=Kubernetes" title=" Kubernetes"> Kubernetes</a>, <a href="https://publications.waset.org/abstracts/search?q=transformers" title=" transformers"> transformers</a> </p> <a href="https://publications.waset.org/abstracts/145048/the-platform-for-digitization-of-georgian-documents" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/145048.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">144</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">23</span> Hate Speech Detection Using Deep Learning and Machine Learning Models</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Nabil%20Shawkat">Nabil Shawkat</a>, <a href="https://publications.waset.org/abstracts/search?q=Jamil%20Saquer"> Jamil Saquer</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Social media has accelerated our ability to engage with others and eliminated many communication barriers. On the other hand, the widespread use of social media resulted in an increase in online hate speech. This has drastic impacts on vulnerable individuals and societies. Therefore, it is critical to detect hate speech to prevent innocent users and vulnerable communities from becoming victims of hate speech. We investigate the performance of different deep learning and machine learning algorithms on three different datasets. Our results show that the BERT model gives the best performance among all the models by achieving an F1-score of 90.6% on one of the datasets and F1-scores of 89.7% and 88.2% on the other two datasets. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=hate%20speech" title="hate speech">hate speech</a>, <a href="https://publications.waset.org/abstracts/search?q=machine%20learning" title=" machine learning"> machine learning</a>, <a href="https://publications.waset.org/abstracts/search?q=deep%20learning" title=" deep learning"> deep learning</a>, <a href="https://publications.waset.org/abstracts/search?q=abusive%20words" title=" abusive words"> abusive words</a>, <a href="https://publications.waset.org/abstracts/search?q=social%20media" title=" social media"> social media</a>, <a href="https://publications.waset.org/abstracts/search?q=text%20classification" title=" text classification"> text classification</a> </p> <a href="https://publications.waset.org/abstracts/164751/hate-speech-detection-using-deep-learning-and-machine-learning-models" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/164751.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">136</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">22</span> Design and Optimization of a 6 Degrees of Freedom Co-Manipulated Parallel Robot for Prostate Brachytherapy</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Aziza%20Ben%20Halima">Aziza Ben Halima</a>, <a href="https://publications.waset.org/abstracts/search?q=Julien%20Bert"> Julien Bert</a>, <a href="https://publications.waset.org/abstracts/search?q=Dimitris%20Visvikis"> Dimitris Visvikis</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, we propose designing and evaluating a parallel co-manipulated robot dedicated to low-dose-rate prostate brachytherapy. We developed 6 degrees of freedom compact and lightweight robot easy to install in the operating room thanks to its parallel design. This robotic system provides a co-manipulation allowing the surgeon to keep control of the needle’s insertion and consequently to improve the acceptability of the plan for the clinic. The best dimension’s configuration was solved by calculating the geometric model and using an optimization approach. The aim was to ensure the whole coverage of the prostate volume and consider the allowed free space around the patient that includes the ultrasound probe. The final robot dimensions fit in a cube of 300 300 300 mm³. A prototype was 3D printed, and the robot workspace was measured experimentally. The results show that the proposed robotic system satisfies the medical application requirements and permits the needle to reach any point within the prostate. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=medical%20robotics" title="medical robotics">medical robotics</a>, <a href="https://publications.waset.org/abstracts/search?q=co-manipulation" title=" co-manipulation"> co-manipulation</a>, <a href="https://publications.waset.org/abstracts/search?q=prostate%20brachytherapy" title=" prostate brachytherapy"> prostate brachytherapy</a>, <a href="https://publications.waset.org/abstracts/search?q=optimization" title=" optimization"> optimization</a> </p> <a href="https://publications.waset.org/abstracts/131084/design-and-optimization-of-a-6-degrees-of-freedom-co-manipulated-parallel-robot-for-prostate-brachytherapy" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/131084.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">205</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">21</span> ViraPart: A Text Refinement Framework for Automatic Speech Recognition and Natural Language Processing Tasks in Persian</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Narges%20Farokhshad">Narges Farokhshad</a>, <a href="https://publications.waset.org/abstracts/search?q=Milad%20Molazadeh"> Milad Molazadeh</a>, <a href="https://publications.waset.org/abstracts/search?q=Saman%20Jamalabbasi"> Saman Jamalabbasi</a>, <a href="https://publications.waset.org/abstracts/search?q=Hamed%20Babaei%20Giglou"> Hamed Babaei Giglou</a>, <a href="https://publications.waset.org/abstracts/search?q=Saeed%20Bibak"> Saeed Bibak</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The Persian language is an inflectional subject-object-verb language. This fact makes Persian a more uncertain language. However, using techniques such as Zero-Width Non-Joiner (ZWNJ) recognition, punctuation restoration, and Persian Ezafe construction will lead us to a more understandable and precise language. In most of the works in Persian, these techniques are addressed individually. Despite that, we believe that for text refinement in Persian, all of these tasks are necessary. In this work, we proposed a ViraPart framework that uses embedded ParsBERT in its core for text clarifications. First, used the BERT variant for Persian followed by a classifier layer for classification procedures. Next, we combined models outputs to output cleartext. In the end, the proposed model for ZWNJ recognition, punctuation restoration, and Persian Ezafe construction performs the averaged F1 macro scores of 96.90%, 92.13%, and 98.50%, respectively. Experimental results show that our proposed approach is very effective in text refinement for the Persian language. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Persian%20Ezafe" title="Persian Ezafe">Persian Ezafe</a>, <a href="https://publications.waset.org/abstracts/search?q=punctuation" title=" punctuation"> punctuation</a>, <a href="https://publications.waset.org/abstracts/search?q=ZWNJ" title=" ZWNJ"> ZWNJ</a>, <a href="https://publications.waset.org/abstracts/search?q=NLP" title=" NLP"> NLP</a>, <a href="https://publications.waset.org/abstracts/search?q=ParsBERT" title=" ParsBERT"> ParsBERT</a>, <a href="https://publications.waset.org/abstracts/search?q=transformers" title=" transformers"> transformers</a> </p> <a href="https://publications.waset.org/abstracts/143102/virapart-a-text-refinement-framework-for-automatic-speech-recognition-and-natural-language-processing-tasks-in-persian" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/143102.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">215</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">20</span> Probing Syntax Information in Word Representations with Deep Metric Learning</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Bowen%20Ding">Bowen Ding</a>, <a href="https://publications.waset.org/abstracts/search?q=Yihao%20Kuang"> Yihao Kuang</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In recent years, with the development of large-scale pre-trained lan-guage models, building vector representations of text through deep neural network models has become a standard practice for natural language processing tasks. From the performance on downstream tasks, we can know that the text representation constructed by these models contains linguistic information, but its encoding mode and extent are unclear. In this work, a structural probe is proposed to detect whether the vector representation produced by a deep neural network is embedded with a syntax tree. The probe is trained with the deep metric learning method, so that the distance between word vectors in the metric space it defines encodes the distance of words on the syntax tree, and the norm of word vectors encodes the depth of words on the syntax tree. The experiment results on ELMo and BERT show that the syntax tree is encoded in their parameters and the word representations they produce. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=deep%20metric%20learning" title="deep metric learning">deep metric learning</a>, <a href="https://publications.waset.org/abstracts/search?q=syntax%20tree%20probing" title=" syntax tree probing"> syntax tree probing</a>, <a href="https://publications.waset.org/abstracts/search?q=natural%20language%20processing" title=" natural language processing"> natural language processing</a>, <a href="https://publications.waset.org/abstracts/search?q=word%20representations" title=" word representations"> word representations</a> </p> <a href="https://publications.waset.org/abstracts/173855/probing-syntax-information-in-word-representations-with-deep-metric-learning" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/173855.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">68</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">19</span> Text2Time: Transformer-Based Article Time Period Prediction</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Karthick%20Prasad%20Gunasekaran">Karthick Prasad Gunasekaran</a>, <a href="https://publications.waset.org/abstracts/search?q=B.%20Chase%20Babrich"> B. Chase Babrich</a>, <a href="https://publications.waset.org/abstracts/search?q=Saurabh%20Shirodkar"> Saurabh Shirodkar</a>, <a href="https://publications.waset.org/abstracts/search?q=Hee%20Hwang"> Hee Hwang</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Construction preparation is crucial for the success of a construction project. By involving project participants early in the construction phase, project managers can plan ahead and resolve issues early, resulting in project success and satisfaction. This study uses quantitative data from construction management projects to determine the relationship between the pre-construction phase, construction schedule, and customer satisfaction. This study examined a total of 65 construction projects and 93 clients per job to (a) identify the relationship between the pre-construction phase and program reduction and (b) the pre-construction phase and customer retention. Based on a quantitative analysis, this study found a negative correlation between pre-construction status and project schedule in 65 construction projects. This finding means that the more preparatory work done on a particular project, the shorter the total construction time. The Net Promoter Score of 93 clients from 65 projects was then used to determine the relationship between construction preparation and client satisfaction. The pre-construction status and the projects were further analyzed, and a positive correlation between them was found. This shows that customers are happier with projects with a higher ready-to-build ratio than projects with less ready-to-build. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=NLP" title="NLP">NLP</a>, <a href="https://publications.waset.org/abstracts/search?q=BERT" title=" BERT"> BERT</a>, <a href="https://publications.waset.org/abstracts/search?q=LLM" title=" LLM"> LLM</a>, <a href="https://publications.waset.org/abstracts/search?q=deep%20learning" title=" deep learning"> deep learning</a>, <a href="https://publications.waset.org/abstracts/search?q=classification" title=" classification"> classification</a> </p> <a href="https://publications.waset.org/abstracts/166130/text2time-transformer-based-article-time-period-prediction" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/166130.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">104</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">18</span> From Parents to Pioneers: Examining Parental Impact on Entrepreneurial Traits in Latin America</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Bert%20Seither">Bert Seither</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Entrepreneurship is a critical driver of economic growth, especially in emerging regions such as Latin America. This study investigates how parental influences, particularly parental individual entrepreneurial orientation (IEO), shape the entrepreneurial traits of Latin American entrepreneurs. By examining key factors like parental IEO, work ethic, parenting style, and family support, this research assesses how much of an entrepreneur's own IEO can be attributed to parental influence. The study also explores how socio-economic status and cultural context moderate the relationship between parental traits and entrepreneurial orientation. Data will be collected from 600 Latin American entrepreneurs via an online survey. This research aims to provide a comprehensive understanding of the intergenerational transmission of entrepreneurial traits and the broader socio-cultural factors that contribute to entrepreneurial success in diverse contexts. Findings from this study will offer valuable insights for policymakers, educators, and business leaders on fostering entrepreneurship across Latin America, providing practical applications for shaping entrepreneurial behavior through family influences. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=entrepreneurial%20orientation" title="entrepreneurial orientation">entrepreneurial orientation</a>, <a href="https://publications.waset.org/abstracts/search?q=parental%20influence" title=" parental influence"> parental influence</a>, <a href="https://publications.waset.org/abstracts/search?q=Latin%20American%20entrepreneurs" title=" Latin American entrepreneurs"> Latin American entrepreneurs</a>, <a href="https://publications.waset.org/abstracts/search?q=socio-economic%20status" title=" socio-economic status"> socio-economic status</a>, <a href="https://publications.waset.org/abstracts/search?q=cultural%20context" title=" cultural context"> cultural context</a> </p> <a href="https://publications.waset.org/abstracts/192155/from-parents-to-pioneers-examining-parental-impact-on-entrepreneurial-traits-in-latin-america" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/192155.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">18</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">17</span> Using Bidirectional Encoder Representations from Transformers to Extract Topic-Independent Sentiment Features for Social Media Bot Detection</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Maryam%20Heidari">Maryam Heidari</a>, <a href="https://publications.waset.org/abstracts/search?q=James%20H.%20Jones%20Jr."> James H. Jones Jr.</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Millions of online posts about different topics and products are shared on popular social media platforms. One use of this content is to provide crowd-sourced information about a specific topic, event or product. However, this use raises an important question: what percentage of information available through these services is trustworthy? In particular, might some of this information be generated by a machine, i.e., a bot, instead of a human? Bots can be, and often are, purposely designed to generate enough volume to skew an apparent trend or position on a topic, yet the consumer of such content cannot easily distinguish a bot post from a human post. In this paper, we introduce a model for social media bot detection which uses Bidirectional Encoder Representations from Transformers (Google Bert) for sentiment classification of tweets to identify topic-independent features. Our use of a Natural Language Processing approach to derive topic-independent features for our new bot detection model distinguishes this work from previous bot detection models. We achieve 94\% accuracy classifying the contents of data as generated by a bot or a human, where the most accurate prior work achieved accuracy of 92\%. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=bot%20detection" title="bot detection">bot detection</a>, <a href="https://publications.waset.org/abstracts/search?q=natural%20language%20processing" title=" natural language processing"> natural language processing</a>, <a href="https://publications.waset.org/abstracts/search?q=neural%20network" title=" neural network"> neural network</a>, <a href="https://publications.waset.org/abstracts/search?q=social%20media" title=" social media"> social media</a> </p> <a href="https://publications.waset.org/abstracts/129049/using-bidirectional-encoder-representations-from-transformers-to-extract-topic-independent-sentiment-features-for-social-media-bot-detection" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/129049.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">116</span> </span> </div> </div> <ul class="pagination"> <li class="page-item disabled"><span class="page-link">‹</span></li> <li class="page-item active"><span class="page-link">1</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=Bert%20Moolman&page=2">2</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=Bert%20Moolman&page=2" rel="next">›</a></li> </ul> </div> </main> <footer> <div id="infolinks" class="pt-3 pb-2"> <div class="container"> <div style="background-color:#f5f5f5;" class="p-3"> <div class="row"> <div class="col-md-2"> <ul class="list-unstyled"> About <li><a href="https://waset.org/page/support">About Us</a></li> <li><a href="https://waset.org/page/support#legal-information">Legal</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/WASET-16th-foundational-anniversary.pdf">WASET celebrates its 16th foundational anniversary</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Account <li><a href="https://waset.org/profile">My Account</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Explore <li><a href="https://waset.org/disciplines">Disciplines</a></li> <li><a href="https://waset.org/conferences">Conferences</a></li> <li><a href="https://waset.org/conference-programs">Conference Program</a></li> <li><a href="https://waset.org/committees">Committees</a></li> <li><a href="https://publications.waset.org">Publications</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Research <li><a href="https://publications.waset.org/abstracts">Abstracts</a></li> <li><a href="https://publications.waset.org">Periodicals</a></li> <li><a href="https://publications.waset.org/archive">Archive</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Open Science <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Philosophy.pdf">Open Science Philosophy</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Award.pdf">Open Science Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Society-Open-Science-and-Open-Innovation.pdf">Open Innovation</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Postdoctoral-Fellowship-Award.pdf">Postdoctoral Fellowship Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Scholarly-Research-Review.pdf">Scholarly Research Review</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Support <li><a href="https://waset.org/page/support">Support</a></li> <li><a href="https://waset.org/profile/messages/create">Contact Us</a></li> <li><a href="https://waset.org/profile/messages/create">Report Abuse</a></li> </ul> </div> </div> </div> </div> </div> <div class="container text-center"> <hr style="margin-top:0;margin-bottom:.3rem;"> <a href="https://creativecommons.org/licenses/by/4.0/" target="_blank" class="text-muted small">Creative Commons Attribution 4.0 International License</a> <div id="copy" class="mt-2">© 2024 World Academy of Science, Engineering and Technology</div> </div> </footer> <a href="javascript:" id="return-to-top"><i class="fas fa-arrow-up"></i></a> <div class="modal" id="modal-template"> <div class="modal-dialog"> <div class="modal-content"> <div class="row m-0 mt-1"> <div class="col-md-12"> <button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">×</span></button> </div> </div> <div class="modal-body"></div> </div> </div> </div> <script src="https://cdn.waset.org/static/plugins/jquery-3.3.1.min.js"></script> <script src="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/js/bootstrap.bundle.min.js"></script> <script src="https://cdn.waset.org/static/js/site.js?v=150220211556"></script> <script> jQuery(document).ready(function() { /*jQuery.get("https://publications.waset.org/xhr/user-menu", function (response) { jQuery('#mainNavMenu').append(response); });*/ jQuery.get({ url: "https://publications.waset.org/xhr/user-menu", cache: false }).then(function(response){ jQuery('#mainNavMenu').append(response); }); }); </script> </body> </html>