CINXE.COM
Search results for: Wenli Hu
<!DOCTYPE html> <html lang="en" dir="ltr"> <head> <!-- Google tag (gtag.js) --> <script async src="https://www.googletagmanager.com/gtag/js?id=G-P63WKM1TM1"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('js', new Date()); gtag('config', 'G-P63WKM1TM1'); </script> <!-- Yandex.Metrika counter --> <script type="text/javascript" > (function(m,e,t,r,i,k,a){m[i]=m[i]||function(){(m[i].a=m[i].a||[]).push(arguments)}; m[i].l=1*new Date(); for (var j = 0; j < document.scripts.length; j++) {if (document.scripts[j].src === r) { return; }} k=e.createElement(t),a=e.getElementsByTagName(t)[0],k.async=1,k.src=r,a.parentNode.insertBefore(k,a)}) (window, document, "script", "https://mc.yandex.ru/metrika/tag.js", "ym"); ym(55165297, "init", { clickmap:false, trackLinks:true, accurateTrackBounce:true, webvisor:false }); </script> <noscript><div><img src="https://mc.yandex.ru/watch/55165297" style="position:absolute; left:-9999px;" alt="" /></div></noscript> <!-- /Yandex.Metrika counter --> <!-- Matomo --> <!-- End Matomo Code --> <title>Search results for: Wenli Hu</title> <meta name="description" content="Search results for: Wenli Hu"> <meta name="keywords" content="Wenli Hu"> <meta name="viewport" content="width=device-width, initial-scale=1, minimum-scale=1, maximum-scale=1, user-scalable=no"> <meta charset="utf-8"> <link href="https://cdn.waset.org/favicon.ico" type="image/x-icon" rel="shortcut icon"> <link href="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/css/bootstrap.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/plugins/fontawesome/css/all.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/css/site.css?v=150220211555" rel="stylesheet"> </head> <body> <header> <div class="container"> <nav class="navbar navbar-expand-lg navbar-light"> <a class="navbar-brand" href="https://waset.org"> <img src="https://cdn.waset.org/static/images/wasetc.png" alt="Open Science Research Excellence" title="Open Science Research Excellence" /> </a> <button class="d-block d-lg-none navbar-toggler ml-auto" type="button" data-toggle="collapse" data-target="#navbarMenu" aria-controls="navbarMenu" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <div class="w-100"> <div class="d-none d-lg-flex flex-row-reverse"> <form method="get" action="https://waset.org/search" class="form-inline my-2 my-lg-0"> <input class="form-control mr-sm-2" type="search" placeholder="Search Conferences" value="Wenli Hu" name="q" aria-label="Search"> <button class="btn btn-light my-2 my-sm-0" type="submit"><i class="fas fa-search"></i></button> </form> </div> <div class="collapse navbar-collapse mt-1" id="navbarMenu"> <ul class="navbar-nav ml-auto align-items-center" id="mainNavMenu"> <li class="nav-item"> <a class="nav-link" href="https://waset.org/conferences" title="Conferences in 2024/2025/2026">Conferences</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/disciplines" title="Disciplines">Disciplines</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/committees" rel="nofollow">Committees</a> </li> <li class="nav-item dropdown"> <a class="nav-link dropdown-toggle" href="#" id="navbarDropdownPublications" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Publications </a> <div class="dropdown-menu" aria-labelledby="navbarDropdownPublications"> <a class="dropdown-item" href="https://publications.waset.org/abstracts">Abstracts</a> <a class="dropdown-item" href="https://publications.waset.org">Periodicals</a> <a class="dropdown-item" href="https://publications.waset.org/archive">Archive</a> </div> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/page/support" title="Support">Support</a> </li> </ul> </div> </div> </nav> </div> </header> <main> <div class="container mt-4"> <div class="row"> <div class="col-md-9 mx-auto"> <form method="get" action="https://publications.waset.org/abstracts/search"> <div id="custom-search-input"> <div class="input-group"> <i class="fas fa-search"></i> <input type="text" class="search-query" name="q" placeholder="Author, Title, Abstract, Keywords" value="Wenli Hu"> <input type="submit" class="btn_search" value="Search"> </div> </div> </form> </div> </div> <div class="row mt-3"> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Commenced</strong> in January 2007</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Frequency:</strong> Monthly</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Edition:</strong> International</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Paper Count:</strong> 7</div> </div> </div> </div> <h1 class="mt-3 mb-3 text-center" style="font-size:1.6rem;">Search results for: Wenli Hu</h1> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">7</span> Performance Comparison of Non-Binary RA and QC-LDPC Codes</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Ni%20Wenli">Ni Wenli</a>, <a href="https://publications.waset.org/abstracts/search?q=He%20Jing"> He Jing</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Repeat–Accumulate (RA) codes are subclass of LDPC codes with fast encoder structures. In this paper, we consider a nonbinary extension of binary LDPC codes over GF(q) and construct a non-binary RA code and a non-binary QC-LDPC code over GF(2^4), we construct non-binary RA codes with linear encoding method and non-binary QC-LDPC codes with algebraic constructions method. And the BER performance of RA and QC-LDPC codes over GF(q) are compared with BP decoding and by simulation over the Additive White Gaussian Noise (AWGN) channels. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=non-binary%20RA%20codes" title="non-binary RA codes">non-binary RA codes</a>, <a href="https://publications.waset.org/abstracts/search?q=QC-LDPC%20codes" title=" QC-LDPC codes"> QC-LDPC codes</a>, <a href="https://publications.waset.org/abstracts/search?q=performance%20comparison" title=" performance comparison"> performance comparison</a>, <a href="https://publications.waset.org/abstracts/search?q=BP%20algorithm" title=" BP algorithm"> BP algorithm</a> </p> <a href="https://publications.waset.org/abstracts/42170/performance-comparison-of-non-binary-ra-and-qc-ldpc-codes" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/42170.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">376</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">6</span> Health Percentage Evaluation for Satellite Electrical Power System Based on Linear Stresses Accumulation Damage Theory</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Lin%20Wenli">Lin Wenli</a>, <a href="https://publications.waset.org/abstracts/search?q=Fu%20Linchun"> Fu Linchun</a>, <a href="https://publications.waset.org/abstracts/search?q=Zhang%20Yi"> Zhang Yi</a>, <a href="https://publications.waset.org/abstracts/search?q=Wu%20Ming"> Wu Ming</a> </p> <p class="card-text"><strong>Abstract:</strong></p> To meet the demands of long-life and high-intelligence for satellites, the electrical power system should be provided with self-health condition evaluation capability. Any over-stress events in operations should be recorded. Based on Linear stresses accumulation damage theory, accumulative damage analysis was performed on thermal-mechanical-electrical united stresses for three components including the solar array, the batteries and the power conditioning unit. Then an overall health percentage evaluation model for satellite electrical power system was built. To obtain the accurate quantity for system health percentage, an automatic feedback closed-loop correction method for all coefficients in the evaluation model was present. The evaluation outputs could be referred as taking earlier fault-forecast and interventions for Ground Control Center or Satellites self. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=satellite%20electrical%20power%20system" title="satellite electrical power system">satellite electrical power system</a>, <a href="https://publications.waset.org/abstracts/search?q=health%20percentage" title=" health percentage"> health percentage</a>, <a href="https://publications.waset.org/abstracts/search?q=linear%20stresses%20accumulation%20damage" title=" linear stresses accumulation damage"> linear stresses accumulation damage</a>, <a href="https://publications.waset.org/abstracts/search?q=evaluation%20model" title=" evaluation model"> evaluation model</a> </p> <a href="https://publications.waset.org/abstracts/41050/health-percentage-evaluation-for-satellite-electrical-power-system-based-on-linear-stresses-accumulation-damage-theory" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/41050.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">411</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">5</span> Virtual Reality Based 3D Video Games and Speech-Lip Synchronization Superseding Algebraic Code Excited Linear Prediction</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=P.%20S.%20Jagadeesh%20Kumar">P. S. Jagadeesh Kumar</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20Meenakshi%20Sundaram"> S. Meenakshi Sundaram</a>, <a href="https://publications.waset.org/abstracts/search?q=Wenli%20Hu"> Wenli Hu</a>, <a href="https://publications.waset.org/abstracts/search?q=Yang%20Yung"> Yang Yung</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In 3D video games, the dominance of production is unceasingly growing with a protruding level of affordability in terms of budget. Afterward, the automation of speech-lip synchronization technique is customarily onerous and has advanced a critical research subject in virtual reality based 3D video games. This paper presents one of these automatic tools, precisely riveted on the synchronization of the speech and the lip movement of the game characters. A robust and precise speech recognition segment that systematized with Algebraic Code Excited Linear Prediction method is developed which unconventionally delivers lip sync results. The Algebraic Code Excited Linear Prediction algorithm is constructed on that used in code-excited linear prediction, but Algebraic Code Excited Linear Prediction codebooks have an explicit algebraic structure levied upon them. This affords a quicker substitute to the software enactments of lip sync algorithms and thus advances the superiority of service factors abridged production cost. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=algebraic%20code%20excited%20linear%20prediction" title="algebraic code excited linear prediction">algebraic code excited linear prediction</a>, <a href="https://publications.waset.org/abstracts/search?q=speech-lip%20synchronization" title=" speech-lip synchronization"> speech-lip synchronization</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20games" title=" video games"> video games</a>, <a href="https://publications.waset.org/abstracts/search?q=virtual%20reality" title=" virtual reality"> virtual reality</a> </p> <a href="https://publications.waset.org/abstracts/78585/virtual-reality-based-3d-video-games-and-speech-lip-synchronization-superseding-algebraic-code-excited-linear-prediction" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/78585.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">474</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">4</span> Ophthalmic Hashing Based Supervision of Glaucoma and Corneal Disorders Imposed on Deep Graphical Model</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=P.%20S.%20Jagadeesh%20Kumar">P. S. Jagadeesh Kumar</a>, <a href="https://publications.waset.org/abstracts/search?q=Yang%20Yung"> Yang Yung</a>, <a href="https://publications.waset.org/abstracts/search?q=Mingmin%20Pan"> Mingmin Pan</a>, <a href="https://publications.waset.org/abstracts/search?q=Xianpei%20Li"> Xianpei Li</a>, <a href="https://publications.waset.org/abstracts/search?q=Wenli%20Hu"> Wenli Hu</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Glaucoma is impelled by optic nerve mutilation habitually represented as cupping and visual field injury frequently with an arcuate pattern of mid-peripheral loss, subordinate to retinal ganglion cell damage and death. Glaucoma is the second foremost cause of blindness and the chief cause of permanent blindness worldwide. Consequently, all-embracing study into the analysis and empathy of glaucoma is happening to escort deep learning based neural network intrusions to deliberate this substantial optic neuropathy. This paper advances an ophthalmic hashing based supervision of glaucoma and corneal disorders preeminent on deep graphical model. Ophthalmic hashing is a newly proposed method extending the efficacy of visual hash-coding to predict glaucoma corneal disorder matching, which is the faster than the existing methods. Deep graphical model is proficient of learning interior explications of corneal disorders in satisfactory time to solve hard combinatoric incongruities using deep Boltzmann machines. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=corneal%20disorders" title="corneal disorders">corneal disorders</a>, <a href="https://publications.waset.org/abstracts/search?q=deep%20Boltzmann%20machines" title=" deep Boltzmann machines"> deep Boltzmann machines</a>, <a href="https://publications.waset.org/abstracts/search?q=deep%20graphical%20model" title=" deep graphical model"> deep graphical model</a>, <a href="https://publications.waset.org/abstracts/search?q=glaucoma" title=" glaucoma"> glaucoma</a>, <a href="https://publications.waset.org/abstracts/search?q=neural%20networks" title=" neural networks"> neural networks</a>, <a href="https://publications.waset.org/abstracts/search?q=ophthalmic%20hashing" title=" ophthalmic hashing"> ophthalmic hashing</a> </p> <a href="https://publications.waset.org/abstracts/78678/ophthalmic-hashing-based-supervision-of-glaucoma-and-corneal-disorders-imposed-on-deep-graphical-model" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/78678.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">250</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3</span> Tensor Deep Stacking Neural Networks and Bilinear Mapping Based Speech Emotion Classification Using Facial Electromyography</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=P.%20S.%20Jagadeesh%20Kumar">P. S. Jagadeesh Kumar</a>, <a href="https://publications.waset.org/abstracts/search?q=Yang%20Yung"> Yang Yung</a>, <a href="https://publications.waset.org/abstracts/search?q=Wenli%20Hu"> Wenli Hu</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Speech emotion classification is a dominant research field in finding a sturdy and profligate classifier appropriate for different real-life applications. This effort accentuates on classifying different emotions from speech signal quarried from the features related to pitch, formants, energy contours, jitter, shimmer, spectral, perceptual and temporal features. Tensor deep stacking neural networks were supported to examine the factors that influence the classification success rate. Facial electromyography signals were composed of several forms of focuses in a controlled atmosphere by means of audio-visual stimuli. Proficient facial electromyography signals were pre-processed using moving average filter, and a set of arithmetical features were excavated. Extracted features were mapped into consistent emotions using bilinear mapping. With facial electromyography signals, a database comprising diverse emotions will be exposed with a suitable fine-tuning of features and training data. A success rate of 92% can be attained deprived of increasing the system connivance and the computation time for sorting diverse emotional states. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=speech%20emotion%20classification" title="speech emotion classification">speech emotion classification</a>, <a href="https://publications.waset.org/abstracts/search?q=tensor%20deep%20stacking%20neural%20networks" title=" tensor deep stacking neural networks"> tensor deep stacking neural networks</a>, <a href="https://publications.waset.org/abstracts/search?q=facial%20electromyography" title=" facial electromyography"> facial electromyography</a>, <a href="https://publications.waset.org/abstracts/search?q=bilinear%20mapping" title=" bilinear mapping"> bilinear mapping</a>, <a href="https://publications.waset.org/abstracts/search?q=audio-visual%20stimuli" title=" audio-visual stimuli"> audio-visual stimuli</a> </p> <a href="https://publications.waset.org/abstracts/78499/tensor-deep-stacking-neural-networks-and-bilinear-mapping-based-speech-emotion-classification-using-facial-electromyography" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/78499.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">254</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">2</span> BodeACD: Buffer Overflow Vulnerabilities Detecting Based on Abstract Syntax Tree, Control Flow Graph, and Data Dependency Graph</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Xinghang%20Lv">Xinghang Lv</a>, <a href="https://publications.waset.org/abstracts/search?q=Tao%20Peng"> Tao Peng</a>, <a href="https://publications.waset.org/abstracts/search?q=Jia%20Chen"> Jia Chen</a>, <a href="https://publications.waset.org/abstracts/search?q=Junping%20Liu"> Junping Liu</a>, <a href="https://publications.waset.org/abstracts/search?q=Xinrong%20Hu"> Xinrong Hu</a>, <a href="https://publications.waset.org/abstracts/search?q=Ruhan%20He"> Ruhan He</a>, <a href="https://publications.waset.org/abstracts/search?q=Minghua%20Jiang"> Minghua Jiang</a>, <a href="https://publications.waset.org/abstracts/search?q=Wenli%20Cao"> Wenli Cao</a> </p> <p class="card-text"><strong>Abstract:</strong></p> As one of the most dangerous vulnerabilities, effective detection of buffer overflow vulnerabilities is extremely necessary. Traditional detection methods are not accurate enough and consume more resources to meet complex and enormous code environment at present. In order to resolve the above problems, we propose the method for Buffer overflow detection based on Abstract syntax tree, Control flow graph, and Data dependency graph (BodeACD) in C/C++ programs with source code. Firstly, BodeACD constructs the function samples of buffer overflow that are available on Github, then represents them as code representation sequences, which fuse control flow, data dependency, and syntax structure of source code to reduce information loss during code representation. Finally, BodeACD learns vulnerability patterns for vulnerability detection through deep learning. The results of the experiments show that BodeACD has increased the precision and recall by 6.3% and 8.5% respectively compared with the latest methods, which can effectively improve vulnerability detection and reduce False-positive rate and False-negative rate. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=vulnerability%20detection" title="vulnerability detection">vulnerability detection</a>, <a href="https://publications.waset.org/abstracts/search?q=abstract%20syntax%20tree" title=" abstract syntax tree"> abstract syntax tree</a>, <a href="https://publications.waset.org/abstracts/search?q=control%20flow%20graph" title=" control flow graph"> control flow graph</a>, <a href="https://publications.waset.org/abstracts/search?q=data%20dependency%20graph" title=" data dependency graph"> data dependency graph</a>, <a href="https://publications.waset.org/abstracts/search?q=code%20representation" title=" code representation"> code representation</a>, <a href="https://publications.waset.org/abstracts/search?q=deep%20learning" title=" deep learning"> deep learning</a> </p> <a href="https://publications.waset.org/abstracts/148239/bodeacd-buffer-overflow-vulnerabilities-detecting-based-on-abstract-syntax-tree-control-flow-graph-and-data-dependency-graph" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/148239.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">170</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">1</span> Proficiency Testing of English for Specific Academic Purpose: Using a Pilot Test in a Taiwanese University as an Example</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Wenli%20Tsou">Wenli Tsou</a>, <a href="https://publications.waset.org/abstracts/search?q=Jessica%20Wu"> Jessica Wu</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Courses of English for specific academic purposes (ESAP) have become popular for higher education in Taiwan; however, no standardized tests have been developed for evaluating learners’ English proficiency in individual designated fields. Assuming a learner’s proficiency in a specific academic area is built up with one’s general proficiency in English with specific knowledge and vocabulary in the content areas, an adequate ESAP proficiency test may be constructed by some selected test items related to the designated academic areas. In this study, through collaboration between a language testing institution and a university in Taiwan, three sets of ESAP tests, covering three disciplinary areas of business and the workplace, science and engineering, and health and medicine majors, were developed and administered to sophomore students (N=1704) who were enrolled in ESAP courses at a university in southern Taiwan. For this study, the courses were grouped into the above-mentioned three disciplines, and students took the specialized proficiency test based on the ESAP course they were taking. Because students were free to select which ESAP course to take, each course had both major and non-major students. Toward the end of the one-semester course, ending in January, 2015, each student took two tests, one of general English (General English Proficiency Test, or GEPT) and the other ESAP. Following each test, students filled out a survey, reporting their test taking experiences. After comparing students’ two test scores, it was found that business majors and health and medical students performed better in ESAP than the non-majors in the class, whereas science and engineering majors did about the same as their non-major counterparts. In addition, test takers with CERF B2 (upper intermediate) level or above performed well in both tests, while students who are below B2 did slightly better in ESAP. The findings suggest that students’ test performance have been enhanced by their specialist content and vocabulary knowledge. Furthermore, results of the survey show that the difficulty levels reported by students are consistent with their test performances. Based on the item analysis, the findings can be used to develop proficiency tests for specific disciplines and to identify ability indicators for college students in their designated fields. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=english%20for%20specific%20academic%20purposes%20%28ESAP%29" title="english for specific academic purposes (ESAP)">english for specific academic purposes (ESAP)</a>, <a href="https://publications.waset.org/abstracts/search?q=general%20english%20proficiency%20test%20%28GEPT%29" title=" general english proficiency test (GEPT)"> general english proficiency test (GEPT)</a>, <a href="https://publications.waset.org/abstracts/search?q=higher%20education" title=" higher education"> higher education</a>, <a href="https://publications.waset.org/abstracts/search?q=proficiency%20test" title=" proficiency test"> proficiency test</a> </p> <a href="https://publications.waset.org/abstracts/30188/proficiency-testing-of-english-for-specific-academic-purpose-using-a-pilot-test-in-a-taiwanese-university-as-an-example" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/30188.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">529</span> </span> </div> </div> </div> </main> <footer> <div id="infolinks" class="pt-3 pb-2"> <div class="container"> <div style="background-color:#f5f5f5;" class="p-3"> <div class="row"> <div class="col-md-2"> <ul class="list-unstyled"> About <li><a href="https://waset.org/page/support">About Us</a></li> <li><a href="https://waset.org/page/support#legal-information">Legal</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/WASET-16th-foundational-anniversary.pdf">WASET celebrates its 16th foundational anniversary</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Account <li><a href="https://waset.org/profile">My Account</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Explore <li><a href="https://waset.org/disciplines">Disciplines</a></li> <li><a href="https://waset.org/conferences">Conferences</a></li> <li><a href="https://waset.org/conference-programs">Conference Program</a></li> <li><a href="https://waset.org/committees">Committees</a></li> <li><a href="https://publications.waset.org">Publications</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Research <li><a href="https://publications.waset.org/abstracts">Abstracts</a></li> <li><a href="https://publications.waset.org">Periodicals</a></li> <li><a href="https://publications.waset.org/archive">Archive</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Open Science <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Philosophy.pdf">Open Science Philosophy</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Award.pdf">Open Science Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Society-Open-Science-and-Open-Innovation.pdf">Open Innovation</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Postdoctoral-Fellowship-Award.pdf">Postdoctoral Fellowship Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Scholarly-Research-Review.pdf">Scholarly Research Review</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Support <li><a href="https://waset.org/page/support">Support</a></li> <li><a href="https://waset.org/profile/messages/create">Contact Us</a></li> <li><a href="https://waset.org/profile/messages/create">Report Abuse</a></li> </ul> </div> </div> </div> </div> </div> <div class="container text-center"> <hr style="margin-top:0;margin-bottom:.3rem;"> <a href="https://creativecommons.org/licenses/by/4.0/" target="_blank" class="text-muted small">Creative Commons Attribution 4.0 International License</a> <div id="copy" class="mt-2">© 2024 World Academy of Science, Engineering and Technology</div> </div> </footer> <a href="javascript:" id="return-to-top"><i class="fas fa-arrow-up"></i></a> <div class="modal" id="modal-template"> <div class="modal-dialog"> <div class="modal-content"> <div class="row m-0 mt-1"> <div class="col-md-12"> <button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">×</span></button> </div> </div> <div class="modal-body"></div> </div> </div> </div> <script src="https://cdn.waset.org/static/plugins/jquery-3.3.1.min.js"></script> <script src="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/js/bootstrap.bundle.min.js"></script> <script src="https://cdn.waset.org/static/js/site.js?v=150220211556"></script> <script> jQuery(document).ready(function() { /*jQuery.get("https://publications.waset.org/xhr/user-menu", function (response) { jQuery('#mainNavMenu').append(response); });*/ jQuery.get({ url: "https://publications.waset.org/xhr/user-menu", cache: false }).then(function(response){ jQuery('#mainNavMenu').append(response); }); }); </script> </body> </html>