CINXE.COM
Search results for: digital video
<!DOCTYPE html> <html lang="en" dir="ltr"> <head> <!-- Google tag (gtag.js) --> <script async src="https://www.googletagmanager.com/gtag/js?id=G-P63WKM1TM1"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('js', new Date()); gtag('config', 'G-P63WKM1TM1'); </script> <!-- Yandex.Metrika counter --> <script type="text/javascript" > (function(m,e,t,r,i,k,a){m[i]=m[i]||function(){(m[i].a=m[i].a||[]).push(arguments)}; m[i].l=1*new Date(); for (var j = 0; j < document.scripts.length; j++) {if (document.scripts[j].src === r) { return; }} k=e.createElement(t),a=e.getElementsByTagName(t)[0],k.async=1,k.src=r,a.parentNode.insertBefore(k,a)}) (window, document, "script", "https://mc.yandex.ru/metrika/tag.js", "ym"); ym(55165297, "init", { clickmap:false, trackLinks:true, accurateTrackBounce:true, webvisor:false }); </script> <noscript><div><img src="https://mc.yandex.ru/watch/55165297" style="position:absolute; left:-9999px;" alt="" /></div></noscript> <!-- /Yandex.Metrika counter --> <!-- Matomo --> <!-- End Matomo Code --> <title>Search results for: digital video</title> <meta name="description" content="Search results for: digital video"> <meta name="keywords" content="digital video"> <meta name="viewport" content="width=device-width, initial-scale=1, minimum-scale=1, maximum-scale=1, user-scalable=no"> <meta charset="utf-8"> <link href="https://cdn.waset.org/favicon.ico" type="image/x-icon" rel="shortcut icon"> <link href="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/css/bootstrap.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/plugins/fontawesome/css/all.min.css" rel="stylesheet"> <link href="https://cdn.waset.org/static/css/site.css?v=150220211555" rel="stylesheet"> </head> <body> <header> <div class="container"> <nav class="navbar navbar-expand-lg navbar-light"> <a class="navbar-brand" href="https://waset.org"> <img src="https://cdn.waset.org/static/images/wasetc.png" alt="Open Science Research Excellence" title="Open Science Research Excellence" /> </a> <button class="d-block d-lg-none navbar-toggler ml-auto" type="button" data-toggle="collapse" data-target="#navbarMenu" aria-controls="navbarMenu" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <div class="w-100"> <div class="d-none d-lg-flex flex-row-reverse"> <form method="get" action="https://waset.org/search" class="form-inline my-2 my-lg-0"> <input class="form-control mr-sm-2" type="search" placeholder="Search Conferences" value="digital video" name="q" aria-label="Search"> <button class="btn btn-light my-2 my-sm-0" type="submit"><i class="fas fa-search"></i></button> </form> </div> <div class="collapse navbar-collapse mt-1" id="navbarMenu"> <ul class="navbar-nav ml-auto align-items-center" id="mainNavMenu"> <li class="nav-item"> <a class="nav-link" href="https://waset.org/conferences" title="Conferences in 2024/2025/2026">Conferences</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/disciplines" title="Disciplines">Disciplines</a> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/committees" rel="nofollow">Committees</a> </li> <li class="nav-item dropdown"> <a class="nav-link dropdown-toggle" href="#" id="navbarDropdownPublications" role="button" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Publications </a> <div class="dropdown-menu" aria-labelledby="navbarDropdownPublications"> <a class="dropdown-item" href="https://publications.waset.org/abstracts">Abstracts</a> <a class="dropdown-item" href="https://publications.waset.org">Periodicals</a> <a class="dropdown-item" href="https://publications.waset.org/archive">Archive</a> </div> </li> <li class="nav-item"> <a class="nav-link" href="https://waset.org/page/support" title="Support">Support</a> </li> </ul> </div> </div> </nav> </div> </header> <main> <div class="container mt-4"> <div class="row"> <div class="col-md-9 mx-auto"> <form method="get" action="https://publications.waset.org/abstracts/search"> <div id="custom-search-input"> <div class="input-group"> <i class="fas fa-search"></i> <input type="text" class="search-query" name="q" placeholder="Author, Title, Abstract, Keywords" value="digital video"> <input type="submit" class="btn_search" value="Search"> </div> </div> </form> </div> </div> <div class="row mt-3"> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Commenced</strong> in January 2007</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Frequency:</strong> Monthly</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Edition:</strong> International</div> </div> </div> <div class="col-sm-3"> <div class="card"> <div class="card-body"><strong>Paper Count:</strong> 3622</div> </div> </div> </div> <h1 class="mt-3 mb-3 text-center" style="font-size:1.6rem;">Search results for: digital video</h1> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3622</span> H.263 Based Video Transceiver for Wireless Camera System</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Won-Ho%20Kim">Won-Ho Kim</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, a design of H.263 based wireless video transceiver is presented for wireless camera system. It uses standard WIFI transceiver and the covering area is up to 100m. Furthermore the standard H.263 video encoding technique is used for video compression since wireless video transmitter is unable to transmit high capacity raw data in real time and the implemented system is capable of streaming at speed of less than 1Mbps using NTSC 720x480 video. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=wireless%20video%20transceiver" title="wireless video transceiver">wireless video transceiver</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20surveillance%20camera" title=" video surveillance camera"> video surveillance camera</a>, <a href="https://publications.waset.org/abstracts/search?q=H.263%20video%20encoding%20digital%20signal%20processing" title=" H.263 video encoding digital signal processing"> H.263 video encoding digital signal processing</a> </p> <a href="https://publications.waset.org/abstracts/12951/h263-based-video-transceiver-for-wireless-camera-system" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/12951.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">364</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3621</span> Going Viral: Constructively Aligning the Use of Digital Video to Effectively Support Faculty Development</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Samuel%20Olugbenga%20King">Samuel Olugbenga King</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This review article, which is a synthesis of the relevant research literature, focuses on the capabilities of digital video to support, facilitate and enhance faculty development. Based on the literature review, faculty development (i.e., academic or educational development) requires the continued adoption of cohesive, theoretical frameworks to guide research and practice; incorporation of relevant tools from analogous fields, such as teacher professional development; systematic program evaluations; and detailed descriptions of practice to further practice and creative development. A cohesive, five-heuristic framework is subsequently outlined to inform the design and evaluation of the use of digital video, so as to address the barriers to advancing faculty development, as identified through the literature review. Alternative impact evaluation approaches are also described, while the limitations of using digital video for faculty development are highlighted. This paper is therefore conceived as one way to meaningfully leverage the educational affordances of digital video to address some lingering gaps in faculty development. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=digital%20video" title="digital video">digital video</a>, <a href="https://publications.waset.org/abstracts/search?q=faculty%2Feducational%20development" title=" faculty/educational development"> faculty/educational development</a>, <a href="https://publications.waset.org/abstracts/search?q=evaluation" title=" evaluation"> evaluation</a>, <a href="https://publications.waset.org/abstracts/search?q=scholarship%20of%20teaching%20and%20learning%20%28SoTL%29" title=" scholarship of teaching and learning (SoTL)"> scholarship of teaching and learning (SoTL)</a> </p> <a href="https://publications.waset.org/abstracts/51184/going-viral-constructively-aligning-the-use-of-digital-video-to-effectively-support-faculty-development" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/51184.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">352</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3620</span> Tackling the Digital Divide: Enhancing Video Consultation Access for Digital Illiterate Patients in the Hospital</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Wieke%20Ellen%20Bouwes">Wieke Ellen Bouwes</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This study aims to unravel which factors enhance accessibility of video consultations (VCs) for patients with low digital literacy. Thirteen in-depth interviews with patients, hospital employees, eHealth experts, and digital support organizations were held. Patients with low digital literacy received in-home support during real-time video consultations and are observed during the set-up of these consultations. Key findings highlight the importance of patient acceptance, emphasizing video consultations benefits and avoiding standardized courses. The lack of a uniform video consultation system across healthcare providers poses a barrier. Familiarity with support organizations – to support patients in usage of digital tools - among healthcare practitioners enhances accessibility. Moreover, considerations regarding the Dutch General Data Protection Regulation (GDPR) law influence support patients receive. Also, provider readiness to use video consultations influences patient access. Further, alignment between learning styles and support methods seems to determine abilities to learn how to use video consultations. Future research could delve into tailored learning styles and technological solutions for remote access to further explore effectiveness of learning methods. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video%20consultations" title="video consultations">video consultations</a>, <a href="https://publications.waset.org/abstracts/search?q=digital%20literacy%20skills" title=" digital literacy skills"> digital literacy skills</a>, <a href="https://publications.waset.org/abstracts/search?q=effectiveness%20of%20support" title=" effectiveness of support"> effectiveness of support</a>, <a href="https://publications.waset.org/abstracts/search?q=intra-%20and%20inter-organizational%20relationships" title=" intra- and inter-organizational relationships"> intra- and inter-organizational relationships</a>, <a href="https://publications.waset.org/abstracts/search?q=patient%20acceptance%20of%20video%20consultations" title=" patient acceptance of video consultations"> patient acceptance of video consultations</a> </p> <a href="https://publications.waset.org/abstracts/173756/tackling-the-digital-divide-enhancing-video-consultation-access-for-digital-illiterate-patients-in-the-hospital" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/173756.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">74</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3619</span> Remote Video Supervision via DVB-H Channels</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Hanen%20Ghabi">Hanen Ghabi</a>, <a href="https://publications.waset.org/abstracts/search?q=Youssef%20Oudhini"> Youssef Oudhini</a>, <a href="https://publications.waset.org/abstracts/search?q=Hassen%20Mnif"> Hassen Mnif</a> </p> <p class="card-text"><strong>Abstract:</strong></p> By reference to recent publications dealing with the same problem, and as a follow-up to this research work already published, we propose in this article a new original idea of tele supervision exploiting the opportunities offered by the DVB-H system. The objective is to exploit the RF channels of the DVB-H network in order to insert digital remote monitoring images dedicated to a remote solar power plant. Indeed, the DVB-H (Digital Video Broadcast-Handheld) broadcasting system was designed and deployed for digital broadcasting on the same platform as the parent system, DVB-T. We claim to be able to exploit this approach in order to satisfy the operator of remote photovoltaic sites (and others) in order to remotely control the components of isolated installations by means of video surveillance. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video%20surveillance" title="video surveillance">video surveillance</a>, <a href="https://publications.waset.org/abstracts/search?q=digital%20video%20broadcast-handheld" title=" digital video broadcast-handheld"> digital video broadcast-handheld</a>, <a href="https://publications.waset.org/abstracts/search?q=photovoltaic%20sites" title=" photovoltaic sites"> photovoltaic sites</a>, <a href="https://publications.waset.org/abstracts/search?q=AVC" title=" AVC"> AVC</a> </p> <a href="https://publications.waset.org/abstracts/147516/remote-video-supervision-via-dvb-h-channels" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/147516.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">184</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3618</span> Extraction of Text Subtitles in Multimedia Systems</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Amarjit%20Singh">Amarjit Singh</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, a method for extraction of text subtitles in large video is proposed. The video data needs to be annotated for many multimedia applications. Text is incorporated in digital video for the motive of providing useful information about that video. So need arises to detect text present in video to understanding and video indexing. This is achieved in two steps. First step is text localization and the second step is text verification. The method of text detection can be extended to text recognition which finds applications in automatic video indexing; video annotation and content based video retrieval. The method has been tested on various types of videos. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video" title="video">video</a>, <a href="https://publications.waset.org/abstracts/search?q=subtitles" title=" subtitles"> subtitles</a>, <a href="https://publications.waset.org/abstracts/search?q=extraction" title=" extraction"> extraction</a>, <a href="https://publications.waset.org/abstracts/search?q=annotation" title=" annotation"> annotation</a>, <a href="https://publications.waset.org/abstracts/search?q=frames" title=" frames"> frames</a> </p> <a href="https://publications.waset.org/abstracts/24441/extraction-of-text-subtitles-in-multimedia-systems" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/24441.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">601</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3617</span> Smartphone Video Source Identification Based on Sensor Pattern Noise</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Raquel%20Ramos%20L%C3%B3pez">Raquel Ramos López</a>, <a href="https://publications.waset.org/abstracts/search?q=Anissa%20El-Khattabi"> Anissa El-Khattabi</a>, <a href="https://publications.waset.org/abstracts/search?q=Ana%20Lucila%20Sandoval%20Orozco"> Ana Lucila Sandoval Orozco</a>, <a href="https://publications.waset.org/abstracts/search?q=Luis%20Javier%20Garc%C3%ADa%20Villalba"> Luis Javier García Villalba</a> </p> <p class="card-text"><strong>Abstract:</strong></p> An increasing number of mobile devices with integrated cameras has meant that most digital video comes from these devices. These digital videos can be made anytime, anywhere and for different purposes. They can also be shared on the Internet in a short period of time and may sometimes contain recordings of illegal acts. The need to reliably trace the origin becomes evident when these videos are used for forensic purposes. This work proposes an algorithm to identify the brand and model of mobile device which generated the video. Its procedure is as follows: after obtaining the relevant video information, a classification algorithm based on sensor noise and Wavelet Transform performs the aforementioned identification process. We also present experimental results that support the validity of the techniques used and show promising results. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=digital%20video" title="digital video">digital video</a>, <a href="https://publications.waset.org/abstracts/search?q=forensics%20analysis" title=" forensics analysis"> forensics analysis</a>, <a href="https://publications.waset.org/abstracts/search?q=key%20frame" title=" key frame"> key frame</a>, <a href="https://publications.waset.org/abstracts/search?q=mobile%20device" title=" mobile device"> mobile device</a>, <a href="https://publications.waset.org/abstracts/search?q=PRNU" title=" PRNU"> PRNU</a>, <a href="https://publications.waset.org/abstracts/search?q=sensor%20noise" title=" sensor noise"> sensor noise</a>, <a href="https://publications.waset.org/abstracts/search?q=source%20identification" title=" source identification"> source identification</a> </p> <a href="https://publications.waset.org/abstracts/70332/smartphone-video-source-identification-based-on-sensor-pattern-noise" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/70332.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">428</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3616</span> A Passive Digital Video Authentication Technique Using Wavelet Based Optical Flow Variation Thresholding</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=R.%20S.%20Remya">R. S. Remya</a>, <a href="https://publications.waset.org/abstracts/search?q=U.%20S.%20Sethulekshmi"> U. S. Sethulekshmi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Detecting the authenticity of a video is an important issue in digital forensics as Video is used as a silent evidence in court such as in child pornography, movie piracy cases, insurance claims, cases involving scientific fraud, traffic monitoring etc. The biggest threat to video data is the availability of modern open video editing tools which enable easy editing of videos without leaving any trace of tampering. In this paper, we propose an efficient passive method for inter-frame video tampering detection, its type and location by estimating the optical flow of wavelet features of adjacent frames and thresholding the variation in the estimated feature. The performance of the algorithm is compared with the z-score thresholding and achieved an efficiency above 95% on all the tested databases. The proposed method works well for videos with dynamic (forensics) as well as static (surveillance) background. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=discrete%20wavelet%20transform" title="discrete wavelet transform">discrete wavelet transform</a>, <a href="https://publications.waset.org/abstracts/search?q=optical%20flow" title=" optical flow"> optical flow</a>, <a href="https://publications.waset.org/abstracts/search?q=optical%20flow%20variation" title=" optical flow variation"> optical flow variation</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20tampering" title=" video tampering"> video tampering</a> </p> <a href="https://publications.waset.org/abstracts/45252/a-passive-digital-video-authentication-technique-using-wavelet-based-optical-flow-variation-thresholding" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/45252.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">359</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3615</span> Key Frame Based Video Summarization via Dependency Optimization</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Janya%20Sainui">Janya Sainui</a> </p> <p class="card-text"><strong>Abstract:</strong></p> As a rapid growth of digital videos and data communications, video summarization that provides a shorter version of the video for fast video browsing and retrieval is necessary. Key frame extraction is one of the mechanisms to generate video summary. In general, the extracted key frames should both represent the entire video content and contain minimum redundancy. However, most of the existing approaches heuristically select key frames; hence, the selected key frames may not be the most different frames and/or not cover the entire content of a video. In this paper, we propose a method of video summarization which provides the reasonable objective functions for selecting key frames. In particular, we apply a statistical dependency measure called quadratic mutual informaion as our objective functions for maximizing the coverage of the entire video content as well as minimizing the redundancy among selected key frames. The proposed key frame extraction algorithm finds key frames as an optimization problem. Through experiments, we demonstrate the success of the proposed video summarization approach that produces video summary with better coverage of the entire video content while less redundancy among key frames comparing to the state-of-the-art approaches. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video%20summarization" title="video summarization">video summarization</a>, <a href="https://publications.waset.org/abstracts/search?q=key%20frame%20extraction" title=" key frame extraction"> key frame extraction</a>, <a href="https://publications.waset.org/abstracts/search?q=dependency%20measure" title=" dependency measure"> dependency measure</a>, <a href="https://publications.waset.org/abstracts/search?q=quadratic%20mutual%20information" title=" quadratic mutual information"> quadratic mutual information</a> </p> <a href="https://publications.waset.org/abstracts/75218/key-frame-based-video-summarization-via-dependency-optimization" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/75218.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">266</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3614</span> Forensic Challenges in Source Device Identification for Digital Videos</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Mustapha%20Aminu%20Bagiwa">Mustapha Aminu Bagiwa</a>, <a href="https://publications.waset.org/abstracts/search?q=Ainuddin%20Wahid%20Abdul%20Wahab"> Ainuddin Wahid Abdul Wahab</a>, <a href="https://publications.waset.org/abstracts/search?q=Mohd%20Yamani%20Idna%20Idris"> Mohd Yamani Idna Idris</a>, <a href="https://publications.waset.org/abstracts/search?q=Suleman%20Khan"> Suleman Khan</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Video source device identification has become a problem of concern in numerous domains especially in multimedia security and digital investigation. This is because videos are now used as evidence in legal proceedings. Source device identification aim at identifying the source of digital devices using the content they produced. However, due to affordable processing tools and the influx in digital content generating devices, source device identification is still a major problem within the digital forensic community. In this paper, we discuss source device identification for digital videos by identifying techniques that were proposed in the literature for model or specific device identification. This is aimed at identifying salient open challenges for future research. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video%20forgery" title="video forgery">video forgery</a>, <a href="https://publications.waset.org/abstracts/search?q=source%20camcorder" title=" source camcorder"> source camcorder</a>, <a href="https://publications.waset.org/abstracts/search?q=device%20identification" title=" device identification"> device identification</a>, <a href="https://publications.waset.org/abstracts/search?q=forgery%20detection" title=" forgery detection "> forgery detection </a> </p> <a href="https://publications.waset.org/abstracts/21641/forensic-challenges-in-source-device-identification-for-digital-videos" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/21641.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">631</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3613</span> The Effect of Video Games on English as a Foreign Language Students' Language Learning Motivation</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Shamim%20Ali">Shamim Ali</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Researchers and teachers have begun developing digital games and model environments for educational purpose; therefore this study examines the effect of a videos game on secondary school students’ language learning motivation. Secondly, it tries to find out the opportunities to develop a decision making process and simultaneously it analyzes the solutions for further implementation in educational setting. Participants were 30 male students randomly assigned to one of the following three treatments: 10 students were assigned to read the game’s story; 10 students were players, who played video game; and, and the last 10 students acted as watchers and observers, their duty was to watch their classmates play the digital video game. A language learning motivation scale was developed and it was given to the participants as a pre- and post-test. Results indicated a significant language learning motivation and the participants were quite motivated in the end. It is, thus, concluded that the use of video games can help enhance high school students’ language learning motivation. It was suggested that video games should be used as a complementary activity not as a replacement for textbook since excessive use of video games can divert the original purpose of learning. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=EFL" title="EFL">EFL</a>, <a href="https://publications.waset.org/abstracts/search?q=English%20as%20a%20Foreign%20Language" title=" English as a Foreign Language"> English as a Foreign Language</a>, <a href="https://publications.waset.org/abstracts/search?q=motivation" title=" motivation"> motivation</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20games" title=" video games"> video games</a>, <a href="https://publications.waset.org/abstracts/search?q=EFL%20learners" title=" EFL learners"> EFL learners</a> </p> <a href="https://publications.waset.org/abstracts/100055/the-effect-of-video-games-on-english-as-a-foreign-language-students-language-learning-motivation" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/100055.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">178</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3612</span> Efficient Storage and Intelligent Retrieval of Multimedia Streams Using H. 265</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=S.%20Sarumathi">S. Sarumathi</a>, <a href="https://publications.waset.org/abstracts/search?q=C.%20Deepadharani"> C. Deepadharani</a>, <a href="https://publications.waset.org/abstracts/search?q=Garimella%20Archana"> Garimella Archana</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20Dakshayani"> S. Dakshayani</a>, <a href="https://publications.waset.org/abstracts/search?q=D.%20Logeshwaran"> D. Logeshwaran</a>, <a href="https://publications.waset.org/abstracts/search?q=D.%20Jayakumar"> D. Jayakumar</a>, <a href="https://publications.waset.org/abstracts/search?q=Vijayarangan%20Natarajan"> Vijayarangan Natarajan</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The need of the hour for the customers who use a dial-up or a low broadband connection for their internet services is to access HD video data. This can be achieved by developing a new video format using H. 265. This is the latest video codec standard developed by ISO/IEC Moving Picture Experts Group (MPEG) and ITU-T Video Coding Experts Group (VCEG) on April 2013. This new standard for video compression has the potential to deliver higher performance than the earlier standards such as H. 264/AVC. In comparison with H. 264, HEVC offers a clearer, higher quality image at half the original bitrate. At this lower bitrate, it is possible to transmit high definition videos using low bandwidth. It doubles the data compression ratio supporting 8K Ultra HD and resolutions up to 8192×4320. In the proposed model, we design a new video format which supports this H. 265 standard. The major areas of applications in the coming future would lead to enhancements in the performance level of digital television like Tata Sky and Sun Direct, BluRay Discs, Mobile Video, Video Conferencing and Internet and Live Video streaming. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=access%20HD%20video" title="access HD video">access HD video</a>, <a href="https://publications.waset.org/abstracts/search?q=H.%20265%20video%20standard" title=" H. 265 video standard"> H. 265 video standard</a>, <a href="https://publications.waset.org/abstracts/search?q=high%20performance" title=" high performance"> high performance</a>, <a href="https://publications.waset.org/abstracts/search?q=high%20quality%20image" title=" high quality image"> high quality image</a>, <a href="https://publications.waset.org/abstracts/search?q=low%20bandwidth" title=" low bandwidth"> low bandwidth</a>, <a href="https://publications.waset.org/abstracts/search?q=new%20video%20format" title=" new video format"> new video format</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20streaming%20applications" title=" video streaming applications"> video streaming applications</a> </p> <a href="https://publications.waset.org/abstracts/1881/efficient-storage-and-intelligent-retrieval-of-multimedia-streams-using-h-265" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/1881.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">354</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3611</span> Video Shot Detection and Key Frame Extraction Using Faber-Shauder DWT and SVD</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Assma%20Azeroual">Assma Azeroual</a>, <a href="https://publications.waset.org/abstracts/search?q=Karim%20Afdel"> Karim Afdel</a>, <a href="https://publications.waset.org/abstracts/search?q=Mohamed%20El%20Hajji"> Mohamed El Hajji</a>, <a href="https://publications.waset.org/abstracts/search?q=Hassan%20Douzi"> Hassan Douzi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Key frame extraction methods select the most representative frames of a video, which can be used in different areas of video processing such as video retrieval, video summary, and video indexing. In this paper we present a novel approach for extracting key frames from video sequences. The frame is characterized uniquely by his contours which are represented by the dominant blocks. These dominant blocks are located on the contours and its near textures. When the video frames have a noticeable changement, its dominant blocks changed, then we can extracte a key frame. The dominant blocks of every frame is computed, and then feature vectors are extracted from the dominant blocks image of each frame and arranged in a feature matrix. Singular Value Decomposition is used to calculate sliding windows ranks of those matrices. Finally the computed ranks are traced and then we are able to extract key frames of a video. Experimental results show that the proposed approach is robust against a large range of digital effects used during shot transition. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=FSDWT" title="FSDWT">FSDWT</a>, <a href="https://publications.waset.org/abstracts/search?q=key%20frame%20extraction" title=" key frame extraction"> key frame extraction</a>, <a href="https://publications.waset.org/abstracts/search?q=shot%20detection" title=" shot detection"> shot detection</a>, <a href="https://publications.waset.org/abstracts/search?q=singular%20value%20decomposition" title=" singular value decomposition"> singular value decomposition</a> </p> <a href="https://publications.waset.org/abstracts/18296/video-shot-detection-and-key-frame-extraction-using-faber-shauder-dwt-and-svd" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/18296.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">397</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3610</span> Video on Demand (VOD) Industry in Iran: Study of Reasons of Increasing Film and Series Platforms</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Narges%20Hamidipour">Narges Hamidipour</a> </p> <p class="card-text"><strong>Abstract:</strong></p> VOD, which stands for "video on demand", is one kind of watching movies and series on web platforms that, by using them, individuals can access lots of video content by paying abonnement. The first platform in Iran was funded in 2014, and in the last 10 years, it has become the main part of the movie and series industry. There are 374 VOD platforms in Iran, but just three of them are in the mainstream. However, in these years, they have been developed and famed in different ways. This article focuses on the reasons for this development in the past years. For the framework, "digital economy", "media industries," and "political economy" have been used with the interview method. In this research, some experts in SATRA (regulatory organization of inclusive audio and video media in Iran), owners or managers of VODs and some others who directly have been in the system conveyed their opinions. By the way, some documents and analysis statistics are invoked to reach complete results. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=digital%20economy" title="digital economy">digital economy</a>, <a href="https://publications.waset.org/abstracts/search?q=political%20economy" title=" political economy"> political economy</a>, <a href="https://publications.waset.org/abstracts/search?q=VOD" title=" VOD"> VOD</a>, <a href="https://publications.waset.org/abstracts/search?q=interview" title=" interview"> interview</a>, <a href="https://publications.waset.org/abstracts/search?q=iran" title=" iran"> iran</a> </p> <a href="https://publications.waset.org/abstracts/177709/video-on-demand-vod-industry-in-iran-study-of-reasons-of-increasing-film-and-series-platforms" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/177709.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">66</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3609</span> Video Summarization: Techniques and Applications</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Zaynab%20El%20Khattabi">Zaynab El Khattabi</a>, <a href="https://publications.waset.org/abstracts/search?q=Youness%20Tabii"> Youness Tabii</a>, <a href="https://publications.waset.org/abstracts/search?q=Abdelhamid%20Benkaddour"> Abdelhamid Benkaddour</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Nowadays, huge amount of multimedia repositories make the browsing, retrieval and delivery of video contents very slow and even difficult tasks. Video summarization has been proposed to improve faster browsing of large video collections and more efficient content indexing and access. In this paper, we focus on approaches to video summarization. The video summaries can be generated in many different forms. However, two fundamentals ways to generate summaries are static and dynamic. We present different techniques for each mode in the literature and describe some features used for generating video summaries. We conclude with perspective for further research. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video%20summarization" title="video summarization">video summarization</a>, <a href="https://publications.waset.org/abstracts/search?q=static%20summarization" title=" static summarization"> static summarization</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20skimming" title=" video skimming"> video skimming</a>, <a href="https://publications.waset.org/abstracts/search?q=semantic%20features" title=" semantic features"> semantic features</a> </p> <a href="https://publications.waset.org/abstracts/27644/video-summarization-techniques-and-applications" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/27644.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">400</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3608</span> Anonymous Editing Prevention Technique Using Gradient Method for High-Quality Video</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Jiwon%20Lee">Jiwon Lee</a>, <a href="https://publications.waset.org/abstracts/search?q=Chanho%20Jung"> Chanho Jung</a>, <a href="https://publications.waset.org/abstracts/search?q=Si-Hwan%20Jang"> Si-Hwan Jang</a>, <a href="https://publications.waset.org/abstracts/search?q=Kyung-Ill%20Kim"> Kyung-Ill Kim</a>, <a href="https://publications.waset.org/abstracts/search?q=Sanghyun%20Joo"> Sanghyun Joo</a>, <a href="https://publications.waset.org/abstracts/search?q=Wook-Ho%20Son"> Wook-Ho Son</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Since the advances in digital imaging technologies have led to development of high quality digital devices, there are a lot of illegal copies of copyrighted video content on the internet. Thus, we propose a high-quality (HQ) video watermarking scheme that can prevent these illegal copies from spreading out. The proposed scheme is applied spatial and temporal gradient methods to improve the fidelity and detection performance. Also, the scheme duplicates the watermark signal temporally to alleviate the signal reduction caused by geometric and signal-processing distortions. Experimental results show that the proposed scheme achieves better performance than previously proposed schemes and it has high fidelity. The proposed scheme can be used in broadcast monitoring or traitor tracking applications which need fast detection process to prevent illegally recorded video content from spreading out. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=editing%20prevention%20technique" title="editing prevention technique">editing prevention technique</a>, <a href="https://publications.waset.org/abstracts/search?q=gradient%20method" title=" gradient method"> gradient method</a>, <a href="https://publications.waset.org/abstracts/search?q=luminance%20change" title=" luminance change"> luminance change</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20watermarking" title=" video watermarking"> video watermarking</a> </p> <a href="https://publications.waset.org/abstracts/42072/anonymous-editing-prevention-technique-using-gradient-method-for-high-quality-video" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/42072.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">456</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3607</span> Digital Image Forensics: Discovering the History of Digital Images</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Gurinder%20Singh">Gurinder Singh</a>, <a href="https://publications.waset.org/abstracts/search?q=Kulbir%20Singh"> Kulbir Singh</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Digital multimedia contents such as image, video, and audio can be tampered easily due to the availability of powerful editing softwares. Multimedia forensics is devoted to analyze these contents by using various digital forensic techniques in order to validate their authenticity. Digital image forensics is dedicated to investigate the reliability of digital images by analyzing the integrity of data and by reconstructing the historical information of an image related to its acquisition phase. In this paper, a survey is carried out on the forgery detection by considering the most recent and promising digital image forensic techniques. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Computer%20Forensics" title="Computer Forensics">Computer Forensics</a>, <a href="https://publications.waset.org/abstracts/search?q=Multimedia%20Forensics" title=" Multimedia Forensics"> Multimedia Forensics</a>, <a href="https://publications.waset.org/abstracts/search?q=Image%20Ballistics" title=" Image Ballistics"> Image Ballistics</a>, <a href="https://publications.waset.org/abstracts/search?q=Camera%20Source%20Identification" title=" Camera Source Identification"> Camera Source Identification</a>, <a href="https://publications.waset.org/abstracts/search?q=Forgery%20Detection" title=" Forgery Detection"> Forgery Detection</a> </p> <a href="https://publications.waset.org/abstracts/76669/digital-image-forensics-discovering-the-history-of-digital-images" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/76669.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">246</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3606</span> Effectiveness of Video Interventions for Perpetrators of Domestic Violence</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Zeynep%20Turhan">Zeynep Turhan</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Digital tools can improve knowledge and awareness of strategies and skills for healthy and respectful intimate relationships. The website of the Healthy and Respectful Relationship Program has been developed and included five key videos about how to build healthy intimate relationships. This study examined the perspectives about informative videos by focusing on how individuals learn new information or challenge their preconceptions or attitudes regarding male privilege and women's oppression. Five individuals who received no-contact orders and attended group intervention were the sample of this study. The observation notes were the major methodology examining how participants responded to video tools. The data analysis method was the interpretative phenomenological analysis. The results showed that many participants found the tools useful in learning the types of violence and communication strategies. Nevertheless, obstacles to implementing some techniques were found in their relationships. These digital tools might enhance healthy and respectful relationships despite some limitations. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=healthy%20relationship" title="healthy relationship">healthy relationship</a>, <a href="https://publications.waset.org/abstracts/search?q=digital%20tools" title=" digital tools"> digital tools</a>, <a href="https://publications.waset.org/abstracts/search?q=intimate%20partner%20violence" title=" intimate partner violence"> intimate partner violence</a>, <a href="https://publications.waset.org/abstracts/search?q=perpetrators" title=" perpetrators"> perpetrators</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20interventions" title=" video interventions"> video interventions</a> </p> <a href="https://publications.waset.org/abstracts/165028/effectiveness-of-video-interventions-for-perpetrators-of-domestic-violence" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/165028.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">95</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3605</span> The Production, Negotiation and Resistance of Short Video Producers</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Cui%20Li">Cui Li</a>, <a href="https://publications.waset.org/abstracts/search?q=Xu%20Yuping"> Xu Yuping</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Based on the question of, "Are short video creators who are digital workers controlled by platform rules?" this study discusses the specific ways of platform rules control and the impact on short video creators. Based on the theory of digital labor, this paper adopts the method of in-depth interview and participant observation and chooses 24 producers of short video content of Tiktok to conduct in-depth interview. At the same time, through entering the short video creation field, the author carries on the four-month field investigation, obtains the creation process related data, and analyzes how the short video creator, as the digital labor, is controlled by the platform rule, as well as the creator in this process of compromise and resistance, a more comprehensive presentation of the short video creators of the labor process. It is found that the short video creators are controlled by the platform rules, mainly in the control of traffic rules, and the creators create content, compromise and resist under the guidance of traffic. First, while the platform seems to offer a flexible and autonomous way for creators to monetize, the threshold for participating in the event is actually very high for creators, and the rules for monetizing the event are vague. Under the influence of the flow rule, the creator is faced unstable incomes and high costs. Therefore, creators have to follow the rules of traffic to guide their own creation, began to flow-oriented content production, mainly reflected in the need to keep up-to-date, the pursuit of traffic to ride on the hot spots, in order to flow regardless, set up people "Born for the show", by the label solidified content creation. Secondly, the irregular working hours lead to the extension and overwork of the working hours, which leads to the internal friction of the short video creators at the spiritual level, and finally leads to the Rat Race of video creation. Thirdly, the video creator has completed the internalization and compromise of the platform rules in practice, which promotes the creator to continue to create independently, and forms the intrinsic motive force of the creator. Finally, the rule-controlled short video creators resist and fight in flexible ways, make use of the mechanism and rules of the platform to carry on the second creation, carry on the routine production, purchase the false flow, transfer the creation position to maintain own creation autonomy. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=short%20videos" title="short videos">short videos</a>, <a href="https://publications.waset.org/abstracts/search?q=tiktok" title=" tiktok"> tiktok</a>, <a href="https://publications.waset.org/abstracts/search?q=production" title=" production"> production</a>, <a href="https://publications.waset.org/abstracts/search?q=digital%20labors" title=" digital labors"> digital labors</a> </p> <a href="https://publications.waset.org/abstracts/167590/the-production-negotiation-and-resistance-of-short-video-producers" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/167590.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">60</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3604</span> Digital Musical Organology: The Audio Games: The Question of “A-Musicological” Interfaces</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Herv%C3%A9%20Z%C3%A9nouda">Hervé Zénouda</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This article seeks to shed light on an emerging creative field: "Audio games," at the crossroads between video games and computer music. Indeed, many applications, which propose entertaining audio-visual experiences with the objective of musical creation, are available today for different supports (game consoles, computers, cell phones). The originality of this field is the use of the gameplay of video games applied to music composition. Thus, composing music using interfaces but also cognitive logics that we qualify as "a-musicological" seem to us particularly interesting from the perspective of musical digital organology. This field raises questions about the representation of sound and musical structures and develops new instrumental gestures and strategies of musical composition. We will try in this article to define the characteristics of this field by highlighting some historical milestones (abstract cinema, game theory in music, actions, and graphic scores) as well as the novelties brought by digital technologies. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=audio-games" title="audio-games">audio-games</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20games" title=" video games"> video games</a>, <a href="https://publications.waset.org/abstracts/search?q=computer%20generated%20music" title=" computer generated music"> computer generated music</a>, <a href="https://publications.waset.org/abstracts/search?q=gameplay" title=" gameplay"> gameplay</a>, <a href="https://publications.waset.org/abstracts/search?q=interactivity" title=" interactivity"> interactivity</a>, <a href="https://publications.waset.org/abstracts/search?q=synesthesia" title=" synesthesia"> synesthesia</a>, <a href="https://publications.waset.org/abstracts/search?q=sound%20interfaces" title=" sound interfaces"> sound interfaces</a>, <a href="https://publications.waset.org/abstracts/search?q=relationships%20image%2Fsound" title=" relationships image/sound"> relationships image/sound</a>, <a href="https://publications.waset.org/abstracts/search?q=audiovisual%20music" title=" audiovisual music"> audiovisual music</a> </p> <a href="https://publications.waset.org/abstracts/152518/digital-musical-organology-the-audio-games-the-question-of-a-musicological-interfaces" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/152518.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">112</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3603</span> Texture-Based Image Forensics from Video Frame</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Li%20Zhou">Li Zhou</a>, <a href="https://publications.waset.org/abstracts/search?q=Yanmei%20Fang"> Yanmei Fang</a> </p> <p class="card-text"><strong>Abstract:</strong></p> With current technology, images and videos can be obtained more easily than ever. It is so easy to manipulate these digital multimedia information when obtained, and that the content or source of the image and video could be easily tampered. In this paper, we propose to identify the image and video frame by the texture-based approach, e.g. Markov Transition Probability (MTP), which is in space domain, DCT domain and DWT domain, respectively. In the experiment, image and video frame database is constructed, and is used to train and test the classifier Support Vector Machine (SVM). Experiment results show that the texture-based approach has good performance. In order to verify the experiment result, and testify the universality and robustness of algorithm, we build a random testing dataset, the random testing result is in keeping with above experiment. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=multimedia%20forensics" title="multimedia forensics">multimedia forensics</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20frame" title=" video frame"> video frame</a>, <a href="https://publications.waset.org/abstracts/search?q=LBP" title=" LBP"> LBP</a>, <a href="https://publications.waset.org/abstracts/search?q=MTP" title=" MTP"> MTP</a>, <a href="https://publications.waset.org/abstracts/search?q=SVM" title=" SVM"> SVM</a> </p> <a href="https://publications.waset.org/abstracts/42936/texture-based-image-forensics-from-video-frame" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/42936.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">427</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3602</span> Reliving Historical Events Using Augmented Reality Techniques</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Josep%20Domenech%20Mingot">Josep Domenech Mingot</a>, <a href="https://publications.waset.org/abstracts/search?q=Francisco%20Javier%20Esclapes%20Jover"> Francisco Javier Esclapes Jover</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The arrival of the age of information and new technologies allowed humanity to see what the future has in store, but occasionally it also brings the opportunity to look through a window to the past, an opportunity to relive history. This paper introduces a prototype of a digital system that lets us peek into our past making use of augmented reality technologies. A 3D scene will be modeled and animated based on an old image, depicting an event of historical significance. From this scene, a video will be rendered, recreating the events that were taking place at the time. Also, a smartphone app will be created. This app will detect the original image with the smartphone’s camera, overlay the rendered video so that it fully covers it and track the detected image, so that the overlaying video can keep covering the image. The recreation of Alicante’s Central Market bombing during the Spanish Civil War is presented as a case study. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=augmented%20reality" title="augmented reality">augmented reality</a>, <a href="https://publications.waset.org/abstracts/search?q=digital%20heritage" title=" digital heritage"> digital heritage</a>, <a href="https://publications.waset.org/abstracts/search?q=history" title=" history"> history</a>, <a href="https://publications.waset.org/abstracts/search?q=multimedia" title=" multimedia"> multimedia</a>, <a href="https://publications.waset.org/abstracts/search?q=smartphone" title=" smartphone"> smartphone</a> </p> <a href="https://publications.waset.org/abstracts/86915/reliving-historical-events-using-augmented-reality-techniques" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/86915.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">220</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3601</span> The Digital Video and Online Media Development for Integrated Marketing Communication and Tourism Promote in Taling Chan District, Bangkok</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Somsak%20Klaysung">Somsak Klaysung</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This study purpose to develop video to promote cultural tourism in Taling Chan District. For qualitative research, the sample size was 40 people from 5 group of the tourism entrepreneur in Taling Chan district, conducted the key informants’ content analysis by using focus group and structures in-depth interview from all stakeholders. Quota sampling was used for this kind of research. The findings indicated that get media video marketing and tourism contribute a set length 11.35 9 minutes there is plenty of social capital in Taling Chan District including detail like local wisdom, knowledge, and way of thinking related to nature, history, historic document, occupation, administration and attribute of local people. Additional research found the new path of travel through the water route according to Khlong Bang Ramat called Route 9 temples that travelers can travel by boat are available in the market in four areas Taling Chan also as well. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=digital%20video" title="digital video">digital video</a>, <a href="https://publications.waset.org/abstracts/search?q=integrated%20marketing%20communication" title=" integrated marketing communication"> integrated marketing communication</a>, <a href="https://publications.waset.org/abstracts/search?q=online%20media%20development" title=" online media development"> online media development</a>, <a href="https://publications.waset.org/abstracts/search?q=Taling%20Chan%20district" title=" Taling Chan district"> Taling Chan district</a> </p> <a href="https://publications.waset.org/abstracts/55288/the-digital-video-and-online-media-development-for-integrated-marketing-communication-and-tourism-promote-in-taling-chan-district-bangkok" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/55288.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">359</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3600</span> Detecting and Disabling Digital Cameras Using D3CIP Algorithm Based on Image Processing</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=S.%20Vignesh">S. Vignesh</a>, <a href="https://publications.waset.org/abstracts/search?q=K.%20S.%20Rangasamy"> K. S. Rangasamy</a> </p> <p class="card-text"><strong>Abstract:</strong></p> The paper deals with the device capable of detecting and disabling digital cameras. The system locates the camera and then neutralizes it. Every digital camera has an image sensor known as a CCD, which is retro-reflective and sends light back directly to its original source at the same angle. The device shines infrared LED light, which is invisible to the human eye, at a distance of about 20 feet. It then collects video of these reflections with a camcorder. Then the video of the reflections is transferred to a computer connected to the device, where it is sent through image processing algorithms that pick out infrared light bouncing back. Once the camera is detected, the device would project an invisible infrared laser into the camera's lens, thereby overexposing the photo and rendering it useless. Low levels of infrared laser neutralize digital cameras but are neither a health danger to humans nor a physical damage to cameras. We also discuss the simplified design of the above device that can used in theatres to prevent piracy. The domains being covered here are optics and image processing. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=CCD" title="CCD">CCD</a>, <a href="https://publications.waset.org/abstracts/search?q=optics" title=" optics"> optics</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20processing" title=" image processing"> image processing</a>, <a href="https://publications.waset.org/abstracts/search?q=D3CIP" title=" D3CIP"> D3CIP</a> </p> <a href="https://publications.waset.org/abstracts/1736/detecting-and-disabling-digital-cameras-using-d3cip-algorithm-based-on-image-processing" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/1736.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">357</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3599</span> Lecture Video Indexing and Retrieval Using Topic Keywords</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=B.%20J.%20Sandesh">B. J. Sandesh</a>, <a href="https://publications.waset.org/abstracts/search?q=Saurabha%20Jirgi"> Saurabha Jirgi</a>, <a href="https://publications.waset.org/abstracts/search?q=S.%20Vidya"> S. Vidya</a>, <a href="https://publications.waset.org/abstracts/search?q=Prakash%20Eljer"> Prakash Eljer</a>, <a href="https://publications.waset.org/abstracts/search?q=Gowri%20Srinivasa"> Gowri Srinivasa</a> </p> <p class="card-text"><strong>Abstract:</strong></p> In this paper, we propose a framework to help users to search and retrieve the portions in the lecture video of their interest. This is achieved by temporally segmenting and indexing the lecture video using the topic keywords. We use transcribed text from the video and documents relevant to the video topic extracted from the web for this purpose. The keywords for indexing are found by applying the non-negative matrix factorization (NMF) topic modeling techniques on the web documents. Our proposed technique first creates indices on the transcribed documents using the topic keywords, and these are mapped to the video to find the start and end time of the portions of the video for a particular topic. This time information is stored in the index table along with the topic keyword which is used to retrieve the specific portions of the video for the query provided by the users. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video%20indexing%20and%20retrieval" title="video indexing and retrieval">video indexing and retrieval</a>, <a href="https://publications.waset.org/abstracts/search?q=lecture%20videos" title=" lecture videos"> lecture videos</a>, <a href="https://publications.waset.org/abstracts/search?q=content%20based%20video%20search" title=" content based video search"> content based video search</a>, <a href="https://publications.waset.org/abstracts/search?q=multimodal%20indexing" title=" multimodal indexing"> multimodal indexing</a> </p> <a href="https://publications.waset.org/abstracts/77066/lecture-video-indexing-and-retrieval-using-topic-keywords" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/77066.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">250</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3598</span> The Chemistry in the Video Game No Man’s Sky</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Diogo%20Santos">Diogo Santos</a>, <a href="https://publications.waset.org/abstracts/search?q=Nelson%20Zagalo"> Nelson Zagalo</a>, <a href="https://publications.waset.org/abstracts/search?q=Carla%20Morais"> Carla Morais</a> </p> <p class="card-text"><strong>Abstract:</strong></p> No Man’s Sky (NMS) is a sci-fi video game about survival and exploration where players fly spaceships, search for elements, and use them to survive. NMS isn’t a serious game, and not all the science in the game is presented with scientific evidence. To find how players felt about the scientific content in the game and how they perceive the chemistry in it, a survey was sent to NMS’s players, from which were collected answers from 124 respondents from 23 countries. Chemophobia is still a phenomenon when chemistry or chemicals are a subject of discussion, but 68,9% of our respondents showed a positive attitude towards the presence of chemistry in NMS, with 57% stating that playing the video game motivated them to know more about science. 8% of the players stated that NMS often prompted conversations about the science in the video game between them and teachers, parents, or friends. These results give us ideas on how an entertainment game can potentially help scientists, educators, and science communicators reach a growing, evolving, vibrant, diverse, and demanding audience. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=digital%20games" title="digital games">digital games</a>, <a href="https://publications.waset.org/abstracts/search?q=science%20communication" title=" science communication"> science communication</a>, <a href="https://publications.waset.org/abstracts/search?q=chemistry" title=" chemistry"> chemistry</a>, <a href="https://publications.waset.org/abstracts/search?q=informal%20learning" title=" informal learning"> informal learning</a>, <a href="https://publications.waset.org/abstracts/search?q=No%20Man%E2%80%99s%20Sky" title=" No Man’s Sky"> No Man’s Sky</a> </p> <a href="https://publications.waset.org/abstracts/156533/the-chemistry-in-the-video-game-no-mans-sky" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/156533.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">110</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3597</span> Distributed Processing for Content Based Lecture Video Retrieval on Hadoop Framework</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=U.%20S.%20N.%20Raju">U. S. N. Raju</a>, <a href="https://publications.waset.org/abstracts/search?q=Kothuri%20Sai%20Kiran"> Kothuri Sai Kiran</a>, <a href="https://publications.waset.org/abstracts/search?q=Meena%20G.%20Kamal"> Meena G. Kamal</a>, <a href="https://publications.waset.org/abstracts/search?q=Vinay%20Nikhil%20Pabba"> Vinay Nikhil Pabba</a>, <a href="https://publications.waset.org/abstracts/search?q=Suresh%20Kanaparthi"> Suresh Kanaparthi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> There is huge amount of lecture video data available for public use, and many more lecture videos are being created and uploaded every day. Searching for videos on required topics from this huge database is a challenging task. Therefore, an efficient method for video retrieval is needed. An approach for automated video indexing and video search in large lecture video archives is presented. As the amount of video lecture data is huge, it is very inefficient to do the processing in a centralized computation framework. Hence, Hadoop Framework for distributed computing for Big Video Data is used. First, step in the process is automatic video segmentation and key-frame detection to offer a visual guideline for the video content navigation. In the next step, we extract textual metadata by applying video Optical Character Recognition (OCR) technology on key-frames. The OCR and detected slide text line types are adopted for keyword extraction, by which both video- and segment-level keywords are extracted for content-based video browsing and search. The performance of the indexing process can be improved for a large database by using distributed computing on Hadoop framework. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video%20lectures" title="video lectures">video lectures</a>, <a href="https://publications.waset.org/abstracts/search?q=big%20video%20data" title=" big video data"> big video data</a>, <a href="https://publications.waset.org/abstracts/search?q=video%20retrieval" title=" video retrieval"> video retrieval</a>, <a href="https://publications.waset.org/abstracts/search?q=hadoop" title=" hadoop"> hadoop</a> </p> <a href="https://publications.waset.org/abstracts/26648/distributed-processing-for-content-based-lecture-video-retrieval-on-hadoop-framework" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/26648.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">533</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3596</span> Towards an Analysis of Rhetoric of Digital Arabic Discourse</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Gameel%20Abdelmageed">Gameel Abdelmageed</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Arabs have a rhetorical heritage which has greatly contributed to the monitoring and analyzing of the rhetoric of the Holy Quran, <em>Hadith</em>, and Arabic texts on poetry and oratory. But Arab scholars - as far as the researcher knows – have not contributed to monitoring and analyzing the rhetoric of digital Arabic discourse although it has prominence, particularly in social media and has strong effectiveness in the political and social life of Arab society. This discourse has made its impact by using very new rhetorical techniques in language, voice, image, painting and video clips which are known as “Multimedia” and belong to “Digital Rhetoric”. This study suggests that it is time to draw the attention of Arab scholars and invite them to monitor and analyze the rhetoric of digital Arabic discourse. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=digital%20discourse" title="digital discourse">digital discourse</a>, <a href="https://publications.waset.org/abstracts/search?q=digital%20rhetoric" title=" digital rhetoric"> digital rhetoric</a>, <a href="https://publications.waset.org/abstracts/search?q=Facebook" title=" Facebook"> Facebook</a>, <a href="https://publications.waset.org/abstracts/search?q=social%20media" title=" social media"> social media</a> </p> <a href="https://publications.waset.org/abstracts/60172/towards-an-analysis-of-rhetoric-of-digital-arabic-discourse" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/60172.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">371</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3595</span> Virtual and Augmented Reality Based Heritage Gamification: Basilica of Smyrna in Turkey</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Tugba%20Saricaoglu">Tugba Saricaoglu</a> </p> <p class="card-text"><strong>Abstract:</strong></p> This study argues about the potential representation and interpretation of Basilica of Smyrna through gamification. Representation can be defined as a key which plays a role as a converter in order to provide interpretation of something according to the person who perceives. Representation of cultural heritage is a hypothetical and factual approach in terms of its sustainable conservation. Today, both site interpreters and public of cultural heritage have varying perspectives due to their different demographic, social, and even cultural backgrounds. Additionally, gamification application offers diversion of methods suchlike video games to improve user perspective of non-game platforms, contexts, and issues. Hence, cultural heritage and video game decided to be analyzed. Moreover, there are basically different ways of representation of cultural heritage such as digital, physical, and virtual methods in terms of conservation. Virtual reality (VR) and augmented reality (AR) technologies are two of the contemporary digital methods of heritage conservation. In this study, 3D documented ruins of the Basilica will be presented in the virtual and augmented reality based technology as a theoretical gamification sample. Also, this paper will focus on two sub-topics: First, evaluation of the video-game platforms applied to cultural heritage sites, and second, potentials of cultural heritage to be represented in video game platforms. The former will cover the analysis of some case(s) with regard to the concepts and representational aspects of cultural heritage. The latter will include the investigation of cultural heritage sites which carry such a potential and their sustainable conversation. Consequently, after mutual collection of information from cultural heritage and video game platforms, a perspective will be provided in terms of interpretation of representation of cultural heritage by sampling that on Basilica of Smyrna by using VR and AR based technologies. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=Basilica%20of%20Smyrna" title="Basilica of Smyrna">Basilica of Smyrna</a>, <a href="https://publications.waset.org/abstracts/search?q=cultural%20heritage" title=" cultural heritage"> cultural heritage</a>, <a href="https://publications.waset.org/abstracts/search?q=digital%20heritage" title=" digital heritage"> digital heritage</a>, <a href="https://publications.waset.org/abstracts/search?q=gamification" title=" gamification"> gamification</a> </p> <a href="https://publications.waset.org/abstracts/57712/virtual-and-augmented-reality-based-heritage-gamification-basilica-of-smyrna-in-turkey" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/57712.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">465</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3594</span> Video Stabilization Using Feature Point Matching</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Shamsundar%20Kulkarni">Shamsundar Kulkarni</a> </p> <p class="card-text"><strong>Abstract:</strong></p> Video capturing by non-professionals will lead to unanticipated effects. Such as image distortion, image blurring etc. Hence, many researchers study such drawbacks to enhance the quality of videos. In this paper, an algorithm is proposed to stabilize jittery videos .A stable output video will be attained without the effect of jitter which is caused due to shaking of handheld camera during video recording. Firstly, salient points from each frame from the input video are identified and processed followed by optimizing and stabilize the video. Optimization includes the quality of the video stabilization. This method has shown good result in terms of stabilization and it discarded distortion from the output videos recorded in different circumstances. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video%20stabilization" title="video stabilization">video stabilization</a>, <a href="https://publications.waset.org/abstracts/search?q=point%20feature%20matching" title=" point feature matching"> point feature matching</a>, <a href="https://publications.waset.org/abstracts/search?q=salient%20points" title=" salient points"> salient points</a>, <a href="https://publications.waset.org/abstracts/search?q=image%20quality%20measurement" title=" image quality measurement"> image quality measurement</a> </p> <a href="https://publications.waset.org/abstracts/57341/video-stabilization-using-feature-point-matching" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/57341.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">313</span> </span> </div> </div> <div class="card paper-listing mb-3 mt-3"> <h5 class="card-header" style="font-size:.9rem"><span class="badge badge-info">3593</span> Structural Analysis on the Composition of Video Game Virtual Spaces</h5> <div class="card-body"> <p class="card-text"><strong>Authors:</strong> <a href="https://publications.waset.org/abstracts/search?q=Qin%20Luofeng">Qin Luofeng</a>, <a href="https://publications.waset.org/abstracts/search?q=Shen%20Siqi"> Shen Siqi</a> </p> <p class="card-text"><strong>Abstract:</strong></p> For the 58 years since the first video game came into being, the video game industry is getting through an explosive evolution from then on. Video games exert great influence on society and become a reflection of public life to some extent. Video game virtual spaces are where activities are taking place like real spaces. And that’s the reason why some architects pay attention to video games. However, compared to the researches on the appearance of games, we observe a lack of theoretical comprehensive on the construction of video game virtual spaces. The research method of this paper is to collect literature and conduct theoretical research about the virtual space in video games firstly. And then analogizing the opinions on the space phenomena from the theory of literature and films. Finally, this paper proposes a three-layer framework for the construction of video game virtual spaces: “algorithmic space-narrative space players space”, which correspond to the exterior, expressive, affective parts of the game space. Also, we illustrate each sub-space according to numerous instances of published video games. Hoping this writing could promote the interactive development of video games and architecture. <p class="card-text"><strong>Keywords:</strong> <a href="https://publications.waset.org/abstracts/search?q=video%20game" title="video game">video game</a>, <a href="https://publications.waset.org/abstracts/search?q=virtual%20space" title=" virtual space"> virtual space</a>, <a href="https://publications.waset.org/abstracts/search?q=narrativity" title=" narrativity"> narrativity</a>, <a href="https://publications.waset.org/abstracts/search?q=social%20space" title=" social space"> social space</a>, <a href="https://publications.waset.org/abstracts/search?q=emotional%20connection" title=" emotional connection"> emotional connection</a> </p> <a href="https://publications.waset.org/abstracts/118519/structural-analysis-on-the-composition-of-video-game-virtual-spaces" class="btn btn-primary btn-sm">Procedia</a> <a href="https://publications.waset.org/abstracts/118519.pdf" target="_blank" class="btn btn-primary btn-sm">PDF</a> <span class="bg-info text-light px-1 py-1 float-right rounded"> Downloads <span class="badge badge-light">267</span> </span> </div> </div> <ul class="pagination"> <li class="page-item disabled"><span class="page-link">‹</span></li> <li class="page-item active"><span class="page-link">1</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=2">2</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=3">3</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=4">4</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=5">5</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=6">6</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=7">7</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=8">8</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=9">9</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=10">10</a></li> <li class="page-item disabled"><span class="page-link">...</span></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=120">120</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=121">121</a></li> <li class="page-item"><a class="page-link" href="https://publications.waset.org/abstracts/search?q=digital%20video&page=2" rel="next">›</a></li> </ul> </div> </main> <footer> <div id="infolinks" class="pt-3 pb-2"> <div class="container"> <div style="background-color:#f5f5f5;" class="p-3"> <div class="row"> <div class="col-md-2"> <ul class="list-unstyled"> About <li><a href="https://waset.org/page/support">About Us</a></li> <li><a href="https://waset.org/page/support#legal-information">Legal</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/WASET-16th-foundational-anniversary.pdf">WASET celebrates its 16th foundational anniversary</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Account <li><a href="https://waset.org/profile">My Account</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Explore <li><a href="https://waset.org/disciplines">Disciplines</a></li> <li><a href="https://waset.org/conferences">Conferences</a></li> <li><a href="https://waset.org/conference-programs">Conference Program</a></li> <li><a href="https://waset.org/committees">Committees</a></li> <li><a href="https://publications.waset.org">Publications</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Research <li><a href="https://publications.waset.org/abstracts">Abstracts</a></li> <li><a href="https://publications.waset.org">Periodicals</a></li> <li><a href="https://publications.waset.org/archive">Archive</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Open Science <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Philosophy.pdf">Open Science Philosophy</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Science-Award.pdf">Open Science Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Open-Society-Open-Science-and-Open-Innovation.pdf">Open Innovation</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Postdoctoral-Fellowship-Award.pdf">Postdoctoral Fellowship Award</a></li> <li><a target="_blank" rel="nofollow" href="https://publications.waset.org/static/files/Scholarly-Research-Review.pdf">Scholarly Research Review</a></li> </ul> </div> <div class="col-md-2"> <ul class="list-unstyled"> Support <li><a href="https://waset.org/page/support">Support</a></li> <li><a href="https://waset.org/profile/messages/create">Contact Us</a></li> <li><a href="https://waset.org/profile/messages/create">Report Abuse</a></li> </ul> </div> </div> </div> </div> </div> <div class="container text-center"> <hr style="margin-top:0;margin-bottom:.3rem;"> <a href="https://creativecommons.org/licenses/by/4.0/" target="_blank" class="text-muted small">Creative Commons Attribution 4.0 International License</a> <div id="copy" class="mt-2">© 2024 World Academy of Science, Engineering and Technology</div> </div> </footer> <a href="javascript:" id="return-to-top"><i class="fas fa-arrow-up"></i></a> <div class="modal" id="modal-template"> <div class="modal-dialog"> <div class="modal-content"> <div class="row m-0 mt-1"> <div class="col-md-12"> <button type="button" class="close" data-dismiss="modal" aria-label="Close"><span aria-hidden="true">×</span></button> </div> </div> <div class="modal-body"></div> </div> </div> </div> <script src="https://cdn.waset.org/static/plugins/jquery-3.3.1.min.js"></script> <script src="https://cdn.waset.org/static/plugins/bootstrap-4.2.1/js/bootstrap.bundle.min.js"></script> <script src="https://cdn.waset.org/static/js/site.js?v=150220211556"></script> <script> jQuery(document).ready(function() { /*jQuery.get("https://publications.waset.org/xhr/user-menu", function (response) { jQuery('#mainNavMenu').append(response); });*/ jQuery.get({ url: "https://publications.waset.org/xhr/user-menu", cache: false }).then(function(response){ jQuery('#mainNavMenu').append(response); }); }); </script> </body> </html>