CINXE.COM

Evaluating Generative Neural Attention WeightsBased Chatbot on Customer Support Twitter Dataset

<?xml version="1.0" encoding="UTF-8"?> <article key="pdf/10013687" mdate="2024-07-02 00:00:00"> <author>Sinarwati Mohamad Suhaili and Naomie Salim and Mohamad Nazim Jambli</author> <title>Evaluating Generative Neural Attention WeightsBased Chatbot on Customer Support Twitter Dataset</title> <pages>346 - 352</pages> <year>2024</year> <volume>18</volume> <number>7</number> <journal>International Journal of Computer and Systems Engineering</journal> <ee>https://publications.waset.org/pdf/10013687</ee> <url>https://publications.waset.org/vol/211</url> <publisher>World Academy of Science, Engineering and Technology</publisher> <abstract>Sequencetosequence (seq2seq) models augmented with attention mechanisms are increasingly important in automated customer service. These models, adept at recognizing complex relationships between input and output sequences, are essential for optimizing chatbot responses. Central to these mechanisms are neural attention weights that determine the model&amp;rsquo;s focus during sequence generation. Despite their widespread use, there remains a gap in the comparative analysis of different attention weighting functions within seq2seq models, particularly in the context of chatbots utilizing the Customer Support Twitter (CST) dataset. This study addresses this gap by evaluating four distinct attentionscoring functions&amp;mdash;dot, multiplicativegeneral, additive, and an extended multiplicative function with a tanh activation parameter &amp;mdash; in neural generative seq2seq models. Using the CST dataset, these models were trained and evaluated over 10 epochs with the AdamW optimizer. Evaluation criteria included validation loss and BLEU scores implemented under both greedy and beam search strategies with a beam size of k 3. Results indicate that the model with the tanhaugmented multiplicative function significantly outperforms its counterparts, achieving the lowest validation loss (1.136484) and the highest BLEU scores (0.438926 under greedy search, 0.443000 under beam search, k 3). These findings emphasize the crucial influence of selecting an appropriate attentionscoring function to enhance the performance of seq2seq models for chatbots, particularly highlighting the model integrating tanh activation as a promising approach to improving chatbot quality in customer support contexts.</abstract> <index>Open Science Index 211, 2024</index> </article>