TY - GEN
T1 - A Comparative Study of Intent Classification Performance in Truncated Consumer Communication using GPT-Neo and GPT-2
AU - Hirway, Chanda
AU - Fallon, Enda
AU - Connolly, Paul
AU - Flanagan, Kieran
AU - Yadav, Deepak
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - This study presents a comparative analysis of intent classification performance using two widely used language models, GPT-Neo and GPT-2, in the context of truncated consumer communications. Generative Pre-Trained Transformer (GPT) is a machine learning technique that has revolutionized the field of natural language processing (NLP). GPT uses a transformer-based neural network architecture that is pre-Trained on large volumes of data to generate highly accurate and versatile NLP models capable of performing various tasks, such as language translation, question-Answering, and text summarization. GPT's ability to generate natural language responses that closely resemble those of humans has greatly enhanced the potential of language processing in the future. The data used in this study was provided by Circana. This data is an essential resource as it includes real world consumer purchases, promotional and e-commerce activity from multiple retail markets, as well as related consumer messaging. The model is trained, evaluated and analyzed on their respective accuracies, precision, recall, and F1 scores. The results indicate that GPT-Neo with BCELoss function outperformed GPT-2 in terms of accuracy and F1 score, with a significant improvement in the reduction of false negative values. These findings demonstrate the potential of GPT-Neo for improving intent classification in consumer communication applications with limited text input.
AB - This study presents a comparative analysis of intent classification performance using two widely used language models, GPT-Neo and GPT-2, in the context of truncated consumer communications. Generative Pre-Trained Transformer (GPT) is a machine learning technique that has revolutionized the field of natural language processing (NLP). GPT uses a transformer-based neural network architecture that is pre-Trained on large volumes of data to generate highly accurate and versatile NLP models capable of performing various tasks, such as language translation, question-Answering, and text summarization. GPT's ability to generate natural language responses that closely resemble those of humans has greatly enhanced the potential of language processing in the future. The data used in this study was provided by Circana. This data is an essential resource as it includes real world consumer purchases, promotional and e-commerce activity from multiple retail markets, as well as related consumer messaging. The model is trained, evaluated and analyzed on their respective accuracies, precision, recall, and F1 scores. The results indicate that GPT-Neo with BCELoss function outperformed GPT-2 in terms of accuracy and F1 score, with a significant improvement in the reduction of false negative values. These findings demonstrate the potential of GPT-Neo for improving intent classification in consumer communication applications with limited text input.
KW - Consumer Communication
KW - GPT Neo
KW - GPT-2
KW - Intent Classification
KW - Natural Language Processing (NLP)
UR - http://www.scopus.com/inward/record.url?scp=85180771120&partnerID=8YFLogxK
U2 - 10.1109/ICETCI58599.2023.10331337
DO - 10.1109/ICETCI58599.2023.10331337
M3 - Conference contribution
AN - SCOPUS:85180771120
SN - 9798350300604
T3 - Proceedings of the 2023 International Conference on Emerging Techniques in Computational Intelligence, ICETCI 2023
SP - 97
EP - 104
BT - Proceedings of the 2023 International Conference on Emerging Techniques in Computational Intelligence, ICETCI 2023
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 3rd International Conference on Emerging Techniques in Computational Intelligence, ICETCI 2023
Y2 - 21 September 2023 through 23 September 2023
ER -