A Testing Framework for AI Linguistic Systems (testFAILS)
Document Type
Conference Proceeding
Publication Date
1-1-2023
Abstract
This paper introduces testFAILS, an innovative testing framework designed for the rigorous evaluation of AI Linguistic Systems, with a particular emphasis on various iterations of ChatGPT. Leveraging orthogonal array coverage, this framework provides a robust mechanism for assessing AI systems, addressing the critical question, 'How should we evaluate AI?' While the Turing test has traditionally been the benchmark for AI evaluation, we argue that current publicly available chatbots, despite their rapid advancements, have yet to meet this standard. However, the pace of progress suggests that achieving Turing test-level performance may be imminent. In the interim, the need for effective AI evaluation and testing methodologies remains paramount. Our research, which is ongoing, has already validated several versions of ChatGPT, and we are currently conducting comprehensive testing on the latest models, including ChatGPT-4, Bard, Bing Bot, and the LLaMA model. The testFAILS framework is designed to be adaptable, ready to evaluate new bot versions as they are released. Additionally, we have tested available chatbot APIs and developed our own application, AIDoctor, utilizing the ChatGPT-4 model and Microsoft Azure AI technologies.
Publication Title
Proceedings - 5th IEEE International Conference on Artificial Intelligence Testing, AITest 2023
First Page Number
51
Last Page Number
54
DOI
10.1109/AITest58265.2023.00017
Recommended Citation
Kumar, Y.; Morreale, P.; Sorial, P.; Delgado, J.; Li, J. Jenny; and Martins, P., "A Testing Framework for AI Linguistic Systems (testFAILS)" (2023). Kean Publications. 284.
https://digitalcommons.kean.edu/keanpublications/284