Researchers evaluated the bots’ performance on EyeQuiz, a platform containing ophthalmology board certification examination practice questions.
They examined the bots’ accuracy, response length, response time and provision of explanations as well as subspecialty-specific performance.
Overall, Google Gemini and Bard both had accuracies of 71% across 150 text-based multiple-choice questions.
Both bots had an acceptable performance on the exam questions, and chatbots also tended to provide a confident explanation even when providing an incorrect answer.
At the Becker's 23rd Annual Spine, Orthopedic and Pain Management-Driven ASC + The Future of Spine Conference, taking place June 11-13 in Chicago, spine surgeons, orthopedic leaders and ASC executives will come together to explore minimally invasive techniques, ASC growth strategies and innovations shaping the future of outpatient spine care. Apply for complimentary registration now.
