Article
Captcha Solving using AI with LIME Explainability
International Journal of Technology & Emerging Research · Published 24 Jul 2025
International Journal of Technology & Emerging Research / Archives
Authors
Sanapala Joshika, Setti Sarika
Sanapala Joshika
Setti Sarika
Abstract
This paper presents an intelligent system for solving CAPTCHA challenges using Artificial Intelligence (AI) integrated with explainable frameworks. CAPTCHAs, designed to differentiate humans from bots, often pose accessibility and usability issues. To address this, we developed a deep learning model capable of accurately recognizing and solving both alphanumeric and math-based text CAPTCHAs. The model utilizes Convolutional Neural Networks (CNNs) for image-based text recognition, trained on synthetically generated CAPTCHA datasets. To enhance transparency and trust in AI predictions, the system incorporates LIME (Local Interpretable Model-agnostic Explanations), which visually explains each character prediction by highlighting important regions of the CAPTCHA image. This interpretability aids developers in validating the model's decisions and ensures robustness against adversarial inputs. The system aims to balance accuracy, security, and explainability, making it suitable for real-world applications where both user experience and AI accountability are critical.
Keywords: Artificial Intelligence (AI), CAPTCHA, Deep Learning, Convolutional Neural Networks (CNN), Explainable AI (XAI), LIME, Model Interpretability