Demo Video Link : https://drive.google.com/file/d/1bDbyg3gjAWFnHNr5yscC28rUvxkyczUH/view
Excited to share a major milestone! We've built an innovative iOS application that combines the power of OCR and ChatGPT to respond to image queries/questions with incredible speed and accuracy.
- Text Recognition: Utilizes Vision Kit for seamless text recognition from images.
- Intelligent Responses: Integrated ChatGPT API to deliver intelligent responses in under 13 seconds.
- Accessibility: Responses are read aloud using Apple’s AVSpeechSynthesis, enhancing accessibility and user experience.
Vision Kit is a powerful framework provided by Apple for text and image analysis. It offers robust tools for detecting and recognizing text within images, which is essential for OCR (Optical Character Recognition) functionality. With Vision Kit, our app can accurately and efficiently extract text from images captured by the user, enabling a wide range of applications from document scanning to real-time translation.
Apple’s AVSpeechSynthesis is a part of the AVFoundation framework that provides a text-to-speech (TTS) engine, enabling applications to generate spoken audio from text. This feature helps create more accessible and interactive applications by reading out content aloud. It offers customizable voices, adjustable speech parameters, and multilingual support, allowing for a personalized and engaging user experience.
- Azure-hosted Python Server: Ensures efficient data processing and streamlined performance. By hosting our backend on Azure, we leverage Infrastructure as a Service (IaaS) to handle the dynamic scaling and robust performance needed for our application's needs.
To get a local copy up and running, follow these simple steps.
- Xcode
- CocoaPods
- An Azure account for the backend server
-
Clone the repo:
git clone https://github.com/DeepBhupatkar/QueryBox-AI-Powered.git
-
Install CocoaPods dependencies:
cd QueryBox-AI-Powered pod install
-
Open the project in Xcode:
open QueryBox.xcworkspace
-
Set up your Azure backend and configure the API endpoints in your project.
- Run the app on the simulator or a physical device.
- Use the app to capture text from images and receive audio responses through AVSpeechSynthesis.
Contributions are what make the open-source community such an amazing place to learn, inspire, and create. Any contributions you make are greatly appreciated.
- Fork the Project
- Create your Feature Branch (`git checkout -b feature/AmazingFeature`)
- Commit your Changes (`git commit -m 'Add some AmazingFeature'`)
- Push to the Branch (`git push origin feature/AmazingFeature`)
- Open a Pull Request
Thank You For Checking out the QueryBox AI Powered iOS Application