- This repository contains llama.cpp based offline android chat application cloned from llama.cpp android example. Install, download model and run completely offline privately.
- The app supports downloading GGUF models from Hugging Face and offers customizable parameters for flexible use.
- Being open-source, it allows for easy modifications and improvements, providing a secure, private, and fully offline experience.
-
Get IRIS (offline GPT) on Google Play:
- Go to releases : https://github.com/nerve-sparks/iris_android/releases
- Download app
- Install app
- Works Offline: Access all features without needing an internet connection.
- Privacy-Focused: All data is processed securely on your device.
- Expandable Models: Download external GGUF models from Hugging Face.
- Open Source: Fully transparent development.
- Customizable Parameters: n_threads, top_k, top_p, and temperature can be adjusted to optimize performance and behavior based on device capabilities and desired output.
- Text To Speech: Support for Text-to-Speech functionality.
- Speech To Text: Support for Speech-to-Text functionality.
- Default Model Selection: Set a default model to load automatically on startup.
The performance of Iris is directly influenced by the size, speed, and compute requirements of the models you use. These factors also impact the overall user experience. For a faster and smoother experience, consider downloading smaller models.
Example Recommendation:
On opening the app, users can download suggested models to optimize performance based on their preferences and device capabilities.
- Smaller models are ideal for quicker interactions but may compromise slightly on response quality.
- Larger models offer more comprehensive responses but require higher compute power and may result in slower speeds.
- Choose a model that best balances speed and quality for your specific use case.
- Iris may produce inaccurate results depending on the complexity of queries and model limitations.
- Performance and accuracy are influenced by the size and type of model selected.
- Download Android studio
- Clone this repository and import into Android Studio
git clone https://github.com/nerve-sparks/iris_android.git
- Clone the llama.cpp repository in the same folder as iris_android
git clone https://github.com/ggerganov/llama.cpp
- Navigate to the llama.cpp directory and checkout a specific commit for proper compatibility:
cd llama.cpp
git checkout 1f922254f0c984a8fb9fbaa0c390d7ffae49aedb
cd ..
- Open developer options on the Android mobile phone.
- Enable developer options.
- Click on developer options and enable wireless debugging.
- In Android Studio, select the drop down on the left side of the 'app' button on the Navbar.
- Select on 'Pair devices using Wi-Fi'. A QR code appears on screen.
- Click on wireless debugging on Android phone. Select 'Pair device with QR code'. Scan the code. (Make sure both devices are on the same Wi-fi)
- You can use Usb Debugging also to connect your phone.
- Once the phone is connected, select the device name in the drop down menu and click on play button.
- In the app, download at least one model from the given options.
- Now you can run the app offline. (In airplane mode as well)
- Fork the repository.
- Create a new feature branch:
git checkout -b my-new-feature
- Commit your changes:
git commit -m 'Add some feature'
- Push your branch to the remote repository:
git push origin my-new-feature
- Open a Pull Request.
- Visit www.nervesparks.com to contact us.