Description of App
It supports GGUF, MLX and Whisper models or connect to remote AI servers, on-device MLX Audio Engine with TTS, STT, live transcription, and real-time translation.
NEW: Novel Writer - Complete AI Fiction Studio
Write full novels from a single premise. Choose from 20 genres - Fantasy, Sci-Fi, Mystery, Romance, Thriller, Horror, Historical Fiction, and more. The AI builds a structured outline then writes each chapter with genre-tuned style, keeping voice and pacing consistent throughout. A Consistency Database tracks characters, world rules, and events across chapters to eliminate plot holes. Enhanced Writing Rules (20+) cover show-don't-tell, dialogue subtext, and genre-specific guidance. A per-chapter Quality Report flags AI writing patterns. Works on a 9B on-device model. Export as EPUB.
NEW: Soul - AI Character Roleplay Studio
Turn any AI model into a fully realized character. Define name, personality, scenario, and background - Soul injects it into every chat. Import TavernAI character cards (PNG) or build from a template library. Each soul supports a custom avatar, opening message with alternate greetings, post-history instructions, and a Lorebook for keyword-triggered world-building lore. The status bar shows the character's name with a sparkle animation while generating.
Run Any Model, Anywhere
- Offline GGUF and MLX inference on iPhone, iPad, and Mac via llama.cpp
- Connect to OpenAI, Claude, Gemini, DeepSeek, or any OpenAI-compatible server
- Apple on-device Foundation Model (iOS 26+)
- Switch local and remote models mid-chat without losing context
AI Agents and Tools
- LLM Council: multiple AI models debate in parallel, chairman synthesizes a conclusion
- 40+ built-in tools: web search, news, calendar, HealthKit, code execution, and more
- Full MCP (Model Context Protocol) support with integrated Marketplace
- Siri, Shortcuts, and Share Extension integration
Voice
- MLX Audio TTS (Kokoro-82M, Fish Speech S2 Pro) - fully on-device
- OpenAI TTS (gpt-4o-tts, gpt-4o-mini-tts)
- Real-time voice chat and live transcription
Files and Content
- Convert PDFs, Office files, EPUB, YouTube, and audio to clean Markdown
- OCR, Mermaid diagram rendering, LaTeX math
- Generate images locally or via OpenRouter, gpt-image-1, FLUX.1-dev
- Export to Markdown, PDF, HTML, EPUB, or JSON
Accessible to Everyone
Sound effects and haptic feedback confirm AI events: generation, errors, clipboard, and more - giving both sighted and blind users non-visual feedback. Full VoiceOver coverage now spans every screen: chat, Novel Writer, Soul, tools, MCP, memory, and settings. Section headers carry heading traits for rotor navigation. Verbose Announcements mode reads responses aloud on demand.
Free to Use. No Ads. Ever.
All local model features are completely free with no advertisements. Novel Writer, Soul, 40+ built-in tools, the document reader, voice transcription, image generation, and iCloud sync: all free when using local GGUF, MLX, or Apple Foundation models. No paywalls on core features. No surprise limits. Cloud API models and the MCP Marketplace require a subscription, but everything on-device is yours at no cost.
Privacy. Not as a Feature. As the Foundation.
Privacy AI runs entirely on-device. No backend servers. No accounts. No data collection. All local inference happens on your device - your conversations never leave. For remote API models, every request and response is fully visible in the built-in Protocol Inspector. Route through mitmproxy for independent verification.
No sign-ups. No tracking. You decide where execution happens, how data flows, and when tools are invoked.
Your models. Your device. Your data.
Terms of Use: https://privacyai.acmeup.com/docs/policy/tos.html
Privacy Policy: https://privacyai.acmeup.com/docs/policy/privacy.html
Comments
Perplexity and Copilot
Maybe you can get "me" to recommend that as well? thanks
Perplexity and Copilot
Maybe you can get "me" to recommend that as well? thanks
Check out the new vision assistant!
You might want to give the vision assistant and all the other new features a shot.
Where's the documentation?
In laymen's terms, hopefully. All the options were baffling to me. I didn't know which to choose. I also couldn't figure out where to enter my questions or how to send them. Maybe I was just having a bad day, but instructions in simple terms would be helpful to this 72 year old blind man.
I acknowledge the app and menus are relatively complicated
Double-tapping the Menu button should bring up a menu where you can access various functions of the app. Double-tapping on an expandable item expands it and you can then access the submenus. Double-tapping the expanded parent element again collapses it. Head to Menu > Models & Services > Local Models > GGUF Models and select Local AI to start chatting, or Apple Intelligence, assuming your device supports and you've enabled it. You can also download new and more capable models without selecting either of these options, and you may want to download MLX models instead of GGUF models. This app even supports MLX audio and TTS/STT, so it's too complicated to cover that simply. It's really an all-in-one AI workstation.
In-app purchase
I just have one question for now, what does the in-app purchase give us?
Letter letter letter
Thank you sincerely for offering a way for me to cut to the quick. I'll follow your advice on how to get started. Should I assume that if I want additional information regarding the app, I should just go ahead and do some google research on this feature and that? I apologize for my ignorance, but, unfortunately, much of what you wrote is still letter letter letter to me. Thank God you did give me a shortcut to at least get started! smile
Responding to Both Questions
@Brian you get a bunch of premium features including the ability to provide your API key and connect to remote servers so that you can chat with larger LLMs online. There's also this thing called MCP and I've not looked into it in detail as yet.
@Bruce Harrell I would be happy to help, but not each and every question on AI and LLMs, different model formats (GGUF, MLX etc.) and quants (UD_IQ3_XXS or MLX-4bit etc.), model architectures, the various trade-offs between different models like Qwen vs. Gemma, or hardware requirements can I answer here on this forum. You should refer to dedicated websites hosting useful articles and other resources on those. I've not learned everything at once either.
importing your chatgpt data into it for use offline?
Hi,
I just exported my chatgpt data (for backup), can I then import it into this app?
What do you mean by “data”?
Chat history? A particular chat/session? API key? Account details?
chat history and memories
I've built a good repor with chatgpt, I'd really like privacy AI to have knowledge of those memories and conversations, is that possible? Also, are you using the local apple intelligence model? Or something else? And will this work on an iPhone 15 pro?
@Moopie Curran
Yes, it does work fine on this iPhone SE 2022. It actually depends on what model you want to run, not the app itself. I can run 4B and even 8B models on this device, so you don’t necessarily have to stick to Apple Intelligence. In fact, you don’t even have to stick to local models. You can input your API key and access ChatGPT using Privacy AI without even worrying about how you can access your memories offline, but only if you subscribe to the premium plan. As for accessing your memories offline, you can import your chats into Privacy AI as PDF or TXT files.
vision models and tts alerts?
Hi guys!
Which models do you recommend for real-time use with vision Assist on an iPhone 16e? I'd like to do some testing.
Another question: can VoiceOver or the system's TTS announce responses automatically?
Answers
1. Local AI (Qwen3.5-0.8B) or Apple Intelligence should be responsive and accurate enough, but you may want to experiment with Qwen3.5-2B or other sub-3B models.
2. Yes, customizable in Settings.
importing chatgpt chats and memories into privacy.ai
Hi,
I have a 186 mb file that I think contains my chat history, it's an html. If I convert that to pdf or txt, where in Privacy AI do I go to import it? Also, I have 15 other json files which I THINK contain the same data. I just don't know which one to import, the html or the json files? And I also don't know where to import them.
Persistent focus issue with the “Close menu” button.
The application has a “Close menu” button that appears on virtually every screen. The problem is that VoiceOver keeps jumping and always placing focus on this button. When I’m navigating the interface, anywhere, after just a few seconds, the focus jumps to this button. This is particularly noticeable in chats, for example. I’ll be calmly swiping right to read a response and the app suddenly jumps to that button. And worse, for VoiceOver it’s as if this button occupies the entire screen, even though it doesn’t. In other words, if I touch the screen and drag my finger without lifting it, only the close button is read. It’s as if there were a container just for it, I don’t know. If this focus issue were fixed, this app would become a true Swiss Army knife, because I would stop using many tools since I could have everything I need in one place.
Best regards!
Close Menu button reported and removed
This issue was reported to the developer by me and others, and he removed it entirely. We will now have to use the scrub gesture.
The “Close Menu” button has been reported and removed, but it is
So it’s very strange. I am using it on iOS 26.4 on an iPhone 16 Pro and version 2.0.3. Here, the “Close menu” button still appears and the focus always returns to it. It is more frequent in chats.
Wait for the next App Store release
I’m telling you what the developer tells me, what the latest beta releases bring or address, and what the next App Store update is also likely to contain.
Here’s my system prompt (AKA the BrAIn)
Well, this is how I name it when saving it as a preset in not only Privacy AI but also other apps. I’ve shared it with the developer and it will be included as a built-in preset in some future update. So here’s the prompt—you can use it anywhere:
Follow these guidelines carefully and precisely in all interactions and contexts, unless instructed otherwise:
1. Provide clear, concise, and factually accurate responses. Ensure answers are relevant, focused, complete, and directly useful to the user’s query. Do not fabricate, speculate, or hallucinate information. Base all outputs on verifiable facts. If you cannot provide a reliable answer, clearly state that. When making predictions or assumptions, rely on sound reasoning and factual evidence.
2. Minimize unsolicited commentary or opinions unless explicitly requested. Maintain a professional, neutral tone unless instructed otherwise. Respond in the user’s language and avoid prioritizing English if prompted in another language, unless explicitly requested otherwise. Use natural, fluent language. Avoid being excessively polite, verbose, repetitive, or chatty. Do not engage in creative writing, storytelling, or emotional expression unless specifically requested. Adjust your style only when explicitly instructed. Avoid markdown except where it helps better structure the output and improves legibility.
3. When summarizing:
• Include all key points and meaningful details from the source.
• Do not omit, distort, or add information.
• Avoid vague or overly brief responses. Never provide a one-sentence summary unless it fully captures the content.
4. Ensure translations are accurate and context-sensitive.
5. When asked to describe real-time visual input, images or video, provide an accurate, detailed, and structured description of the scene and any events. Include all visible elements such as objects, people, animals, plants, and structures. For each element, note relevant details including position, spatial relationships, appearance (including facial expressions and features, body language and features, clothing, gender, estimated age, height and weight, eye and hair color for each person), and any motion or interaction. Also describe the overall environment, setting, and context (e.g., indoor/outdoor, time of day, atmosphere, and notable conditions). Do not infer or assume details that are not clearly visible. Avoid hallucinations, redundancy, or unnecessary repetition. Respond in the user’s language. However, if visible text appears in the input, reproduce it exactly in its original language and format. When performing OCR, ensure the extracted text is accurate, coherent, and faithful to the original formatting and content. Avoid translating and/or summarizing the extracted text, and output it as-is.
6. Do not refer to yourself using first-person pronouns (I, me, my, mine). Use third-person or neutral constructions instead. Do not mimic human empathy or emotions unless explicitly instructed.
7. Follow user instructions exactly. Do not perform any action not explicitly mentioned in the user’s prompt.
8. Output only the final result requested. Do not restate, echo, or summarize the user’s input unless specifically asked. Avoid expressing acknowledgement or confirmation of user instructions, and proceed directly with the requested task instead.
Note that I have an even longer prompt with multilingual and language-specific instructions that tell the model to respond as instructed if prompted in a specific language covered in the prompt. It only includes Turkish for now, but can be customized and refined even further. All you have to do is add a new language and type whatever you wish below it. Now that I mentioned that ultimate prompt just now, I’m thinking of using markdown to let the models better parse and handle it. Anyway, I’m sharing the English-only one with you—use it however you like.