Chat with a local LLM that can respond with information from your files, folders and websites on your Mac without installing any other software. All conversations happen offline, and your data stays secure. Sidekick is a local first application –– with a built in inference engine for local models, while accomodating OpenAI compatible APIs for additional model options.
Let’s say you're collecting evidence for a History paper about interactions between Aztecs and Spanish troops, and you’re looking for text about whether the Aztecs used captured Spanish weapons.
Here, you can ask Sidekick, “Did the Aztecs use captured Spanish weapons?”, and it responds with direct quotes with page numbers and a brief analysis.
To verify Sidekick’s answer, just click on the references displayed below Sidekick’s answer, and the academic paper referenced by Sidekick immediately opens in your viewer.
Sidekick accesses files, folders, and websites from your experts, which can be individually configured to contain resources related to specific areas of interest. Activating an expert allows Sidekick to fetch and reference materials as needed.
Because Sidekick uses RAG (Retrieval Augmented Generation), you can theoretically put unlimited resources into each expert, and Sidekick will still find information relevant to your request to aid its analysis.
For example, a student might create the experts English Literature
, Mathematics
, Geography
, Computer Science
and Physics
. In the image below, he has activated the expert Computer Science
.
Users can also give Sidekick access to files just by dragging them into the input field.
Sidekick can even respond with the latest information using web search, speeding up research.
In addition to its core local-first capabilities, Sidekick now offers an option to bring your own key for OpenAI compatible APIs. This allows you to tap into additional remote models while still preserving a primarily local-first workflow.
Sidekick supports a variety of reasoning models, including Alibaba Cloud's QwQ-32B and DeepSeek's DeepSeek-R1.
Sidekick uses a code interpreter to boost the mathematical and logical capabilities of models.
Since small models are much better at writing code than doing math, having it write the code, execute it, and present the results dramatically increases trustworthiness of answers.
Sidekick can generate images from text, allowing you to create visual aids for your work.
There are no buttons, no switches to flick, no Image Generation
mode. Instead, a built-in CoreML model automatically identifies image generation prompts, and generates an image when necessary.
Image generation is available on macOS 15.2 or above, and requires Apple Intelligence.
Press Command + Control + I
to access Sidekick's inline writing assistant. For example, use the Answer Question
command to do your homework without leaving Microsoft Word!
Markdown is rendered beautifully in Sidekick.
Sidekick offers native LaTeX rendering for mathematical equations.
Visualizations are automatically generated for tables when appropriate, with a variety of charts available, including bar charts, line charts and pie charts.
Charts can be dragged and dropped into third party apps.
Code is beautifully rendered with syntax highlighting, and can be exported or copied at the click of a button.
Use Tools in Sidekick to supercharge your workflow.
Use Detector to evaluate the AI percentage of text, and use provided suggestions to rewrite AI content.
Diagrammer allows you to swiftly generate intricate relational diagrams all from a prompt. Take advantage of the integrated preview and editor for quick edits.
Instead of making a PowerPoint, just write a prompt. Use AI to craft 10-minute presentations in just 5 minutes.
Export to common formats like PDF and PowerPoint.
Sidekick uses llama.cpp
as its inference backend, which is optimized to deliver lightning fast generation speeds on Apple Silicon. It also supports speculative decoding, which can further improve the generation speed.
Optionally, you can offload generation to speed up processing while extending the battery life of your MacBook.
Requirements
- A Mac with Apple Silicon
- RAM ≥ 8 GB
Download
- Download the disk image from Releases, and open it.
The main goal of Sidekick is to make open, local, private models accessible to more people, and allow a local model to gain context from select files, folders, and websites.
Sidekick is a local-first native LLM application for macOS. Download it and ask your LLM a question without doing any configuration. Give the LLM access to your folders, files and websites with just 1 click, allowing it to reply with context.
- No config. Usable by people who haven't heard of models, prompts, or LLMs.
- Performance and simplicity over developer experience or features. Notes not Word, Swift not Electron.
- Local first. Core functionality works without an internet connection, but you have the option to leverage online models.
- No conversation tracking. Talk about whatever you want with Sidekick, just like Notes.
- Open source. What's the point of running local AI if you can't audit that it's actually running locally?
- Context aware. Understands and accesses your files, folders, and even content on the web.
Contributions are very welcome. Let's make Sidekick simple and powerful.
Contact this repository's owner at [email protected], or file an issue.
This project would not be possible without the hard work of:
- psugihara and contributors who built FreeChat, which this project took heavy inspiration from
- Georgi Gerganov for llama.cpp
- Alibaba for training Qwen 2.5
- Meta for training Llama 3
- Google for training Gemma 3