lilbee
unlistedby tobocop2
Search, chat, and sync your local knowledge base powered by lilbee
lilbee for Obsidian
Beta — feedback and bug reports welcome. Open an issue.
If you delete a file from your vault, it will still show up in search results. Removing deleted files from the index is coming soon.
Chat with your documents privately, entirely on your own machine. Ask questions about your notes, PDFs, code, spreadsheets, and images — and get answers grounded in what you've actually written, with source citations. Save conversations back to your vault as markdown. No cloud services, no API keys, no data leaves your computer.
Demo
Scanned PDF → vision OCR → chat (click to expand)
Attaching a scanned 1998 Star Wars: X-Wing Collector's Edition manual (PDF with no extractable text), indexing it with vision OCR, and chatting about the dev team credits — entirely local.
Recording sped up 5.5x. Real time ~4 min on M1 Pro / 32 GB. Most time is vision OCR.

What you need
Ollama
Ollama is a free app that runs AI models locally on your computer. lilbee uses it behind the scenes to understand your documents and answer your questions — nothing is sent to the cloud.
- Download and install Ollama from ollama.com
- Open it — it runs in the background (you'll see a llama icon in your menu bar on macOS or system tray on Windows/Linux)
That's it. The plugin takes care of downloading the specific models it needs. You don't need to use the Ollama terminal or know any commands.
Models — what they are and why you need them
A "model" is an AI brain that runs on your computer. lilbee uses three kinds:
| Model type | What it does | Do I need to set it up? |
|---|---|---|
| Embedding model | Reads your documents and converts them into a searchable format so lilbee can find the right passages when you ask a question. This is what makes search work. | No — the plugin downloads this automatically the first time you sync. |
| Chat model | The AI that reads the relevant passages and writes an answer in plain language. This is what you're talking to in the chat sidebar. | The plugin shows you a list of recommended models and downloads your pick with one click. |
| Vision model (optional) | Can "read" images and scanned PDFs that don't have selectable text — think photographed pages, screenshots, or old scanned documents. It converts them to text so they become searchable. | Only needed if you want to index images or scanned PDFs. You can enable it in settings whenever you're ready. |
Models are large files (a few GB each) and take a few minutes to download the first time. After that they're cached on your machine and load in seconds.
Hardware note: Models run on your CPU or GPU. A Mac with Apple Silicon (M1/M2/M3/M4) or a PC with an NVIDIA GPU will give the best performance. 8 GB of RAM is the minimum; 16–32 GB is recommended for a smooth experience.
Quick start
- Install and open Ollama
- Install BRAT in Obsidian (Settings → Community plugins → Browse → search "BRAT" → Install → Enable)
- Open the command palette (
Cmd/Ctrl + P) → BRAT: Plugins: Add a beta plugin for testing → pastetobocop2/obsidian-lilbee→ Add Plugin - Enable lilbee in Settings → Community plugins
- Go to Settings → lilbee (or click the gear icon next to lilbee in Community plugins) to choose a chat model — pick one from the catalog and it downloads with one click. Optionally enable a vision model for images and scanned PDFs.
The plugin downloads and manages the lilbee server automatically — no terminal commands, no Python, no manual setup. Wait for the status bar to show lilbee: ready, then open the chat.
Open the chat
Once the status bar shows lilbee: ready:
| Platform | How to open chat |
|---|---|
| macOS | Cmd + P → type lilbee: Open chat → Enter |
| Windows / Linux | Ctrl + P → type lilbee: Open chat → Enter |
The chat panel opens in the sidebar. From there you can start asking questions, attach files, or run Sync vault (Cmd/Ctrl + P → "lilbee: Sync vault") to index everything at once.
How it works
On first launch, the plugin downloads the lilbee server and runs it in the background. When you sync your vault or attach files in the chat, this server breaks your documents into passages and uses Ollama to create searchable embeddings. When you ask a question, it finds the most relevant passages and sends them to the chat model, which writes an answer grounded in your actual documents — with links back to the sources.
Everything stays on your machine. The server, the models, the search index, and your documents all live locally. Like all Obsidian plugins, lilbee is installed per vault — each vault runs its own server instance with its own index, so there is no shared global store. If you already run your own lilbee server, you can point the plugin at it by overriding the server URL in Settings → lilbee.
macOS users: The server binary is unsigned (Apple charges $99/year for that). The plugin clears the quarantine flag automatically. If macOS still blocks it, go to System Settings → Privacy & Security and click "Allow Anyway". See the lilbee source if you want to audit the build.
Updating the plugin
Go to Settings → BRAT → Beta Plugin List, click the edit (pencil) icon next to lilbee, and change the version to the latest release tag. BRAT will download the new version. Restart Obsidian after the update for the new version to take effect.
Updating the server
The plugin tracks the installed lilbee server version. Go to Settings → lilbee → Check for updates. If a newer release is available, the button changes to Update to vX.Y.Z — one click stops the running server, downloads the new version, and restarts.
Documentation
See Usage Guide for the full reference — all commands, settings, chat features, supported formats, troubleshooting, and advanced configuration.
Build your own integration
lilbee exposes a REST API that isn't tied to any specific model. The search endpoint returns relevant passages without calling an LLM — so you can build your own tools on top of it, or integrate document search into other apps. This plugin is a full working example of a client built on that API.
See the lilbee README for the API docs.
License
MIT
For plugin developers
Search results and similarity scores are powered by semantic analysis of your plugin's README. If your plugin isn't appearing for searches you'd expect, try updating your README to clearly describe your plugin's purpose, features, and use cases.