A Python CLI that extracts payment details from invoice PDFs.
This is a practical example of building local AI tools and apps with
- No cloud costs
- No network latency
- No data privacy loss
- What's inside?
- Understanding the architecture
- Environment setup
- How to run it?
- Results
- Next steps
- Need help?
In this example, you will learn how to:
- Set up local AI inference using llama.cpp to run Liquid models entirely on your machine without requiring cloud services or API keys
- Build a file monitoring system that automatically processes new files dropped into a directory
- Extract structured output from images using LFM2.5-VL-1.6B, a small vision-language model.
You will need
- llama.cpp to serve the Language Models locally.
- uv to manage Python dependencies and run the application efficiently without creating virtual environments manually.
Click to see installation instructions for your platform
macOS:
brew install llama.cppLinux (build from source):
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build --config Release
# Add build/bin to your PATHWindows: Follow the instructions in the llama.cpp repository.
Verify that llama-server is available:
which llama-serverClick to see installation instructions for your platform
macOS/Linux:
curl -LsSf https://astral.sh/uv/install.sh | shWindows:
powershell -ExecutionPolicy ByPass -c "irm https://astral.sh/uv/install.ps1 | iex"Let's start by cloning the repository:
git clone https://github.com/Liquid4All/cookbook.git
cd cookbook/examples/invoice-parserThe tool supports two modes: watch (continuous monitoring) and process (one-shot).
The tool automatically starts and stops llama-server for you — no need to run it separately.
Run it as a background service that continuously monitors a directory and automatically parses invoice images as they land in the folder:
uv run python src/invoice_parser/main.py watch \
--dir invoices/ \
--image-model LiquidAI/LFM2.5-VL-1.6B-GGUF:Q8_0 \
--output bills.csv \
--process-existingProcess specific files or folders and exit:
# Process an entire folder
uv run python src/invoice_parser/main.py process \
--image-model LiquidAI/LFM2.5-VL-1.6B-GGUF:Q8_0 \
invoices/
# Process specific files
uv run python src/invoice_parser/main.py process \
--image-model LiquidAI/LFM2.5-VL-1.6B-GGUF:Q8_0 \
invoices/water_australia.png invoices/british_gas.png
# Save results to a CSV file
uv run python src/invoice_parser/main.py process \
--image-model LiquidAI/LFM2.5-VL-1.6B-GGUF:Q8_0 \
--output bills.csv \
invoices/Feel free to modify the path to the invoices directory and the model IDs to suit your needs.
If you have make installed, you can run the application with the following commands:
make run # watch mode
make process # one-shot process modeYou can run the tool with a sample of images under invoices/ with
uv run python src/invoice_parser/main.py process \
--image-model LiquidAI/LFM2.5-VL-1.6B-GGUF:Q8_0 \
invoices/
or, if you have Make installed in your system, just do
make process
Results are then printed on console. The model correctly extracts all 4 out of 4 invoices:
| File | Utility | Amount | Currency |
|---|---|---|---|
| water_australia.png | water | 68.46 | AUD |
| Sample-electric-Bill-2023.jpg | electricity | 28.32 | USD |
| castlewater1.png | water | 436.55 | GBP |
| british_gas.png | electricity | 81.31 | GBP |
The model works perfectly out-of-the-box on our sample of invoices. However, depending on your specific invoice formats and layouts, you may encounter cases where the extraction is not accurate enough.
In those cases, you can fine-tune the model on your own dataset to improve accuracy. Check out the fine-tuning notebook for Vision Language Models to learn how.
Join the Liquid AI Discord Community and ask.
