33This directory contains example boilerplates that demonstrate how easy it is to integrate popular LLM libraries with the
44router by simply switching the host.
55
6+ ---
7+
68## Available Examples
79
810- ** [ LlamaIndex] ( llamaindex_example.py ) ** – Integration with LlamaIndex (GPT Index)
@@ -24,6 +26,8 @@ router will automatically:
24264 . ✅ Supply monitoring and metrics
25275 . ✅ Handle streaming and non‑streaming responses
2628
29+ ---
30+
2731## Quick Start
2832
2933Each example can be run directly:
@@ -45,6 +49,8 @@ python examples/litellm_example.py
4549python examples/haystack_example.py
4650```
4751
52+ ---
53+
4854## Example Structure
4955
5056Each example includes:
@@ -54,6 +60,25 @@ Each example includes:
54603 . ** Non‑streaming** – handling full responses
55614 . ** Error handling** – managing errors
5662
63+ ---
64+
65+ ## Full Stack with Local Models
66+
67+ The quick‑start guides for running the full stack with ** local models** are included in the repository:
68+
69+ - ** Gemma 3 12B‑IT** – [ README] ( quickstart/google-gemma3-12b-it/README.md )
70+ - ** Bielik 11B‑v2.3‑Instruct‑FP8** – [ README] ( quickstart/speakleash-bielik-11b-v2_3-Instruct/README.md )
71+
72+ These guides walk you through:
73+
74+ 1 . Installing ** vLLM** and the respective model.
75+ 2 . Setting up ** LLM‑Router** with the provided ` models-config.json ` .
76+ 3 . Testing the end‑to‑end flow (router → vLLM).
77+
78+ Follow the linked README files for step‑by‑step instructions to launch a complete stack locally.
79+
80+ ---
81+
5782## Additional Information
5883
5984Learn more about the router:
0 commit comments