Files
stack/apps
Jason Woltje 94afeb67e3 feat(#123): port Ollama LLM provider
Implemented first concrete LLM provider following the provider interface pattern.

Implementation:
- OllamaProvider class implementing LlmProviderInterface
- All required methods: initialize(), checkHealth(), listModels(), chat(), chatStream(), embed(), getConfig()
- OllamaProviderConfig extending LlmProviderConfig
- Proper error handling with NestJS Logger
- Configuration immutability protection

Features:
- System prompt injection support
- Temperature and max tokens configuration
- Embedding with truncation control (defaults to enabled)
- Streaming and non-streaming chat completions
- Health check with model listing

Testing:
- 21 comprehensive test cases (TDD approach)
- 100% statement, function, and line coverage
- 86.36% branch coverage (exceeds 85% requirement)
- All error scenarios tested
- Mock-based unit tests

Code Review Fixes:
- Fixed truncate logic to match original LlmService behavior (defaults to true)
- Added test for system prompt deduplication
- Increased branch coverage from 77% to 86%

Quality Gates:
-  All 21 tests passing
-  Linting clean
-  Type checking passed
-  Code review approved

Fixes #123

Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>
2026-01-31 12:10:43 -06:00
..
2026-01-31 12:10:43 -06:00