discourse-ai/app
Sam d07cf51653
FEATURE: llm quotas (#1047)
Adds a comprehensive quota management system for LLM models that allows:

- Setting per-group (applied per user in the group) token and usage limits with configurable durations
- Tracking and enforcing token/usage limits across user groups
- Quota reset periods (hourly, daily, weekly, or custom)
-  Admin UI for managing quotas with real-time updates

This system provides granular control over LLM API usage by allowing admins
to define limits on both total tokens and number of requests per group.
Supports multiple concurrent quotas per model and automatically handles
quota resets.


Co-authored-by: Keegan George <kgeorge13@gmail.com>
2025-01-14 15:54:09 +11:00
..
controllers/discourse_ai FEATURE: llm quotas (#1047) 2025-01-14 15:54:09 +11:00
helpers/discourse_ai/ai_bot FIX: automatically bust cache for share ai assets (#942) 2024-11-22 11:23:15 +11:00
jobs REFACTOR: Separation of concerns for embedding generation. (#1027) 2024-12-16 09:55:39 -03:00
mailers FEATURE: support sending AI report to an email address (#368) 2023-12-19 17:51:49 +11:00
models FEATURE: llm quotas (#1047) 2025-01-14 15:54:09 +11:00
serializers FEATURE: llm quotas (#1047) 2025-01-14 15:54:09 +11:00
services FEATURE: Add periodic problem checks for each LLM in use (#1020) 2024-12-16 15:00:05 -08:00
views FIX: automatically bust cache for share ai assets (#942) 2024-11-22 11:23:15 +11:00