🤖 Agents System
Locentra OS includes a fully integrated agent layer—self-optimizing background logic designed to monitor prompt quality, trigger training, adapt system behavior, and rewrite input intelligently.
Agents are modular and live at:
They can run synchronously with inference or as asynchronous background processes.
🧬 Agent Modules Overview
AutoTrainer
Detect low-quality responses, retrain automatically
FeedbackLoop
Monitor user behavior + corrections → flag for training
PromptOptimizer
Rewrite noisy or ambiguous input before inference
All agents are:
Plugged directly into the LLM lifecycle
Configurable
Loggable
Extendable with custom logic
📊 System Flow (High-Level)
🔁 AutoTrainer
Core logic for self-correction.
Hooks into
POST /api/llm/query
Scores model output via
utils/eval_metrics.py
If the response score falls below threshold:
Stores prompt in vector memory
Pushes to
models/trainer.py
for live fine-tuningFlags prompt metadata for later analysis
You can control training thresholds via:
The model fine-tunes itself—based on how well it performs.
🧠 FeedbackLoop
Monitors indirect indicators of model failure:
⏱ Response delay (user waits or edits)
✏️ Manual prompt rewrites
🔁 Repeated queries on the same topic
If the pattern matches “low confidence” behavior, the agent:
Embeds the prompt semantically
Stores it in persistent memory
Queues retraining
Flags it for potential prompt optimization
📝 PromptOptimizer
Some prompts just suck. This agent rewrites them before inference.
Functionality:
Splits compound queries
Normalizes slang, abbreviations
Restructures the question to be more LLM-friendly
Preserves semantic intent while improving clarity
Example:
Powered by:
Zero-shot classification
Regex-based refactoring
Embedding distance for ambiguity detection
🔄 Example: Live Agent Loop
User sends:
"wtf is zk-snarks vs starks?"
Model gives weak/unclear response
AutoTrainer
flags score < thresholdFeedbackLoop
captures session metadataPromptOptimizer
rewrites prompt:"Compare zk-SNARKs and zk-STARKs in zero-knowledge cryptography."
Fine-tuning is queued
Memory is updated with embedded vector
The system adapts. No user or dev needed.
📦 Build Your Own Agents
Create a custom agent by extending the base class:
Register it in:
Agents can:
Write logs
Trigger webhooks
Modify memory
Queue training jobs
Even call other agents
They run async, and can be chained, filtered, or scheduled.
🧠 Why This Matters
With agents:
Your model gets better the more it’s used
No manual fine-tuning needed
Prompts evolve
Output quality stabilizes
System becomes self-sustaining
Last updated