Model Name | Architecture | Provider | Purpose | Input | Output | Usage Context |
---|---|---|---|---|---|---|
DriverPredict-XGB | XGBoost (Gradient Boosting Trees) | Custom (in-house) | Predicts driver availability based on historical behavior and app data | Driver ID, activity history, shifts, app pings | Availability probability for time slots | Used in planning to pre-allocate or reassign deliveries |
DispatchGPT | Large Language Model (GPT-4 Turbo) | OpenAI | Interprets driver messages, suggests replies, flags risk | Free-text message, dispatcher context | Suggested response, summarized intent, risk flags | Live assistant in dispatcher UI |
DocOCR-Validator | TrOCR + LayoutLMv3 | Microsoft / HuggingFace | Processes and validates delivery documents | Scans, PDFs, images | Structured data, validation results, compliance status | Document processing for proof-of-delivery and compliance workflows |
Model Name | Avg Inference Time | Retraining Frequency | Context Volume | Explainability | Real-Time Compatible |
---|---|---|---|---|---|
DriverPredict-XGB | ~200 ms | Weekly | ~50 variables (driver activity patterns) | SHAP feature importance | Yes |
DispatchGPT | ~500–700 ms | N/A (prompt-based) | ~300–500 tokens (text + metadata) | Natural-language explanation (LLM) | Yes |
DocOCR-Validator | ~2–3 sec | Rare (base models), rules updated quarterly | ~1 document page (image + form structure) | Partial (OCR confidence + heuristics) | Batch or async preferred |
Model | Response Time | Output Quality | Strengths | Weaknesses |
---|---|---|---|---|
GPT-4 Turbo (Chosen) | ~500–700 ms | High (natural, contextual, actionable) | Fast, cost-effective, highly contextual | Limited to ~128k tokens; sometimes verbose |
GPT-4 (original) | ~2–3 sec | Very High (nuanced, detailed, reliable) | Deep understanding, great in edge cases | High latency and cost for frequent UI interactions |
Model Name | Accuracy | Training Time | Inference Time | Explainability | Strengths | Weaknesses Time |
---|---|---|---|---|---|---|
XGBoost (Chosen) | High (great for structured behavioral data) | Fast | ~200 ms | SHAP feature importance | Interpretable, robust, effective for tabular data | May need manual feature engineering |
LightGBM | Comparable to XGBoost | Very fast | ~150 ms | SHAP | Fast training, scalable on large datasets | Slightly harder to tune for accuracy |
Temporal Fusion Transformer | Very high for time-series | Slow | ~1–2 sec | Attention-based insights | Captures time dependencies, interpretable via attention | Complex infra, slow inference, overkill |