Building the future of edge AI with transparency and open research.
Maaza-MLM-135M-JSON-v1 and Maaza-SLM-360M-JSON-v1 are state-of-the-art small language models designed for structured JSON generation on edge devices. Maaza NLM Orchestrator handles tool routing and orchestration for MCP workflows.
Trained on the EdgeJSON benchmark and MCPBodega production tools, these models demonstrate that task-specific small models can outperform much larger general-purpose models on structured output generation and tool orchestration.
Production-ready JSON extraction from real-world documents.
Built-in reliability layer ensures consistent accuracy.
Fast (~2-3s), accurate (75.7% on EdgeJSON v3), and 2-4× more cost-effective than GPT-4 Turbo.
curl -X POST http://199.68.217.31:50784/v1/extract \
-H 'Content-Type: application/json' \
-d '{
"text": "From: [email protected]\nSubject: Q4 Meeting\nHi team, lets meet on Dec 15th at 2pm in conference room B.\nThanks,\nSarah Johnson\nSenior Manager\n(555) 123-4567",
"schema": {
"type": "object",
"properties": {
"sender_email": {"type": "string"},
"sender_name": {"type": "string"},
"meeting_date": {"type": "string"},
"meeting_time": {"type": "string"}
}
}
}'
Maaza API offers predictable, flat-rate pricing for JSON extraction tasks, compared to token-based pricing from general-purpose LLMs.
| Provider | Model | Cost per 10k Extractions* | Notes |
|---|---|---|---|
| Maaza API | SLM-360M | $29-49/month | Predictable flat rate |
| OpenAI | GPT-4 Turbo | ~$110 | Token-based ($10/$30 per 1M) |
| Anthropic | Claude 3.5 Sonnet | ~$45 | Token-based ($3/$15 per 1M) |
| OpenAI | GPT-4o | ~$32.50 | Token-based ($2.50/$10 per 1M) |
Cost estimates based on typical JSON extraction workload: 500 input tokens + 200 output tokens per request. Calculations use standard on-demand API pricing as of November 2025. Does not include volume discounts, batch API pricing, or prompt caching features which may reduce costs further.
Sources: OpenAI, Anthropic, and Google official pricing pages. Verified November 29, 2025. For detailed methodology and calculations, see our pricing research documentation .
@techreport{maaza2025,
title={Task-Specialized Micro Language Models Outperform Larger
Zero-Shot Models on Structured Data Extraction},
author={CycleCore Technologies},
institution={CycleCore Technologies},
year={2025},
type={Technical Report},
url={https://cyclecore.ai/research}
}