AXCXEPT
Get Early Access
RAGOps Services News About Contact Get Early Access

AXCXEPT Inc. — Sapporo, Japan

AI Consulting &
LLM Engineering

22 years of enterprise IT. ¥50M+ delivered. 25+ open-source LLMs.
From infrastructure to AI — we build, ship, and prove it works.

¥50M+
Enterprise contracts (6 months)
25+
LLMs on HuggingFace
9.08
MT-Bench (EZO-8B)
2,819
Pages processed (+15% precision)
22yr
Enterprise IT experience
Book a Consultation Explore RAGOps ↓
Coming Soon — Spring 2026

RAGOps

Open-source RAG diagnostic engine.
Find what's broken. Fix it. Prove it worked.

"Your RAG is broken. We show you WHY and HOW TO FIX IT."

RAGOps Dashboard — root cause analysis and before/after proof for RAG pipelines

73% of AI projects never reach production

The #1 reason: nobody can fix it when it breaks.

42%
of enterprises abandoned majority of AI projects
S&P Global 2025
95%
of GenAI pilots fail to deliver P&L impact
MIT NANDA (n=300+)
70%
of RAG systems have no evaluation framework
NStarX 2025

Everyone shows THAT accuracy dropped.
Nobody shows WHY.

Langfuse / Arize / Datadog
"What happened?"Metrics dashboard
"What's broken?"You figure it out
"How to fix?"You guess
"Did it work?"Manual re-run
RAGOps
"What happened?"Metrics dashboard
"What's broken?""These 26 chunks from index X"
"How to fix?""Re-chunk with 512 tokens, re-index"
"Did it work?"Automated before/after proof

3 Steps to Root Cause

From integration to proof-of-improvement in minutes.

1

Connect

3 lines of code. 2 minutes. Zero risk to your production pipeline.

from ragops import Tracer
tracer = Tracer(api_key="...")
chain = RetrievalQA(..., callbacks=[tracer])
2

Diagnose

Deterministic root cause analysis. Not LLM guessing.

HIGH Chunk Quality

26 chunks from 'policy_docs' have avg relevance 0.18. Appear in 142 of 167 queries but never contribute to correct answers.

3

Fix & Prove

Prescriptive fix suggestions with before/after quantified proof.

0.45 🔴 0.83 🟢

Context Precision improved. Root cause "Chunk Quality" resolved.

What RAGOps Does

Deterministic diagnosis. Not dashboard-only observability.

📈

5 Evaluation Metrics

RAGAS-aligned, reference-free evaluation. Faithfulness, Context Precision, Answer Relevancy, Context Relevance, Response Groundedness. Scored 0.0–1.0 with color-coded health badges.

🔍

6 Root Cause Rules

Deterministic rule engine, not LLM guessing. Chunk Quality, Query-Chunk Mismatch, Index Freshness, Retrieval Distribution Anomaly, Permission Scope Leak, Reranking Inefficiency.

🛠

Prescriptions

Each root cause comes with verification tasks — not automatic fixes. AI suggests, human decides. Concrete steps to resolve the issue.

Before/After Proof

Quantified improvement evidence. Re-evaluate after fixing and see the score change. Prove the fix worked with data, not hope.

Built on Open Source

axcxept-eval: MIT license. 5 metrics. pip install.

axcxept-eval

Open-source RAG evaluation library. RAGAS-aligned 5-metric scoring. Use standalone or integrate with the RAGOps platform for full diagnosis.

  • ✓ MIT License — free forever
  • 5 evaluation metrics, reference-free
  • pip install axcxept-eval
  • Works with any LLM framework
Star on GitHub
$ pip install axcxept-eval

# Evaluate your RAG traces
from axcxept_eval import evaluate

result = evaluate(
    query="What is the refund policy?",
    contexts=[chunk1, chunk2, chunk3],
    answer="The refund policy states..."
)

# result.faithfulness    → 0.91 🟢
# result.precision       → 0.45 🔴
# result.relevancy       → 0.78 🟡

Pricing

Start free. Scale when ready.

Free
$0

1 full diagnostic cycle. Connect → Diagnose → Prescribe → Re-evaluate. Experience the entire value loop.

Coming Soon
Team
$199/mo

HITL workflow, regression tests, up to 10 members, full evaluation history.

Coming Soon
Enterprise
$999/mo

Evidence Pack, SSO/RBAC, SLA, custom integrations, dedicated support.

Coming Soon

Services & Products

22 years of enterprise IT. Cloud architecture, LLM development, and AI implementation.

🤖

EZO LLM Series

Lightweight, high-performance Japanese LLMs. MT-Bench 9.08 / J-MT-Bench 8.87 — GPT-4o class on 24GB VRAM. 25+ models on HuggingFace. OSS published.

💻

Enterprise AI Solutions

End-to-end AI implementation: cloud architecture, custom LLM training, API integration, MLOps. We start from business problems, not models.

  • Cloud infrastructure (Azure / AWS / GCP)
  • Custom LLM fine-tuning & deployment
  • On-premise / VPC isolation support
🚀

SaaS Products

AI-powered products built in-house, serving users globally.

Looking for enterprise AI consulting? We accept engagements for RAG optimization, LLM deployment, and cloud AI architecture.

Book a Consultation

News

2025.04.15

Released "Pseudo GRPO/PPO approach for low-cost Japanese LLM performance improvement"

2025.04.10

Released ultra-compact LLM "QwQ-32B-Distill-Qwen-1.5B-Alpha" specialized for math reasoning

2025.03.26

Panel discussion at ASIAWORLD-EXPO, Hong Kong (Jumpstarter)

2024.12.10

Speaker at ALIBABA CLOUD TECH DAY TOKYO

About

CompanyAXCXEPT Inc.
CEOKazuya Hodatsu
BusinessAI/LLM Consulting & Implementation, LLM Development (EZO Series), SaaS/App Development, RAGOps Platform
LocationSapporo, Hokkaido, Japan

Founder

KH

Kazuya Hodatsu

CEO & Founder, AXCXEPT Inc.

22 years in enterprise IT — from infrastructure to AI. Built teams, shipped products, and led multi-million dollar projects at leading global technology companies.

Credentials

Azure Solutions Architect Expert CKA / CKAD 25+ LLMs on HuggingFace

Track Record

Enterprise R&D for a major Japanese corporation — \u00A550M+ delivered in 6 months, solo. 2,819 pages processed, +15% precision, +10.8% faithfulness.
EZO-70B: Top-scoring Japanese OSS model on ElyzaTasks-100. EZO-8B: MT-Bench 9.08.
Built the entire RAGOps platform (backend + frontend + SDK) in 5 weeks — solo founder, 80 tests passing.

Recognition

Speaker, JUMPSTARTER 2025 (AsiaWorld-Expo, Hong Kong)
Speaker, Alibaba Cloud Tech Day Tokyo (2024)

Research

Pseudo GRPO/PPO approach for low-cost Japanese LLM performance improvement (2025)
Math reasoning specialized ultra-compact LLM via knowledge distillation (2025)

"LLM engineering, cloud architecture, and enterprise delivery — I bring all three to every engagement."

Contact

Questions, partnerships, or just want to chat about RAG systems.