4 Commits

Author SHA1 Message Date
b202e024a4 v1.14.3
Some checks failed
Docker (tags) / security (push) Successful in 29s
Docker (tags) / test (push) Failing after 1m46s
Docker (tags) / release (push) Has been skipped
Docker (tags) / metadata (push) Has been skipped
2026-01-20 00:55:24 +00:00
2210611f70 fix(repo): no changes detected in the diff; no files modified and no release required 2026-01-20 00:55:24 +00:00
d8bdb18841 fix(test): add JSON validation and retry logic to invoice extraction
- Add tryExtractJson function to validate JSON before accepting
- Use orchestrator.continueTask() to request correction if JSON is invalid
- Retry up to 2 times for malformed JSON responses
- Remove duplicate parseJsonToInvoice function
2026-01-20 00:45:30 +00:00
d384c1d79b feat(tests): integrate smartagent DualAgentOrchestrator with streaming support
- Update test.invoices.nanonets.ts to use DualAgentOrchestrator for JSON extraction
- Enable streaming token callback for real-time progress visibility
- Add markdown caching to avoid re-running Nanonets OCR for cached files
- Update test.bankstatements.minicpm.ts and test.invoices.minicpm.ts with streaming
- Update dependencies to @push.rocks/smartai@0.11.1 and @push.rocks/smartagent@1.2.8
2026-01-20 00:39:36 +00:00
6 changed files with 1467 additions and 226 deletions

View File

@@ -1,5 +1,12 @@
# Changelog # Changelog
## 2026-01-20 - 1.14.3 - fix(repo)
no changes detected in the diff; no files modified and no release required
- Diff contained no changes
- No files were added, removed, or modified
- No code, dependency, or documentation updates to release
## 2026-01-19 - 1.14.2 - fix(readme) ## 2026-01-19 - 1.14.2 - fix(readme)
update README to document Nanonets-OCR2-3B (replaces Nanonets-OCR-s), adjust VRAM and context defaults, expand feature docs, and update examples/test command update README to document Nanonets-OCR2-3B (replaces Nanonets-OCR-s), adjust VRAM and context defaults, expand feature docs, and update examples/test command

View File

@@ -1,6 +1,6 @@
{ {
"name": "@host.today/ht-docker-ai", "name": "@host.today/ht-docker-ai",
"version": "1.14.2", "version": "1.14.3",
"type": "module", "type": "module",
"private": false, "private": false,
"description": "Docker images for AI vision-language models including MiniCPM-V 4.5", "description": "Docker images for AI vision-language models including MiniCPM-V 4.5",
@@ -14,7 +14,9 @@
}, },
"devDependencies": { "devDependencies": {
"@git.zone/tsrun": "^2.0.1", "@git.zone/tsrun": "^2.0.1",
"@git.zone/tstest": "^3.1.5" "@git.zone/tstest": "^3.1.5",
"@push.rocks/smartagent": "^1.2.8",
"@push.rocks/smartai": "^0.11.1"
}, },
"repository": { "repository": {
"type": "git", "type": "git",

1134
pnpm-lock.yaml generated

File diff suppressed because it is too large Load Diff

View File

@@ -1,9 +1,9 @@
/** /**
* Bank statement extraction using MiniCPM-V (visual extraction) * Bank statement extraction using MiniCPM-V (visual extraction)
* *
* JSON per-page approach: * JSON per-page approach with streaming output:
* 1. Ask for structured JSON of all transactions per page * 1. Ask for structured JSON of all transactions per page
* 2. Consensus: extract twice, compare, retry if mismatch * 2. Single pass extraction (no consensus)
*/ */
import { tap, expect } from '@git.zone/tstest/tapbundle'; import { tap, expect } from '@git.zone/tstest/tapbundle';
import * as fs from 'fs'; import * as fs from 'fs';
@@ -66,11 +66,11 @@ function convertPdfToImages(pdfPath: string): string[] {
} }
/** /**
* Query for JSON extraction * Query for JSON extraction with streaming output
*/ */
async function queryJson(image: string, queryId: string): Promise<string> { async function queryJson(image: string, queryId: string): Promise<string> {
console.log(` [${queryId}] Sending request to ${MODEL}...`);
const startTime = Date.now(); const startTime = Date.now();
process.stdout.write(` [${queryId}] `);
const response = await fetch(`${OLLAMA_URL}/api/chat`, { const response = await fetch(`${OLLAMA_URL}/api/chat`, {
method: 'POST', method: 'POST',
@@ -82,25 +82,50 @@ async function queryJson(image: string, queryId: string): Promise<string> {
content: JSON_PROMPT, content: JSON_PROMPT,
images: [image], images: [image],
}], }],
stream: false, stream: true,
options: { options: {
num_ctx: 32768,
num_predict: 4000, num_predict: 4000,
temperature: 0.1, temperature: 0.1,
}, },
}), }),
}); });
const elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
if (!response.ok) { if (!response.ok) {
console.log(` [${queryId}] ERROR: ${response.status} (${elapsed}s)`); const elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
process.stdout.write(`ERROR: ${response.status} (${elapsed}s)\n`);
throw new Error(`Ollama API error: ${response.status}`); throw new Error(`Ollama API error: ${response.status}`);
} }
const data = await response.json(); let content = '';
const content = (data.message?.content || '').trim(); const reader = response.body!.getReader();
console.log(` [${queryId}] Response received (${elapsed}s, ${content.length} chars)`); const decoder = new TextDecoder();
return content;
try {
while (true) {
const { done, value } = await reader.read();
if (done) break;
const chunk = decoder.decode(value, { stream: true });
for (const line of chunk.split('\n').filter(l => l.trim())) {
try {
const json = JSON.parse(line);
const token = json.message?.content || '';
if (token) {
process.stdout.write(token);
content += token;
}
} catch {
// Ignore parse errors for partial chunks
}
}
}
} finally {
const elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
process.stdout.write(` (${elapsed}s)\n`);
}
return content.trim();
} }
/** /**
@@ -284,102 +309,29 @@ function parseAmount(value: unknown): number {
} }
/** /**
* Compare two transaction arrays for consensus * Extract transactions from a single page (single pass)
*/
function transactionArraysMatch(a: ITransaction[], b: ITransaction[]): boolean {
if (a.length !== b.length) return false;
for (let i = 0; i < a.length; i++) {
const dateMatch = a[i].date === b[i].date;
const amountMatch = Math.abs(a[i].amount - b[i].amount) < 0.01;
if (!dateMatch || !amountMatch) return false;
}
return true;
}
/**
* Compare two transaction arrays and log differences
*/
function compareAndLogDifferences(txs1: ITransaction[], txs2: ITransaction[], pageNum: number): void {
if (txs1.length !== txs2.length) {
console.log(` [Page ${pageNum}] Length mismatch: Q1=${txs1.length}, Q2=${txs2.length}`);
return;
}
for (let i = 0; i < txs1.length; i++) {
const dateMatch = txs1[i].date === txs2[i].date;
const amountMatch = Math.abs(txs1[i].amount - txs2[i].amount) < 0.01;
if (!dateMatch || !amountMatch) {
console.log(` [Page ${pageNum}] Tx ${i + 1} differs:`);
console.log(` Q1: ${txs1[i].date} | ${txs1[i].amount}`);
console.log(` Q2: ${txs2[i].date} | ${txs2[i].amount}`);
}
}
}
/**
* Extract transactions from a single page with consensus
*/ */
async function extractTransactionsFromPage(image: string, pageNum: number): Promise<ITransaction[]> { async function extractTransactionsFromPage(image: string, pageNum: number): Promise<ITransaction[]> {
const MAX_ATTEMPTS = 5;
console.log(`\n ======== Page ${pageNum} ========`); console.log(`\n ======== Page ${pageNum} ========`);
console.log(` [Page ${pageNum}] Starting JSON extraction...`);
for (let attempt = 1; attempt <= MAX_ATTEMPTS; attempt++) { const queryId = `P${pageNum}`;
console.log(`\n [Page ${pageNum}] --- Attempt ${attempt}/${MAX_ATTEMPTS} ---`); const response = await queryJson(image, queryId);
const transactions = parseJsonResponse(response, queryId);
// Extract twice in parallel console.log(` [Page ${pageNum}] Extracted ${transactions.length} transactions:`);
const q1Id = `P${pageNum}A${attempt}Q1`; for (let i = 0; i < transactions.length; i++) {
const q2Id = `P${pageNum}A${attempt}Q2`; const tx = transactions[i];
const [response1, response2] = await Promise.all([
queryJson(image, q1Id),
queryJson(image, q2Id),
]);
const txs1 = parseJsonResponse(response1, q1Id);
const txs2 = parseJsonResponse(response2, q2Id);
console.log(` [Page ${pageNum}] Results: Q1=${txs1.length} txs, Q2=${txs2.length} txs`);
if (txs1.length > 0 && transactionArraysMatch(txs1, txs2)) {
console.log(` [Page ${pageNum}] ✓ CONSENSUS REACHED: ${txs1.length} transactions`);
console.log(` [Page ${pageNum}] Transactions:`);
for (let i = 0; i < txs1.length; i++) {
const tx = txs1[i];
console.log(` ${(i + 1).toString().padStart(2)}. ${tx.date} | ${tx.counterparty.substring(0, 30).padEnd(30)} | ${tx.amount >= 0 ? '+' : ''}${tx.amount.toFixed(2)}`);
}
return txs1;
}
console.log(` [Page ${pageNum}] ✗ NO CONSENSUS`);
compareAndLogDifferences(txs1, txs2, pageNum);
if (attempt < MAX_ATTEMPTS) {
console.log(` [Page ${pageNum}] Retrying...`);
}
}
// Fallback: use last response
console.log(`\n [Page ${pageNum}] === FALLBACK (no consensus after ${MAX_ATTEMPTS} attempts) ===`);
const fallbackId = `P${pageNum}FALLBACK`;
const fallbackResponse = await queryJson(image, fallbackId);
const fallback = parseJsonResponse(fallbackResponse, fallbackId);
console.log(` [Page ${pageNum}] ~ FALLBACK RESULT: ${fallback.length} transactions`);
for (let i = 0; i < fallback.length; i++) {
const tx = fallback[i];
console.log(` ${(i + 1).toString().padStart(2)}. ${tx.date} | ${tx.counterparty.substring(0, 30).padEnd(30)} | ${tx.amount >= 0 ? '+' : ''}${tx.amount.toFixed(2)}`); console.log(` ${(i + 1).toString().padStart(2)}. ${tx.date} | ${tx.counterparty.substring(0, 30).padEnd(30)} | ${tx.amount >= 0 ? '+' : ''}${tx.amount.toFixed(2)}`);
} }
return fallback;
return transactions;
} }
/** /**
* Extract all transactions from bank statement * Extract all transactions from bank statement
*/ */
async function extractTransactions(images: string[]): Promise<ITransaction[]> { async function extractTransactions(images: string[]): Promise<ITransaction[]> {
console.log(` [Vision] Processing ${images.length} page(s) with ${MODEL} (JSON consensus)`); console.log(` [Vision] Processing ${images.length} page(s) with ${MODEL} (single pass)`);
const allTransactions: ITransaction[] = []; const allTransactions: ITransaction[] = [];
@@ -527,7 +479,7 @@ tap.test('summary', async () => {
console.log(`\n======================================================`); console.log(`\n======================================================`);
console.log(` Bank Statement Summary (${MODEL})`); console.log(` Bank Statement Summary (${MODEL})`);
console.log(`======================================================`); console.log(`======================================================`);
console.log(` Method: JSON per-page + consensus`); console.log(` Method: JSON per-page (single pass)`);
console.log(` Passed: ${passedCount}/${total}`); console.log(` Passed: ${passedCount}/${total}`);
console.log(` Failed: ${failedCount}/${total}`); console.log(` Failed: ${failedCount}/${total}`);
console.log(`======================================================\n`); console.log(`======================================================\n`);

View File

@@ -67,9 +67,12 @@ const JSON_PROMPT = `Extract invoice data from this image. Return ONLY a JSON ob
Return only the JSON, no explanation.`; Return only the JSON, no explanation.`;
/** /**
* Query MiniCPM-V for JSON output (fast, no thinking) * Query MiniCPM-V for JSON output (fast, no thinking) with streaming
*/ */
async function queryJsonFast(images: string[]): Promise<string> { async function queryJsonFast(images: string[]): Promise<string> {
const startTime = Date.now();
process.stdout.write(` [Fast] `);
const response = await fetch(`${OLLAMA_URL}/api/chat`, { const response = await fetch(`${OLLAMA_URL}/api/chat`, {
method: 'POST', method: 'POST',
headers: { 'Content-Type': 'application/json' }, headers: { 'Content-Type': 'application/json' },
@@ -80,8 +83,9 @@ async function queryJsonFast(images: string[]): Promise<string> {
content: JSON_PROMPT, content: JSON_PROMPT,
images: images, images: images,
}], }],
stream: false, stream: true,
options: { options: {
num_ctx: 32768,
num_predict: 1000, num_predict: 1000,
temperature: 0.1, temperature: 0.1,
}, },
@@ -92,14 +96,44 @@ async function queryJsonFast(images: string[]): Promise<string> {
throw new Error(`Ollama API error: ${response.status}`); throw new Error(`Ollama API error: ${response.status}`);
} }
const data = await response.json(); let content = '';
return (data.message?.content || '').trim(); const reader = response.body!.getReader();
const decoder = new TextDecoder();
try {
while (true) {
const { done, value } = await reader.read();
if (done) break;
const chunk = decoder.decode(value, { stream: true });
for (const line of chunk.split('\n').filter(l => l.trim())) {
try {
const json = JSON.parse(line);
const token = json.message?.content || '';
if (token) {
process.stdout.write(token);
content += token;
}
} catch {
// Ignore parse errors for partial chunks
}
}
}
} finally {
const elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
process.stdout.write(` (${elapsed}s)\n`);
}
return content.trim();
} }
/** /**
* Query MiniCPM-V for JSON output with thinking enabled (slower, more accurate) * Query MiniCPM-V for JSON output with thinking enabled (slower, more accurate) with streaming
*/ */
async function queryJsonWithThinking(images: string[]): Promise<string> { async function queryJsonWithThinking(images: string[]): Promise<string> {
const startTime = Date.now();
process.stdout.write(` [Think] `);
const response = await fetch(`${OLLAMA_URL}/api/chat`, { const response = await fetch(`${OLLAMA_URL}/api/chat`, {
method: 'POST', method: 'POST',
headers: { 'Content-Type': 'application/json' }, headers: { 'Content-Type': 'application/json' },
@@ -110,8 +144,9 @@ async function queryJsonWithThinking(images: string[]): Promise<string> {
content: `Think carefully about this invoice image, then ${JSON_PROMPT}`, content: `Think carefully about this invoice image, then ${JSON_PROMPT}`,
images: images, images: images,
}], }],
stream: false, stream: true,
options: { options: {
num_ctx: 32768,
num_predict: 2000, num_predict: 2000,
temperature: 0.1, temperature: 0.1,
}, },
@@ -122,8 +157,56 @@ async function queryJsonWithThinking(images: string[]): Promise<string> {
throw new Error(`Ollama API error: ${response.status}`); throw new Error(`Ollama API error: ${response.status}`);
} }
const data = await response.json(); let content = '';
return (data.message?.content || '').trim(); let thinkingContent = '';
let thinkingStarted = false;
let outputStarted = false;
const reader = response.body!.getReader();
const decoder = new TextDecoder();
try {
while (true) {
const { done, value } = await reader.read();
if (done) break;
const chunk = decoder.decode(value, { stream: true });
for (const line of chunk.split('\n').filter(l => l.trim())) {
try {
const json = JSON.parse(line);
// Stream thinking tokens
const thinking = json.message?.thinking || '';
if (thinking) {
if (!thinkingStarted) {
process.stdout.write(`THINKING: `);
thinkingStarted = true;
}
process.stdout.write(thinking);
thinkingContent += thinking;
}
// Stream content tokens
const token = json.message?.content || '';
if (token) {
if (!outputStarted) {
if (thinkingStarted) process.stdout.write('\n [Think] ');
process.stdout.write(`OUTPUT: `);
outputStarted = true;
}
process.stdout.write(token);
content += token;
}
} catch {
// Ignore parse errors for partial chunks
}
}
}
} finally {
const elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
process.stdout.write(` (${elapsed}s)\n`);
}
return content.trim();
} }
/** /**

View File

@@ -12,6 +12,8 @@ import * as path from 'path';
import { execSync } from 'child_process'; import { execSync } from 'child_process';
import * as os from 'os'; import * as os from 'os';
import { ensureNanonetsOcr, ensureMiniCpm, isContainerRunning } from './helpers/docker.js'; import { ensureNanonetsOcr, ensureMiniCpm, isContainerRunning } from './helpers/docker.js';
import { SmartAi } from '@push.rocks/smartai';
import { DualAgentOrchestrator } from '@push.rocks/smartagent';
const NANONETS_URL = 'http://localhost:8000/v1'; const NANONETS_URL = 'http://localhost:8000/v1';
const NANONETS_MODEL = 'nanonets/Nanonets-OCR2-3B'; const NANONETS_MODEL = 'nanonets/Nanonets-OCR2-3B';
@@ -19,8 +21,24 @@ const NANONETS_MODEL = 'nanonets/Nanonets-OCR2-3B';
const OLLAMA_URL = 'http://localhost:11434'; const OLLAMA_URL = 'http://localhost:11434';
const EXTRACTION_MODEL = 'gpt-oss:20b'; const EXTRACTION_MODEL = 'gpt-oss:20b';
// Temp directory for storing markdown between stages // Persistent cache directory for storing markdown between runs
const TEMP_MD_DIR = path.join(os.tmpdir(), 'nanonets-invoices-markdown'); const MD_CACHE_DIR = path.join(process.cwd(), '.nogit/invoices-md');
// SmartAi instance for Ollama with optimized settings
const smartAi = new SmartAi({
ollama: {
baseUrl: OLLAMA_URL,
model: EXTRACTION_MODEL,
defaultOptions: {
num_ctx: 32768, // Larger context for long invoices + thinking
temperature: 0, // Deterministic for JSON extraction
},
defaultTimeout: 600000, // 10 minute timeout for large documents
},
});
// DualAgentOrchestrator for structured task execution
let orchestrator: DualAgentOrchestrator;
interface IInvoice { interface IInvoice {
invoice_number: string; invoice_number: string;
@@ -71,10 +89,13 @@ RULES:
7. total_amount: Final total with tax 7. total_amount: Final total with tax
JSON only: JSON only:
{"invoice_number":"X","invoice_date":"YYYY-MM-DD","vendor_name":"X","currency":"EUR","net_amount":0,"vat_amount":0,"total_amount":0}`; {"invoice_number":"X","invoice_date":"YYYY-MM-DD","vendor_name":"X","currency":"EUR","net_amount":0,"vat_amount":0,"total_amount":0}
Double check for valid JSON syntax.
`;
// Constants for smart batching // Constants for smart batching
const MAX_VISUAL_TOKENS = 28000; // ~32K context minus prompt/output headroom
const PATCH_SIZE = 14; // Qwen2.5-VL uses 14x14 patches const PATCH_SIZE = 14; // Qwen2.5-VL uses 14x14 patches
/** /**
@@ -318,16 +339,20 @@ function extractCurrency(s: string | undefined): string {
} }
/** /**
* Extract JSON from response * Try to extract valid JSON from a response string
*/ */
function extractJsonFromResponse(response: string): Record<string, unknown> | null { function tryExtractJson(response: string): Record<string, unknown> | null {
let cleanResponse = response.replace(/<think>[\s\S]*?<\/think>/g, '').trim(); // Remove thinking tags
const codeBlockMatch = cleanResponse.match(/```(?:json)?\s*([\s\S]*?)```/); let clean = response.replace(/<think>[\s\S]*?<\/think>/g, '').trim();
const jsonStr = codeBlockMatch ? codeBlockMatch[1].trim() : cleanResponse;
// Try code block
const codeBlockMatch = clean.match(/```(?:json)?\s*([\s\S]*?)```/);
const jsonStr = codeBlockMatch ? codeBlockMatch[1].trim() : clean;
try { try {
return JSON.parse(jsonStr); return JSON.parse(jsonStr);
} catch { } catch {
// Try to find JSON object
const jsonMatch = jsonStr.match(/\{[\s\S]*\}/); const jsonMatch = jsonStr.match(/\{[\s\S]*\}/);
if (jsonMatch) { if (jsonMatch) {
try { try {
@@ -341,113 +366,92 @@ function extractJsonFromResponse(response: string): Record<string, unknown> | nu
} }
/** /**
* Parse JSON response into IInvoice * Extract invoice from markdown using smartagent DualAgentOrchestrator
*/ * Validates JSON and retries if invalid
function parseJsonToInvoice(response: string): IInvoice | null {
const parsed = extractJsonFromResponse(response);
if (!parsed) return null;
return {
invoice_number: extractInvoiceNumber(String(parsed.invoice_number || '')),
invoice_date: extractDate(String(parsed.invoice_date || '')),
vendor_name: String(parsed.vendor_name || '').replace(/\*\*/g, '').replace(/`/g, '').trim(),
currency: extractCurrency(String(parsed.currency || '')),
net_amount: parseAmount(parsed.net_amount as string | number),
vat_amount: parseAmount(parsed.vat_amount as string | number),
total_amount: parseAmount(parsed.total_amount as string | number),
};
}
/**
* Extract invoice from markdown using GPT-OSS 20B (streaming)
*/ */
async function extractInvoiceFromMarkdown(markdown: string, queryId: string): Promise<IInvoice | null> { async function extractInvoiceFromMarkdown(markdown: string, queryId: string): Promise<IInvoice | null> {
const startTime = Date.now(); const startTime = Date.now();
const maxRetries = 2;
console.log(` [${queryId}] Invoice: ${markdown.length} chars, Prompt: ${JSON_EXTRACTION_PROMPT.length} chars`); console.log(` [${queryId}] Invoice: ${markdown.length} chars`);
const response = await fetch(`${OLLAMA_URL}/api/chat`, { // Build the extraction task with document context
method: 'POST', const taskPrompt = `Extract the invoice data from this document and output ONLY the JSON:
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify({
model: EXTRACTION_MODEL,
messages: [
{ role: 'user', content: 'Hi there, how are you?' },
{ role: 'assistant', content: 'Good, how can I help you today?' },
{ role: 'user', content: `Here is an invoice document:\n\n${markdown}` },
{ role: 'assistant', content: 'I have read the invoice document you provided. I can see all the text content. What would you like me to do with it?' },
{ role: 'user', content: JSON_EXTRACTION_PROMPT },
],
stream: true,
options: {
num_ctx: 32768, // Larger context for long invoices + thinking
temperature: 0, // Deterministic for JSON extraction
},
}),
signal: AbortSignal.timeout(600000), // 10 minute timeout for large documents
});
if (!response.ok) { ${markdown}
const elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
console.log(` [${queryId}] ERROR: ${response.status} (${elapsed}s)`);
throw new Error(`Ollama API error: ${response.status}`);
}
// Stream the response ${JSON_EXTRACTION_PROMPT}`;
let content = '';
let thinkingContent = '';
let thinkingStarted = false;
let outputStarted = false;
const reader = response.body!.getReader();
const decoder = new TextDecoder();
try { try {
while (true) { let result = await orchestrator.run(taskPrompt);
const { done, value } = await reader.read(); let elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
if (done) break; console.log(` [${queryId}] Status: ${result.status}, Iterations: ${result.iterations} (${elapsed}s)`);
const chunk = decoder.decode(value, { stream: true }); // Try to parse JSON from result
let jsonData: Record<string, unknown> | null = null;
let responseText = result.result || '';
// Each line is a JSON object if (result.success && responseText) {
for (const line of chunk.split('\n').filter(l => l.trim())) { jsonData = tryExtractJson(responseText);
try { }
const json = JSON.parse(line);
// Stream thinking tokens // Fallback: try parsing from history
const thinking = json.message?.thinking || ''; if (!jsonData && result.history?.length > 0) {
if (thinking) { const lastMessage = result.history[result.history.length - 1];
if (!thinkingStarted) { if (lastMessage?.content) {
process.stdout.write(` [${queryId}] THINKING: `); responseText = lastMessage.content;
thinkingStarted = true; jsonData = tryExtractJson(responseText);
} }
process.stdout.write(thinking); }
thinkingContent += thinking;
}
// Stream content tokens // If JSON is invalid, retry with correction request
const token = json.message?.content || ''; let retries = 0;
if (token) { while (!jsonData && retries < maxRetries) {
if (!outputStarted) { retries++;
if (thinkingStarted) process.stdout.write('\n'); console.log(` [${queryId}] Invalid JSON, requesting correction (retry ${retries}/${maxRetries})...`);
process.stdout.write(` [${queryId}] OUTPUT: `);
outputStarted = true; result = await orchestrator.continueTask(
} `Your response was not valid JSON. Please output ONLY the JSON object with no markdown, no explanation, no thinking tags. Just the raw JSON starting with { and ending with }. Format:
process.stdout.write(token); {"invoice_number":"X","invoice_date":"YYYY-MM-DD","vendor_name":"X","currency":"EUR","net_amount":0,"vat_amount":0,"total_amount":0}`
content += token; );
}
} catch { elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
// Ignore parse errors for partial chunks console.log(` [${queryId}] Retry ${retries}: ${result.status} (${elapsed}s)`);
responseText = result.result || '';
if (responseText) {
jsonData = tryExtractJson(responseText);
}
if (!jsonData && result.history?.length > 0) {
const lastMessage = result.history[result.history.length - 1];
if (lastMessage?.content) {
responseText = lastMessage.content;
jsonData = tryExtractJson(responseText);
} }
} }
} }
} finally {
if (thinkingStarted || outputStarted) process.stdout.write('\n'); if (!jsonData) {
console.log(` [${queryId}] Failed to get valid JSON after ${retries} retries`);
return null;
}
console.log(` [${queryId}] Valid JSON extracted`);
return {
invoice_number: extractInvoiceNumber(String(jsonData.invoice_number || '')),
invoice_date: extractDate(String(jsonData.invoice_date || '')),
vendor_name: String(jsonData.vendor_name || '').replace(/\*\*/g, '').replace(/`/g, '').trim(),
currency: extractCurrency(String(jsonData.currency || '')),
net_amount: parseAmount(jsonData.net_amount as string | number),
vat_amount: parseAmount(jsonData.vat_amount as string | number),
total_amount: parseAmount(jsonData.total_amount as string | number),
};
} catch (error) {
const elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
console.log(` [${queryId}] ERROR: ${error} (${elapsed}s)`);
throw error;
} }
const elapsed = ((Date.now() - startTime) / 1000).toFixed(1);
console.log(` [${queryId}] Done: ${thinkingContent.length} thinking chars, ${content.length} output chars (${elapsed}s)`);
return parseJsonToInvoice(content);
} }
/** /**
@@ -556,23 +560,45 @@ function findTestCases(): ITestCase[] {
const testCases = findTestCases(); const testCases = findTestCases();
console.log(`\nFound ${testCases.length} invoice test cases\n`); console.log(`\nFound ${testCases.length} invoice test cases\n`);
// Ensure temp directory exists // Ensure cache directory exists
if (!fs.existsSync(TEMP_MD_DIR)) { if (!fs.existsSync(MD_CACHE_DIR)) {
fs.mkdirSync(TEMP_MD_DIR, { recursive: true }); fs.mkdirSync(MD_CACHE_DIR, { recursive: true });
} }
// -------- STAGE 1: OCR with Nanonets -------- // -------- STAGE 1: OCR with Nanonets --------
tap.test('Stage 1: Setup Nanonets', async () => { tap.test('Stage 1: Convert invoices to markdown (with caching)', async () => {
console.log('\n========== STAGE 1: Nanonets OCR ==========\n'); console.log('\n========== STAGE 1: Nanonets OCR ==========\n');
const ok = await ensureNanonetsOcr();
expect(ok).toBeTrue();
});
tap.test('Stage 1: Convert all invoices to markdown', async () => { // Check which invoices need OCR conversion
console.log('\n Converting all invoice PDFs to markdown with Nanonets-OCR-s...\n'); const needsConversion: ITestCase[] = [];
let cachedCount = 0;
for (const tc of testCases) { for (const tc of testCases) {
const mdPath = path.join(MD_CACHE_DIR, `${tc.name}.md`);
if (fs.existsSync(mdPath)) {
cachedCount++;
tc.markdownPath = mdPath;
console.log(` [CACHED] ${tc.name} - using cached markdown`);
} else {
needsConversion.push(tc);
}
}
console.log(`\n Summary: ${cachedCount} cached, ${needsConversion.length} need conversion\n`);
if (needsConversion.length === 0) {
console.log(' All invoices already cached, skipping Nanonets OCR\n');
return;
}
// Start Nanonets only if there are files to convert
console.log(' Starting Nanonets for OCR conversion...\n');
const ok = await ensureNanonetsOcr();
expect(ok).toBeTrue();
// Convert only the invoices that need conversion
for (const tc of needsConversion) {
console.log(`\n === ${tc.name} ===`); console.log(`\n === ${tc.name} ===`);
const images = convertPdfToImages(tc.pdfPath); const images = convertPdfToImages(tc.pdfPath);
@@ -580,13 +606,13 @@ tap.test('Stage 1: Convert all invoices to markdown', async () => {
const markdown = await convertDocumentToMarkdown(images, tc.name); const markdown = await convertDocumentToMarkdown(images, tc.name);
const mdPath = path.join(TEMP_MD_DIR, `${tc.name}.md`); const mdPath = path.join(MD_CACHE_DIR, `${tc.name}.md`);
fs.writeFileSync(mdPath, markdown); fs.writeFileSync(mdPath, markdown);
tc.markdownPath = mdPath; tc.markdownPath = mdPath;
console.log(` Saved: ${mdPath}`); console.log(` Saved: ${mdPath}`);
} }
console.log('\n Stage 1 complete: All invoices converted to markdown\n'); console.log(`\n Stage 1 complete: ${needsConversion.length} invoices converted to markdown\n`);
}); });
tap.test('Stage 1: Stop Nanonets', async () => { tap.test('Stage 1: Stop Nanonets', async () => {
@@ -605,6 +631,42 @@ tap.test('Stage 2: Setup Ollama + GPT-OSS 20B', async () => {
const extractionOk = await ensureExtractionModel(); const extractionOk = await ensureExtractionModel();
expect(extractionOk).toBeTrue(); expect(extractionOk).toBeTrue();
// Initialize SmartAi and DualAgentOrchestrator
console.log(' [SmartAgent] Starting SmartAi...');
await smartAi.start();
console.log(' [SmartAgent] Creating DualAgentOrchestrator...');
orchestrator = new DualAgentOrchestrator({
smartAiInstance: smartAi,
defaultProvider: 'ollama',
guardianPolicyPrompt: `
JSON EXTRACTION POLICY:
- APPROVE all JSON extraction tasks
- This is a read-only operation - no file system or network access needed
- The task is to extract structured data from document text
`,
driverSystemMessage: `You are a precise JSON extraction assistant. Your only job is to extract invoice data from documents.
CRITICAL RULES:
1. Output ONLY valid JSON - no markdown, no explanations, no thinking
2. Use the exact format requested
3. If you cannot find a value, use empty string "" or 0 for numbers
When done, wrap your JSON in <task_complete></task_complete> tags.`,
maxIterations: 3,
// Enable streaming for real-time progress visibility
onToken: (token, source) => {
if (source === 'driver') {
process.stdout.write(token);
}
},
});
// No tools needed for JSON extraction
console.log(' [SmartAgent] Starting orchestrator...');
await orchestrator.start();
console.log(' [SmartAgent] Ready for extraction');
}); });
let passedCount = 0; let passedCount = 0;
@@ -619,7 +681,7 @@ for (const tc of testCases) {
const startTime = Date.now(); const startTime = Date.now();
const mdPath = path.join(TEMP_MD_DIR, `${tc.name}.md`); const mdPath = path.join(MD_CACHE_DIR, `${tc.name}.md`);
if (!fs.existsSync(mdPath)) { if (!fs.existsSync(mdPath)) {
throw new Error(`Markdown not found: ${mdPath}. Run Stage 1 first.`); throw new Error(`Markdown not found: ${mdPath}. Run Stage 1 first.`);
} }
@@ -649,6 +711,14 @@ for (const tc of testCases) {
} }
tap.test('Summary', async () => { tap.test('Summary', async () => {
// Cleanup orchestrator and SmartAi
if (orchestrator) {
console.log('\n [SmartAgent] Stopping orchestrator...');
await orchestrator.stop();
}
console.log(' [SmartAgent] Stopping SmartAi...');
await smartAi.stop();
const totalInvoices = testCases.length; const totalInvoices = testCases.length;
const accuracy = totalInvoices > 0 ? (passedCount / totalInvoices) * 100 : 0; const accuracy = totalInvoices > 0 ? (passedCount / totalInvoices) * 100 : 0;
const totalTimeMs = processingTimes.reduce((a, b) => a + b, 0); const totalTimeMs = processingTimes.reduce((a, b) => a + b, 0);
@@ -658,7 +728,7 @@ tap.test('Summary', async () => {
console.log(` Invoice Summary (Nanonets + GPT-OSS 20B)`); console.log(` Invoice Summary (Nanonets + GPT-OSS 20B)`);
console.log(`========================================`); console.log(`========================================`);
console.log(` Stage 1: Nanonets-OCR-s (doc -> md)`); console.log(` Stage 1: Nanonets-OCR-s (doc -> md)`);
console.log(` Stage 2: GPT-OSS 20B (md -> JSON)`); console.log(` Stage 2: GPT-OSS 20B + SmartAgent (md -> JSON)`);
console.log(` Passed: ${passedCount}/${totalInvoices}`); console.log(` Passed: ${passedCount}/${totalInvoices}`);
console.log(` Failed: ${failedCount}/${totalInvoices}`); console.log(` Failed: ${failedCount}/${totalInvoices}`);
console.log(` Accuracy: ${accuracy.toFixed(1)}%`); console.log(` Accuracy: ${accuracy.toFixed(1)}%`);
@@ -666,14 +736,7 @@ tap.test('Summary', async () => {
console.log(` Total time: ${(totalTimeMs / 1000).toFixed(1)}s`); console.log(` Total time: ${(totalTimeMs / 1000).toFixed(1)}s`);
console.log(` Avg per inv: ${avgTimeSec.toFixed(1)}s`); console.log(` Avg per inv: ${avgTimeSec.toFixed(1)}s`);
console.log(`========================================\n`); console.log(`========================================\n`);
console.log(` Cache location: ${MD_CACHE_DIR}\n`);
// Cleanup temp files
try {
fs.rmSync(TEMP_MD_DIR, { recursive: true, force: true });
console.log(` Cleaned up temp directory: ${TEMP_MD_DIR}\n`);
} catch {
// Ignore
}
}); });
export default tap.start(); export default tap.start();