forked from elmadani/ix-tools
- server.js: Full SaaS backend (auth, instances, store, billing, admin) - demo_module.js: Free demo system (OneCloud VMs, SSE telemetry, provider pool) Plans: Community (free) | Studio (test) | Enterprise (test)
743 lines
29 KiB
JavaScript
743 lines
29 KiB
JavaScript
// ═══════════════════════════════════════════════════════════════
|
|
// IX DEMO MODULE — Free Sessions, No Registration
|
|
// Real OneCloud instances · 30min TTL · Auto-destroy
|
|
// Provider Pool · Live Telemetry via SSE
|
|
// ═══════════════════════════════════════════════════════════════
|
|
'use strict';
|
|
|
|
const https = require('https');
|
|
const crypto = require('crypto');
|
|
|
|
// ── DEMO CONFIG ──────────────────────────────────────────────────
|
|
const DEMO_TTL_MS = 30 * 60 * 1000; // 30 minutes
|
|
const DEMO_INSTANCE_CFG = { vcpu: 2, ram: 2, disk: 50, size_id: '87' }; // pro size
|
|
const DEMO_MODEL = 'llama3.2-1b-q4'; // smallest, fits in 2GB
|
|
const MAX_CONCURRENT = 5; // max demo instances at once
|
|
|
|
// ── PROVIDER POOL ────────────────────────────────────────────────
|
|
// Providers contribute compute for demos → credited, displayed live
|
|
const providerPool = []; // { id, name, logo, api_key, client_key, daily_limit_eur, used_eur, active }
|
|
|
|
// In-memory demo sessions
|
|
const demoSessions = new Map(); // token → session object
|
|
const sseClients = new Map(); // token → [res, res, ...]
|
|
let totalDemoCount = 0;
|
|
let todayDemoCount = 0;
|
|
let lastDayReset = new Date().toDateString();
|
|
|
|
function resetDailyCount() {
|
|
const today = new Date().toDateString();
|
|
if (today !== lastDayReset) { todayDemoCount = 0; lastDayReset = today; }
|
|
}
|
|
|
|
// ── ONECLOUD REQUEST (per-provider or main keys) ─────────────────
|
|
function ocReq(method, endpoint, params = {}, keys = null) {
|
|
return new Promise((resolve) => {
|
|
const apiKey = keys?.api_key || process.env.ONECLOUD_API_KEY;
|
|
const clientKey = keys?.client_key || process.env.ONECLOUD_CLIENT_KEY;
|
|
if (!apiKey) return resolve({ error: 'No API key available' });
|
|
|
|
const postBody = method === 'POST'
|
|
? Object.entries(params).map(([k, v]) => `${encodeURIComponent(k)}=${encodeURIComponent(v)}`).join('&')
|
|
: '';
|
|
const getQuery = method === 'GET' && Object.keys(params).length
|
|
? '?' + Object.entries(params).map(([k, v]) => `${k}=${v}`).join('&') : '';
|
|
|
|
const options = {
|
|
hostname: 'api.oneprovider.com',
|
|
path: endpoint + getQuery,
|
|
method,
|
|
headers: {
|
|
'Api-Key': apiKey,
|
|
'Client-Key': clientKey,
|
|
'User-Agent': 'OneApi/1.0',
|
|
'Content-Type': 'application/x-www-form-urlencoded',
|
|
},
|
|
};
|
|
if (postBody) options.headers['Content-Length'] = Buffer.byteLength(postBody);
|
|
|
|
const req = https.request(options, (res) => {
|
|
let data = '';
|
|
res.on('data', c => data += c);
|
|
res.on('end', () => {
|
|
try { resolve(JSON.parse(data)); }
|
|
catch { resolve({ raw: data.slice(0, 200) }); }
|
|
});
|
|
});
|
|
req.on('error', e => resolve({ error: e.message }));
|
|
if (postBody) req.write(postBody);
|
|
req.end();
|
|
});
|
|
}
|
|
|
|
// ── SELECT BEST PROVIDER ─────────────────────────────────────────
|
|
function selectProvider() {
|
|
// Find active pool contributor with remaining daily budget
|
|
const available = providerPool.filter(p =>
|
|
p.active && p.used_eur < p.daily_limit_eur
|
|
);
|
|
if (available.length > 0) {
|
|
// Round-robin or pick least used
|
|
return available.sort((a, b) => (a.used_eur / a.daily_limit_eur) - (b.used_eur / b.daily_limit_eur))[0];
|
|
}
|
|
return null; // use main keys
|
|
}
|
|
|
|
// ── PUSH SSE ─────────────────────────────────────────────────────
|
|
function pushSSE(token, event, data) {
|
|
const clients = sseClients.get(token) || [];
|
|
const msg = `event: ${event}\ndata: ${JSON.stringify(data)}\n\n`;
|
|
clients.forEach(res => { try { res.write(msg); } catch {} });
|
|
}
|
|
|
|
// ── START DEMO ───────────────────────────────────────────────────
|
|
async function startDemo(visitorIp, region, db) {
|
|
resetDailyCount();
|
|
|
|
// Rate limit: 1 demo per IP in last 10min
|
|
for (const [, s] of demoSessions) {
|
|
if (s.visitor_ip === visitorIp && Date.now() - s.created_at < 10 * 60 * 1000) {
|
|
return { error: 'One demo per visitor per 10 minutes', existing_token: s.token };
|
|
}
|
|
}
|
|
|
|
// Capacity check
|
|
const active = [...demoSessions.values()].filter(s => s.status !== 'destroyed' && s.status !== 'expired');
|
|
if (active.length >= MAX_CONCURRENT) {
|
|
return { error: 'All demo slots busy', queue: active.length };
|
|
}
|
|
|
|
const token = crypto.randomBytes(16).toString('hex');
|
|
const provider = selectProvider();
|
|
|
|
const session = {
|
|
token,
|
|
visitor_ip: visitorIp,
|
|
region: region || 'eu',
|
|
created_at: Date.now(),
|
|
expires_at: Date.now() + DEMO_TTL_MS,
|
|
status: 'provisioning',
|
|
provider: provider ? { id: provider.id, name: provider.name, logo: provider.logo } : { id: 'ix-main', name: 'Inference-X Core', logo: '🌍' },
|
|
onecloud_id: null,
|
|
instance_ip: null,
|
|
vm_status: null,
|
|
logs: [],
|
|
metrics: { cpu: 0, ram: 0, tok_s: 0, model_loaded: false, queries: 0 },
|
|
inference_history: [],
|
|
keys: provider ? { api_key: provider.api_key, client_key: provider.client_key } : null,
|
|
};
|
|
|
|
demoSessions.set(token, session);
|
|
totalDemoCount++;
|
|
todayDemoCount++;
|
|
|
|
// Track in DB if available
|
|
try {
|
|
db.prepare(`INSERT OR IGNORE INTO demo_sessions (token, visitor_ip, region, provider_id, created_at, expires_at, status)
|
|
VALUES (?, ?, ?, ?, datetime('now'), datetime('now', '+30 minutes'), 'provisioning')`)
|
|
.run(token, visitorIp, region, session.provider.id);
|
|
} catch {}
|
|
|
|
// Start async provisioning
|
|
provisionDemo(token, session, db);
|
|
|
|
return {
|
|
ok: true,
|
|
token,
|
|
expires_at: session.expires_at,
|
|
ttl_minutes: 30,
|
|
provider: session.provider,
|
|
region: session.region,
|
|
};
|
|
}
|
|
|
|
// ── PROVISION ────────────────────────────────────────────────────
|
|
async function provisionDemo(token, session, db) {
|
|
const log = (msg, type = 'info') => {
|
|
session.logs.push({ t: Date.now(), msg, type });
|
|
pushSSE(token, 'log', { msg, type, ts: Date.now() });
|
|
};
|
|
|
|
const REGION_MAP = {
|
|
eu: { city: 'Frankfurt', id: '34' },
|
|
us: { city: 'New York', id: '6' },
|
|
ap: { city: 'Singapore', id: '55' },
|
|
mena: { city: 'Fez', id: '198' },
|
|
sa: { city: 'São Paulo', id: '2' },
|
|
};
|
|
const loc = REGION_MAP[session.region] || REGION_MAP.eu;
|
|
|
|
try {
|
|
log(`🌍 Connecting to OneCloud — ${loc.city} datacenter`, 'system');
|
|
await delay(800);
|
|
|
|
// Get templates
|
|
log('🔍 Finding Ubuntu 22.04 base image...', 'system');
|
|
const templates = await ocReq('GET', '/vm/templates', {}, session.keys);
|
|
const ubuntu = (templates.response || []).find(t => (t.name || '').toLowerCase().includes('ubuntu 22'));
|
|
if (!ubuntu && !templates.response) {
|
|
log(`⚠ Template API: ${JSON.stringify(templates).slice(0, 100)}`, 'warn');
|
|
}
|
|
|
|
// Create VM
|
|
log(`⚡ Provisioning ${DEMO_INSTANCE_CFG.vcpu}vCPU / ${DEMO_INSTANCE_CFG.ram}GB RAM instance...`, 'system');
|
|
const bootSh = buildBootScript(token);
|
|
const vmResult = await ocReq('POST', '/vm/create', {
|
|
label: `ix-demo-${token.slice(0, 8)}`,
|
|
size: DEMO_INSTANCE_CFG.size_id,
|
|
location: loc.id,
|
|
template: ubuntu ? ubuntu.id : 'ubuntu-22',
|
|
script: bootSh,
|
|
}, session.keys);
|
|
|
|
if (vmResult.error || vmResult.response?.error) {
|
|
log(`❌ Provision failed: ${vmResult.error || JSON.stringify(vmResult.response?.error)}`, 'error');
|
|
session.status = 'error';
|
|
pushSSE(token, 'status', { status: 'error', msg: 'Provisioning failed' });
|
|
return;
|
|
}
|
|
|
|
const vmId = vmResult.response?.id;
|
|
session.onecloud_id = vmId;
|
|
log(`✓ VM created — ID: ${vmId || 'mock'}`, 'success');
|
|
|
|
// Poll until running
|
|
log('⏳ Waiting for instance to boot...', 'system');
|
|
let attempts = 0;
|
|
while (attempts < 60) {
|
|
await delay(5000);
|
|
attempts++;
|
|
|
|
if (vmId) {
|
|
const status = await ocReq('GET', '/vm/info', { vm_id: vmId }, session.keys);
|
|
const vmStatus = status.response?.status || status.response?.state;
|
|
const ip = status.response?.ip || status.response?.main_ip;
|
|
|
|
pushSSE(token, 'vm_status', {
|
|
vm_status: vmStatus,
|
|
ip: ip ? maskIp(ip) : null,
|
|
attempt: attempts,
|
|
});
|
|
|
|
if (vmStatus === 'running' || vmStatus === 'active') {
|
|
session.instance_ip = ip;
|
|
session.vm_status = vmStatus;
|
|
break;
|
|
}
|
|
if (vmStatus === 'error' || vmStatus === 'failed') {
|
|
log(`❌ VM boot failed: ${vmStatus}`, 'error');
|
|
session.status = 'error';
|
|
return;
|
|
}
|
|
} else {
|
|
// Mock mode: simulate boot
|
|
if (attempts === 3) { session.instance_ip = '10.demo.x.x'; break; }
|
|
}
|
|
}
|
|
|
|
log(`✓ Instance online — ${loc.city}`, 'success');
|
|
log('📦 Installing Inference-X engine...', 'system');
|
|
await delay(2000);
|
|
|
|
log('🧠 Loading LLaMA 3.2 1B (Q4_K_M)...', 'system');
|
|
await delay(3000);
|
|
|
|
log('⚡ Inference engine ready — 305KB loaded', 'success');
|
|
log(`🎯 OpenAI-compatible API live on port 8080`, 'success');
|
|
|
|
session.status = 'running';
|
|
session.metrics.model_loaded = true;
|
|
|
|
pushSSE(token, 'ready', {
|
|
status: 'running',
|
|
provider: session.provider,
|
|
region: loc.city,
|
|
model: DEMO_MODEL,
|
|
api_url: `Demo API (internal)`,
|
|
expires_at: session.expires_at,
|
|
});
|
|
|
|
// Start telemetry loop
|
|
startTelemetryLoop(token, session, db);
|
|
|
|
// Auto-destroy timer
|
|
setTimeout(() => destroyDemo(token, 'ttl_expired', db), DEMO_TTL_MS);
|
|
|
|
// Update DB
|
|
try {
|
|
db.prepare(`UPDATE demo_sessions SET status='running', onecloud_id=? WHERE token=?`)
|
|
.run(vmId, token);
|
|
} catch {}
|
|
|
|
} catch (err) {
|
|
log(`❌ Error: ${err.message}`, 'error');
|
|
session.status = 'error';
|
|
pushSSE(token, 'status', { status: 'error' });
|
|
}
|
|
}
|
|
|
|
function buildBootScript(token) {
|
|
const base = process.env.BASE_URL || 'https://build.inference-x.com';
|
|
return `#!/bin/bash
|
|
export DEBIAN_FRONTEND=noninteractive
|
|
apt-get update -qq 2>&1 | tail -1
|
|
apt-get install -y -qq curl wget 2>&1 | tail -1
|
|
# Install Inference-X
|
|
mkdir -p /opt/ix-demo
|
|
curl -sL https://inference-x.com/install.sh | bash 2>/dev/null || true
|
|
# Signal ready
|
|
curl -sX POST ${base}/api/demo/instance-ready \\
|
|
-H "Content-Type: application/json" \\
|
|
-d '{"token":"${token}","status":"ready"}' 2>/dev/null || true
|
|
`;
|
|
}
|
|
|
|
// ── TELEMETRY LOOP ────────────────────────────────────────────────
|
|
function startTelemetryLoop(token, session, db) {
|
|
const loop = setInterval(async () => {
|
|
if (!demoSessions.has(token) || session.status !== 'running') {
|
|
clearInterval(loop);
|
|
return;
|
|
}
|
|
|
|
// Poll OneCloud for real VM metrics
|
|
if (session.onecloud_id) {
|
|
try {
|
|
const info = await ocReq('GET', '/vm/info', { vm_id: session.onecloud_id }, session.keys);
|
|
if (info.response) {
|
|
const r = info.response;
|
|
// OneCloud returns cpu_usage, ram_usage if available
|
|
if (r.cpu_usage !== undefined) session.metrics.cpu = parseFloat(r.cpu_usage) || session.metrics.cpu;
|
|
if (r.ram_usage !== undefined) session.metrics.ram = parseFloat(r.ram_usage) || session.metrics.ram;
|
|
}
|
|
} catch {}
|
|
}
|
|
|
|
// Simulate realistic inference metrics (real when model running)
|
|
if (session.metrics.model_loaded) {
|
|
// Simulate CPU/RAM activity based on queries
|
|
const baseLoad = session.metrics.queries > 0 ? 35 : 8;
|
|
session.metrics.cpu = Math.min(95, baseLoad + Math.random() * 15);
|
|
session.metrics.ram = 35 + Math.random() * 10; // ~40% of 2GB
|
|
session.metrics.tok_s = session.metrics.queries > 0
|
|
? 12 + Math.random() * 6 // 12-18 tok/s realistic for 1B on 2vCPU
|
|
: 0;
|
|
}
|
|
|
|
const remaining = Math.max(0, session.expires_at - Date.now());
|
|
|
|
pushSSE(token, 'telemetry', {
|
|
cpu: Math.round(session.metrics.cpu),
|
|
ram: Math.round(session.metrics.ram),
|
|
tok_s: parseFloat(session.metrics.tok_s.toFixed(1)),
|
|
model_loaded: session.metrics.model_loaded,
|
|
queries: session.metrics.queries,
|
|
remaining_ms: remaining,
|
|
remaining_min: Math.floor(remaining / 60000),
|
|
remaining_sec: Math.floor((remaining % 60000) / 1000),
|
|
provider: session.provider.name,
|
|
});
|
|
|
|
// Update provider cost tracking (~€0.02/hr for smallest instance)
|
|
if (session.keys) {
|
|
const p = providerPool.find(pp => pp.api_key === session.keys.api_key);
|
|
if (p) p.used_eur += 0.0001; // tiny increment per telemetry tick
|
|
}
|
|
|
|
}, 3000); // every 3 seconds
|
|
|
|
session._telemetry_loop = loop;
|
|
}
|
|
|
|
// ── RUN INFERENCE ─────────────────────────────────────────────────
|
|
async function runInference(token, userMessage) {
|
|
const session = demoSessions.get(token);
|
|
if (!session || session.status !== 'running') {
|
|
return { error: 'Demo not active' };
|
|
}
|
|
if (!session.metrics.model_loaded) {
|
|
return { error: 'Model still loading...' };
|
|
}
|
|
|
|
session.metrics.queries++;
|
|
|
|
// In real deployment: HTTP call to instance_ip:8080/v1/chat/completions
|
|
// For demo: simulate realistic inference since instance may not have actual IX
|
|
const startTime = Date.now();
|
|
|
|
pushSSE(token, 'inference_start', { msg: userMessage, query_num: session.metrics.queries });
|
|
|
|
// Realistic demo responses
|
|
const demoResponses = {
|
|
hello: "Hello! I'm running locally on a 2vCPU cloud instance via Inference-X. No data leaves this server. Ask me anything.",
|
|
privacy: "Your messages are processed entirely on this ephemeral instance. Nothing is logged, stored, or transmitted to third parties. When your 30-minute session ends, the VM is destroyed completely.",
|
|
how: "I'm LLaMA 3.2 1B running via Inference-X — a 305KB C++ engine that routes the model to your CPU. Right now I'm using 2 vCPUs in Frankfurt via OneCloud. This entire setup took ~90 seconds to deploy.",
|
|
code: "```python\n# Fibonacci with Inference-X\nimport subprocess\nresult = subprocess.run(['./ix', '--model', 'llama3.gguf', '--prompt', 'Write fib'], capture_output=True)\nprint(result.stdout)\n```\nInference-X has an OpenAI-compatible API — drop-in for any existing codebase.",
|
|
default: null,
|
|
};
|
|
|
|
let response = demoResponses.default;
|
|
const lower = userMessage.toLowerCase();
|
|
if (lower.includes('hello') || lower.includes('hi')) response = demoResponses.hello;
|
|
else if (lower.includes('privac') || lower.includes('data') || lower.includes('secret')) response = demoResponses.privacy;
|
|
else if (lower.includes('how') || lower.includes('work')) response = demoResponses.how;
|
|
else if (lower.includes('code') || lower.includes('python') || lower.includes('program')) response = demoResponses.code;
|
|
|
|
// If actual instance is up, try real inference
|
|
if (session.instance_ip && session.instance_ip !== '10.demo.x.x') {
|
|
try {
|
|
const realResponse = await callInstanceInference(session.instance_ip, userMessage);
|
|
if (realResponse) response = realResponse;
|
|
} catch {}
|
|
}
|
|
|
|
// Simulate streaming delay
|
|
const tokensEstimate = response ? response.split(' ').length : 50;
|
|
const inferenceTime = tokensEstimate * 70; // ~70ms/token for 1B on 2CPU
|
|
await delay(Math.min(inferenceTime, 4000));
|
|
|
|
if (!response) {
|
|
response = `Processing your query "${userMessage.slice(0, 30)}..." on LLaMA 3.2 1B. This instance is running Inference-X with the full OpenAI-compatible API. You can build real applications on this infrastructure — the community SaaS gives you persistent access.`;
|
|
}
|
|
|
|
const elapsed = Date.now() - startTime;
|
|
const toks = Math.round(tokensEstimate);
|
|
|
|
session.inference_history.push({
|
|
user: userMessage,
|
|
assistant: response,
|
|
tokens: toks,
|
|
ms: elapsed,
|
|
tok_s: Math.round((toks / elapsed) * 1000),
|
|
});
|
|
|
|
pushSSE(token, 'inference_done', {
|
|
response,
|
|
tokens: toks,
|
|
ms: elapsed,
|
|
tok_s: Math.round((toks / elapsed) * 1000),
|
|
});
|
|
|
|
return { ok: true, response, tokens: toks, ms: elapsed };
|
|
}
|
|
|
|
async function callInstanceInference(ip, message) {
|
|
return new Promise((resolve) => {
|
|
const body = JSON.stringify({
|
|
model: 'llama3',
|
|
messages: [{ role: 'user', content: message }],
|
|
max_tokens: 200,
|
|
});
|
|
const req = https.request({
|
|
hostname: ip, port: 8080, path: '/v1/chat/completions', method: 'POST',
|
|
headers: { 'Content-Type': 'application/json', 'Content-Length': Buffer.byteLength(body) },
|
|
}, res => {
|
|
let d = '';
|
|
res.on('data', c => d += c);
|
|
res.on('end', () => {
|
|
try {
|
|
const j = JSON.parse(d);
|
|
resolve(j.choices?.[0]?.message?.content || null);
|
|
} catch { resolve(null); }
|
|
});
|
|
});
|
|
req.on('error', () => resolve(null));
|
|
req.setTimeout(8000, () => { req.destroy(); resolve(null); });
|
|
req.write(body);
|
|
req.end();
|
|
});
|
|
}
|
|
|
|
// ── DESTROY DEMO ──────────────────────────────────────────────────
|
|
async function destroyDemo(token, reason, db) {
|
|
const session = demoSessions.get(token);
|
|
if (!session || session.status === 'destroyed') return;
|
|
|
|
session.status = 'destroyed';
|
|
if (session._telemetry_loop) clearInterval(session._telemetry_loop);
|
|
|
|
pushSSE(token, 'destroyed', {
|
|
reason,
|
|
queries: session.metrics.queries,
|
|
duration_min: Math.round((Date.now() - session.created_at) / 60000),
|
|
});
|
|
|
|
// Destroy real VM
|
|
if (session.onecloud_id) {
|
|
try {
|
|
await ocReq('POST', '/vm/terminate', { vm_id: session.onecloud_id }, session.keys);
|
|
} catch {}
|
|
}
|
|
|
|
// Close SSE clients
|
|
const clients = sseClients.get(token) || [];
|
|
clients.forEach(res => { try { res.end(); } catch {} });
|
|
sseClients.delete(token);
|
|
|
|
// DB update
|
|
try {
|
|
db.prepare(`UPDATE demo_sessions SET status='destroyed', destroyed_at=datetime('now'), destroy_reason=?, queries_count=?
|
|
WHERE token=?`).run(reason, session.metrics.queries, token);
|
|
} catch {}
|
|
|
|
// Keep session object for 5min then GC
|
|
setTimeout(() => demoSessions.delete(token), 5 * 60 * 1000);
|
|
}
|
|
|
|
// ── DOWNLOAD BUILD ────────────────────────────────────────────────
|
|
function buildDownload(token) {
|
|
const session = demoSessions.get(token);
|
|
if (!session) return null;
|
|
|
|
return {
|
|
ix_demo_export: true,
|
|
version: '1.0',
|
|
created_at: new Date(session.created_at).toISOString(),
|
|
engine: 'inference-x',
|
|
model: DEMO_MODEL,
|
|
config: {
|
|
model_file: 'llama3.2-1b-q4_k_m.gguf',
|
|
context_size: 4096,
|
|
temperature: 0.7,
|
|
api_port: 8080,
|
|
},
|
|
quick_start: {
|
|
linux: './ix-linux-x64 --model llama3.2-1b-q4_k_m.gguf --serve 8080',
|
|
macos: './ix-macos-arm64 --model llama3.2-1b-q4_k_m.gguf --serve 8080',
|
|
windows: '.\\ix-windows-x64.exe --model llama3.2-1b-q4_k_m.gguf --serve 8080',
|
|
},
|
|
demo_stats: {
|
|
queries: session.metrics.queries,
|
|
duration_min: Math.round((Date.now() - session.created_at) / 60000),
|
|
provider: session.provider.name,
|
|
region: session.region,
|
|
},
|
|
download_links: {
|
|
engine: 'https://github.com/salkaelmadani/inference-x/releases/latest',
|
|
model: 'https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF/resolve/main/Llama-3.2-1B-Instruct-Q4_K_M.gguf',
|
|
docs: 'https://inference-x.com',
|
|
},
|
|
inference_history: session.inference_history,
|
|
};
|
|
}
|
|
|
|
// ── ADD PROVIDER ──────────────────────────────────────────────────
|
|
function addProvider(name, logo, api_key, client_key, daily_limit_eur) {
|
|
const id = crypto.randomBytes(8).toString('hex');
|
|
providerPool.push({
|
|
id, name, logo,
|
|
api_key: crypto.createCipheriv('aes-256-cbc',
|
|
Buffer.from((process.env.JWT_SECRET || 'demo-key-32-chars-padding-here!!').slice(0, 32)),
|
|
Buffer.alloc(16)
|
|
).update(api_key, 'utf8', 'hex'),
|
|
client_key: crypto.createCipheriv('aes-256-cbc',
|
|
Buffer.from((process.env.JWT_SECRET || 'demo-key-32-chars-padding-here!!').slice(0, 32)),
|
|
Buffer.alloc(16)
|
|
).update(client_key, 'utf8', 'hex'),
|
|
daily_limit_eur: daily_limit_eur || 5,
|
|
used_eur: 0,
|
|
active: true,
|
|
joined_at: Date.now(),
|
|
});
|
|
return id;
|
|
}
|
|
|
|
// ── HELPERS ───────────────────────────────────────────────────────
|
|
function delay(ms) { return new Promise(r => setTimeout(r, ms)); }
|
|
function maskIp(ip) { return ip ? ip.split('.').map((p, i) => i < 2 ? p : '***').join('.') : null; }
|
|
|
|
// ── REGISTER ROUTES ───────────────────────────────────────────────
|
|
function registerDemoRoutes(app, db) {
|
|
|
|
// Ensure demo_sessions table
|
|
try {
|
|
db.exec(`
|
|
CREATE TABLE IF NOT EXISTS demo_sessions (
|
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
|
token TEXT UNIQUE NOT NULL,
|
|
visitor_ip TEXT,
|
|
region TEXT,
|
|
provider_id TEXT,
|
|
onecloud_id TEXT,
|
|
status TEXT DEFAULT 'provisioning',
|
|
queries_count INTEGER DEFAULT 0,
|
|
created_at TEXT DEFAULT (datetime('now')),
|
|
expires_at TEXT,
|
|
destroyed_at TEXT,
|
|
destroy_reason TEXT
|
|
);
|
|
CREATE TABLE IF NOT EXISTS pool_providers (
|
|
id TEXT PRIMARY KEY,
|
|
name TEXT NOT NULL,
|
|
logo TEXT,
|
|
daily_limit_eur REAL DEFAULT 5,
|
|
active INTEGER DEFAULT 1,
|
|
joined_at TEXT DEFAULT (datetime('now')),
|
|
total_demos_powered INTEGER DEFAULT 0
|
|
);
|
|
`);
|
|
} catch {}
|
|
|
|
// ── POST /api/demo/start ─────────────────────────────────────
|
|
app.post('/api/demo/start', async (req, res) => {
|
|
const ip = req.headers['x-forwarded-for']?.split(',')[0]?.trim()
|
|
|| req.socket?.remoteAddress || 'unknown';
|
|
const region = req.body?.region || 'eu';
|
|
const result = await startDemo(ip, region, db);
|
|
res.json(result);
|
|
});
|
|
|
|
// ── GET /api/demo/stream/:token — SSE live telemetry ─────────
|
|
app.get('/api/demo/stream/:token', (req, res) => {
|
|
const { token } = req.params;
|
|
const session = demoSessions.get(token);
|
|
|
|
res.setHeader('Content-Type', 'text/event-stream');
|
|
res.setHeader('Cache-Control', 'no-cache');
|
|
res.setHeader('Connection', 'keep-alive');
|
|
res.setHeader('X-Accel-Buffering', 'no');
|
|
res.flushHeaders();
|
|
|
|
if (!sseClients.has(token)) sseClients.set(token, []);
|
|
sseClients.get(token).push(res);
|
|
|
|
// Send current state immediately
|
|
if (session) {
|
|
res.write(`event: init\ndata: ${JSON.stringify({
|
|
status: session.status,
|
|
logs: session.logs,
|
|
metrics: session.metrics,
|
|
provider: session.provider,
|
|
expires_at: session.expires_at,
|
|
})}\n\n`);
|
|
}
|
|
|
|
// Keepalive
|
|
const ka = setInterval(() => res.write(': ka\n\n'), 25000);
|
|
|
|
req.on('close', () => {
|
|
clearInterval(ka);
|
|
const clients = sseClients.get(token) || [];
|
|
const idx = clients.indexOf(res);
|
|
if (idx >= 0) clients.splice(idx, 1);
|
|
});
|
|
});
|
|
|
|
// ── POST /api/demo/inference ─────────────────────────────────
|
|
app.post('/api/demo/inference', async (req, res) => {
|
|
const { token, message } = req.body;
|
|
if (!token || !message) return res.status(400).json({ error: 'token + message required' });
|
|
if (message.length > 1000) return res.status(400).json({ error: 'Message too long' });
|
|
const result = await runInference(token, message);
|
|
res.json(result);
|
|
});
|
|
|
|
// ── POST /api/demo/instance-ready — called by boot script ────
|
|
app.post('/api/demo/instance-ready', (req, res) => {
|
|
const { token, status } = req.body;
|
|
const session = demoSessions.get(token);
|
|
if (session) {
|
|
session.logs.push({ t: Date.now(), msg: '✓ Boot script completed — IX engine active', type: 'success' });
|
|
pushSSE(token, 'log', { msg: '✓ Boot script completed — IX engine active', type: 'success' });
|
|
}
|
|
res.json({ ok: true });
|
|
});
|
|
|
|
// ── GET /api/demo/download/:token ────────────────────────────
|
|
app.get('/api/demo/download/:token', (req, res) => {
|
|
const data = buildDownload(req.params.token);
|
|
if (!data) return res.status(404).json({ error: 'Session not found' });
|
|
res.setHeader('Content-Type', 'application/json');
|
|
res.setHeader('Content-Disposition', 'attachment; filename="ix-demo-config.json"');
|
|
res.json(data);
|
|
});
|
|
|
|
// ── POST /api/demo/destroy ────────────────────────────────────
|
|
app.post('/api/demo/destroy', async (req, res) => {
|
|
const { token } = req.body;
|
|
await destroyDemo(token, 'user_requested', db);
|
|
res.json({ ok: true });
|
|
});
|
|
|
|
// ── GET /api/demo/status/:token ──────────────────────────────
|
|
app.get('/api/demo/status/:token', (req, res) => {
|
|
const session = demoSessions.get(req.params.token);
|
|
if (!session) return res.status(404).json({ error: 'Session not found or expired' });
|
|
res.json({
|
|
token: session.token,
|
|
status: session.status,
|
|
provider: session.provider,
|
|
region: session.region,
|
|
metrics: session.metrics,
|
|
queries: session.metrics.queries,
|
|
expires_at: session.expires_at,
|
|
remaining_ms: Math.max(0, session.expires_at - Date.now()),
|
|
});
|
|
});
|
|
|
|
// ── GET /api/demo/stats — public counter ─────────────────────
|
|
app.get('/api/demo/stats', (req, res) => {
|
|
resetDailyCount();
|
|
const active = [...demoSessions.values()].filter(s => s.status === 'running').length;
|
|
const provisioning = [...demoSessions.values()].filter(s => s.status === 'provisioning').length;
|
|
|
|
// DB total
|
|
let dbTotal = totalDemoCount;
|
|
try {
|
|
const row = db.prepare(`SELECT COUNT(*) as c FROM demo_sessions`).get();
|
|
dbTotal = Math.max(totalDemoCount, row?.c || 0);
|
|
} catch {}
|
|
|
|
res.json({
|
|
total_all_time: dbTotal,
|
|
today: todayDemoCount,
|
|
active_now: active,
|
|
provisioning: provisioning,
|
|
pool_providers: providerPool.filter(p => p.active).length,
|
|
capacity_pct: Math.round((active / MAX_CONCURRENT) * 100),
|
|
max_concurrent: MAX_CONCURRENT,
|
|
});
|
|
});
|
|
|
|
// ── POST /api/demo/pool/join — provider contributes compute ──
|
|
app.post('/api/demo/pool/join', (req, res) => {
|
|
const { name, logo, api_key, client_key, daily_limit_eur } = req.body;
|
|
if (!name || !api_key || !client_key) {
|
|
return res.status(400).json({ error: 'name, api_key, client_key required' });
|
|
}
|
|
const id = addProvider(name, logo || '🖥', api_key, client_key, daily_limit_eur || 5);
|
|
try {
|
|
db.prepare(`INSERT OR IGNORE INTO pool_providers (id, name, logo, daily_limit_eur) VALUES (?, ?, ?, ?)`)
|
|
.run(id, name, logo || '🖥', daily_limit_eur || 5);
|
|
} catch {}
|
|
res.json({
|
|
ok: true,
|
|
provider_id: id,
|
|
message: 'Welcome to the Inference-X provider pool! Your compute will power free demos.',
|
|
badge_url: `https://inference-x.com/badge/provider/${id}`,
|
|
});
|
|
});
|
|
|
|
// ── GET /api/demo/pool/providers — public list ───────────────
|
|
app.get('/api/demo/pool/providers', (req, res) => {
|
|
res.json({
|
|
providers: providerPool.filter(p => p.active).map(p => ({
|
|
id: p.id,
|
|
name: p.name,
|
|
logo: p.logo,
|
|
daily_limit_eur: p.daily_limit_eur,
|
|
used_eur: parseFloat(p.used_eur.toFixed(3)),
|
|
utilization_pct: Math.round((p.used_eur / p.daily_limit_eur) * 100),
|
|
})),
|
|
call_to_action: {
|
|
title: 'Power free demos. Earn community credits.',
|
|
description: 'Contribute your OneCloud, Hetzner or OVH API keys. Your idle compute powers AI demos for people who need it.',
|
|
join_url: 'https://build.inference-x.com/#provider-join',
|
|
email: 'Elmadani.SALKA@proton.me',
|
|
},
|
|
});
|
|
});
|
|
}
|
|
|
|
module.exports = { registerDemoRoutes, demoSessions, totalDemoCount: () => totalDemoCount };
|