11 Commits

Author SHA1 Message Date
d84674b0c6 Merge branch 'dev' into feat/jobs-system
All checks were successful
CI / test (pull_request) Successful in 13s
CI / build-dev (pull_request) Has been skipped
2026-03-28 20:20:03 -04:00
7999f46ca2 fix: millisecond precision timestamps and correct history ordering
All checks were successful
CI / test (pull_request) Successful in 21s
CI / build-dev (pull_request) Has been skipped
datetime('now') only stores to the second, making same-second events
indistinguishable. Switched all instance_history and job_runs writes
to strftime('%Y-%m-%dT%H:%M:%f', 'now') for millisecond precision.

Reverted getInstanceHistory to ORDER BY changed_at DESC, id DESC so
newest events appear at the top and instance creation (lowest id,
earliest timestamp) is always at the bottom.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-03-28 20:19:42 -04:00
307c5cf9e8 Merge pull request 'fix: initialize jobs nav dot on every page load' (#57) from feat/jobs-system into dev
All checks were successful
CI / test (push) Successful in 14s
CI / build-dev (push) Successful in 28s
Reviewed-on: #57
2026-03-28 20:16:02 -04:00
34af8e5a8f Merge branch 'dev' into feat/jobs-system
All checks were successful
CI / test (pull_request) Successful in 14s
CI / build-dev (pull_request) Has been skipped
2026-03-28 20:15:37 -04:00
76d2bffb4f fix: initialize jobs nav dot on every page load
All checks were successful
CI / test (pull_request) Successful in 14s
CI / build-dev (pull_request) Has been skipped
Previously the dot only updated when visiting the Jobs page.
Now a jobs fetch runs at bootstrap so the dot reflects status
immediately on any page, including after a hard refresh.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-03-28 20:14:53 -04:00
64de0e432c Merge pull request 'fix: queue on-create jobs sequentially and fix history ordering' (#56) from feat/jobs-system into dev
All checks were successful
CI / test (push) Successful in 13s
CI / build-dev (push) Successful in 28s
Reviewed-on: #56
2026-03-28 20:12:31 -04:00
a5b409a348 Merge branch 'dev' into feat/jobs-system
All checks were successful
CI / test (pull_request) Successful in 14s
CI / build-dev (pull_request) Has been skipped
2026-03-28 20:09:59 -04:00
8f35724bde fix: queue on-create jobs sequentially and fix history ordering
All checks were successful
CI / test (pull_request) Successful in 13s
CI / build-dev (pull_request) Has been skipped
runJobsOnCreate now awaits each job before starting the next,
ensuring they don't stomp each other's DB writes in parallel.

getInstanceHistory changed to ORDER BY changed_at ASC, id ASC so
the creation event (lowest id) is always first regardless of
same-second timestamps.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-03-28 20:09:32 -04:00
cec82a3347 Merge pull request 'feat: run jobs on instance creation when run_on_create is enabled' (#54) from feat/jobs-system into dev
All checks were successful
CI / test (push) Successful in 17s
CI / build-dev (push) Successful in 34s
Reviewed-on: #54
2026-03-28 20:01:53 -04:00
883e59789b Merge branch 'dev' into feat/jobs-system
All checks were successful
CI / test (pull_request) Successful in 17s
CI / build-dev (pull_request) Has been skipped
2026-03-28 20:01:20 -04:00
817fdaef13 feat: run jobs on instance creation when run_on_create is enabled
All checks were successful
CI / test (pull_request) Successful in 18s
CI / build-dev (pull_request) Has been skipped
Jobs with run_on_create=true in their config fire automatically
after a new instance is created. Runs fire-and-forget so they don't
delay the 201 response. Option exposed as a checkbox in each job's
detail panel.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-03-28 20:00:45 -04:00
6 changed files with 55 additions and 11 deletions

View File

@@ -53,4 +53,6 @@ if (VERSION) {
document.getElementById('nav-version').textContent = label; document.getElementById('nav-version').textContent = label;
} }
fetch('/api/jobs').then(r => r.json()).then(_updateJobsNavDot).catch(() => {});
handleRoute(); handleRoute();

View File

@@ -463,6 +463,13 @@ async function loadJobDetail(jobId) {
<label class="form-label" for="job-schedule">Poll interval (minutes)</label> <label class="form-label" for="job-schedule">Poll interval (minutes)</label>
<input class="form-input" id="job-schedule" type="number" min="1" value="${job.schedule}" style="max-width:100px"> <input class="form-input" id="job-schedule" type="number" min="1" value="${job.schedule}" style="max-width:100px">
</div> </div>
<div class="form-group">
<label class="form-label" style="display:flex;align-items:center;gap:8px;cursor:pointer">
<input type="checkbox" id="job-run-on-create" ${cfg.run_on_create ? 'checked' : ''}
style="accent-color:var(--accent);width:13px;height:13px">
Run on instance creation
</label>
</div>
${_renderJobConfigFields(job.key, cfg)} ${_renderJobConfigFields(job.key, cfg)}
<div class="job-actions"> <div class="job-actions">
<button class="btn btn-secondary" onclick="saveJobDetail(${job.id})">Save</button> <button class="btn btn-secondary" onclick="saveJobDetail(${job.id})">Save</button>
@@ -525,6 +532,8 @@ async function saveJobDetail(jobId) {
if (apiKey) cfg.api_key = apiKey.value; if (apiKey) cfg.api_key = apiKey.value;
if (apiUrl) cfg.api_url = apiUrl.value.trim(); if (apiUrl) cfg.api_url = apiUrl.value.trim();
if (apiToken) cfg.api_token = apiToken.value; if (apiToken) cfg.api_token = apiToken.value;
const runOnCreate = document.getElementById('job-run-on-create');
if (runOnCreate) cfg.run_on_create = runOnCreate.checked;
const res = await fetch(`/api/jobs/${jobId}`, { const res = await fetch(`/api/jobs/${jobId}`, {
method: 'PUT', method: 'PUT',
headers: { 'Content-Type': 'application/json' }, headers: { 'Content-Type': 'application/json' },

View File

@@ -173,7 +173,8 @@ export function createInstance(data) {
@tailscale, @andromeda, @tailscale_ip, @hardware_acceleration) @tailscale, @andromeda, @tailscale_ip, @hardware_acceleration)
`).run(data); `).run(data);
db.prepare( db.prepare(
`INSERT INTO instance_history (vmid, field, old_value, new_value) VALUES (?, 'created', NULL, NULL)` `INSERT INTO instance_history (vmid, field, old_value, new_value, changed_at)
VALUES (?, 'created', NULL, NULL, strftime('%Y-%m-%dT%H:%M:%f', 'now'))`
).run(data.vmid); ).run(data.vmid);
} }
@@ -184,12 +185,13 @@ export function updateInstance(vmid, data) {
name=@name, state=@state, stack=@stack, vmid=@newVmid, name=@name, state=@state, stack=@stack, vmid=@newVmid,
atlas=@atlas, argus=@argus, semaphore=@semaphore, patchmon=@patchmon, atlas=@atlas, argus=@argus, semaphore=@semaphore, patchmon=@patchmon,
tailscale=@tailscale, andromeda=@andromeda, tailscale_ip=@tailscale_ip, tailscale=@tailscale, andromeda=@andromeda, tailscale_ip=@tailscale_ip,
hardware_acceleration=@hardware_acceleration, updated_at=datetime('now') hardware_acceleration=@hardware_acceleration, updated_at=strftime('%Y-%m-%dT%H:%M:%f', 'now')
WHERE vmid=@vmid WHERE vmid=@vmid
`).run({ ...data, newVmid: data.vmid, vmid }); `).run({ ...data, newVmid: data.vmid, vmid });
const newVmid = data.vmid; const newVmid = data.vmid;
const insertEvt = db.prepare( const insertEvt = db.prepare(
`INSERT INTO instance_history (vmid, field, old_value, new_value) VALUES (?, ?, ?, ?)` `INSERT INTO instance_history (vmid, field, old_value, new_value, changed_at)
VALUES (?, ?, ?, ?, strftime('%Y-%m-%dT%H:%M:%f', 'now'))`
); );
for (const field of HISTORY_FIELDS) { for (const field of HISTORY_FIELDS) {
const oldVal = String(old[field] ?? ''); const oldVal = String(old[field] ?? '');
@@ -227,7 +229,7 @@ export function importInstances(rows, historyRows = []) {
export function getInstanceHistory(vmid) { export function getInstanceHistory(vmid) {
return db.prepare( return db.prepare(
'SELECT * FROM instance_history WHERE vmid = ? ORDER BY changed_at DESC' 'SELECT * FROM instance_history WHERE vmid = ? ORDER BY changed_at DESC, id DESC'
).all(vmid); ).all(vmid);
} }
@@ -309,12 +311,14 @@ export function updateJob(id, { enabled, schedule, config }) {
} }
export function createJobRun(jobId) { export function createJobRun(jobId) {
return Number(db.prepare('INSERT INTO job_runs (job_id) VALUES (?)').run(jobId).lastInsertRowid); return Number(db.prepare(
`INSERT INTO job_runs (job_id, started_at) VALUES (?, strftime('%Y-%m-%dT%H:%M:%f', 'now'))`
).run(jobId).lastInsertRowid);
} }
export function completeJobRun(runId, status, result) { export function completeJobRun(runId, status, result) {
db.prepare(` db.prepare(`
UPDATE job_runs SET ended_at=datetime('now'), status=@status, result=@result WHERE id=@id UPDATE job_runs SET ended_at=strftime('%Y-%m-%dT%H:%M:%f', 'now'), status=@status, result=@result WHERE id=@id
`).run({ id: runId, status, result }); `).run({ id: runId, status, result });
} }

View File

@@ -129,6 +129,15 @@ export async function runJob(jobId) {
const _intervals = new Map(); const _intervals = new Map();
export async function runJobsOnCreate() {
for (const job of getJobs()) {
const cfg = JSON.parse(job.config || '{}');
if (cfg.run_on_create) {
try { await runJob(job.id); } catch (e) { console.error(`runJobsOnCreate job ${job.id}:`, e); }
}
}
}
export function restartJobs() { export function restartJobs() {
for (const iv of _intervals.values()) clearInterval(iv); for (const iv of _intervals.values()) clearInterval(iv);
_intervals.clear(); _intervals.clear();

View File

@@ -5,7 +5,7 @@ import {
getConfig, setConfig, getJobs, getJob, updateJob, getJobRuns, getConfig, setConfig, getJobs, getJob, updateJob, getJobRuns,
getAllJobs, getAllJobRuns, importJobs, getAllJobs, getAllJobRuns, importJobs,
} from './db.js'; } from './db.js';
import { runJob, restartJobs } from './jobs.js'; import { runJob, restartJobs, runJobsOnCreate } from './jobs.js';
export const router = Router(); export const router = Router();
@@ -102,6 +102,7 @@ router.post('/instances', (req, res) => {
createInstance(data); createInstance(data);
const created = getInstance(data.vmid); const created = getInstance(data.vmid);
res.status(201).json(created); res.status(201).json(created);
runJobsOnCreate().catch(() => {});
} catch (e) { } catch (e) {
handleDbError('POST /api/instances', e, res); handleDbError('POST /api/instances', e, res);
} }

View File

@@ -623,6 +623,25 @@ describe('POST /api/jobs/:id/run', () => {
expect(res.status).toBe(500) expect(res.status).toBe(500)
}) })
it('run_on_create: triggers matching jobs when an instance is created', async () => {
createJob({ ...testJob, config: JSON.stringify({ api_key: 'k', tailnet: 't', run_on_create: true }) })
const id = (await request(app).get('/api/jobs')).body[0].id
vi.stubGlobal('fetch', vi.fn().mockResolvedValue({ ok: true, json: async () => ({ devices: [] }) }))
await request(app).post('/api/instances').send(base)
await new Promise(r => setImmediate(r))
const detail = await request(app).get(`/api/jobs/${id}`)
expect(detail.body.runs).toHaveLength(1)
expect(detail.body.runs[0].status).toBe('success')
})
it('run_on_create: does not trigger jobs without the flag', async () => {
createJob(testJob)
const id = (await request(app).get('/api/jobs')).body[0].id
await request(app).post('/api/instances').send(base)
await new Promise(r => setImmediate(r))
expect((await request(app).get(`/api/jobs/${id}`)).body.runs).toHaveLength(0)
})
it('semaphore_sync: parses ansible inventory and updates instances', async () => { it('semaphore_sync: parses ansible inventory and updates instances', async () => {
const semaphoreJob = { const semaphoreJob = {
key: 'semaphore_sync', name: 'Semaphore Sync', description: 'test', key: 'semaphore_sync', name: 'Semaphore Sync', description: 'test',