feat: split web and scraper into separate Docker images
All checks were successful
Build and Deploy / Build & Push (push) Successful in 3m4s
All checks were successful
Build and Deploy / Build & Push (push) Successful in 3m4s
- Dockerfile: replace single runner stage with web + scraper named targets - web: Next.js standalone only — no playwright, tsx, or scripts - scraper: scripts/lib/node_modules/playwright only — no Next.js output - docker-compose.yml: each service pulls its dedicated image tag - .gitea/workflows/deploy.yml: build both targets on push to main - lib/db.ts: STALE_AFTER_MS reads PARK_HOURS_STALENESS_HOURS env var (default 72h) - lib/park-meta.ts: COASTER_STALE_MS reads COASTER_STALENESS_HOURS env var (default 720h) Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
This commit is contained in:
@@ -4,26 +4,14 @@ on:
|
|||||||
push:
|
push:
|
||||||
branches:
|
branches:
|
||||||
- main
|
- main
|
||||||
tags:
|
|
||||||
- 'v*'
|
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
build-push:
|
build-push:
|
||||||
name: Build & Push
|
name: Build & Push
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
if: github.event_name == 'push'
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Docker metadata
|
|
||||||
id: meta
|
|
||||||
uses: docker/metadata-action@v5
|
|
||||||
with:
|
|
||||||
images: ${{ vars.REGISTRY }}/${{ gitea.repository_owner }}/sixflagssupercalendar
|
|
||||||
tags: |
|
|
||||||
type=semver,pattern={{version}}
|
|
||||||
type=raw,value=latest,enable={{is_default_branch}}
|
|
||||||
|
|
||||||
- name: Log in to Gitea registry
|
- name: Log in to Gitea registry
|
||||||
uses: docker/login-action@v3
|
uses: docker/login-action@v3
|
||||||
with:
|
with:
|
||||||
@@ -31,10 +19,18 @@ jobs:
|
|||||||
username: ${{ gitea.actor }}
|
username: ${{ gitea.actor }}
|
||||||
password: ${{ secrets.REGISTRY_TOKEN }}
|
password: ${{ secrets.REGISTRY_TOKEN }}
|
||||||
|
|
||||||
- name: Build and push
|
- name: Build and push web image
|
||||||
uses: docker/build-push-action@v6
|
uses: docker/build-push-action@v6
|
||||||
with:
|
with:
|
||||||
context: .
|
context: .
|
||||||
|
target: web
|
||||||
push: true
|
push: true
|
||||||
tags: ${{ steps.meta.outputs.tags }}
|
tags: ${{ vars.REGISTRY }}/${{ gitea.repository_owner }}/sixflagssupercalendar:web
|
||||||
labels: ${{ steps.meta.outputs.labels }}
|
|
||||||
|
- name: Build and push scraper image
|
||||||
|
uses: docker/build-push-action@v6
|
||||||
|
with:
|
||||||
|
context: .
|
||||||
|
target: scraper
|
||||||
|
push: true
|
||||||
|
tags: ${{ vars.REGISTRY }}/${{ gitea.repository_owner }}/sixflagssupercalendar:scraper
|
||||||
|
|||||||
63
Dockerfile
63
Dockerfile
@@ -1,4 +1,4 @@
|
|||||||
# Stage 1: Install all dependencies (dev included — scripts need tsx + playwright)
|
# Stage 1: Install all dependencies (dev included — scraper needs tsx + playwright)
|
||||||
FROM node:22-bookworm-slim AS deps
|
FROM node:22-bookworm-slim AS deps
|
||||||
RUN apt-get update && apt-get install -y --no-install-recommends python3 make g++ && \
|
RUN apt-get update && apt-get install -y --no-install-recommends python3 make g++ && \
|
||||||
rm -rf /var/lib/apt/lists/*
|
rm -rf /var/lib/apt/lists/*
|
||||||
@@ -11,47 +11,60 @@ FROM deps AS builder
|
|||||||
COPY . .
|
COPY . .
|
||||||
RUN npm run build
|
RUN npm run build
|
||||||
|
|
||||||
# Stage 3: Production runner
|
# ── web ──────────────────────────────────────────────────────────────────────
|
||||||
FROM node:22-bookworm-slim AS runner
|
# Minimal Next.js runner. No playwright, no tsx, no scripts.
|
||||||
|
# next build --output standalone bundles its own node_modules (incl. better-sqlite3).
|
||||||
|
FROM node:22-bookworm-slim AS web
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
|
|
||||||
ENV NODE_ENV=production
|
ENV NODE_ENV=production
|
||||||
ENV NEXT_TELEMETRY_DISABLED=1
|
ENV NEXT_TELEMETRY_DISABLED=1
|
||||||
# Store Playwright browser in a predictable path inside the image
|
|
||||||
ENV PLAYWRIGHT_BROWSERS_PATH=/app/.playwright
|
|
||||||
|
|
||||||
# Create non-root user before copying files so --chown works
|
|
||||||
RUN addgroup --system --gid 1001 nodejs && \
|
RUN addgroup --system --gid 1001 nodejs && \
|
||||||
adduser --system --uid 1001 nextjs
|
adduser --system --uid 1001 nextjs
|
||||||
|
|
||||||
# Copy Next.js standalone output
|
|
||||||
COPY --from=builder --chown=nextjs:nodejs /app/.next/standalone ./
|
COPY --from=builder --chown=nextjs:nodejs /app/.next/standalone ./
|
||||||
COPY --from=builder --chown=nextjs:nodejs /app/.next/static ./.next/static
|
COPY --from=builder --chown=nextjs:nodejs /app/.next/static ./.next/static
|
||||||
COPY --from=builder --chown=nextjs:nodejs /app/public ./public
|
COPY --from=builder --chown=nextjs:nodejs /app/public ./public
|
||||||
|
|
||||||
# Copy scripts + library source (needed for npm run discover/scrape via tsx)
|
|
||||||
COPY --from=builder --chown=nextjs:nodejs /app/scripts ./scripts
|
|
||||||
COPY --from=builder --chown=nextjs:nodejs /app/lib ./lib
|
|
||||||
COPY --from=builder --chown=nextjs:nodejs /app/package.json ./package.json
|
|
||||||
COPY --from=builder --chown=nextjs:nodejs /app/tsconfig.json ./tsconfig.json
|
|
||||||
|
|
||||||
# Replace standalone's minimal node_modules with full deps
|
|
||||||
# (includes tsx, playwright, and all devDependencies)
|
|
||||||
COPY --from=deps --chown=nextjs:nodejs /app/node_modules ./node_modules
|
|
||||||
|
|
||||||
# Install Playwright Chromium browser + all required system libraries.
|
|
||||||
# Runs as root so apt-get works; browser lands in PLAYWRIGHT_BROWSERS_PATH.
|
|
||||||
RUN npx playwright install --with-deps chromium && \
|
|
||||||
chown -R nextjs:nodejs /app/.playwright
|
|
||||||
|
|
||||||
# SQLite data directory — mount a named volume here for persistence
|
|
||||||
RUN mkdir -p /app/data && chown nextjs:nodejs /app/data
|
RUN mkdir -p /app/data && chown nextjs:nodejs /app/data
|
||||||
VOLUME ["/app/data"]
|
VOLUME ["/app/data"]
|
||||||
|
|
||||||
USER nextjs
|
USER nextjs
|
||||||
|
|
||||||
EXPOSE 3000
|
EXPOSE 3000
|
||||||
ENV PORT=3000
|
ENV PORT=3000
|
||||||
ENV HOSTNAME="0.0.0.0"
|
ENV HOSTNAME="0.0.0.0"
|
||||||
|
|
||||||
CMD ["node", "server.js"]
|
CMD ["node", "server.js"]
|
||||||
|
|
||||||
|
# ── scraper ───────────────────────────────────────────────────────────────────
|
||||||
|
# Scraper-only image. No Next.js output. Runs on a nightly schedule via
|
||||||
|
# scripts/scrape-schedule.sh. Staleness windows are configurable via env vars:
|
||||||
|
# PARK_HOURS_STALENESS_HOURS (default: 72)
|
||||||
|
# COASTER_STALENESS_HOURS (default: 720 = 30 days)
|
||||||
|
FROM node:22-bookworm-slim AS scraper
|
||||||
|
WORKDIR /app
|
||||||
|
|
||||||
|
ENV NODE_ENV=production
|
||||||
|
ENV PLAYWRIGHT_BROWSERS_PATH=/app/.playwright
|
||||||
|
|
||||||
|
RUN addgroup --system --gid 1001 nodejs && \
|
||||||
|
adduser --system --uid 1001 nextjs
|
||||||
|
|
||||||
|
COPY --from=builder --chown=nextjs:nodejs /app/scripts ./scripts
|
||||||
|
COPY --from=builder --chown=nextjs:nodejs /app/lib ./lib
|
||||||
|
COPY --from=builder --chown=nextjs:nodejs /app/tests ./tests
|
||||||
|
COPY --from=builder --chown=nextjs:nodejs /app/package.json ./package.json
|
||||||
|
COPY --from=builder --chown=nextjs:nodejs /app/tsconfig.json ./tsconfig.json
|
||||||
|
|
||||||
|
# Full node_modules — includes tsx, playwright, better-sqlite3, all devDeps
|
||||||
|
COPY --from=deps --chown=nextjs:nodejs /app/node_modules ./node_modules
|
||||||
|
|
||||||
|
# Install Playwright Chromium + system libraries (runs as root, then fixes ownership)
|
||||||
|
RUN npx playwright install --with-deps chromium && \
|
||||||
|
chown -R nextjs:nodejs /app/.playwright
|
||||||
|
|
||||||
|
RUN mkdir -p /app/data && chown nextjs:nodejs /app/data
|
||||||
|
VOLUME ["/app/data"]
|
||||||
|
|
||||||
|
USER nextjs
|
||||||
|
CMD ["sh", "/app/scripts/scrape-schedule.sh"]
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
services:
|
services:
|
||||||
web:
|
web:
|
||||||
image: gitea.thewrightserver.net/josh/sixflagssupercalendar:latest
|
image: gitea.thewrightserver.net/josh/sixflagssupercalendar:web
|
||||||
ports:
|
ports:
|
||||||
- "3000:3000"
|
- "3000:3000"
|
||||||
volumes:
|
volumes:
|
||||||
@@ -10,13 +10,14 @@ services:
|
|||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|
||||||
scraper:
|
scraper:
|
||||||
image: gitea.thewrightserver.net/josh/sixflagssupercalendar:latest
|
image: gitea.thewrightserver.net/josh/sixflagssupercalendar:scraper
|
||||||
volumes:
|
volumes:
|
||||||
- park_data:/app/data
|
- park_data:/app/data
|
||||||
environment:
|
environment:
|
||||||
- NODE_ENV=production
|
- NODE_ENV=production
|
||||||
- TZ=America/New_York # set your local timezone so "3am" is 3am your time
|
- TZ=America/New_York
|
||||||
command: sh /app/scripts/scrape-schedule.sh
|
- PARK_HOURS_STALENESS_HOURS=72
|
||||||
|
- COASTER_STALENESS_HOURS=720
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|
||||||
volumes:
|
volumes:
|
||||||
|
|||||||
@@ -167,7 +167,8 @@ export function getMonthCalendar(
|
|||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
const STALE_AFTER_MS = 72 * 60 * 60 * 1000; // 72 hours
|
const STALE_AFTER_MS =
|
||||||
|
parseInt(process.env.PARK_HOURS_STALENESS_HOURS ?? "72", 10) * 60 * 60 * 1000;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Returns true when the scraper should skip this park+month.
|
* Returns true when the scraper should skip this park+month.
|
||||||
|
|||||||
@@ -44,7 +44,8 @@ export function defaultParkMeta(): ParkMeta {
|
|||||||
return { rcdb_id: null, coasters: [], coasters_scraped_at: null };
|
return { rcdb_id: null, coasters: [], coasters_scraped_at: null };
|
||||||
}
|
}
|
||||||
|
|
||||||
const COASTER_STALE_MS = 30 * 24 * 60 * 60 * 1000; // 30 days
|
const COASTER_STALE_MS =
|
||||||
|
parseInt(process.env.COASTER_STALENESS_HOURS ?? "720", 10) * 60 * 60 * 1000;
|
||||||
|
|
||||||
/** Returns true when the coaster list needs to be re-scraped from RCDB. */
|
/** Returns true when the coaster list needs to be re-scraped from RCDB. */
|
||||||
export function areCoastersStale(entry: ParkMeta): boolean {
|
export function areCoastersStale(entry: ParkMeta): boolean {
|
||||||
|
|||||||
Reference in New Issue
Block a user