Josh Wright d84a15ad64
All checks were successful
Build and Deploy / Build & Push (push) Successful in 54s
fix: restore 240px park column width — clamp() unreliable in col elements
The actual overflow fix was removing whiteSpace:nowrap from the td.
With that gone, 240px is sufficient and content wraps naturally when tight.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-04-05 14:48:55 -04:00
2026-04-03 22:06:54 -04:00
2026-04-04 15:26:00 -04:00

Thoosie Calendar

A week-by-week calendar showing operating hours for all Six Flags Entertainment Group theme parks — including the former Cedar Fair parks. Data is scraped from the Six Flags internal API and stored locally in SQLite. Click any park to see its full month calendar and live ride status with current wait times.

Parks

24 theme parks across the US, Canada, and Mexico, grouped by region:

Region Parks
Northeast Great Adventure (NJ), New England (MA), Great Escape (NY), Darien Lake (NY), Dorney Park (PA), Canada's Wonderland (ON)
Southeast Over Georgia, Carowinds (NC), Kings Dominion (VA)
Midwest Great America (IL), St. Louis (MO), Cedar Point (OH), Kings Island (OH), Valleyfair (MN), Worlds of Fun (MO), Michigan's Adventure (MI)
Texas & South Over Texas, Fiesta Texas (TX), Frontier City (OK)
West & International Magic Mountain (CA), Discovery Kingdom (CA), Knott's Berry Farm (CA), California's Great America (CA), Mexico

Tech Stack

  • Next.js 15 — App Router, Server Components, standalone output
  • Tailwind CSS v4@theme {} CSS variables, no config file
  • SQLite via better-sqlite3 — persisted in /app/data/parks.db
  • Playwright — one-time headless browser run to discover each park's internal API ID
  • Six Flags CloudFront APIhttps://d18car1k0ff81h.cloudfront.net/operating-hours/park/{id}?date=YYYYMM
  • Queue-Times.com API — live ride open/closed status and wait times, updated every 5 minutes

Ride Status

The park detail page shows ride open/closed status using a two-tier approach:

  1. Live data (Queue-Times.com) — when a park is operating, ride status and wait times are fetched from the Queue-Times.com API and cached for 5 minutes. All 24 parks are mapped. Displays a Live badge with per-ride wait times.

  2. Schedule fallback (Six Flags API) — the Six Flags operating-hours API drops the current day from its response once a park opens. When Queue-Times data is unavailable, the app falls back to the nearest upcoming date from the Six Flags schedule API as an approximation.

Roller Coaster Filter

When live data is shown, a Coasters only toggle appears if roller coaster data has been populated for that park. Coaster lists are sourced from RCDB and stored in data/park-meta.json. To populate them:

  1. Open data/park-meta.json and set rcdb_id for each park to the numeric RCDB park ID (visible in the URL: https://rcdb.com/4529.htm4529).
  2. Run npm run scrape — coaster lists are fetched from RCDB and stored in the JSON file. They refresh automatically every 30 days on subsequent scrapes.

Local Development

Prerequisites: Node.js 22+, npm

npm install
npx playwright install chromium

Seed the database

Run once to discover each park's internal API ID (opens a headless browser per park):

npm run discover

Scrape operating hours for the full year:

npm run scrape

Force a full re-scrape (ignores the staleness window):

npm run scrape:force

Debug a specific park + date

Inspect raw API data and parsed output for any park and date:

npm run debug -- --park kingsisland --date 2026-06-15

Output is printed to the terminal and saved to debug/{parkId}_{date}.txt.

Run tests

npm test

Run the dev server

npm run dev

Open http://localhost:3000. Navigate weeks with the / buttons, or pass ?week=YYYY-MM-DD directly. Click any park name to open its detail page.


Deployment

The app ships as two separate Docker images that share a named volume for the SQLite database:

Image Tag Purpose
Next.js web server :web Reads DB, serves content. No scraping tools.
Scraper + scheduler :scraper Nightly data refresh. No web server.

Images are built and pushed automatically by CI on every push to main.

First-time setup

1. Pull the images

docker pull gitea.thewrightserver.net/josh/sixflagssupercalendar:web
docker pull gitea.thewrightserver.net/josh/sixflagssupercalendar:scraper

2. Discover park API IDs

This one-time step opens a headless browser for each park to find its internal Six Flags API ID. Run it against the scraper image so Playwright is available:

docker run --rm -v root_park_data:/app/data \
  gitea.thewrightserver.net/josh/sixflagssupercalendar:scraper \
  npm run discover

3. Set RCDB IDs for the coaster filter

Open data/park-meta.json in the Docker volume and set rcdb_id for each park to the numeric ID from the RCDB URL (e.g. https://rcdb.com/4529.htm4529). You can curl it directly from the repo:

curl -o /var/lib/docker/volumes/root_park_data/_data/park-meta.json \
  https://gitea.thewrightserver.net/josh/SixFlagsSuperCalendar/raw/branch/main/data/park-meta.json

4. Run the initial scrape

docker run --rm -v root_park_data:/app/data \
  gitea.thewrightserver.net/josh/sixflagssupercalendar:scraper \
  npm run scrape

5. Start services

docker compose up -d

Both services start. The scraper runs nightly at 3 AM (container timezone, set via TZ).

Updating

docker compose pull && docker compose up -d

Scraper environment variables

Set these in docker-compose.yml under the scraper service to override defaults:

Variable Default Description
TZ UTC Timezone for the nightly 3 AM run (e.g. America/New_York)
PARK_HOURS_STALENESS_HOURS 72 Hours before park schedule data is re-fetched
COASTER_STALENESS_HOURS 720 Hours before RCDB coaster lists are re-fetched (720 = 30 days)

Manual scrape

To trigger a scrape outside the nightly schedule:

docker compose exec scraper npm run scrape

Force re-scrape of all data (ignores staleness):

docker compose exec scraper npm run scrape:force

Data Refresh

The scraper skips any park + month already scraped within the staleness window (PARK_HOURS_STALENESS_HOURS, default 72h). Past dates are never overwritten — once a day occurs, the API stops returning data for it, so the record written when it was a future date is preserved forever. The nightly scraper handles refresh automatically.

Roller coaster lists (from RCDB) are refreshed per COASTER_STALENESS_HOURS (default 720h = 30 days) for parks with a configured rcdb_id.

Description
No description provided
Readme 731 KiB
Languages
TypeScript 93.7%
CSS 3.2%
Dockerfile 1.8%
Shell 1%
JavaScript 0.3%