Let the park name side flex-shrink (minWidth:0, flex:1) so the ride count always fits in the row without overflowing its column. Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
Thoosie Calendar
A week-by-week calendar showing operating hours for all Six Flags Entertainment Group theme parks — including the former Cedar Fair parks. Data is scraped from the Six Flags internal API and stored locally in SQLite. Click any park to see its full month calendar and live ride status with current wait times.
Parks
24 theme parks across the US, Canada, and Mexico, grouped by region:
| Region | Parks |
|---|---|
| Northeast | Great Adventure (NJ), New England (MA), Great Escape (NY), Darien Lake (NY), Dorney Park (PA), Canada's Wonderland (ON) |
| Southeast | Over Georgia, Carowinds (NC), Kings Dominion (VA) |
| Midwest | Great America (IL), St. Louis (MO), Cedar Point (OH), Kings Island (OH), Valleyfair (MN), Worlds of Fun (MO), Michigan's Adventure (MI) |
| Texas & South | Over Texas, Fiesta Texas (TX), Frontier City (OK) |
| West & International | Magic Mountain (CA), Discovery Kingdom (CA), Knott's Berry Farm (CA), California's Great America (CA), Mexico |
Tech Stack
- Next.js 15 — App Router, Server Components, standalone output
- Tailwind CSS v4 —
@theme {}CSS variables, no config file - SQLite via
better-sqlite3— persisted in/app/data/parks.db - Playwright — one-time headless browser run to discover each park's internal API ID
- Six Flags CloudFront API —
https://d18car1k0ff81h.cloudfront.net/operating-hours/park/{id}?date=YYYYMM - Queue-Times.com API — live ride open/closed status and wait times, updated every 5 minutes
Ride Status
The park detail page shows ride open/closed status using a two-tier approach:
-
Live data (Queue-Times.com) — when a park is operating, ride status and wait times are fetched from the Queue-Times.com API and cached for 5 minutes. All 24 parks are mapped. Displays a Live badge with per-ride wait times.
-
Schedule fallback (Six Flags API) — the Six Flags operating-hours API drops the current day from its response once a park opens. When Queue-Times data is unavailable, the app falls back to the nearest upcoming date from the Six Flags schedule API as an approximation.
Roller Coaster Filter
When live data is shown, a Coasters only toggle appears if roller coaster data has been populated for that park. Coaster lists are sourced from RCDB and stored in data/park-meta.json. To populate them:
- Open
data/park-meta.jsonand setrcdb_idfor each park to the numeric RCDB park ID (visible in the URL:https://rcdb.com/4529.htm→4529). - Run
npm run scrape— coaster lists are fetched from RCDB and stored in the JSON file. They refresh automatically every 30 days on subsequent scrapes.
Local Development
Prerequisites: Node.js 22+, npm
npm install
npx playwright install chromium
Seed the database
Run once to discover each park's internal API ID (opens a headless browser per park):
npm run discover
Scrape operating hours for the full year:
npm run scrape
Force a full re-scrape (ignores the staleness window):
npm run scrape:force
Debug a specific park + date
Inspect raw API data and parsed output for any park and date:
npm run debug -- --park kingsisland --date 2026-06-15
Output is printed to the terminal and saved to debug/{parkId}_{date}.txt.
Run tests
npm test
Run the dev server
npm run dev
Open http://localhost:3000. Navigate weeks with the ← / → buttons, or pass ?week=YYYY-MM-DD directly. Click any park name to open its detail page.
Deployment
The app ships as two separate Docker images that share a named volume for the SQLite database:
| Image | Tag | Purpose |
|---|---|---|
| Next.js web server | :web |
Reads DB, serves content. No scraping tools. |
| Scraper + scheduler | :scraper |
Nightly data refresh. No web server. |
Images are built and pushed automatically by CI on every push to main.
First-time setup
1. Pull the images
docker pull gitea.thewrightserver.net/josh/sixflagssupercalendar:web
docker pull gitea.thewrightserver.net/josh/sixflagssupercalendar:scraper
2. Discover park API IDs
This one-time step opens a headless browser for each park to find its internal Six Flags API ID. Run it against the scraper image so Playwright is available:
docker run --rm -v root_park_data:/app/data \
gitea.thewrightserver.net/josh/sixflagssupercalendar:scraper \
npm run discover
3. Set RCDB IDs for the coaster filter
Open data/park-meta.json in the Docker volume and set rcdb_id for each park to the numeric ID from the RCDB URL (e.g. https://rcdb.com/4529.htm → 4529). You can curl it directly from the repo:
curl -o /var/lib/docker/volumes/root_park_data/_data/park-meta.json \
https://gitea.thewrightserver.net/josh/SixFlagsSuperCalendar/raw/branch/main/data/park-meta.json
4. Run the initial scrape
docker run --rm -v root_park_data:/app/data \
gitea.thewrightserver.net/josh/sixflagssupercalendar:scraper \
npm run scrape
5. Start services
docker compose up -d
Both services start. The scraper runs nightly at 3 AM (container timezone, set via TZ).
Updating
docker compose pull && docker compose up -d
Scraper environment variables
Set these in docker-compose.yml under the scraper service to override defaults:
| Variable | Default | Description |
|---|---|---|
TZ |
UTC |
Timezone for the nightly 3 AM run (e.g. America/New_York) |
PARK_HOURS_STALENESS_HOURS |
72 |
Hours before park schedule data is re-fetched |
COASTER_STALENESS_HOURS |
720 |
Hours before RCDB coaster lists are re-fetched (720 = 30 days) |
Manual scrape
To trigger a scrape outside the nightly schedule:
docker compose exec scraper npm run scrape
Force re-scrape of all data (ignores staleness):
docker compose exec scraper npm run scrape:force
Data Refresh
The scraper skips any park + month already scraped within the staleness window (PARK_HOURS_STALENESS_HOURS, default 72h). Past dates are never overwritten — once a day occurs, the API stops returning data for it, so the record written when it was a future date is preserved forever. The nightly scraper handles refresh automatically.
Roller coaster lists (from RCDB) are refreshed per COASTER_STALENESS_HOURS (default 720h = 30 days) for parks with a configured rcdb_id.