Files
egg 71c8102de6 feat: dataset cache for hold/resource history + slow connection migration
Two changes combined:

1. historical-query-slow-connection: Migrate all historical query pages
   to read_sql_df_slow with semaphore concurrency control (max 3),
   raise DB slow timeout to 300s, gunicorn timeout to 360s, and
   unify frontend timeouts to 360s for all historical pages.

2. hold-resource-history-dataset-cache: Convert hold-history and
   resource-history from multi-query to single-query + dataset cache
   pattern (L1 ProcessLevelCache + L2 Redis parquet/base64, TTL=900s).
   Replace old GET endpoints with POST /query + GET /view two-phase
   API. Frontend auto-retries on 410 cache_expired.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-02-25 13:15:02 +08:00

3.5 KiB

MODIFIED Requirements

Requirement: Hold History API SHALL provide daily trend data with Redis caching

The Hold History API SHALL return trend, reason-pareto, duration, and list data from a single cached dataset via a two-phase query pattern (POST /query + GET /view). The old independent GET endpoints for trend, reason-pareto, duration, and list SHALL be replaced.

Scenario: Primary query endpoint

  • WHEN POST /api/hold-history/query is called with { start_date, end_date, hold_type }
  • THEN the service SHALL execute a single Oracle query (or read from cache) via hold_dataset_cache.execute_primary_query()
  • THEN the response SHALL return { success: true, data: { query_id, trend, reason_pareto, duration, list, summary } }
  • THEN list SHALL contain page 1 with default per_page of 50

Scenario: Supplementary view endpoint

  • WHEN GET /api/hold-history/view?query_id=...&hold_type=...&reason=...&page=...&per_page=... is called
  • THEN the service SHALL read the cached DataFrame and derive filtered views via hold_dataset_cache.apply_view()
  • THEN no Oracle query SHALL be executed
  • THEN the response SHALL return { success: true, data: { trend, reason_pareto, duration, list } }

Scenario: Cache expired on view request

  • WHEN GET /view is called with an expired query_id
  • THEN the response SHALL return { success: false, error: "cache_expired" } with HTTP 410

Scenario: Trend uses shift boundary at 07:30

  • WHEN daily aggregation is calculated
  • THEN transactions with time >= 07:30 SHALL be attributed to the next calendar day
  • THEN transactions with time < 07:30 SHALL be attributed to the current calendar day

Scenario: Trend hold type classification

  • WHEN trend data is aggregated by hold type
  • THEN quality classification SHALL use the same NON_QUALITY_HOLD_REASONS set as existing hold endpoints

Requirement: Hold History API SHALL provide reason Pareto data

The reason Pareto data SHALL be derived from the cached dataset, not from a separate Oracle query.

Scenario: Reason Pareto from cache

  • WHEN reason Pareto is requested via GET /view with hold_type filter
  • THEN the cached DataFrame SHALL be filtered by hold_type and grouped by HOLDREASONNAME
  • THEN each item SHALL contain { reason, count, qty, pct, cumPct }
  • THEN items SHALL be sorted by count descending

Requirement: Hold History API SHALL provide hold duration distribution

The duration distribution SHALL be derived from the cached dataset.

Scenario: Duration from cache

  • WHEN duration is requested via GET /view
  • THEN the cached DataFrame SHALL be filtered to released holds only
  • THEN 4 buckets SHALL be computed: <4h, 4-24h, 1-3d, >3d

Requirement: Hold History API SHALL provide paginated detail list

The detail list SHALL be paginated from the cached dataset.

Scenario: List pagination from cache

  • WHEN list is requested via GET /view with page and per_page params
  • THEN the cached DataFrame SHALL be filtered and paginated in-memory
  • THEN response SHALL include items and pagination metadata

Requirement: Hold History API SHALL keep department endpoint as separate query

The department endpoint SHALL remain as a separate Oracle query due to its unique person-level aggregation.

Scenario: Department endpoint unchanged

  • WHEN GET /api/hold-history/department is called
  • THEN it SHALL continue to execute its own Oracle query
  • THEN it SHALL NOT use the dataset cache