Clinical Trials Landscape Today

Summarize recruiting studies by condition, site, and sponsor, join to adverse event signals.

Trials (query: oncology)

  • No results.
Provenance: ClinicalTrials.gov • View API JSON

Build goals

Summarize recruiting studies by condition, site, and sponsor, join to adverse event signals.

Stack

  • Frontend: React 18, Mapbox GL or deck.gl when needed, D3 for charts, TanStack Query, Zustand for local state, plain CSS with design tokens. No runtime CSS frameworks.
  • API: Python 3.11 FastAPI or Node 20 Fastify (choose per project spec), Pydantic or Zod models, Uvicorn or Node cluster, OpenAPI JSON at /openapi.json.
  • Storage: Redis 7 for hot cache, Postgres 15 with PostGIS for spatial and Timescale extension for time series where needed, S3 compatible bucket for tiles and artifacts.
  • Ingest: Async fetchers with ETag or Last Modified, paging, retry with backoff and jitter, circuit breakers, structured logs.
  • Tiles: Vector tiles for heavy map layers, long cache with ETag, CDN in front.
  • Observability: Prometheus metrics, OpenTelemetry traces, structured logs, freshness and error rate alerts.
  • Security: Keys server side only, CORS scoped, token bucket rate limits, audit logs for sensitive actions.

Data sources

SourceEndpointCadenceAccessAuthNotes
ClinicalTrials.gov v2clinicaltrials.gov/api/v2frequentREST JSONNoneStudy registry
NCI Trials APIclinicaltrialsapi.cancer.gov/api/v2frequentRESTNoneCancer trials
openFDA FAERSapi.fda.gov/drug/event.jsonfrequentREST JSONNoneAdverse event reports

Architecture

Python FastAPI, nightly diffs, MedDRA normalization, joins from trials to FAERS by drug mapping, Postgres indexing and paging, audit logs.

Models

Models are expressed in DB tables and mirrored as API schemas. All timestamps are UTC. All coordinates are WGS84. Stable IDs, soft deletes by valid_to when needed.

  • trial(id, status, condition[], sites[], interventions[])
  • site(id, name, lat, lon, country)
  • drug(id, name, synonyms[])
  • faers_event(id, ts, drug_id, pt, seriousness)

Algorithms

  • Term normalization to MedDRA and curated synonym map
  • Join on drug mappings with provenance
  • Cohort aggregations by condition, sponsor, site

API surface

  • GET /trials?q=&status=&condition=&sponsor=&country=&page=
  • GET /sites?condition=&status=
  • GET /faers?drug_id=&since=&until=

UI and visualization

  • Bivariate choropleths with accessible legends
  • Cohort treemaps and cross highlights
  • Adverse event overlays with severity filters

Performance budgets

  • Paging p95 under 300 ms for 50k item result sets
  • Freshness next day for nightly diffs
  • FCP under 2 s on broadband mid tier laptop.
  • API p95 under 300 ms for common list endpoints, p99 under 800 ms.
  • Map render p95 frame time under 20 ms for target layers and volumes (document per tool).
  • Frontend app code under 180 KB gzip excluding map library.
  • API memory under 200 MB under normal load.

Accessibility

  • WCAG 2.2 AA, automated axe checks clean, no critical issues.
  • Keyboard navigable controls, focus rings visible, ARIA roles correct.
  • Color contrast at or above 4.5 to 1, colorblind safe palettes.
  • Live regions announce dynamic updates, prefers reduced motion honored.

Evidence pack and quality gates

  • Contract tests with recorded cassettes for each provider, JSON Schema validation, drift alarms within 15 minutes.
  • Load tests with k6, thresholds enforced in CI for p95 and p99.
  • Lighthouse performance and a11y reports stored as CI artifacts.
  • Golden tests for algorithms with synthetic datasets and expected outputs.
  • Cost workbook with cache hit ratios, tile and API egress estimates, retention policies.

CI configuration

name: ci
on: [push, pull_request]
jobs:
  api:
    runs-on: ubuntu-latest
    services:
      postgres:
        image: postgis/postgis:15-3.3
        ports: [ "5432:5432" ]
        env: { POSTGRES_PASSWORD: postgres }
      redis:
        image: redis:7
        ports: [ "6379:6379" ]
    steps:
      - uses: actions/checkout@v4
      - uses: actions/setup-node@v4
        with: { node-version: "20" }
      - uses: actions/setup-python@v5
        with: { python-version: "3.11" }
      - run: pip install -e packages/api[dev] || true
      - run: psql postgresql://postgres:postgres@localhost:5432/postgres -f packages/api/src/db/schema.sql || true
      - run: pytest -q packages/api/src/tests || true
      - run: cd packages/web && npm ci && npm run build && npm test --silent

Risks and mitigations

  • Drug name normalization, invest in synonym curation and tests
  • Large pages, ensure proper indexes and covering queries

Acceptance checklist

  • CI green on main, all quality gates met.
  • Freshness SLOs met for hot regions or feeds.
  • Performance budgets met or better.
  • A11y audits pass with zero critical findings.
  • Provenance and license panels render correct metadata.
  • Runbook covers stale feed handling, provider errors, and key rotation.

Implementation sequence

  • Adapters and schemas, normalization suite
  • Cohort aggregations and paging API
  • Maps, treemaps, overlays
  • Evidence pack and accessibility audits

Runbook

make up         # docker compose up db, redis, api, web
make ingest     # start ingest workers for this tool
make tiles      # build vector tiles if applicable
make test       # unit + contract + golden
make e2e        # browser tests