← Back to dashboard

What’s Next

Context windows, AI infrastructure, and construction technologies to watch.

Compared context window sizes from 256,000 to 10,000,000 tokens on a log-scaled bar.

Context Window Scale

More context can let one model review more of a spec book, submittal package, contract, or project email history at once.

Available today

  • xAI Grok 4

    Current/API

    256,000 tokens

    Fits many whole drawing indexes, RFIs, or discipline packages β€” not the entire multi-year project file tree.

  • Google Gemini Pro

    Current/product

    1,000,000 tokens

    Enough breadth for a large project text corpus in one pass β€” still check citations and drawings.

  • Anthropic Claude (Opus 4.7 / 4.6, Sonnet 4.6)

    Current/API

    1,000,000 tokens

    Holds very large submittal or spec stacks for review β€” reasoning quality is still task-dependent.

  • OpenAI GPT-4.1

    Current/API

    1,047,576 tokens

    Near–1M-token class: strong fit for bundled OCR text, specs, and narrative scope in one request.

  • Google Gemini 1.5 Pro (API / developers)

    Current/API-developer

    2,000,000 tokens

    Supports multiple fat manuals or long correspondence threads together β€” still validate against source files.

  • xAI Grok 4 Fast

    Current/API

    2,000,000 tokens

    Same 2M class as other developer offerings: room for big packages, not automatic completeness.

Emerging / frontier

  • Meta Llama 4 Scout

    Current/open model frontier

    10,000,000 tokens

    Open-weight frontier for how much raw text can ride along β€” ops and compliance paths differ from vendor APIs.

  • Google Gemini 1.5 (research test)

    Research frontier

    10,000,000 tokens

    Shows where labs are pushing memory β€” production workflows usually sit far below this today.

Practical scale

  • 256KLarge spec package or project document set
  • 1MVery large text volume β€” on the order of ~1,500 pages for some products' framing
  • 2MMultiple large project manuals or a long email / document history
  • 10MResearch / frontier-scale memory β€” not a typical enterprise daily workflow yet

Larger context means the model can accept more material at once. It does not guarantee it will use every detail correctly.

Infrastructure scale

Headline vendor and outlook figures β€” what they imply for power, builds, and on-device throughput.

AI chips

NVIDIA Rubin-class inference

Up to 50 PFLOPS NVFP4 inference

Accelerated inference on next-gen data center GPUs.

More inference capacity per chip can make project assistants and batch document work faster and cheaper.

AI clusters

Google Ironwood TPU pod

Up to 42.5 exaflops at pod scale

Hyperscale inference/train pods built from Ironwood TPUs.

Large pods mean more warehouse-scale builds: power routing, liquid cooling, and fit-out in tight sequences.

AI clusters

AWS Trainium3 UltraServer

Up to 362 FP8 PFLOPS

Dense AI accelerator servers for training and inference in AWS regions.

Regional capacity additions show up as data hall retrofits and greenfield campuses your trade partners may be chasing.

Data center power

Data center electricity (IEA outlook)

~945 TWh by 2030

IEA projection for electricity use linked to AI and data center demand.

AI growth ties to utility-scale power, substations, and on-site generation β€” direct civil, electrical, and commissioning demand.

Construction demand

Data center capacity (McKinsey scenarios)

~220 GW by 2030 (scenarios)

Illustrative industry capacity path β€” highly scenario- and region-dependent.

Megawatt-scale rollouts drive shell, MEP, and mission-critical trades; timing varies by market.

Construction demand

Data center build-out (McKinsey scenarios)

~$7 trillion by 2030 (scenarios)

Aggregate capital signaled for infrastructure that supports AI-scale compute.

Large capex bands mean sustained demand for contractors on power, cooling, shell, and site civil work β€” if scenarios materialize.

Construction technologies to watch

Field and site-adjacent signals β€” directional, not a timeline.

Data center growth

Regional pipeline still beats global headlines β€” permitting, power, and parcels decide what you bid.

Track your market’s interconnection queue and campus announcements; pair with the capacity figures above for scale.

Humanoid robotics

Vendors emphasize structured industrial tasks first; unpredictable field conditions lag controlled environments.

Prefab, logistics, and repetitive handling are earlier bets than open jobsite autonomy.

AR / XR

Forecasts vary by segment β€” treat as adoption signal, not a precise market timetable.

Mature overlays could speed layout QA, inspections, and install guidance from BIM.

Wearable AI

Construction wearables sit in the low tens of billions USD across analyst taxonomies.

Safety coaching, proximity alerts, and richer field notes β€” still with competent supervision.

Robotic layout / field automation

Robotic total stations target repeatable layout vs. manual baselines in vendor benchmarks.

Layout, scanning, and progress capture are practical wins before full-site robotics.

Vendor limits, product tiers, and analyst scenarios change β€” treat these figures as directional snapshots, not contract specs.