10.5 Future Challenges
Oceanography faces grand challenges: closing observation gaps in the deep ocean, polar regions, and boundary currents; predicting marine heatwaves and deoxygenation; harnessing AI/ML for data-driven discovery; developing ocean carbon dioxide removal strategies; building a digital twin of the ocean; and governing deep-sea mining. The UN Decade of Ocean Science (2021โ2030) frames these priorities for "the ocean we need for the future we want."
UN Decade of Ocean Science (2021โ2030)
The UN Ocean Decade aims to generate transformative ocean science to support sustainable development. Seven desired outcomes define "The Ocean We Want":
Marine Heatwaves & Deoxygenation
A marine heatwave (MHW) is defined as a discrete, prolonged anomalously warm water event. Following Hobday et al. (2016), a MHW occurs when SST exceeds the 90th percentile of the climatological distribution for at least 5 consecutive days:
$$\text{MHW if } T(t) > T_{90}(d) \text{ for } \geq 5 \text{ consecutive days}$$
$T_{90}(d)$ = 90th percentile threshold for day-of-year $d$ from 30-year climatology
MHW Intensity Metrics
Maximum intensity: $I_{\max} = \max(T(t) - T_{\text{clim}}(d))$ during event. Cumulative intensity: $I_{\text{cum}} = \sum (T(t) - T_{\text{clim}}(d)) \cdot \Delta t$ (ยฐCยทdays). MHW frequency has doubled since the 1980s.
Ocean Deoxygenation
Global ocean oxygen content has decreased by ~2% since 1960. Oxygen minimum zones (OMZs) are expanding. Driven by warming (reduced solubility: $\partial C_{\text{sat}}/\partial T < 0$) and increased stratification (reduced ventilation).
AI/ML Revolution in Ocean Science
Artificial intelligence and machine learning are transforming oceanography across scales. Physics-informed neural networks (PINNs) encode governing equations as soft constraints in the loss function:
$$\mathcal{L} = \mathcal{L}_{\text{data}} + \lambda \mathcal{L}_{\text{physics}} = \frac{1}{N}\sum_i \|u_\theta(x_i) - u_i^{\text{obs}}\|^2 + \lambda \frac{1}{M}\sum_j \|\mathcal{N}[u_\theta](x_j)\|^2$$
$\mathcal{N}$ = differential operator from governing PDE,$\lambda$ = physics weight, $u_\theta$ = neural network prediction
Digital Twin of the Ocean
Real-time, data-assimilating ocean models coupled with ML emulators. Copernicus Digital Twin Ocean and NOAA initiatives. Enables rapid scenario testing.
Exascale Ocean Modeling
Next-generation models at ~1 km global resolution. GPU-accelerated. Explicitly resolve mesoscale eddies and boundary currents worldwide.
Data-Driven Discovery
Unsupervised learning reveals new ocean regimes. Causal inference identifies drivers of variability. Foundation models trained on petabytes of ocean data.
Observing System Gaps
Deep ocean (>2000 m): ~10% observed. Under-ice Arctic/Antarctic. Western boundary currents. Biogeochemistry. Deep Argo and autonomous platforms are expanding coverage.
Ocean Carbon Dioxide Removal (CDR)
Achieving net-zero emissions will likely require active CO&sub2; removal from the ocean-atmosphere system. Ocean-based CDR approaches exploit the ocean's vast capacity as a carbon reservoir:
Ocean Alkalinity Enhancement (OAE)
Adding alkaline minerals (olivine, lime) to the ocean increases $[\text{CO}_3^{2-}]$, shifting carbonate equilibrium to absorb more atmospheric CO&sub2;. The reaction:$\text{CaO} + \text{CO}_2 + \text{H}_2\text{O} \rightarrow \text{Ca}^{2+} + 2\text{HCO}_3^-$
Macroalgae (Kelp) Farming & Sinking
Growing kelp in open-ocean farms and sinking it to the deep ocean for long-term carbon sequestration. Kelp productivity: $\sim 1\text{--}3$ kg C/mยฒ/yr. Challenges: MRV, ecological impact, permanence.
Direct Ocean Capture (DOC)
Electrochemical processes that extract dissolved CO&sub2; from seawater, enabling the ocean to absorb more from the atmosphere. Energy-intensive but potentially scalable.
Derivation: Climate Projection Uncertainty Quantification
Step 1: Sources of Uncertainty in Climate Projections
Total uncertainty in a climate projection $Y(t)$ (e.g., SST in 2100) comes from three independent sources: scenario (forcing) uncertainty, model (structural) uncertainty, and internal (natural) variability:
$$\text{Var}[Y(t)] = \sigma_{\text{scenario}}^2(t) + \sigma_{\text{model}}^2(t) + \sigma_{\text{internal}}^2(t)$$
Step 2: ANOVA Decomposition (Hawkins & Sutton)
Following Hawkins and Sutton (2009), decompose the multi-model ensemble into these components. For $M$ models each run under $S$ scenarios with $R$ realisations, the total variance is partitioned by ANOVA:
$$Y_{msr}(t) = \mu(t) + \alpha_s(t) + \beta_m(t) + \epsilon_{msr}(t)$$
Step 3: Estimate Each Variance Component
The grand mean $\mu(t)$ is the forced response. Scenario uncertainty $\sigma_{\text{scenario}}^2$ is the variance of scenario means, model uncertainty $\sigma_{\text{model}}^2$ is the variance of model means within a scenario, and internal variability $\sigma_{\text{internal}}^2$ is the residual:
$$\sigma_{\text{scenario}}^2 = \frac{1}{S-1}\sum_s (\bar{Y}_{s\cdot\cdot} - \bar{Y}_{\cdot\cdot\cdot})^2, \quad \sigma_{\text{model}}^2 = \frac{1}{S(M-1)}\sum_{s,m}(\bar{Y}_{sm\cdot} - \bar{Y}_{s\cdot\cdot})^2$$
Step 4: Time Dependence of Uncertainty Fractions
Near-term (2020--2040): internal variability dominates (~60%). Mid-century (2040--2060): model uncertainty is largest (~50%). End-of-century (2080--2100): scenario uncertainty dominates (~70%) as forcing pathways diverge. This informs both science priorities and policy relevance:
$$f_{\text{source}}(t) = \frac{\sigma_{\text{source}}^2(t)}{\sigma_{\text{total}}^2(t)} \times 100\%$$
Derivation: Multi-Model Ensemble Weighting Methods
Step 1: Simple Multi-Model Mean (Democracy)
The simplest ensemble approach gives equal weight to each model. The multi-model mean (MMM) and its uncertainty are:
$$\bar{Y} = \frac{1}{M}\sum_{m=1}^{M} Y_m, \quad \sigma_{\bar{Y}} = \frac{1}{M}\sqrt{\sum_m (Y_m - \bar{Y})^2}$$
Step 2: Performance-Based Weighting
Models can be weighted by their skill in reproducing historical observations. Using a distance metric $D_m$ between model $m$ and observations, weights are assigned inversely proportional to squared distance:
$$w_m = \frac{\exp(-D_m^2 / (2\sigma_D^2))}{\sum_{k=1}^{M} \exp(-D_k^2 / (2\sigma_D^2))}, \quad \bar{Y}_w = \sum_m w_m Y_m$$
Step 3: Independence Weighting (ClimWIP)
Many CMIP models share code and development history, violating independence assumptions. The Climate model Weighting by Independence and Performance (ClimWIP) method down-weights similar models using an inter-model distance $S_{mk}$:
$$w_m \propto \frac{\exp(-D_m^2 / \sigma_D^2)}{\sum_{k \ne m} \exp(-S_{mk}^2 / \sigma_S^2) + 1}$$
Step 4: Bayesian Model Averaging (BMA)
BMA treats each model as a hypothesis and computes posterior weights from the likelihood of the observations given each model, multiplied by a prior weight:
$$p(Y|\text{obs}) = \sum_{m=1}^{M} p(Y|M_m) \cdot p(M_m|\text{obs}), \quad p(M_m|\text{obs}) \propto p(\text{obs}|M_m) \cdot p(M_m)$$
Step 5: Ensemble Spread vs. Skill
A well-calibrated ensemble has the property that its spread matches its actual prediction error, tested by the rank histogram. The reliability ratio $R = \sigma_{\text{ensemble}}^2 / \text{MSE}$ should be near 1. Under-dispersive ensembles ($R < 1$) underestimate uncertainty, which is common in CMIP projections and motivates the weighting approaches above.
Python: Marine Heatwave Detection, O&sub2; Trends & Simple PINN
Python: Marine Heatwave Detection, O&sub2; Trends & Simple PINN
Python!/usr/bin/env python3
Click Run to execute the Python code
Code will be executed with Python 3 on the server
Fortran: Climate Projection Bias Correction & Downscaling
This program implements quantile mapping bias correction for ocean climate model output. The model CDF is mapped to the observed CDF to correct systematic biases in SST projections while preserving the climate change signal:
$$T_{\text{corrected}} = F_{\text{obs}}^{-1}\!\left(F_{\text{model}}(T_{\text{raw}})\right)$$
$F$ = cumulative distribution function; subscripts denote observed vs model distributions
Fortran: Climate Projection Bias Correction & Downscaling
FortranQuantile mapping bias correction for ocean climate projections
Click Run to execute the Fortran code
Code will be compiled with gfortran and executed on the server