Scott Steinschneider’s research while affiliated with Cornell University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (141)


The main methodological components of the study. In Step 1, we create an ensemble of synthetic weather simulations over the Late Renaissance Megadrought that are conditioned on tree‐ring‐based natural variability and include plausible signals of climate change. In Step 2, we use the ensembles of generated surface weather to force hydrologic models to develop streamflow ensembles for each inflow point into the water systems model. In Step 3, we force the model with the streamflow ensembles and then quantify vulnerabilities in different parts of the system.
California Food‐Energy‐Water System model storage and conveyance network, which includes natural channels, canals, reservoirs, irrigation districts, and groundwater banks. Nodes in the network that are focused on in this study are highlighted with a colored outline. Pink outlines correspond to Central Valley Project structures and yellow outlines correspond State Water Project structures. Shared structures are outline in both colors.
(a) The 12 climate scenarios for this study are developed by applying respective temperature increases or precipitation scaling to the baseline paleo‐based weather. (b) Step changes between 1°C and 4°C are added to each grid cell's simulated temperature. (c) Precipitation scaling of 3.5%, 5.25%, and 7% are applied to the upper and lower tail of the distribution, while the mean is left unchanged. Each climate scenario is comprised of 50 stochastic ensemble members. The orange climate scenario in panel (a) is selected for further exploration in the results section.
The 15‐year rolling annual average full natural flow (FNF) in the (a) Sacramento San‐Joaquin Basin and (b) the Tulare Basin through the extent of the paleo‐period (1400–2017 CE). The dark brown line represents the median FNF and the shading represents the 5th and 95th percentile confidence bounds. The 16th century megadrought period (which lasted approximately from 1520 to 1580 CE) is highlighted in gray. Panels (c) and (d) are violin plots of annual accumulated flow compared between the California Data Exchange Center (CDEC) observed data set (1997–2016) and the most severe 30‐year part (1550–1580 CE) of the 16th century megadrought. Panels (e) and (f) show similar results for monthly minimum flow. Panels (g) and (h) show the minimum 3‐, 5‐, and 10‐year rolling average flow across the 30‐year period of each member of the ensemble. The aqua dot corresponds to the respective minimum rolling average flow for the CDEC data set. The left and right columns show results for the Sacramento San‐Joaquin Basin and Tulare Basin, respectively.
Quantile plots for monthly storage at (a) Shasta and (b) Oroville. The shading represents the percentage of months in the 50‐member ensemble between 1550 and 1580 that lie below a given storage (values on the y‐axis). Lines denoting the period‐of‐record average and minimum from California Food‐Energy‐Water System model forced with California Data Exchange Center observations of full natural flow are shown in black and red, respectively. Gray lines denote maximum capacity and dead pool for each reservoir.

+7

Exploring Water System Vulnerabilities in California's Central Valley Under the Late Renaissance Megadrought and Climate Change
  • Article
  • Full-text available

April 2025

·

39 Reads

·

1 Citation

·

Scott Steinschneider

·

California faces cycles of drought and flooding that are projected to intensify, but these extremes may impact water users across the state differently due to the region's natural hydroclimate variability and complex institutional framework governing water deliveries. To assess these risks, this study introduces a novel exploratory modeling framework informed by paleo and climate‐change based scenarios to better understand how impacts propagate through the Central Valley's complex water system. A stochastic weather generator, conditioned on tree‐ring data, produces a large ensemble of daily weather sequences conditioned on drought and flood conditions under the Late Renaissance Megadrought period (1550–1580 CE). Regional climate changes are applied to this weather data and drive hydrologic projections for the Sacramento, San Joaquin, and Tulare Basins. The resulting streamflow ensembles are used in an exploratory stress test using the California Food‐Energy‐Water System model, a highly resolved, daily model of water storage and conveyance throughout California's Central Valley. Results show that megadrought conditions lead to unprecedented reductions in inflows and storage at major California reservoirs. Both junior and senior water rights holders experience multi‐year periods of curtailed water deliveries and complete drawdowns of groundwater assets. When megadrought dynamics are combined with climate change, risks for unprecedented depletion of reservoir storage and sustained curtailment of water deliveries across multiple years increase. Asymmetries in risk emerge depending on water source, rights, and access to groundwater banks.

Download

A novel hybrid fine-tuning method for supercharging deep learning model development for hydrological prediction

March 2025

·

13 Reads

·

·

Chaopeng Shen

·

[...]

·

Jan Adamowski

This study proposes a novel hybrid method that substantially accelerates and improves deep learning (DL) model development for streamflow prediction. The method leverages a combination of a long short-term memory (LSTM) network and random forests. A hybrid encoder-decoder model is designed, where a pre-trained LSTM is utilized as an encoder to extract temporal features from the input data. Subsequently, the random forest decoder processes the encoded information to make streamflow predictions. Our method was tested on 421 catchments in the continental United States and 324 in Germany, both selected from two CAMELS datasets. The hybrid method has several benefits. First, it is much more efficient and robust than training LSTMs on each catchment individually (~14x faster). Second, it is much less computationally expensive than LSTM fine-tuning (i.e., feasible on a CPU-based workstation). Third, it achieves superior accuracy compared to a catchment-wise training strategy (e.g., 9.2 % improvement in the median in Nash-Sutcliffe Efficiency (NSE)), shows competitive performance compared to regional LSTM models when trained with fewer data, and through fine-tuning, improves regional LSTM performance in out-of-training samples by 13.13 % (median NSE). To our knowledge, this is the first decision-tree model integrated within a DL workflow to enhance fine-tuning efficiency of pre-trained models in new locations. This hybrid approach holds significant promise for future applications in hydrological modeling, particularly considering the imminent rise of geospatial foundation models in hydrology that will rely on transfer learning techniques for effective deployment.


A novel hybrid fine-tuning method for supercharging deep learning model development for hydrological prediction

March 2025

·

25 Reads

This study proposes a novel hybrid method that substantially accelerates and improves deep learning (DL) model development for streamflow prediction. The method leverages a combination of a long short-term memory (LSTM) network and random forests. A hybrid encoder-decoder model is designed, where a pre-trained LSTM is utilized as an encoder to extract temporal features from the input data. Subsequently, the random forest decoder processes the encoded information to make streamflow predictions. Our method was tested on 421 catchments in the continental United States and 324 in Germany, both selected from two CAMELS datasets. The hybrid method has several benefits. First, it is much more efficient and robust than training LSTMs on each catchment individually (~14x faster). Second, it is much less computationally expensive than LSTM fine-tuning (i.e., feasible on a CPU-based workstation). Third, it achieves superior accuracy compared to a catchment-wise training strategy (e.g., 9.2 % improvement in the median in Nash-Sutcliffe Efficiency (NSE)), shows competitive performance compared to regional LSTM models when trained with fewer data, and through fine-tuning, improves regional LSTM performance in out-of-training samples by 13.13 % (median NSE). To our knowledge, this is the first decision-tree model integrated within a DL workflow to enhance fine-tuning efficiency of pre-trained models in new locations. This hybrid approach holds significant promise for future applications in hydrological modeling, particularly considering the imminent rise of geospatial foundation models in hydrology that will rely on transfer learning techniques for effective deployment.


A novel hybrid fine-tuning method for supercharging deep learning model development for hydrological prediction

March 2025

·

28 Reads

This study proposes a novel hybrid method that substantially accelerates and improves deep learning (DL) model development for streamflow prediction. The method leverages a combination of a long short-term memory (LSTM) network and random forests. A hybrid encoder-decoder model is designed, where a pre-trained LSTM is utilized as an encoder to extract temporal features from the input data. Subsequently, the random forest decoder processes the encoded information to make streamflow predictions. Our method was tested on 421 catchments in the continental United States and 324 in Germany, both selected from two CAMELS datasets. The hybrid method has several benefits. First, it is much more efficient and robust than training LSTMs on each catchment individually (~14x faster). Second, it is much less computationally expensive than LSTM fine-tuning (i.e., feasible on a CPU-based workstation). Third, it achieves superior accuracy compared to a catchment-wise training strategy (e.g., 9.2 % improvement in the median in Nash-Sutcliffe Efficiency (NSE)), shows competitive performance compared to regional LSTM models when trained with fewer data, and through fine-tuning, improves regional LSTM performance in out-of-training samples by 13.13 % (median NSE). To our knowledge, this is the first decision-tree model integrated within a DL workflow to enhance fine-tuning efficiency of pre-trained models in new locations. This hybrid approach holds significant promise for future applications in hydrological modeling, particularly considering the imminent rise of geospatial foundation models in hydrology that will rely on transfer learning techniques for effective deployment.



Compound coastal flooding in San Francisco Bay under climate change

January 2025

·

102 Reads

·

3 Citations

The risk of compound coastal flooding in the San Francisco Bay Area is increasing due to climate change yet remains relatively underexplored. Using a novel hybrid statistical-dynamical downscaling approach, this study investigates the impacts of climate change induced sea-level rise and higher river discharge on the magnitude and frequency of flooding events as well as the relative importance of various forcing drivers to compound flooding within the Bay. Results reveal that rare occurrences of flooding under the present-day climate are projected to occur once every few hundred years under climate change with relatively low sea-level rise (0.5 m) but would become annual events under climate change with high sea-level rise (1.0 to 1.5 m). Results also show that extreme water levels that are presently dominated by tides will be dominated by sea-level rise in most locations of the Bay in the future. The dominance of river discharge to the non-tidal and non-sea-level rise driven water level signal in the North Bay is expected to extend ~15 km further seaward under extreme climate change. These findings are critical for informing climate adaptation and coastal resilience planning in San Francisco Bay.



A hybrid statistical–dynamical framework for compound coastal flooding analysis

December 2024

·

117 Reads

·

1 Citation

Compound coastal flooding due to astronomic, atmospheric, oceanographic, and hydrologic drivers poses severe threats to coastal communities. While physics-driven approaches are able to dynamically simulate temporally and spatially varying compound flooding generated by multiple drivers with correlations between some of them, computational burdens limit their capability to explore the full range of conditions that contribute to compound coastal hazards. Data-driven statistical approaches address some of these computational challenges; however, they are also unable to explore all possible forcing combinations due to short observational records, and projections are typically limited to a few locations. This study proposes a hybrid statistical–dynamical framework for compound coastal flooding analysis that integrates a stochastic generator of compound flooding drivers, a hydrodynamic model, and machine learning-based surrogate models. The framework was demonstrated in San Francisco Bay (SF) over the past 500 years with accuracy similar to the physics-driven approach but with much higher computational efficiency. The stochastic generator of compound flooding drivers is developed by coupling a sea surface temperature (SST) reconstruction model with a climate emulator, weather generator, and model of the hydrological and reservoir system. Using reconstructed SSTs as input, the generator of compound flooding drivers is employed to simulate time series of the forcing factors contributing to compound flooding (e.g. surge, waves, river discharge, etc) in SF Bay. A process-based hydrodynamic model is built to predict total water levels varying in time and space throughout SF Bay based on stochastically generated drivers. The machine learning-based surrogate models are then developed from a relatively small library (several hundred) of hydrodynamic model simulations to efficiently predict water levels for compound flooding analysis under the full range of stochastic drivers. This study contributes a hybrid statistical–dynamical framework to better understand the spatial distribution and temporal evolution of compound coastal-fluvial flooding, along with the relative contributions of drivers in complex nearshore, estuarine, and river environments for centennial timescales under past, present, and future climates.



Bridging the Gap Between Top‐Down and Bottom‐Up Climate Vulnerability Assessments: Process Informed Exploratory Scenarios Identify System‐Based Water Resource Vulnerabilities

November 2024

·

55 Reads

·

3 Citations

The threat that climate change poses to water resource systems has led to a substantial and growing number of impact studies. These studies follow two approaches: (a) top‐down studies are driven by projections of future climate change provided by downscaled general circulation models (GCMs); and (b) bottom‐up studies are driven by the systematic evaluation of exploratory scenarios. Top‐down approaches produce realistic scenarios rooted in the simulation of thermodynamic and dynamic processes represented in GCMs, but the internal resolution of these processes make it difficult to link vulnerabilities to discrete components of change. Bottom‐up approaches link vulnerabilities to discrete components of change through the structured evaluation of exploratory scenarios, but the lack of insight rooted in climate change processes can lead to the development of implausible scenarios. This paper evaluates exploratory scenarios developed through thermodynamic and dynamical guided perturbations motivated by GCM‐bound insights. The result is a hybrid approach that bridges a gap between top‐down and bottom‐up approaches. This yields several advantages. First, emerging vulnerabilities are linked to distinct thermodynamic and dynamic processes that are modeled in GCMs with differential likelihoods and plausible ranges of change. Second, the structured evaluation of process‐informed exploratory scenarios link system vulnerabilities to distinct components of climate change. An illustrative case study demonstrates perturbations linked to thermodynamic and dynamical processes have a large impact on stakeholder‐defined flood and drought performance, and the structured evaluation of process‐informed exploratory scenarios find nuanced infrastructure‐specific vulnerabilities that would be difficult to identify using an alternative approach.


Citations (75)


... As the uncertainty around how to refer to compound events and how to incorporate them into existing disaster frameworks decreased, there still existed a lack of quantitative methods developed around real-world applications of these frameworks. It is now possible to see this trend shift on an international scale, especially in the case of flood-related compound events 13,14 . As the body of research on these topics grow, there are many perspectives from which the quantitative methods can be applied. ...

Reference:

Quantifying the compounding effects of natural hazard events: a case study on wildfires and floods in California
Compound coastal flooding in San Francisco Bay under climate change

... Although yielding important findings, these studies focused primarily on single regions with limited data samples. The literature also includes studies with a considerable of number watershed samples, such as Jiang (2001), who analyzed 307 watersheds and 24 plots across the United States, although without representation in the Western regions, and Neelam et al. (2024), who studied 333 watersheds also within the United States. Again, these studies focus on a single country and often treated the frequency-matched approach as a secondary aim. ...

Improved Regionalization of the CN Method for Extreme Events at Ungauged Sites across the US
  • Citing Article
  • December 2024

Journal of Hydrologic Engineering

... Alternative hybrid modeling approaches have been developed by merging statistical and numerical modeling to generate the response of dynamical approaches under a full range of possible forcing combinations 40,45,54-56 . This study uses the hybrid statistical-dynamical approach developed in ref. 56 to analyze compound coastal flooding in SF Bay under a range of climate change scenarios. By combining a stochastic generator of compound flooding drivers, a high-fidelity hydrodynamic simulator, and machine learning-based surrogate models; the hybrid framework can investigate the full range of plausible forcing combinations for compound flooding analysis, applied to SF Bay. ...

A hybrid statistical–dynamical framework for compound coastal flooding analysis

... To reduce the bias in the simulation of extreme precipitation events that have in the past been poorly represented by GCMs, statistical models have been used which represent regional climate patterns with reasonable skill (Farnham et al. 2018). Process-informed stochastic models that formalised the link between causal pathways in the climate system, local weather scenarios and water system vulnerabilities (Kucharski et al. 2024;Rahat et al. 2022) have been introduced (Steinschneider et al. 2019). ...

Bridging the Gap Between Top‐Down and Bottom‐Up Climate Vulnerability Assessments: Process Informed Exploratory Scenarios Identify System‐Based Water Resource Vulnerabilities

... WRs are identified (typically, and in this case) by NHMMs clustering of the principal components (PC) of daily 500 hPa geopotential height anomalies following the methodology of Steinschneider et al. (2019), Najibi et al. (2021), and Najibi et al. (2024) for the atmospheric circulation data selection, and a clustering method to identify the WR states. For efficient timeindependent clustering on PCs, K-means can be used, whilst NHMM is preferred for WR identification that accounts for the persistence of WRs. ...

A statewide, weather-regime based stochastic weather generator for process-based bottom-up climate risk assessments in California – Part I: Model evaluation

Climate Services

... Hybrid statistical-dynamical framework The hybrid statistical-dynamical framework for compound coastal flooding analysis is developed by integrating a stochastic generator of compound flooding drivers, a physics-based high fidelity hydrodynamic model, and machine learning-based surrogate models 56 (Fig. S6). The generator of compound flooding drivers can simulate time series of joint astronomic, atmospheric, oceanographic, and hydrologic forcings of compound coastal flooding in SF Bay by combining a sea surface temperature (SST) reconstruction model 72 , a stochastic climate emulator 73 , a stochastic weather generator [63][64][65] , and a hybrid physics-based and data-driven hydrologic and reservoir system model 66,67 ). The SST reconstruction model 72 creates the annual principal components (APCs) of SST anomalies, which are passed into the climate emulator TESLA (Time-varying Emulator for Short-and Long-Term analysis) 73 to generate synthetic annual weather types (AWTs). ...

A statewide, weather-regime based stochastic weather generator for process-based bottom-up climate risk assessments in California – Part II: Thermodynamic and dynamic climate change scenarios

Climate Services

... The GAMLSS model typically employs the maximum likelihood method to estimate time-varying parameter models [42][43][44][45]. However, the validity of parameter estimation methods for P-III distribution based on the likelihood function, such as the maximum likelihood and Bayesian methods, is restricted to samples with absolute skewness not greater than two [46]. ...

A Hybrid, Non‐Stationary Stochastic Watershed Model (SWM) for Uncertain Hydrologic Simulations Under Climate Change

... Uncertainty modeling is a common approach to represent the uncertainty of RES generation, market prices, and changes in daily electrical load, rather than assuming a certain knowledge of known variables in the deterministic method. In this case, random distributions are used as input to the stochastic optimization problem to mimic the probability characteristics of the variables [39].The integration of RES, such as wind and solar power, into power systems has introduced new challenges related to uncertainty. RES generation is highly dependent on weather conditions, which can be variable and difficult to predict accurately. ...

Quantifying the impact of multi-scale climate variability on electricity prices in a renewable-dominated power grid
  • Citing Article
  • March 2024

Renewable Energy

... Streamflow data plays a critical role in various aspects of water resources management, including flood risk or flood hazard assessment (Do et al. 2017;Tran et al. 2020Tran et al. , 2023, climate change impact assessment (Wang et al. 2024), and reservoir design and operation (Brodeur et al. 2024;Vogel and Stedinger 1988). However, despite its fundamental importance, there has been a concerning trend of declining stream gauge networks worldwide in recent years (Chen et al. 2023). ...

Synthetic Forecast Ensembles for Evaluating Forecast Informed Reservoir Operations

... All modeling and tuning were conducted in Python using scikitlearn and specialized ML libraries. This procedure follows established best practices for hydrological modeling (Wi & Steinschneider 2024), (Jiang et al. 2023). Cross-validation was applied to ensure model robustness, prevent overfitting, ...

On the need for physical constraints in deep learning rainfall–runoff projections under climate change: a sensitivity analysis to warming and shifts in potential evapotranspiration