Loading content ...

A theoretical synthesis for minimum–time control of vehicles operating in strong and dynamic flows is provided. The synthesis relies on the computation of the forward reachable set of states. Using ideas rooted in the theory of non–smooth calculus, we prove that this set is governed by the viscosity solution of an unsteady Hamilton-Jacobi (HJ) equation. We show that the minimum arrival time satisfies a static HJ equation, when a special local controllability condition holds. Results are exemplified by a realistic application to autonomous underwater gliders operating in the Sulu Archipelago.

Subramani, D.N., T. Lolla, P.J. Haley and P.F.J Lermusiaux, 2014. A Stochastic Optimization Method for Energy-based Path Planning, In: Dynamic Data-driven Environmental Systems Science Conference (Eds. Ravela, Sandu et al), Springer Lecture Notes In Computer Science, In Press. Final Publication will be available at http://link.springer.com/

The level set methodology for time-optimal path planning is employed to predict collision-free and fastest time trajectories for swarms of underwater vehicles deployed in the Philippine Archipelago region.
To simulate the multiscale ocean flows in this complex region, a data-assimilative primitive-equation ocean modeling system is employed with
telescoping domains that are interconnected by implicit two-way nesting.
These data-driven multiresolution simulations provide a
realistic flow environment, including variable large-scale currents,
strong jets, eddies, wind-driven currents and tides.
The properties and capabilities of the rigorous level set methodology are
illustrated and assessed quantitatively for several vehicle types and mission scenarios.
Feasibility studies of all-to-all broadcast missions, leading to minimal time transmission between source and receiver locations, are performed using a large number of vehicles.
The results with gliders and faster propelled vehicles are compared.
Reachability studies, i.e.~determining the boundaries of regions that can be reached by vehicles for exploratory missions, are then exemplified and analyzed.
Finally, the methodology is used to determine the optimal strategies
for fastest time pick-up of deployed gliders by means of
underway surface vessels or stationary platforms.
The results highlight the complex effects of multiscale flows on the optimal paths,
the need to utilize the ocean environment for more efficient autonomous
missions and the benefits of including ocean forecasts in the planning of time-optimal paths.

We develop an accurate partial differential equation based methodology that predicts the time-optimal paths of autonomous vehicles navigating in any continuous, strong and dynamic ocean currents, obviating the need for heuristics. The goal is to predict a sequence of steering directions so that vehicles can best utilize or avoid currents to minimize their travel time. Inspired by the level set method, we derive and demonstrate that a modified level set equation governs the time-optimal path in any continuous flow. We show that our algorithm is computationally efficient and apply it to a number of experiments. First, we validate our approach through a simple benchmark application in a Rankine vortex flow for which an analytical solution is available. Next, we apply our methodology to more complex, simulated flow-fields such as unsteady double-gyre flows driven by wind stress and flows behind a circular island. These examples show that time-optimal paths for multiple vehicles can be planned, even in the presence of complex flows in domains with obstacles. Finally, we present, and support through illustrations, several remarks that describe specific features of our methodology.

Lermusiaux P.F.J, T. Lolla, P.J. Haley. Jr., K. Yigit, M.P. Ueckermann, T. Sondergaard and W.G. Leslie, 2014. *Science of Autonomy: Time-Optimal Path Planning and Adaptive Sampling for Swarms of Ocean Vehicles*. Chapter 11, Springer Handbook of Ocean Engineering: Autonomous Ocean Vehicles, Subsystems and Control, Tom Curtin (Ed.). In press.

The properties and capabilities of the GMM-DO filter are assessed and exemplified by applications
to two dynamical systems: (1) the Double Well Diffusion and (2) Sudden Expansion flows; both
of which admit far-from-Gaussian statistics. The former test case, or twin experiment, validates
the use of the EM algorithm and Bayesian Information Criterion with Gaussian Mixture Models
in a filtering context; the latter further exemplifies its ability to efficiently handle state vectors of
non-trivial dimensionality and dynamics with jets and eddies. For each test case, qualitative and
quantitative comparisons are made with contemporary filters. The sensitivity to input parameters
is illustrated and discussed. Properties of the filter are examined and its estimates are described,
including: the equation-based and adaptive prediction of the probability densities; the evolution
of the mean field, stochastic subspace modes and stochastic coefficients; the fitting of Gaussian
Mixture Models; and, the efficient and analytical Bayesian updates at assimilation times and the
corresponding data impacts. The advantages of respecting nonlinear dynamics and preserving
non-Gaussian statistics are brought to light. For realistic test cases admitting complex distributions
and with sparse or noisy measurements, the GMM-DO filter is shown to fundamentally improve the
filtering skill, outperforming simpler schemes invoking the Gaussian parametric distribution.

This work introduces and derives an efficient, data-driven assimilation scheme, focused on a
time-dependent stochastic subspace, that respects nonlinear dynamics and captures non-Gaussian
statistics as it occurs. The motivation is to obtain a filter that is applicable to realistic geophysical
applications but that also rigorously utilizes the governing dynamical equations with information
theory and learning theory for efficient Bayesian data assimilation. Building on the foundations of
classical filters, the underlying theory and algorithmic implementation of the new filter are developed
and derived. The stochastic Dynamically Orthogonal (DO) field equations and their adaptive
stochastic subspace are employed to predict prior probabilities for the full dynamical state, effectively
approximating the Fokker-Planck equation. At assimilation times, the DO realizations are fit to
semiparametric Gaussian mixture models (GMMs) using the Expectation-Maximization algorithm
and the Bayesian Information Criterion. Bayes’ Law is then efficiently carried out analytically within
the evolving stochastic subspace. The resulting GMM-DO filter is illustrated in a very simple example.
Variations of the GMM-DO filter are also provided along with comparisons with related schemes.

The quantification of uncertainties is critical when systems are nonlinear and have uncertain terms in their governing equations or are constrained by limited knowledge of initial
and boundary conditions. Such situations are common in multiscale, intermittent and non-
homogeneous fluid and ocean flows. The Dynamically Orthogonal (DO) field equations
provide an efficient time-dependent adaptive methodology to predict the probability density functions of such flows. The present work derives efficient computational schemes for
the DO methodology applied to unsteady stochastic Navier-Stokes and Boussinesq equations, and illustrates and studies the numerical aspects of these schemes. Semi-implicit
projection methods are developed for the mean and for the orthonormal modes that define
a basis for the evolving DO subspace, and time-marching schemes of first to fourth order
are used for the stochastic coefficients. Conservative second-order finite-volumes are employed in physical space with Total Variation Diminishing schemes for the advection terms.
Other results specific to the DO equations include: (i) the definition of pseudo-stochastic
pressures to obtain a number of pressure equations that is linear in the subspace size in-
stead of quadratic; (ii) symmetric Total Variation Diminishing-based advection schemes
for the stochastic velocities; (iii) the use of generalized inversion to deal with singular
subspace covariances or deterministic modes; and (iv) schemes to maintain orthonormal
modes at the numerical level. To verify the correctness of our implementation and study
the properties of our schemes and their variations, a set of stochastic flow benchmarks are
defined including asymmetric Dirac and symmetric lock-exchange flows, lid-driven cavity flows, and flows past objects in a confined channel. Different Reynolds number and
Grashof number regimes are employed to illustrate robustness. Optimal convergence under both time and space refinements is shown as well as the convergence of the probability
density functions with the number of stochastic realizations.

We develop and illustrate an efficient but rigorous
methodology that predicts the time-optimal paths of ocean
vehicles in dynamic continuous flows. The goal is to best
utilize or avoid currents, without limitation on these currents
nor on the number of vehicles. The methodology employs a
new modified level set equation to evolve a wavefront from
the starting point of vehicles until they reach their desired
goal locations, combining flow advection with nominal vehicle
motions. The optimal paths of vehicles are then computed
by solving particle tracking equations backwards in time.
The computational cost is linear with the number of vehicles
and geometric with spatial dimensions. The methodology is
applicable to any continuous flows and many vehicles scenarios.
Present illustrations consist of the crossing of a canonical
uniform jet and its validation with an optimization problem,
as well as more complex time varying 2D flow fields, including
jets, eddies and forbidden regions.

Rixen, M., P.F.J. Lermusiaux and J. Osler, 2012. *Quantifying, Predicting and Exploiting Uncertainties in Marine Environments*, Ocean Dynamics, 62(3):495–499, doi: 10.1007/s10236-012-0526-8.

Following the scientific, technical and field trial initiatives ongoing since the Maritime Rapid Environmental Assessment (MREA) conferences in 2003, 2004 and 2007, the MREA10 conference provided a timely opportunity to review the progress on various aspects of MREA, with a particular emphasis on marine environmental uncertainty management. A key objective of the conference was to review the present state-of-the art in Quantifying, Predicting and Exploiting (QPE) marine environmental uncertainties. The integration of emerging environmental monitoring and modeling techniques into data assimilation streams and their subsequent exploitation at an operational level involves a complex chain of non-linear uncertainty transfers, including human factors. Accordingly the themes for the MREA10 conference were selected to develop a better understanding of uncertainty, from its inception in the properties being measured and instrumentation employed, to its eventual impact in the applications that rely upon environmental information.

Contributions from the scientific community were encouraged on all aspects of environmental uncertainties: their quantification, prediction, understanding and exploitation. Contributions from operational communities, the consumers of environmental information who have to cope with uncertainty, were also encouraged. All temporal and spatial scales were relevant: tactical, operational, and strategic, including uncertainty studies for topics with long-term implications. Manuscripts reporting new technical and theoretical developments in MREA, but acknowledging effects of uncertainties to be accounted for in future research, were also included.

The response was excellent with 87 oral presentations (11 of which were invited keynote speakers) and 24 poster presentations during the conference. A subset of these presentations was submitted to this topical issue and 22 manuscripts have been published by Ocean Dynamics.

We estimate and study the evolution of the dominant dimensionality of
dynamical systems with uncertainty governed by stochastic partial differential
equations, within the context of dynamically orthogonal (DO) field equations.
Transient nonlinear dynamics, irregular data and non-stationary statistics are
typical in a large range of applications such as oceanic and atmospheric flow
estimation. To efficiently quantify uncertainties in such systems, it is
essential to vary the dimensionality of the stochastic subspace with time. An
objective here is to provide criteria to do so, working directly with the
original equations of the dynamical system under study and its DO
representation. We first analyze the scaling of the computational cost of
these DO equations with the stochastic dimensionality and show that unlike
many other stochastic methods the DO equations do not suffer from the curse of
dimensionality. Subsequently, we present the new adaptive criteria for the
variation of the stochastic dimensionality based on instantaneous i) stability
arguments and ii) Bayesian data updates. We then illustrate the capabilities
of the derived criteria to resolve the transient dynamics of two 2D stochastic
fluid flows, specifically a double-gyre wind-driven circulation and a
lid-driven cavity flow in a basin. In these two applications, we focus on the
growth of uncertainty due to internal instabilities in deterministic flows. We
consider a range of flow conditions described by varied Reynolds numbers and
we study and compare the evolution of the uncertainty estimates under these
varied conditions.

Wang, D., P.F.J. Lermusiaux, P.J. Haley, D. Eickstedt, W.G. Leslie and H. Schmidt, 2009. *Acoustically Focused Adaptive Sampling and On-board Routing for Marine Rapid Environmental Assessment.* Special issue of Journal of Marine Systems on "Coastal processes: challenges for monitoring and prediction", Drs. J.W. Book, Prof. M. Orlic and Michel Rixen (Guest Eds), 78, S393-S407, doi: 10.1016/j.jmarsys.2009.01.037.

In this work we derive an exact, closed set of evolution equations for general continuous stochastic fields
described by a Stochastic Partial Differential Equation (SPDE). By hypothesizing a decomposition of the
solution field into a mean and stochastic dynamical component, we derive a system of field equations
consisting of a Partial Differential Equation (PDE) for the mean field, a family of PDEs for the orthonormal
basis that describe the stochastic subspace where the stochasticity `lives’ as well as a system of Stochastic
Differential Equations that defines how the stochasticity evolves in the time varying stochastic subspace.
These new evolution equations are derived directly from the original SPDE, using nothing more than
a dynamically orthogonal condition on the representation of the solution. If additional restrictions are
assumed on the form of the representation, we recover both the Proper Orthogonal Decomposition
equations and the generalized Polynomial Chaos equations. We apply this novel methodology to two
cases of two-dimensional viscous fluid flows described by the NavierStokes equations and we compare
our results with Monte Carlo simulations.

The goal of adaptive sampling in the ocean is to predict
the types and locations of additional ocean measurements that
would be most useful to collect. Quantitatively, what is most useful
is defined by an objective function and the goal is then to optimize
this objective under the constraints of the available observing network.
Examples of objectives are better oceanic understanding, to
improve forecast quality, or to sample regions of high interest. This
work provides a new path-planning scheme for the adaptive sampling
problem. We define the path-planning problem in terms of
an optimization framework and propose a method based on mixed
integer linear programming (MILP). The mathematical goal is to
find the vehicle path that maximizes the line integral of the uncertainty
of field estimates along this path. Sampling this path can improve
the accuracy of the field estimates the most. While achieving
this objective, several constraints must be satisfied and are implemented.
They relate to vehicle motion, intervehicle coordination,
communication, collision avoidance, etc. The MILP formulation is
quite powerful to handle different problem constraints and flexible
enough to allow easy extensions of the problem. The formulation
covers single- and multiple-vehicle cases as well as singleand
multiple-day formulations. The need for a multiple-day formulation
arises when the ocean sampling mission is optimized for
several days ahead. We first introduce the details of the formulation,
then elaborate on the objective function and constraints, and
finally, present a varied set of examples to illustrate the applicability
of the proposed method.

Lermusiaux, P.F.J, 2007. *Adaptive Modeling, Adaptive Data Assimilation and Adaptive Sampling.* Refereed invited manuscript. Special issue on "Mathematical Issues and Challenges in Data Assimilation for Geophysical Systems: Interdisciplinary Perspectives". C.K.R.T. Jones and K. Ide, Eds. Physica D, Vol 230, 172-196, doi:
10.1016/j.physd.2007.02.014.

The problem of how to optimally deploy a suite of sensors to estimate the oceanographic
environment is addressed. An optimal way to estimate (nowcast) and predict (forecast)
the ocean environment is to assimilate measurements from dynamic and uncertain regions
into a dynamical ocean model. In order to determine the sensor deployment strategy
that optimally samples the regions of uncertainty, a Genetic Algorithm (GA) approach
is presented. The scalar cost function is defined as a weighted combination of a sensor
suite’s sampling of the ocean variability, ocean dynamics, transmission loss sensitivity,
modeled temperature uncertainty (and others). The benefit of the GA approach is that the
user can determine “optimal” via a weighting of constituent cost functions, which can
include ocean dynamics, acoustics, cost, time, etc. A numerical example with three gliders,
two powered AUVs, and three moorings is presented to illustrate the optimization
approach in the complex shelfbreak region south of New England.

Variabilities in the coastal ocean environment span
a wide range of spatial and temporal scales. From an acoustic
viewpoint, the limited oceanographic measurements and today’s
ocean modeling capabilities can’t always provide oceanic-acoustic
predictions in sufficient detail and with enough accuracy. Adaptive
Rapid Environmental Assessment (AREA) is a new adaptive sampling
concept being developed in connection with the emergence
of the Autonomous Ocean Sampling Network (AOSN) technology.
By adaptively and optimally deploying in-situ measurement
resources and assimilating these data in coupled nested ocean
and acoustic models, AREA can dramatically improve the ocean
estimation that matters for acoustic predictions and so be
essential for such predictions. These concepts are outlined and
preliminary methods are developed and illustrated based on
the Focused Acoustic Forecasting-05 (FAF05) exercise. During
FAF05, AREA simulations were run in real-time and engineering
tests carried out, within the context of an at-sea experiment
with Autonomous Underwater Vehicles (AUV) in the northern
Tyrrhenian sea, on the eastern side of the Corsican channel.

Yilmaz, N.K., C. Evangelinos, N.M. Patrikalakis, P.F.J. Lermusiaux, P.J. Haley, W.G. Leslie, A.R. Robinson, D. Wang and H. Schmidt, 2006a. *Path Planning Methods for Adaptive Sampling of Environmental and Acoustical Ocean Fields*, Oceans 2006, 6pp, Boston, MA, 18-21 Sept. 2006, doi: 10.1109/OCEANS.2006.306841.

THIS REPORT summarizes goals,
activities, and recommendations of a
workshop on data assimilation held in
Williamsburg, Virginia on September
9-11, 2003, and sponsored by the U.S.
Office of Naval Research (ONR) and National
Science Foundation (NSF). The
overall goal of the workshop was to synthesize
research directions for ocean data
assimilation (DA) and outline efforts
required during the next 10 years and
beyond to evolve DA into an integral and
sustained component of global, regional,
and coastal ocean science and observing
and prediction systems. The workshop
built on the success of recent and existing
DA activities such as those sponsored
by the National Oceanographic Partnership
Program (NOPP) and NSF-Information
Technology Research (NSF-ITR).
DA is a quantitative approach to optimally
combine models and observations.
The combination is usually consistent
with model and data uncertainties, which
need to be represented. Ocean DA can
extract maximum knowledge from the
sparse and expensive measurements of
the highly variable ocean dynamics. The
ultimate goal is to better understand and
predict these dynamics on multiple spatial
and temporal scales, including interactions
with other components of the
climate system. There are many applications
that involve DA or build on its results,
including: coastal, regional, seasonal,
and inter-annual ocean and climate
dynamics; carbon and biogeochemical
cycles; ecosystem dynamics; ocean engineering;
observing-system design; coastal
management; fisheries; pollution control;
naval operations; and defense and security.
These applications have different requirements
that lead to variations in the
DA schemes utilized. For literature on
DA, we refer to Ghil and Malanotte-Rizzoli
(1991), the National Research Council
(1991), Bennett (1992), Malanotte-
Rizzoli (1996), Wunsch (1996), Robinson
et al. (1998), Robinson and Lermusiaux
(2002), and Kalnay (2003). We also refer
to the U.S. Global Ocean Data Assimilation
Experiment (GODAE) workshop on
Global Ocean Data Assimilation: Prospects
and Strategies (Rienecker et al., 2001);
U.S. National Oceanic and Atmospheric
Administration-Office of Global Programs
(NOAA-OGP) workshop on Coupled
Data Assimilation (Rienecker, 2003);
and, NOAA-NASA-NSF workshop on
Ongoing Analysis of the Climate System
(Arkin et al., 2003).

Scientific computations for the quantification, estimation and prediction of uncertainties for ocean dynamics are developed
and exemplified. Primary characteristics of ocean data, models and uncertainties are reviewed and quantitative data
assimilation concepts defined. Challenges involved in realistic data-driven simulations of uncertainties for four-dimensional
interdisciplinary ocean processes are emphasized. Equations governing uncertainties in the Bayesian probabilistic
sense are summarized. Stochastic forcing formulations are introduced and a new stochastic-deterministic ocean model
is presented. The computational methodology and numerical system, Error Subspace Statistical Estimation, that is used
for the efficient estimation and prediction of oceanic uncertainties based on these equations is then outlined. Capabilities
of the ESSE system are illustrated in three data-assimilative applications: estimation of uncertainties for physical-biogeochemical
fields, transfers of ocean physics uncertainties to acoustics, and real-time stochastic ensemble predictions with
assimilation of a wide range of data types. Relationships with other modern uncertainty quantification schemes and promising
research directions are discussed.

Lermusiaux, P.F.J., C.-S. Chiu, G.G. Gawarkiewicz, P. Abbot, A.R. Robinson, R.N. Miller, P.J. Haley, W.G. Leslie, S.J. Majumdar, A. Pang and F. Lekien, 2006. *Quantifying Uncertainties in Ocean Predictions.* Refereed invited manuscript. Oceanography, Special issue on "Advances in Computational Oceanography", T. Paluszkiewicz and S. Harper (Office of Naval Research), Eds., 19, 1, 92-105, doi: 10.5670/oceanog.2006.93.

Lermusiaux, P.F.J, C. Evangelinos, R. Tian, P.J. Haley, J.J. McCarthy, N.M. Patrikalakis, A.R. Robinson and H. Schmidt, 2004. *Adaptive Coupled Physical and Biogeochemical Ocean Predictions: A Conceptual Basis.* Refereed invited manuscript, F. Darema (Ed.), Lecture Notes in Computer Science, 3038, 685-692.

The International Lie`ge Colloquium on Ocean
Dynamics is organized annually. The topic differs
from year to year in an attempt to address, as much
as possible, recent problems and incentive new subjects
in oceanography.
Assembling a group of active and eminent scientists
from various countries and often different disciplines,
the Colloquia provide a forum for discussion
and foster a mutually beneficial exchange of information
opening on to a survey of recent discoveries,
essential mechanisms, impelling question marks and
valuable recommendations for future research.
The objective of the 2001 Colloquium was to
evaluate the progress of data assimilation methods in
marine science and, in particular, in coupled hydrodynamic,
ecological and bio-geo-chemical models of
the ocean.
The past decades have seen important advances
in the understanding and modelling of key processes
of the ocean circulation and bio-geo-chemical
cycles. The increasing capabilities of data and
models, and their combination, are allowing the
study of multidisciplinary interactions that occur
dynamically, in multiple ways, on multiscales and
with feedbacks.
The capacity of dynamical models to simulate interdisciplinary
ocean processes over specific space-
time windows and thus forecast their evolution over
predictable time scales is also conditioned upon the
availability of relevant observations to: initialise and
continually update the physical and bio-geo-chemical
sectors of the ocean state; provide relevant atmospheric
and boundary forcing; calibrate the parameterizations
of sub-grid scale processes, growth rates and
reaction rates; construct interdisciplinary and multiscale
correlation and feature models; identify and
estimate the main sources of errors in the models;
control or correct for mis-represented or neglected
processes.
The access to multivariate data sets requires the
implementation, exploitation and management of dedicated
ocean observing and prediction systems. However,
the available data are often limited and, for
instance, seldom in a form to be directly compatible
or directly inserted into the numerical models. To relate
the data to the ocean state on all scales and regions that
matter, evolving three-dimensional and multivariate
(measurement) models are becoming important.
Equally significant is the reduction of observational
requirements by design of sampling strategies via
Observation System Simulation Experiments and
adaptive sampling.
Data assimilation is a quantitative approach to
extract adequate information content from the data
and to improve the consistency between data sets and
model estimates. It is also a methodology to dynamically
interpolate between data scattered in space and
time, allowing comprehensive interpretation of multivariate
observations.
In general, the goals of data assimilation are to:
control the growth of predictability errors; correct
dynamical deficiencies; estimate model parameters,
including the forcings, initial and boundary conditions;
characterise key processes by analysis of four-
0924-7963/03/$ – see front matter D 2003 Elsevier Science B.V. All rights reserved.
doi:10.1016/S0924-7963(03)00027-7
www.elsevier.com/locate/jmarsys
The use of data assimilation in coupled hydrodynamic, ecological and
bio-geo-chemical models of the ocean
Journal of Marine Systems 40-41 (2003) 1-3
dimensional fields and their statistics (balances of
terms, etc.); carry out advanced sensitivity studies
and Observation System Simulation Experiments,
and conduct efficient operations, management and
monitoring.
The theoretical framework of data assimilation
for marine sciences is now relatively well established,
routed in control theory, estimation theory or inverse
techniques, from variational to sequential approaches.
Ongoing research efforts of special importance for
interdisciplinary applications include the: stochastic
representation of processes and determination of
model and data errors; treatment of (open) boundary
conditions and strong nonlinearities; space-time,
multivariate extrapolation of limited and noisy data
and determination of measurement models; demonstration
that bio-geo-chemical models are valid
enough and of adequate structures for their deficiencies
to be controlled by data assimilation; and finally,
ability to provide accurate estimates of fields, parameters,
variabilities and errors, with large and complex
dynamical models and data sets.
Operationally, major engineering and computational
challenges for the coming years include the:
development of theoretically sound methods into
useful, practical and reliable techniques at affordable
costs; implementation of scalable, seamless and automated
systems linking observing systems, numerical
models and assimilation schemes; adequate mix of
integrated and distributed (Web-based) networks; construction
of user-friendly architectures and establishment
of standards for the description of data and
software (metadata) for efficient communication, dissemination
and management.
In addition to addressing the above items, the 33rd
Lie`ge Colloquium has offered the opportunity to:
- review the status and current progress of data
assimilation methodologies utilised in the physical,
acoustical, optical and bio-geo-chemical
scientific communities;
- demonstrate the potentials of data assimilation
systems developed for coupled physical/ecosystem
models, from scientific to management inquiries;
- examine the impact of data assimilation and
inverse modelling in improving model parameterisations;
- discuss the observability and controllability properties
of, and identify the missing gaps in current
observing and prediction systems; and
exchange the results of and the learnings from preoperational
marine exercises.
The presentations given during the Colloquium
lead to discussions on a series of topics organized
within the following sections: (1) Interdisciplinary
research progress and issues: data, models, data
assimilation criteria. (2) Observations for interdisciplinary
data assimilation. (3) Advanced fields estimation
for interdisciplinary systems. (4) Estimation of
interdisciplinary parameters and model structures. (5)
Assimilation methodologies for physical and interdisciplinary
systems. (6) Toward operational interdisciplinary
oceanography and data assimilation. A subset
of these presentations is reported in the present
Special Issue.
As was pointed out during the Colloquium, coupled
biological-physical data assimilation is in its infancy
and much can be accomplished now by the immediate
application of existing methods. Data assimilation
intimately links dynamical models and observations,
and it can play a critical role in the important area of
fundamental biological oceanographic dynamical
model development and validation over a hierarchy
of complexities. Since coupled assimilation for coupled
processes is challenging and can be complicated, care
must be exercised in understanding, modeling and
controlling errors and in performing sensitivity analyses
to establish the robustness of results. Compatible
interdisciplinary data sets are essential and data assimilation
should iteratively define data impact and data
requirements.
Based on the results presented during the Colloquium,
data assimilation is expected to enable future
marine technologies and naval operations otherwise
impossible or not feasible. Interdisciplinary predictability
research, multiscale in both space and time, is
required. State and parameter estimation via data
assimilation is central to the successful establishment
of advanced interdisciplinary ocean observing and
prediction systems which, functioning in real time,
will contribute to novel and efficient capabilities to
manage, and to operate in our oceans.
The Scientific Committee and the participants to
the 33rd Lie`ge Colloquium wish to express their
2 Preface
gratitude to the Ministe`re de l’Enseignement Supe’rieur
et de la Recherche Scientifique de la Communaute
- Francaise de Belgique, the Fonds National de
la Recherche Scientifique de Belgique (F.N.R.S.,
Belgium), the Ministe`re de l’Emploi et de la Formation
du Gouvernement Wallon, the University of
Lie`ge, the Commission of European Union, the
Scientific Committee on Oceanographic Research
(SCOR), the International Oceanographic Commission
of the UNESCO, the US Office of Naval
Research, the National Science Foundation (NSF,
USA) and the International Association for the
Physical Sciences of the Ocean (IAPSO) for their
most valuable support.

The estimation of oceanic environmental and acoustical fields is considered as a single coupled data assimilation problem. The four-dimensional data assimilation methodology employed is Error Subspace Statistical Estimation. Environmental fields and their dominant uncertainties are predicted by an ocean dynamical model and transferred to acoustical fields and uncertainties by an acoustic propagation model. The resulting coupled dominant uncertainties define the error subspace. The available physical and acoustical data are then assimilated into the predicted fields in accord with the error subspace and all data uncertainties. The criterion for data assimilation is presently to correct the predicted fields such that the total error variance in the error subspace is minimized. The approach is exemplified for the New England continental shelfbreak region, using data collected during the 1996 Shelfbreak Primer Experiment. The methodology is discussed, computational issues are outlined and the assimilation of model-simulated acoustical data is carried out. Results are encouraging and provide some insights into the dominant variability and uncertainty properties of acoustical fields.

The efficient interdisciplinary 4D data assimilation with nonlinear models via Error Subspace Statistical Estimation (ESSE) is reviewed and exemplified. ESSE is based on evolving an error subspace, of variable size, that spans and tracks the scales and processes where the dominant errors occur. A specific focus here is the use of ESSE in interdisciplinary smoothing which allows the correction of past estimates based on future data, dynamics and model errors. ESSE is useful for a wide range of purposes which are illustrated by three investigations: (i) smoothing estimation of physical ocean fields in the Eastern Mediterranean, (ii) coupled physical-acoustical data assimilation in the Middle Atlantic Bight shelfbreak, and (iii) coupled physical-biological smoothing and dynamics in Massachusetts Bay.

Data assimilation is a modern methodology of relating natural data and dynamical
models. The general dynamics of a model is combined or melded with a set of observations.
All dynamical models are to some extent approximate, and all data sets are
finite and to some extent limited by error bounds. The purpose of data assimilation
is to provide estimates of nature which are better estimates than can be obtained by
using only the observational data or the dynamical model. There are a number of
specific approaches to data assimilation which are suitable for estimation of the state
of nature, including natural parameters, and for evaluation of the dynamical approximations.
Progress is accelerating in understanding the dynamics of real ocean biological-
physical interactive processes. Although most biophysical processes in the sea await
discovery, new techniques and novel interdisciplinary studies are evolving ocean science
to a new level of realism. Generally, understanding proceeds from a quantitative
description of four-dimensional structures and events, through the identification of
specific dynamics, to the formulation of simple generalizations. The emergence of
realistic interdisciplinary four-dimensional data assimilative ocean models and systems
is contributing significantly and increasingly to this progress.

The effects of a priori parameters on the error subspace estimation and mapping methodology introduced by
P. F. J. Lermusiaux et al. is investigated. The approach is three-dimensional, multivariate, and multiscale. The
sensitivities of the subspace and a posteriori fields to the size of the subspace, scales considered, and nonlinearities
in the dynamical adjustments are studied. Applications focus on the mesoscale to subbasin-scale physics in the
northwestern Levantine Sea during 10 February-15 March and 19 March-16 April 1995. Forecasts generated
from various analyzed fields are compared to in situ and satellite data. The sensitivities to size show that the
truncation to a subspace is efficient. The use of criteria to determine adequate sizes is emphasized and a backof-
the-envelope rule is outlined. The sensitivities to scales confirm that, for a given region, smaller scales usually
require larger subspaces because of spectral redness. However, synoptic conditions are also shown to strongly
influence the ordering of scales. The sensitivities to the dynamical adjustment reveal that nonlinearities can
modify the variability decomposition, especially the dominant eigenvectors, and that changes are largest for the
features and regions with high shears. Based on the estimated variability variance fields, eigenvalue spectra,
multivariate eigenvectors and (cross)-covariance functions, dominant dynamical balances and the spatial distribution
of hydrographic and velocity characteristic scales are obtained for primary regional features. In particular,
the Ierapetra Eddy is found to be close to gradient-wind balance and coastal-trapped waves are anticipated to
occur along the northern escarpment of the basin.

An interdisciplinary team of scientists is collaborating to enhance the understanding of the uncertainty in the ocean environment, including the sea bottom, and characterize its impact on tactical system performance. To accomplish these goals quantitatively an end-to-end system approach is necessary. The conceptual basis of this approach and the framework of the end-to-end system, including its components, is the subject of this presentation. Specifically, we present a generic approach to characterize variabilities and uncertainties arising from regional scales and processes, construct uncertainty models for a generic sonar system, and transfer uncertainties from the acoustic environment to the sonar and its signal processing. Illustrative examples are presented to highlight recent progress toward the development of the methodology and components of the system.

Data assimilation is a novel, versatile methodology
for estimating oceanic variables. The estimation of
a quantity of interest via data assimilation involves
the combination of observational data with the underlying
dynamical principles governing the system
under observation. The melding of data and dynamics
is a powerful methodology which makes possible
efRcient, accurate, and realistic estimations otherwise
not feasible. It is providing rapid advances in
important aspects of both basic ocean science and
applied marine technology and operations.
The following sections introduce concepts, describe
purposes, present applications to regional dynamics
and forecasting, overview formalism and
methods, and provide a selected range of examples.

Lermusiaux, P.F.J., 2001. *Evolving the subspace of the three-dimensional multiscale ocean variability: Massachusetts Bay.* Journal of Marine Systems, Special issue on "Three-dimensional ocean circulation: Lagrangian measurements and diagnostic analyses", 29/1-4, 385-422, doi: 10.1016/S0924-7963(01)00025-2.

A basis is outlined for the first-guess spatial mapping of three-dimensional multivariate and multiscale
geophysical fields and their dominant errors. The a priori error statistics are characterized by covariance matrices
and the mapping obtained by solving a minimum-error-variance estimation problem. The size of the problem is
reduced efficiently by focusing on the error subspace, here the dominant eigendecomposition of the a priori error
covariance. The first estimate of this a priori error subspace is constructed in two parts. For the “observed” portions
of the subspace, the covariance of the a priori missing variability is directly specified and eigendecomposed.
For the “non-observed” portions, an ensemble of adjustment dynamical integrations is utilized, building the nonobserved
covariances in statistical accord with the observed ones. This error subspace construction is exemplified
and studied in a Middle Atlantic Bight simulation and in the eastern Mediterranean. Its use allows an accurate,
global, multiscale and multivariate, three-dimensional analysis of primitive-equation fields and their errors, in real
time. The a posteriori error covariance is computed and indicates complex data-variability influences. The error
and variability subspaces obtained can also confirm or reveal the features of dominant variability, such as the
Ierapetra Eddy in the Levantine basin.

Identical twin experiments are utilized to assess and exemplify the capabilities of error subspace statistical
estimation (ESSE). The experiments consists of nonlinear, primitive equation-based, idealized Middle Atlantic
Bight shelfbreak front simulations. Qualitative and quantitative comparisons with an optimal interpolation (OI)
scheme are made. Essential components of ESSE are illustrated. The evolution of the error subspace, in agreement
with the initial conditions, dynamics, and data properties, is analyzed. The three-dimensional multivariate minimum
variance melding in the error subspace is compared to the OI melding. Several advantages and properties
of ESSE are discussed and evaluated. The continuous singular value decomposition of the nonlinearly evolving
variations of variability and the possibilities of ESSE for dominant process analysis are illustrated and emphasized.

A rational approach is used to identify efficient schemes for data assimilation in nonlinear ocean-atmosphere
models. The conditional mean, a minimum of several cost functionals, is chosen for an optimal estimate. After
stating the present goals and describing some of the existing schemes, the constraints and issues particular to
ocean-atmosphere data assimilation are emphasized. An approximation to the optimal criterion satisfying the
goals and addressing the issues is obtained using heuristic characteristics of geophysical measurements and
models. This leads to the notion of an evolving error subspace, of variable size, that spans and tracks the scales
and processes where the dominant errors occur. The concept of error subspace statistical estimation (ESSE) is
defined. In the present minimum error variance approach, the suboptimal criterion is based on a continued and
energetically optimal reduction of the dimension of error covariance matrices. The evolving error subspace is
characterized by error singular vectors and values, or in other words, the error principal components and
coefficients.
Schemes for filtering and smoothing via ESSE are derived. The data-forecast melding minimizes variance in
the error subspace. Nonlinear Monte Carlo forecasts integrate the error subspace in time. The smoothing is
based on a statistical approximation approach. Comparisons with existing filtering and smoothing procedures
are made. The theoretical and practical advantages of ESSE are discussed. The concepts introduced by the
subspace approach are as useful as the practical benefits. The formalism forms a theoretical basis for the
intercomparison of reduced dimension assimilation methods and for the validation of specific assumptions for
tailored applications. The subspace approach is useful for a wide range of purposes, including nonlinear field
and error forecasting, predictability and stability studies, objective analyses, data-driven simulations, model
improvements, adaptive sampling, and parameter estimation.