Supervised Classification of Sound Speed Profiles via Dictionary Learning

Jhon A. Castro-Correa aDepartment of Electrical and Computer Engineering, University of Delaware, Newark, Delaware

Search for other papers by Jhon A. Castro-Correa in
Current site
Google Scholar
PubMed
Close
,
Stephanie A. Arnett bDepartment of Physics and Astronomy, Brigham Young University, Provo, Utah

Search for other papers by Stephanie A. Arnett in
Current site
Google Scholar
PubMed
Close
,
Tracianne B. Neilsen bDepartment of Physics and Astronomy, Brigham Young University, Provo, Utah

Search for other papers by Tracianne B. Neilsen in
Current site
Google Scholar
PubMed
Close
,
Lin Wan aDepartment of Electrical and Computer Engineering, University of Delaware, Newark, Delaware

Search for other papers by Lin Wan in
Current site
Google Scholar
PubMed
Close
, and
Mohsen Badiey aDepartment of Electrical and Computer Engineering, University of Delaware, Newark, Delaware

Search for other papers by Mohsen Badiey in
Current site
Google Scholar
PubMed
Close
Free access

Abstract

The presence of internal waves (IWs) in the ocean alters the isotropic properties of sound speed profiles (SSPs) in the water column. Changes in the SSPs affect underwater acoustics since most of the energy is dissipated into the seabed due to the downward refraction of sound waves. In consequence, variations in the SSP must be considered when modeling acoustic propagation in the ocean. Empirical orthogonal functions (EOFs) are regularly employed to model and represent SSPs using a linear combination of basis functions that capture the sound speed variability. A different approach is to use dictionary learning to obtain a learned dictionary (LD) that generates a nonorthogonal set of basis functions (atoms) that generate a better sparse representation. In this paper, the performance of EOFs and LDs are evaluated for sparse representation of SSPs affected by the passing of IWs. In addition, an LD-based supervised framework is presented for SSP classification and is compared with classical learning models. The algorithms presented in this work are trained and tested on data collected from the Shallow Water Experiment 2006. Results show that LDs yield lower reconstruction error than EOFs when using the same number of bases. In addition, overcomplete LDs are demonstrated to be a robust method to classify SSPs during low, medium, and high IW activity, reporting accuracy that is comparable to and sometimes higher than that of standard supervised classification methods.

© 2023 American Meteorological Society. For information regarding reuse of this content and general copyright information, consult the AMS Copyright Policy (www.ametsoc.org/PUBSReuseLicenses).

Corresponding author: Jhon A. Castro-Correa, jcastro@udel.edu

Abstract

The presence of internal waves (IWs) in the ocean alters the isotropic properties of sound speed profiles (SSPs) in the water column. Changes in the SSPs affect underwater acoustics since most of the energy is dissipated into the seabed due to the downward refraction of sound waves. In consequence, variations in the SSP must be considered when modeling acoustic propagation in the ocean. Empirical orthogonal functions (EOFs) are regularly employed to model and represent SSPs using a linear combination of basis functions that capture the sound speed variability. A different approach is to use dictionary learning to obtain a learned dictionary (LD) that generates a nonorthogonal set of basis functions (atoms) that generate a better sparse representation. In this paper, the performance of EOFs and LDs are evaluated for sparse representation of SSPs affected by the passing of IWs. In addition, an LD-based supervised framework is presented for SSP classification and is compared with classical learning models. The algorithms presented in this work are trained and tested on data collected from the Shallow Water Experiment 2006. Results show that LDs yield lower reconstruction error than EOFs when using the same number of bases. In addition, overcomplete LDs are demonstrated to be a robust method to classify SSPs during low, medium, and high IW activity, reporting accuracy that is comparable to and sometimes higher than that of standard supervised classification methods.

© 2023 American Meteorological Society. For information regarding reuse of this content and general copyright information, consult the AMS Copyright Policy (www.ametsoc.org/PUBSReuseLicenses).

Corresponding author: Jhon A. Castro-Correa, jcastro@udel.edu

1. Introduction

Internal waves (IWs) are caused by differences in the temperature of water in the ocean. The presence of such anomalies in the seawater creates time-dependent and spatial variations in the sound speed profiles (SSPs) and affects underwater sound propagation due to the downward refraction of sound waves (Katsnelson et al. 2021; Rouseff 2001). Usually, sparse representations of SSPs via empirical orthogonal functions (EOFs) are used to support inversion algorithms for sound speed (North 1984). However, effective sparse representation of SSPs can be compromised due to high perturbations in the water column. Recently, Bianco and Gerstoft (2017) have shown that dictionary learning (DL), an unsupervised machine learning method, is better suited to sparsely represent SSPs. In this paper, DL and EOFs are used to model and classify measured SSPs affected by IWs during the Shallow Water Experiment 2006 (SW06).

Internal waves can be thought of as four-dimensional phenomena because of their effects in both 3D spatial (x, y, z) and temporal (t) domains. Characterization of the behavior and statistical properties of IWs have been carried out using 3D mapping techniques (Badiey et al. 2013, 2016), resulting in the general understanding of the regimens in the propagation of IWs. Several studies have shown that internal waves create significant variability in the speed of sound, affecting how sound propagates through the ocean, due to the fluctuations in the acoustic modal behavior (Flatté and Tappert 1975; Rouseff 2001; Helfrich and Melville 2006). The variation due to the passing of IW packets affects the propagation and reception of acoustic signals underwater because of drastic changes in the acoustic channel (Huang et al. 2008).

EOF analysis has been commonly used to represent SSPs as a linear combination of few orthogonal basis functions that describe the statistics of the sound speed uncertainty (Xu and Schmidt 2006; Abiva et al. 2019). Those resulting sparse representations are employed to aid inversion procedures (Huang et al. 2008). However, strong variations in the SSPs occurring because the passing of IWs yield a dramatic decay in the reconstruction accuracy of SSPs using EOFs (Sun and Zhao 2020; Roundy 2015). As a result, different approaches such as 3D mapping (Badiey et al. 2013) or learning methods (Jain and Ali 2006; Bianco and Gerstoft 2017; Sun and Zhao 2020) have been studied for modeling and reconstruction of SSPs.

Dictionary learning, an unsupervised learning method, aims to find a set of nonorthogonal basis functions (referred to as atoms) that can sparsely reconstruct signals (Zhang et al. 2015). The DL framework has been extensively applied to dimensionality reduction (Tošić and Frossard 2011), pattern recognition (Wright et al. 2010) and sparse representation modeling (Rubinstein et al. 2010). Recent studies have been conducted to sparsely represent SSPs using learned dictionaries (LDs) (Kuo and Kiang 2020). Bianco and Gerstoft (2017) showed that LDs are well-suited to generate sparse representations of SSPs using few basis functions. Sun and Zhao (2020) tested the effectiveness of LDs using HYCOM data and supported the results found by Bianco and Gerstoft (2017), concluding that nonorthogonal atoms allow for more flexible dictionaries and produce better sparse representations of SSPs.

Because of the relaxation of the orthogonal requirements and the possibility of generating optimal sparse representations, LDs have also been applied to clustering (Sprechmann and Sapiro 2010) and classification tasks (Tang et al. 2019; Suo et al. 2014). In this approach, specific dictionaries are trained to retain most of the meaningful information of each class. Then, testing data are classified by selecting the dictionary yielding the sparse representation that generates the lowest error (Ramirez et al. 2010; Zhao et al. 2018). Here, classification via dictionary learning is extended to label data containing SSPs affected by low, medium, and high IW activity collected during the SW06 experiment.

In this paper, there are two main contributions. First, a comparison of the ability to sparsely model SSPs with few basis functions is made between EOFs, complete LDs, and overcomplete LDs. Second, a DL-based SSPs classification setting is proposed and assessed via LD. The proposed framework is compared with standard classification algorithms such as support vector machine (SVM) and k-nearest neighbors (KNN) classifier. The dictionary atoms are calculated with online dictionary learning, a stochastic gradient-descent approach introduced by Mairal et al. (2009), while sparse coding is performed using the orthogonal matching pursuit algorithm and least absolute shrinkage and selection operator (lasso) convex optimization. This paper is structured as follows: In section 2 preliminary concepts and notations are presented. Section 3 introduces the data collected during the SW06 experiment used in this paper. Both EOF analysis and dictionary learning frameworks are introduced in section 4, and results for modeling and classification of SSPs are shown in section 5, followed by the conclusions in section 6.

2. Preliminaries and notation

In this paper matrices are denoted in sans serif (except for greek letters) uppercase boldface type, vectors are in lowercase boldface type, and scalars are in lowercase and uppercase italic. The p-norm of a vector or matrix is represented as ǁ⋅ǁp, with 1 ≤ p ≤ ∞. For any X, if p = 2, ǁ⋅ǁp takes the name of Frobenius norm ǁ⋅ǁF. Notice that these p-norms are entrywise rather than the induced norms.

Any square diagonalizable matrix X ∈ Ρn×n can be factorized into its canonical form via eigendecomposition satisfying the linear equation X = QΛQT, where Q = [q1, …, qn] ∈ Ρn×n is the matrix containing eigenvectors and Λ = diag(λ1, …, λn) ∈ Ρn×n is a diagonal matrix with the set of eigenvalues. Similarly, any rectangular diagonizable matrix X ∈ Ρm×n accepts the factorization X = UΣVT via singular value decomposition (SVD) (Stewart 1993), where U ∈ Ρm×m and V ∈ Ρn×n are the matrices of left and right singular vectors, respectively, and Σ ∈ Ρm×n contains the singular values. For the cases XXT and XTX, the diagonal matrix of the nonzero eigenvalues is related to the singular values of X, such that Λ = Σ2 = ΣΣT = ΣTΣ.

3. Experimental data

The raw data introduced in this work were collected during the SW06 experiment performed off the coast of New Jersey from mid-July to mid-September in 2006 (Tang et al. 2007; Newhall et al. 2007). During the experiment, 62 acoustic and oceanographic moorings were deployed in a “T” geometry as sketched in Fig. 1. This T mooring conformation measured data on an almost constant bathymetry of 80 m in the along-shelf direction and a bathymetry across shelf starting at 600 m going shoreward to 60-m depth. The intersection of the two paths in the T was populated with a cluster of 16 moorings to measure the 3D environment. Most of the environmental moorings in the area consisted of temperature, conductivity, and pressure sensors that measured the physical oceanography in the water column.

Fig. 1.
Fig. 1.

Some of the SW06 moorings along shelf and across shelf conforming to a T geometry. SSPs used in this work are derived from data collected at SW30 station, which is marked with a white star.

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

This paper uses data from mooring SW30, deployed at 39°1.501′N, 73°4.007′W, to study the time-evolving SSPs via dictionary learning. The SW30 station was part of the 16-mooring cluster located at the intersection of the T geometry deployed in the SW06 experiment. The location of mooring SW30 inside the cluster is marked with a white star in Fig. 1. The SW30 station had 11 unevenly spaced sensors collecting conductivity and temperature profiles from 14 to 83.3 m in a water column with seafloor at 86-m depth. For this study, temperature, conductivity, and pressure data were extracted from 0000:00 UTC 1 August to 1600:00 UTC 5 September 2006 at the SW30 location.

During the SW06 experiment, IW activities were reported as initiating at the shelf break and propagating toward the shore after 17 August 2006 (Badiey et al. 2013). The transition of internal waves over the area caused highly anisotropic SSPs such as the ones depicted in Fig. 2a, where temperature profiles exhibited significant variations across the water column and caused notable changes in both the acoustic channel and sound transmission.

Fig. 2.
Fig. 2.

IW event spotted from 2100:00 UTC 17 to 1000:00 UTC 18 Aug 2006. (a) Temporal evolution of temperature profiles at mooring SW30. (b) The square of the buoyancy frequency N2, in terms of depth, divided into four sections corresponding to different regimens. (c) Standard deviation (shading) and mean of the SSPs presented in (a). (d) Individual SSP samples at each geotime tgi, with colors matching the vertical dashed lines in (a). (e) Buoyancy frequency N2 for each of the four regimens shown in (b).

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

The study of IWs is a difficult task because of the spatial resolution of the measurements in experiment areas (Badiey et al. 2016). In this paper, only temporal IW anomalies are used for the analyses and applications presented in subsequent sections. As an example, the temporal displacement of an IW event spotted from 2100:00 UTC 17 to 1000:00 UTC 18 August 2006 at the SW30 location is presented in Fig. 2. The temperature variability across the water column produced by the passing of IWs provokes changes in the acoustic duct and degrades the acoustic propagation underwater. In Fig. 2a, the beginning of each stage (approaching, onset, propagation, and tail) of the IW event is marked with a dashed line in Fig. 2a and labeled with a geotime tgi (i.e., tg0, tg1, tg2, and tg3). These abrupt changes in temperature alter the isotropic properties of the SSPs and produce drastic variations in the SSPs as shown in Figs. 2c and 2d. Figure 2c shows the mean μyi and standard deviation σyi values of the SSPs over the entire time window in Fig. 2a, whereas Fig. 2d depicts individual SSPs at geotimes tgi, with colors matching the vertical lines in Fig. 2a.

An additional consequence of the IW passing is the instability of vertical displacements that affects acoustic propagation. The periodicity at which a vertically displaced small volume of water oscillates is measured by the buoyancy frequency N (s−1). The oscillations in the water column are expressed as the square of the buoyancy frequency N2 to obtain real values, as shown in Fig. 2b. The time interval has been divided into four sections, each categorized by the regimens 1–4 describing the IW behavior. These regimens are 1) the approaching, 2) the onset, 3) the propagation, and 4) the tail of the IW event. The squared buoyancy frequency N2 is given by g2ρ(βdSAαdΘ)/dP, where g is the gravitational acceleration (m s−2), ρ is the density (kg m−3), SA is the absolute salinity (g kg−1), Θ is the conservative temperature (°C), P is the pressure (Pa), and β and α are respectively the saline contraction and thermal expansion coefficients evaluated at the average values of SA, Θ, and P. The mean value of N2 for each of the four regimens indicated in Fig. 2b is shown in Fig. 2e. The behavior of the buoyancy frequency clearly shows the magnitude of oscillations at different depths during the IW passing.

The sound speed profiles used in this work were derived from environmental measurements at mooring SW30. The temperature, conductivity, and pressure values collected at mooring SW30 were employed to generate water salinity profiles using the equation introduced by Fofonoff and Millard (1983). Then, a discretized version of SSPs in terms of depth was obtained using Eq. (1), which is referred to as the nine-term equation derived by Mackenzie (1981). The nine-term equation utilizes depth, temperature, and salinity profiles to compute discrete samples of SSPs at times ti, with i = 0, …, n as
cz=1448.96+4.591T5.304×102T2+2.374×104T3+1.340(S35)+1.630×102Z+1.675×107Z21.025×102T(S35)7.139×1013TZ3,
where cz is the discretized sound speed profile (m s−1), T is the temperature (°C), Z is the depth (m), and S is the salinity ratio as defined by the Practical Salinity Scale 1978 (Lewis and Perkin 1981). Even though the salinity ratio is a dimensionless value, it is commonly reported as parts per thousand (ppt).

The computed SSPs were structured as a group of sample vectors yi such that Y = [y1, …, yn] ∈ Ρm×n, with m discrete depth points (features) and n SSP samples in time. SSP samples were generated every 15 s during the entire period. To obtain a better representation of the sound speed in the water column, shape-preserving piecewise cubic interpolation was used along the depths from 14 to 40 m with a spacing of 0.5 m, yielding m = 53 depth-dependent features. The above resulted in n = 205 547 SSP samples for the entire dataset (from 0000:00 UTC 1 August to 1600:00 UTC 5 September 2006).

As shown in Figs. 2 and 3, IW events can be categorized as phenomena with low, medium, and high effects in the water column by considering the amplitude of the internal wave (variation of temperature vs depth). To distinguish between IW activities with low, medium, and high incidence, the extracted SSPs were labeled in four different classes (1–4), as detailed in Table 1. These resulting labels serve for the supervised framework presented in the following sections. The classes were inferred via k-medoids clustering with Euclidean distance as the measuring metric (Park and Jun 2009). This approach is more robust to noise and outliers as compared with k-means because it minimizes a sum of pairwise dissimilarities instead of a sum of squared Euclidean distances. The clustering strategy is composed of two steps, the build step and the swap step. In the build step each of k clusters is associated with a potential medoid, whereas in the swap step each point is tested as a potential medoid by checking the sum of within-cluster distances to define a new medoid. At each iteration, every point is then assigned to the cluster with the closest medoid until convergence.

Table 1

Four resulting classes after applying the k-medoids to SSP data extracted from SW30 (from 0000:00 UTC 1 Aug to 1600:00 UTC 5 Sep 2006). The columns in the table show the number of SSP samples per class, the mean (μY), and the standard deviation (σY) for the SSPs in each class.

Table 1

The aforementioned process led to a full-labeled dataset with four distinguishable subsets Yi, one per class: i ∈ [1, …, 4]. Of the total number of samples in the dataset (n = 205 547), class 1 has n = 32 992 SSPs, class 2 has n = 59 630 samples, class 3 has n = 67 536 SSPs, and class 4 has n = 45 389 samples. The distribution of classes can be observed in Fig. 4. Basic descriptive statistics were calculated for each subset, as shown in Table 1. The mean μYi and the standard deviation σYi of all elements in class i were computed as
μYi=1mni=1mj=1nyij and
σYi=1(n1)(m1)i=1mj=1n(yijμYi)2,
respectively. The values of μYi and σYi for each class subset provide a general view of the magnitude and variability of the SSPs within class i. As shown in Table 1 and Fig. 3, those classes with higher μYi and σYi present SSPs with the larger fluctuations in depth (classes 1 and 4), whereas classes with lower values of μYi and σYi exhibit less-variable SSPs (classes 2 and 3).
Fig. 3.
Fig. 3.

(a)–(e) Sound speed profiles from SW30 station calculated using the nine-term equation (Mackenzie 1981). The SSPs are depth dependent as indicated with the y axis. The x axis shows the number of SSPs in each panel. These SSPs are extracted every 15 s from 0000:00 UTC 1 Aug to 1600:00 UTC 5 Sep 2006. In (a), a total of n = 205 547 SSPs are extracted during the period described. In this work, k-medoids clustering is used to label the data. The resulting classes from 1 to 4 that are shown in (b), (c), (d), and (e), respectively. Notice that the SSPs on each class are not ordered in time.

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

The entire labeled dataset presented in Fig. 3a was split into training/testing sets to be used for supervised classification of SSPs. Uniform random sampling was used to split the data, where 80% of data were destined for training while the remaining 20% were for testing, and resulted in 164 438 and 41 109 samples, respectively. The training data are meant to train the classification algorithms, and the testing data are used to perform classification and measure the performance of each model. The distribution of training and testing sets for each class is shown in Fig. 4.

Fig. 4.
Fig. 4.

Distribution of data into the four classes calculated using a k-medoids algorithm. The original data are split into training and testing sets, whose number of samples per class are also depicted in the plot.

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

4. Sparse representation of SSPs

Sparse representation aims to describe a dataset as a linear combination of few elements (basis) (Rubinstein et al. 2010). These elements capture the relevant statistical information that best describes the data and are combined with a matrix of few nonzero coefficients calculated by imposing an p constraint that controls the sparsity level or nonzero elements (Zhang et al. 2015; Wright et al. 2010). In ocean acoustics, SSPs inversions are often regularized by considering a sparse representation of SSPs using EOF analysis (Gerstoft and Gingras 1996; Huang et al. 2008) or DL (Bianco and Gerstoft 2017). This section introduces a method to implement EOF analysis and DL to represent SSPs as a linear combination of basis functions using measured data from the SW06 experiment (see section 3).

a. EOFs

EOF analysis is employed to reduce the dimensionality and identify meaningful underlying features from a dataset. In statistics, EOF analysis is also known as principal component analysis and is described as the way of transforming correlated variables into a smaller number of uncorrelated variables (Abdi and Williams 2010). EOF analysis can simplify a spatial–temporal dataset by transforming it to spatial patterns of variability and temporal projections of these patterns (Weare et al. 1976). These spatial patterns are called EOFs and can be thought of as basis functions in terms of variance. The associated temporal projections are the principal component scores and are the temporal coefficients of the EOF patterns.

EOFs are computed via an SVD in which a dataset is decomposed in terms of orthogonal basis functions to generate a compressed version of the data. Here, EOF analysis is carried out on a collection of SSPs Y = [y1, …, yn] ∈ Ρm×n with m features corresponding to depths and n time samples. Initially, the dataset Y is centered by removing time-mean value across the rows to capture only the variance of each depth feature. Subsequently, an SVD is applied to the zero-mean version of Y.

The SVD of Y is given by Y = UΣV T. Since Y is a centered matrix with zero mean, the covariance matrix CY can be expressed as
CY=1m1YYT=1m1(UΣVT)(UΣVT)T=UΣ2m1UT.
The last expression is analogous to eigendecomposition, where the terms U and Σ2/(m − 1) are the nonzero eigenvectors and eigenvalues of CY, respectively. This formulation yields the definition of EOFs. Let Q be the matrix whose columns contain the EOFs that represent the eigenvectors of the matrix YYT, such that
YYT=UΣ2UT=QΛQT,
where Q = [q1, …, qn] ∈ Ρm×m are the eigenvectors (EOFs) of YYT, and Λ = Σ2 ∈ Ρm×m shows the variances of the respective EOF qi, with i = 1, …, m.
Since the EOFs in Q are ordered from highest to lowest variance, dimension reduction of Y is addressed using only the first km leading-order EOFs (denoted by columns of Q). This new low-dimensional space retains meaningful properties from the original data. Typically, k = 5 EOFs can explain most of the variance in Y (Bianco and Gerstoft 2017). In the same way, EOFs are also used for data compression, using the basis functions in the dictionary Q. A sparse representation of Y is achieved by expressing the dataset as the product of the EOFs in Q and a coefficient matrix C such that
Y^=QC,
where Y^ is the sparse representation of the original dataset Y, and C = [c1, …, cn] ∈ Ρm×n. A full representation of individual SSPs yi can be obtained using y^i=Qci,i=1,,n. A sparse representation y^i is computed by using the k leading-order EOFs such that y^i=Qkci, with Qk ∈ Ρm×k and a coefficient vector ci ∈ Ρk. Here, a dictionary is defined as a matrix whose columns are composed of basis functions that retrieve meaningful information from a dataset. In consequence, Qk can be thought of as a dictionary containing the first k orthogonal basis functions (EOFs).

b. Sparse coding

Any dataset Y ∈ Ρm×n containing m features and n samples can be successfully reconstructed with acceptable error ϵ utilizing a combination of basis vectors in a dictionary D ∈ Ρm×k, where k represents number of basis functions contained by the dictionary D. If k < m the dictionary is undercomplete, if k = m the dictionary is referred to as complete, whereas if k > m the dictionary is considered overcomplete. Data described with sparse coding are assumed to be a linear combination of basis functions di ∈ Ρm and sparse coding vectors ci ∈ Ρk from the matrix C ∈ Ρk×n, responsible for defining the synthesis of the data using basis functions.

An optimal sparse solution of the coefficients ci such that i = 1,…, k can be obtained by solving an optimization scheme with an 0-norm constraint, which is used to limit the number of nonzero entries in the vector ci. The optimization problem with the 0 penalty is formulated as
c^i=arg minciRkyiDci22subject to ‖ci0nnz,
where nnz is the number of nonzero elements in ci. The optimization problem in Eq. (4) yields an exact solution. However, the 0-norm constraint brings in a nonconvex and nondeterministic polynomial-time (NP)-hard problem whose solution is practically unreachable (Tošić and Frossard 2011; Ramirez et al. 2010). Approximation algorithms exist to find a suboptimal solution for the NP-hard problem. Greedy algorithms such as matching pursuit (MP) and orthogonal MP (OMP) (Mallat and Zhang 1993) iteratively solve the nonconvex problem imposed by the 0-norm constraint. These algorithms find the suboptimal sparse coefficients ci that best approximate the global minimum. Note that the optimization problem presented in Eq. (4) can also be convex-relaxed by changing the 0-norm term to an 1 regularization term (Tošić and Frossard 2011); targeting the solution to
c^i=arg minciRkyiDci22+λci1,
where λ is the regularization parameter. The minimization in Eq. (5) can be solved efficiently as an 1 least-squares problem using approaches based on convex relaxation such as basis pursuit (BP) denoising (Chen et al. 2001) or lasso (Hans 2009). In this approach, the exact number of nonzero elements in the vector ci is not controlled by the 1 penalty as in the 0-norm case.

One good way to compare the performance between EOFs and LDs for sparse representation is by fixing the number of nonzero coefficients nnz in ci. Since nnz is fixed, the sparse representation of SSPs must be performed with the k-leading basis from EOFs and LDs. This setting permits focus mostly on the ability of DL and EOF to capture relevant features in the data using very few basis functions. As a result, the OMP algorithm with 0-norm is used to fairly compare the ability of EOFs and DLs to sparsely represent SSPs.

c. DL

Contrary to EOF analysis in which the basis functions are computed via SVD, DL aims to find a dictionary D = [d1, …, dk] ∈ Ρm×k that is learned directly from a dataset Y = [y1, …, yn] ∈ Ρm×n. The basis functions contained in the columns of the dictionary D are called atoms and are not required to be orthogonal as in EOF analysis. With DL, each signal in Y can be sparsely represented as a linear combination of few atoms. Sparse coding is used during the process to guarantee the sparse representation of the dataset Y, where an p constraint is imposed to the sparse coefficients in C = [c1, …, cn] ∈ Ρk×n as shown in section 4b.

In this method, not only the dictionary atoms di but also the sparse coefficients ci have to simultaneously be learned from data. This problem is addressed using iterative algorithms that solve for the coefficient matrix C and the dictionary D separately and alternate the solutions until convergence. Some efficient algorithms to learn dictionaries are the method of optimal directions (MOD) (Engan et al. 1999) and the K-SVD algorithm (Aharon et al. 2006), based on k-means clustering. Even tough K-SVD converges faster than MOD, K-SVD is computationally expensive and relies in high memory use when the dataset Y is large. In this paper, the stochastic online learning algorithm proposed by Mairal et al. (2009) is used to learn dictionaries. This optimization strategy overcomes the computational issues of the K-SVD algorithm and can be applied to large-scale contexts. Here, the learned dictionary D is chosen from a convex set C that contains dictionaries whose columns are 2-normalized:
C:={DRm×k s.t. i=1,,k,diTdi1}.
The approach introduced by Mairal et al. (2009) uses stochastic gradient descent to update the dictionary Dt sequentially by accessing one training sample at a time and using the previous dictionary Dt−1 as a warm restart during the iteration t, as shown in algorithm 1 [see the appendix for the steps of algorithm 1]. The optimization problem to solve for D and C, using the 1 regularization term, is formulated as
D^,C^=arg minDC,CRk×nYDC22+λC1=arg minDC,CRk×n1ni=1n(yiDci22+λci1),
where λ is the regularization parameter applied to the 1-norm constraint, n is the number of samples in the dataset Y ∈ Ρm×n, D is the dictionary in the convex set C, and C is the matrix with the sparse coefficients. The convex optimization induced by the 1-norm can be solved efficiently using the least-angle regression lasso (LARS-lasso).

d. Dictionary learning for SSP classification

A dictionary D ∈ Ρm×k containing a set of k basis functions di ∈ Ρm can retain many of the statistical properties of a zero-mean dataset Y = [y1, …, yi] ∈ Ρm×n, i ∈ [1, …, n], composing m features and n samples. Similarly, a sparse coefficient vector ciRk is employed to reconstruct samples yiY, with acceptable error ϵ, using a linear combination of the basis functions di (see sections 4b and 4c). Let Φ={ϕ1,,ϕl}j=1l denote a set of labeled classes, then dictionary learning can be extended to supervised classification tasks using a measuring metric R^(y,y^) to classify unlabeled testing data to a specific class ϕj. Here, R^() is a metric to measure the dissimilarity between the sample yi and the reconstruction y^i=Dci. Usually, when using the OMP algorithm for sparse coding the dissimilarity metric is defined as R^(y,y^)=yDc22. However, cases exist in which the OMP algorithm can be unstable for classification tasks since small variations in the input signal (i.e., sound speed profiles) yield irregular sparse codes that difficult classification (Ramirez et al. 2010).

In consequence, for classification contexts, the 0-norm constraint is replaced by an 1-norm regularization term to balance the trade-off between reconstruction error and sparsity, mitigating in this way undesirable outcomes resulting from OMP. The use of 1-norm constraint leads to the updated dissimilarity metric
R^(y,y^)=yDc22+λc1,
which is, in fact, equal to Eq. (5). The terms in Eq. (8) takes into account the reconstruction error (yDc22) and the complexity of the sparse decomposition (λǁcǁ1). In other words, the reconstruction error measures the quality of the approximation while the complexity is measured by the 1-norm of the optimal c.
In addition, the minimization problem to learn dictionaries [refer to Eq. (7)] is modified to train the class-specific dictionaries, following the scheme introduced by Ramirez et al. (2010):
min{Dj,Cj}j=1,,lj=1l{YjDjCj22+λj=1kcjj1}+ηjjDjTDjTF2,
where j and j′ are subindices to denote elements from different classes ϕj, ϕjΦ and η is a regularization parameter that penalizes the coherence between interclass dictionaries Dj and Dj. The last term in Eq. (9) promotes incoherence between the different class dictionaries by weakening the ability of a dictionary Dj to classify correctly for other classes j′ ≠ j. In other words, the class-specific dictionary Dj for class j is trained using Eq. (7) while considering the coherence with dictionaries from other classes at the same time (i.e., dictionaries Dj s.t. j′ ≠ j Dj s.t. jj). This optimization process yields class-specific dictionaries than are very different from each other and contains the most relevant information from the data. For further details, refer to Ramirez et al. (2010).
The classification setting presented here is composed of five steps: Step 1 is dictionary training. class-specific dictionaries Dj are trained using labeled data in each class ϕj; these class-specific learned dictionaries are meant to capture meaningful information about data in the class ϕj. This step is performed by combining algorithm 1 with Eq. (9). Step 2 is sparse coding on testing data. Sparse coding via LARS-lasso algorithm with 1 is applied to each class-specific dictionary Dj and testing data yitest to compute the sparse coefficients citest. Step 3 is reconstruction. The jth class dictionary Dj and set of sparse coefficients citest are used to compute the ith reconstruction y^ij=Djcitest. As a result of this, a sparse representation y^ij is generated for each class ϕjΦ, j = 1, …, l. Step 4 is error calculation. Dissimilarity between each reconstruction y^ij and testing data yitest is computed via R^(ytest,y^ij). Step 5 is labeling. The testing dataset yitest is assigned to the class j^ that generates the lowest error R^() as follows:
j^=arg minj=1,,lR^(yitest,y^ij),
where j^ is the class assigned to the unlabeled testing data sample yitest.

5. Results and discussion

To illustrate the practicability of the methods presented in section 4, both DL and EOF analysis are applied to data collected at SW30 station during the SW06 experiment (section 2). In this section, DL and EOFs are implemented to sparsely represent SSPs altered by the passing of IWs using OMP algorithm for sparse coding. In addition, dictionary learning is later employed for SSP classification following the scheme described in section 4d, using the LARS-lasso algorithm for sparse coding.

The matrix Q ∈ Ρm×k containing the EOFs is calculated via SVD [Eq. (2)], and the learned dictionary D ∈ Ρm×k is obtained using the convex optimization shown in Eq. (7), via online dictionary learning algorithm. Equation (7) is solved iteratively using the LARS algorithm with a tolerance parameter δ = 1 × 10−8, number of iterations Tmax = 2000, and regularization parameter λ = 1 (Pedregosa et al. 2011). Furthermore, sparse coding via OMP algorithm with Tmax = 2000 is used to compute the coefficient matrix C ∈ Ρk×n to compare performance of EOFs and DL (section 5a). The number of nonzero elements in ci, computed via the OMP algorithm, are set to exactly nnz = 3 nonzero coefficients. For classification, only DL is used and sparse coding is done via the LARS-lasso algorithm with λ = 1, η = 1, Tmax = 2000, and δ = 1 × 10−8 (see section 5).

a. Dictionary learning and EOF analysis for SSPs sparse representation

Because of the reward for sparsity and the absence of an orthogonality requirement, LDs can provide an alternative for a sparse representation that might yield more accurate reconstructions than traditional methods, such as EOFs. Bianco and Gerstoft (2017) showed that LDs are suitable for representing sound speed profiles in underwater environments, even outperforming EOFs with few basis functions.

To test the versatility of dictionary atoms in highly fluctuating environments such as those with internal waves, both EOFs and DL are compared using a portion of data collected from 0000:00 UTC 17 to 2359:59 UTC 18 August 2006 at SW30 station during high internal wave activity. The data used to compare both methods are shown between the two black lines drawn in Fig. 4a, with n = 11 520 SSPs sampled every 15 s measured at m = 53 different depths. EOFs are calculated for k = 53 basis functions. Similarly, a complete dictionary with k = 53 and an overcomplete dictionary with k = 100 are also computed. Sparse coding based on both EOFs and LDs is performed using nnz = 3 nonzero coefficients via OMP algorithm. The resulting EOFs and atoms from the LD are shown in Fig. 5. Each EOF/atom contains meaningful information about the sound speed variability in terms of depth. Notice only the leading-order EOFs in Fig. 5a capture the variability of SSPs, whereas the SSP variance is distributed on all atoms of both complete (Fig. 5b) and overcomplete dictionary (Fig. 5c).

Fig. 5.
Fig. 5.

Basis functions computed via (a) EOF analysis, (b) complete DL, and (c) overcomplete DL. Only the first leading-order EOFs describe variability in the SSPs, whereas variance is distributed along all atoms in the LDs.

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

The explained variance ratio per EOF/atom is shown in Fig. 6 and complements the findings presented in Fig. 5. Most of the variance of the SSPs is mainly concentrated in the first k = 5 leading-order EOFs (Fig. 6a), while in the case of LDs, the variance is shared among most of the basis functions (Figs. 6a,b). Given these findings, SSPs can be reconstructed effectively using only the first leading-order EOFs, while in the case of LDs, the SSPs can be sparsely represented by a sparse combination of almost any atom in the dictionary.

Fig. 6.
Fig. 6.

SSP explained variance ratio for (a) EOFs and (b) complete LD entries with k = 53 and for (c) overcomplete LD with k = 100 atoms using nnz = 3 nonzero coefficients.

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

As stated in section 4, the EOFs in the matrix Q are computed via SVD and have the property of being normalized orthogonal basis functions. Conversely, atoms in a dictionary D are not required to be orthogonal and are restricted to have unit 2-norm as stated in Eq. (6). Coherence among column entries in Q and D can be calculated by computing the Gram matrix GQ = | QTQ | for EOFs, and GD = | DTD | for LDs. For the EOF analysis, GQ = I, as shown in Fig. 7a, because of the orthogonal properties of EOFs. In contrast, atoms in the complete and overcomplete LDs are not necessarily orthogonal, as displayed in Figs. 7b and 7c.

Fig. 7.
Fig. 7.

Coherence matrices for (a) EOFs, (b) complete LD entries, and (c) overcomplete LD (see Fig. 5) using nnz = 3 nonzero coefficients. Eigenvectors in the EOF dictionary are orthogonal, whereas those of the LDs are not.

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

The relaxation of the orthogonality requirement for LD atoms leads to more flexible dictionaries and richer data representation. As a result of this, better compression of data is achieved using dictionary learning. For both LDs and EOFs, increasing the number of basis functions k and nonzero coefficients nnz yields lower representation errors. To study sparse representation of SSPs, the OMP algorithm is used to compute sparse codes with 1 ≤ nnz ≤ 12 for EOFs and LDs. The root-mean-square error (RMSE) between the sparse representation Y^i and the original data Y is measured as
RMSE=i=1nj=1m(yijy^ij)2mn,
where m are depths, and n is the number of SSPs in Y. The RMSE values as a function of nnz, shown in Fig. 8, provide a method for comparing the different sparse representations. It is clear that, for nnz ≥ 8, the EOFs, complete LD, and overcomplete LD give the almost the same RMSE. For nnz < 7, however, the RMSE using EOFs is larger than for LDs, with the overcomplete LD yielding the lowest RMSE for small values of nnz.
Fig. 8.
Fig. 8.

Root-mean-square error of EOFs and complete LD, both with k = 53, and overcomplete LD with k = 100 when different numbers of nonzero coefficients nnz are used for sparse coding.

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

To demonstrate the efficacy of both EOFs and LDs to sparsely represent SSPs, six individual random samples yi are chosen from the dataset Y and are depicted with solid green lines in Fig. 9. The sparse representation of SSPs y^i of the six samples using the EOFs and the complete LD, both with k = 53 and nnz = 3 are shown in Fig. 9. The dashed orange line presents the time-mean value from the entire dataset, dotted blue line shows the SSP reconstructed using EOFs, whereas dash–dotted red line the reconstruction using LDs. In addition, the absolute error between each real sample yi and the reconstruction y^i in terms of depth was computed as |yiy^i|, and is shown with color bars (to the right of each line plot) for both LD and EOFs. Dark colors in the bars correspond to small errors, whereas light colors represent high absolute errors. For each case, only with nnz = 3 basis functions both EOFs complete LD yield a good representation of SSPs. Even with internal wave events passing by, complete LD provides a nearly perfect representation of the six SSPs and outperforms conventional EOF analysis in the sparse reconstruction of sound speed profiles.

Fig. 9.
Fig. 9.

Sparse representation of six SSP samples using complete DL and EOF with nnz = 3 nonzero coefficients and k = 53 atoms and EOFs. For each sample, color bars present the absolute error between the sample and the reconstruction in terms of depth for LD and EOFs, respectively.

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

b. Classification of SSPs via dictionary learning

In previous sections is mentioned that learned dictionaries D can retain meaningful information from a dataset Y ∈ Ρm×n that can or not be labeled in a class ϕj, j = 1, …, l. However, limitations for data classification arise when few nnz atoms are used because of the impossibility of representing the variability in the data. Overcomplete dictionaries, because of their redundant nature, are suitable for classification since they can perform data analysis with functions that are likely to match the characteristics in different classes of data.

In this section, the dictionary learning framework is used to label sound speed profiles following the classification setting introduced in section 4d. Complete and overcomplete LDs are employed to classify data with low, medium, and high occurrence of internal wave events in the ocean. For the best analysis, the data introduced in section 2 and Table 1 are used for to test the algorithms in this section.

As the extracted SSPs from the experiment are not initially labeled, each SSP sample is labeled into four classes (1–4) depending on the internal wave level using k-medoids algorithm, as shown in section 3. Once all data samples are labeled, the effectiveness of LD atoms to infer a class type for SSPs is studied using the training/testing sets described in section 3. Here, class-specific dictionaries are learned using training data, to then classify unlabeled SSPs samples in the testing set utilizing a dissimilarity metric R^(y,y^) [Eq. (10)]. The scheme introduced in section 4d is implemented for complete (k = 53) and overcomplete (k = 100 and k = 300) dictionaries. In addition, dictionary learning is compared with SVM and KNN algorithms, both commonly used for supervised classification tasks. KNN is implemented using k = 100 neighbors and Euclidean distance as metric, whereas 2 regularized SVM with Gaussian kernel is employed with a one-vs-the-rest configuration for multiclass classification.

The LDs models, KNN, and SVM are trained using the training set with n = 164 438 samples and four different classes and then tested on the remaining n = 41 109 unseen samples, whose labels are inferred by the classification model. That is, the trained classifier hH aims to assign a label to unlabeled data yitest, such that j^i=h(yitest),j^=1,,l. Here, the accuracy is reported for each model and accounts for the number of correct classifications over the total number of samples in the testing set. This metric is calculated as
accuracy=i=1n1(j^i=ji)n×100%,
where 1(⋅) is the indicator function, n is the total number of samples in the dataset, ji is label or ground truth for the ith sample, and j^i is the inferred class for the ith SSP, such ℓ that j^i=h(yitest).

The classification results for KNN, SVM, and LDs are shown in Table 2, where accuracy is reported for all the models using the entire testing data along with the performance for individual subsets corresponding to each independent class (j). As demonstrated in section 5a, the overcomplete LD yields lower reconstruction errors than complete LDs. This fact is also corroborated by results in Table 2, where the overcomplete LD with k = 300 reaches higher accuracy than LDs with k = 53 and k = 100 for all the four classes.

Table 2

Classification results for KNN, SVM, complete LD, and overcomplete LDs for testing data. Accuracy (%) is reported for each classification model. The first four rows in the table correspond to independent classes in the testing set, and the last row presents the overall accuracy for each model in the complete testing set. Boldface type indicates which model performed best for each class/set.

Table 2

It is notable that results of the overcomplete LD with k = 300 are comparable to SVM and KNN and are even better at differentiating between classes with high internal wave activity (classes 1 and 4). Notice that the misclassification of SSPs by LDs is due to the possible lack of sufficient information about the variability retrieved by the k atoms. It is clear to see that the more atoms the dictionary uses, the more the accuracy will be. Therefore, the classification results reported for LDs in Table 2 can be improved by increasing the number of atoms used. This study shows that classification of SSPs via dictionary learning is feasible and can be extended to large-context scenarios as long as exists labeled data to compute specific dictionaries for each class.

From results in Table 2, it is possible to conclude overcomplete dictionary learning offers a good alternative for classifying SSPs with high internal wave activity if using sufficient k atoms. The relaxation in the orthogonal constraint in the basis functions allows DLs to capture the most representative information from data. Each class-specific dictionary Dj tends to have different patterns than others from different classes (Dj). The coherence GDj for class-specific learned dictionaries with k = 53, k = 100, and k = 300 atoms is shown in Fig. 10. Notice that for different values of k and class, each class-specific dictionary presents different and distinctive patterns than are being learned from data.

Fig. 10.
Fig. 10.

Coherence of individual class-specific learned dictionaries for classification of SSPs: (a)–(d) coherence for the complete LD (k = 53), (e)–(h) coherence of overcomplete LDs with k = 100 atoms for all the classes, and (i)–(l) coherence for LDs with k = 300 for classes (left) 1, (left center) 2, (right center) 3, and (right) 4.

Citation: Journal of Atmospheric and Oceanic Technology 40, 1; 10.1175/JTECH-D-21-0090.1

It is important to remark that if the number of k atoms increases, the dictionary will lead to higher accuracy with a cost of an increment in the complexity of the convex optimization. Therefore, it is important to consider the trade-off between accuracy and complexity when training LDs, as is done with any learning model, such as neural networks. In cases where there are not sufficient labeled training data, it is possible to apply data processing techniques such as data augmentation to increase the variability and the number of data samples within a class (Castro-Correa et al. 2021).

Similar to previous research (Bianco and Gerstoft 2017; Sun and Zhao 2020), overcomplete LDs perform well as a sparse representation algorithm for SSPs. Dictionary learning outperforms EOFs because it tends to distribute the SSPs energy among all the atoms and its ability to generate nonorthogonal functions. Furthermore, as few as nnz nonzero coefficients are needed to consistently provide a complete enough representation to achieve very low training error even when dictionaries are trained on SSPs in the presence of internal waves. Results demonstrate competitive performance of LDs with respect to standard classification models. The amount of data, the sparsity level provided by nnz, and the number of atoms k are crucial factors to obtain optimal sparse representations of SSPs.

6. Conclusions

Both dictionary learning and empirical orthogonal functions were implemented to sparsely represent SSPs disturbed by the passing of internal waves in the SW06 experiment. Because of their redundant nature and their ability to generate nonorthogonal basis functions, overcomplete learned dictionaries showed better performance for reconstructing SSPs than the EOFs and the complete LD when using the best nnz = 3 combination of basis functions.

The presence of internal waves in the water column causes highly anisotropic SSPs. The variability in the SSPs induced by IWs was reflected in higher errors when both EOF and dictionary learning frameworks were applied. Under those circumstances, overcomplete dictionaries with k > m atoms were shown to achieve even better compression of SSPs than EOFs and the complete LD k = m. The improvement in the reconstruction of SSPs was produced due to the relaxation of the orthogonal requirements, and the number of atoms used in the dictionary.

In this paper, the classification of SSPs via dictionary learning was introduced. Here, specific dictionaries were built for each class of internal wave activity, and testing datasets were classified by finding the dictionary corresponding to the most accurate sparse representation. Results demonstrated that overcomplete learned dictionaries trained on labeled data are suitable to classify SSPs successfully. When the training data are representative and are labeled by a class, the resulting overcomplete dictionary can be effectively applied to other datasets to classify SSPs.

This work provides insights into the application of learned dictionaries for the representation and classification of SSPs. Further analyses are required to find the optimal number of nonzero coefficients nnz and k atoms used to obtain the best possible sparse representation of SSPs. Future research needs to be conducted to find a better-suited dissimilarity cost R^() that can yield higher classification performance, while further studies are required to determine if undercomplete LDs (k < m) are suitable for sparse representation and dimensionality reduction of SSPs.

Sparse representation of sound speed profiles using dictionary learning promotes and expedites research into internal waves. A representative dictionary of basis functions is an efficient way to store and generate thousands of unique sound speed profiles, given that the dictionary succinctly and effectively represents the small-scale variability to model. Using learned dictionaries to generate realistic and variable training datasets may further the progress of machine learning in many undersea applications.

Acknowledgments.

This research was supported by Office of Naval Research Grants N00014-21-1-2424 and N00014-21-1-2760.

Data availability statement.

Data used in this work were collected during the Shallow Water Experiment 2006 (SW06). Because of privacy and ethical concerns, neither the data nor the source of the data can be made available.

APPENDIX

Algorithm 1

This appendix gives the steps to the online dictionary learning algorithm that we refer to here in this paper as algorithm 1. The algorithm has been adapted from Mairal et al. (2009). The steps are as follows.

  • Require:

  • Y ∈ Ρm×nρ(y): training data

  • T: number of iterations

  • λ: sparsity level

  • D0 ∈ Ρm×k: initial dictionary

1: A0 ← 0, B0 ← 0 #Initialize A0 and B0

2: for t = 1 to T do

8: end for

9: Return DT (learned dictionary)

REFERENCES

  • Abdi, H., and L. J. Williams, 2010: Principal component analysis. Wiley Interdiscip. Rev.: Comput. Stat., 2, 433459, https://doi.org/10.1002/wics.101.

    • Search Google Scholar
    • Export Citation
  • Abiva, J., T. Fabbri, and R. Vicen-Bueno, 2019: Automatic classification of sound speed profiles using PCA and self-organizing map techniques. OCEANS 2019, Marseille, France, IEEE, https://doi.org/10.1109/OCEANSE.2019.8867526.

  • Aharon, M., M. Elad, and A. Bruckstein, 2006: K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process., 54, 43114322, https://doi.org/10.1109/TSP.2006.881199.

    • Search Google Scholar
    • Export Citation
  • Badiey, M., L. Wan, and J. F. Lynch, 2013: Statistics of internal waves measured during the Shallow Water 2006 experiment. J. Acoust. Soc. Amer., 134, 40354035, https://doi.org/10.1121/1.4830734.

    • Search Google Scholar
    • Export Citation
  • Badiey, M., L. Wan, and J. F. Lynch, 2016: Statistics of nonlinear internal waves during the Shallow Water 2006 experiment. J. Atmos. Oceanic Technol., 33, 839846, https://doi.org/10.1175/JTECH-D-15-0221.1.

    • Search Google Scholar
    • Export Citation
  • Bianco, M., and P. Gerstoft, 2017: Dictionary learning of sound speed profiles. J. Acoust. Soc. Amer., 141, 17491758, https://doi.org/10.1121/1.4977926.

    • Search Google Scholar
    • Export Citation
  • Castro-Correa, J. A., M. Badiey, T. B. Neilsen, D. P. Knobles, and W. S. Hodgkiss, 2021: Impact of data augmentation on supervised learning for a moving mid-frequency source. J. Acoust. Soc. Amer., 150, 39143928, https://doi.org/10.1121/10.0007284.

  • Chen, S. S., D. L. Donoho, and M. A. Saunders, 2001: Atomic decomposition by basis pursuit. SIAM Rev., 43, 129159, https://doi.org/10.1137/S003614450037906X.

    • Search Google Scholar
    • Export Citation
  • Engan, K., S. O. Aase, and J. H. Husoy, 1999: Frame based signal compression using method of optimal directions (MOD). 1999 IEEE Int. Symp. on Circuits and Systems, Orlando, FL, IEEE, https://doi.org/10.1109/ISCAS.1999.779928.

  • Flatté, S. M., and F. D. Tappert, 1975: Calculation of the effect of internal waves on oceanic sound transmission. J. Acoust. Soc. Amer., 58, 11511159, https://doi.org/10.1121/1.380798.

    • Search Google Scholar
    • Export Citation
  • Fofonoff, N. P., and R. Millard Jr., 1983: Algorithms for the computation of fundamental properties of seawater. UNESCO Tech. Paper in Marine Science 44, 53 pp., https://doi.org/10.25607/OBP-1450.

  • Gerstoft, P., and D. F. Gingras, 1996: Parameter estimation using multifrequency range-dependent acoustic data in shallow water. J. Acoust. Soc. Amer., 99, 28392850, https://doi.org/10.1121/1.414818.

    • Search Google Scholar
    • Export Citation
  • Hans, C., 2009: Bayesian lasso regression. Biometrika, 96, 835845, https://doi.org/10.1093/biomet/asp047.

  • Helfrich, K. R., and W. K. Melville, 2006: Long nonlinear internal waves. Annu. Rev. Fluid Mech., 38, 395425, https://doi.org/10.1146/annurev.fluid.38.050304.092129.

    • Search Google Scholar
    • Export Citation
  • Huang, C.-F., P. Gerstoft, and W. S. Hodgkiss, 2008: Effect of ocean sound speed uncertainty on matched-field geoacoustic inversion. J. Acoust. Soc. Amer., 123, EL162EL168, https://doi.org/10.1121/1.2908406.

    • Search Google Scholar
    • Export Citation
  • Jain, S., and M. Ali, 2006: Estimation of sound speed profiles using artificial neural networks. IEEE Geosci. Remote Sens. Lett., 3, 467470, https://doi.org/10.1109/LGRS.2006.876221.

    • Search Google Scholar
    • Export Citation
  • Katsnelson, B. G., O. A. Godin, and Q. Zhang, 2021: Observations of acoustic noise bursts accompanying nonlinear internal gravity waves on the continental shelf off New Jersey. J. Acoust. Soc. Amer., 149, 16091622, https://doi.org/10.1121/10.0003624.

    • Search Google Scholar
    • Export Citation
  • Kuo, Y.-H., and J.-F. Kiang, 2020: Estimation of range-dependent sound-speed profile with dictionary learning method. IET Radar Sonar Navig., 14, 194199, https://doi.org/10.1049/iet-rsn.2019.0107.

    • Search Google Scholar
    • Export Citation
  • Lewis, E., and R. Perkin, 1981: The Practical Salinity Scale 1978: Conversion of existing data. Deep-Sea Res., 28A, 307328, https://doi.org/10.1016/0198-0149(81)90002-9.

    • Search Google Scholar
    • Export Citation
  • Mackenzie, K. V., 1981: Nine-term equation for sound speed in the oceans. J. Acoust. Soc. Amer., 70, 807812, https://doi.org/10.1121/1.386920.

    • Search Google Scholar
    • Export Citation
  • Mairal, J., F. Bach, J. Ponce, and G. Sapiro, 2009: Online dictionary learning for sparse coding. Proc. 26th Annual Int. Conf. on Machine Learning, New York, NY, ACM, 689696, https://doi.org/10.1145/1553374.1553463.

  • Mallat, S. G., and Z. Zhang, 1993: Matching pursuits with time-frequency dictionaries. IEEE Trans. Signal Process., 41, 33973415, https://doi.org/10.1109/78.258082.

    • Search Google Scholar
    • Export Citation
  • Newhall, A. E. , and Coauthors, 2007: Acoustic and oceanographic observations and configuration information for the WHOI moorings from the SW06 experiment. Woods Hole Oceanographic Institution Tech. Rep. WHOI-2007-04, 116 pp., https://doi.org/10.1575/1912/1826.

  • North, G. R., 1984: Empirical orthogonal functions and normal modes. J. Atmos. Sci., 41, 879887, https://doi.org/10.1175/1520-0469(1984)041<0879:EOFANM>2.0.CO;2.

    • Search Google Scholar
    • Export Citation
  • Park, H.-S., and C.-H. Jun, 2009: A simple and fast algorithm for K-medoids clustering. Expert Syst. Appl., 36, 33363341, https://doi.org/10.1016/j.eswa.2008.01.039.

    • Search Google Scholar
    • Export Citation
  • Pedregosa, F., and Coauthors, 2011: Scikit-learn: Machine learning in Python. J. Mach. Learn. Res., 12, 28252830, https://www.jmlr.org/papers/v12/pedregosa11a.html.

  • Ramirez, I., P. Sprechmann, and G. Sapiro, 2010: Classification and clustering via dictionary learning with structured incoherence and shared features. 2010 IEEE Computer Society Conf. on Computer Vision and Pattern Recognition, San Francisco, CA, IEEE, 35013508, https://doi.org/10.1109/CVPR.2010.5539964.

  • Roundy, P. E., 2015: On the interpretation of EOF analysis of ENSO, atmospheric Kelvin waves, and the MJO. J. Climate, 28, 11481165, https://doi.org/10.1175/JCLI-D-14-00398.1.

    • Search Google Scholar
    • Export Citation
  • Rouseff, D., 2001: Effect of shallow water internal waves on ocean acoustic striation patterns. Waves Random Media, 11, 377393, https://doi.org/10.1088/0959-7174/11/4/302.

    • Search Google Scholar
    • Export Citation
  • Rubinstein, R., A. M. Bruckstein, and M. Elad, 2010: Dictionaries for sparse representation modeling. Proc. IEEE, 98, 10451057, https://doi.org/10.1109/JPROC.2010.2040551.

    • Search Google Scholar
    • Export Citation
  • Sprechmann, P., and G. Sapiro, 2010: Dictionary learning and sparse coding for unsupervised clustering. 2010 IEEE Int. Conf. on Acoustics, Speech and Signal Processing. Dallas, TX, IEEE, 20422045, https://doi.org/10.1109/ICASSP.2010.5494985.

  • Stewart, G. W., 1993: On the early history of the singular value decomposition. SIAM Rev., 35, 551566, https://doi.org/10.1137/1035134.

    • Search Google Scholar
    • Export Citation
  • Sun, S., and H. Zhao, 2020: Sparse representation of sound speed profiles based on dictionary learning. 2020 13th Int. Congress on Image and Signal Processing, Biomedical Engineering and Informatics, Chengdu, China, IEEE, 484488, https://doi.org/10.1109/CISP-BMEI51763.2020.9263627.

  • Suo, Y., M. Dao, U. Srinivas, V. Monga, and T. D. Tran, 2014: Structured dictionary learning for classification. arXiv, 1406.1943, https://doi.org/10.48550/arXiv.1406.1943.

    • Search Google Scholar
    • Export Citation
  • Tang, D., and Coauthors, 2007: Shallow Water ’06: A joint acoustic propagation/nonlinear internal wave physics experiment. Oceanography, 20 (4), 156167, https://doi.org/10.5670/oceanog.2007.16.

    • Search Google Scholar
    • Export Citation
  • Tang, W., A. Panahi, H. Krim, and L. Dai, 2019: Analysis dictionary learning based classification: Structure for robustness. IEEE Trans. Image Process., 28, 60356046, https://doi.org/10.1109/TIP.2019.2919409.

    • Search Google Scholar
    • Export Citation
  • Tošić, I., and P. Frossard, 2011: Dictionary learning. IEEE Signal Process. Mag., 28, 2738, https://doi.org/10.1109/MSP.2010.939537.

    • Search Google Scholar
    • Export Citation
  • Weare, B. C., A. R. Navato, and R. E. Newell, 1976: Empirical orthogonal analysis of Pacific sea surface temperatures. J. Phys. Oceanogr., 6, 671678, https://doi.org/10.1175/1520-0485(1976)006<0671:EOAOPS>2.0.CO;2.

    • Search Google Scholar
    • Export Citation
  • Wright, J., Y. Ma, J. Mairal, G. Sapiro, T. S. Huang, and S. Yan, 2010: Sparse representation for computer vision and pattern recognition. Proc. IEEE, 98, 10311044, https://doi.org/10.1109/JPROC.2010.2044470.

    • Search Google Scholar
    • Export Citation
  • Xu, W., and H. Schmidt, 2006: System-orthogonal functions for sound speed profile perturbation. IEEE J. Oceanic Eng., 31, 156169, https://doi.org/10.1109/JOE.2006.872221.

    • Search Google Scholar
    • Export Citation
  • Zhang, Z., Y. Xu, J. Yang, X. Li, and D. Zhang, 2015: A survey of sparse representation: Algorithms and applications. IEEE Access, 3, 490530, https://doi.org/10.1109/ACCESS.2015.2430359.

    • Search Google Scholar
    • Export Citation
  • Zhao, C., Z. Feng, X. Wei, and Y. Qin, 2018: Sparse classification based on dictionary learning for planet bearing fault identification. Expert Syst. Appl., 108, 233245, https://doi.org/10.1016/j.eswa.2018.05.012.

    • Search Google Scholar
    • Export Citation
Save
  • Abdi, H., and L. J. Williams, 2010: Principal component analysis. Wiley Interdiscip. Rev.: Comput. Stat., 2, 433459, https://doi.org/10.1002/wics.101.

    • Search Google Scholar
    • Export Citation
  • Abiva, J., T. Fabbri, and R. Vicen-Bueno, 2019: Automatic classification of sound speed profiles using PCA and self-organizing map techniques. OCEANS 2019, Marseille, France, IEEE, https://doi.org/10.1109/OCEANSE.2019.8867526.

  • Aharon, M., M. Elad, and A. Bruckstein, 2006: K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process., 54, 43114322, https://doi.org/10.1109/TSP.2006.881199.

    • Search Google Scholar
    • Export Citation
  • Badiey, M., L. Wan, and J. F. Lynch, 2013: Statistics of internal waves measured during the Shallow Water 2006 experiment. J. Acoust. Soc. Amer., 134, 40354035, https://doi.org/10.1121/1.4830734.

    • Search Google Scholar
    • Export Citation
  • Badiey, M., L. Wan, and J. F. Lynch, 2016: Statistics of nonlinear internal waves during the Shallow Water 2006 experiment. J. Atmos. Oceanic Technol., 33, 839846, https://doi.org/10.1175/JTECH-D-15-0221.1.

    • Search Google Scholar
    • Export Citation
  • Bianco, M., and P. Gerstoft, 2017: Dictionary learning of sound speed profiles. J. Acoust. Soc. Amer., 141, 17491758, https://doi.org/10.1121/1.4977926.

    • Search Google Scholar
    • Export Citation
  • Castro-Correa, J. A., M. Badiey, T. B. Neilsen, D. P. Knobles, and W. S. Hodgkiss, 2021: Impact of data augmentation on supervised learning for a moving mid-frequency source. J. Acoust. Soc. Amer., 150, 39143928, https://doi.org/10.1121/10.0007284.

  • Chen, S. S., D. L. Donoho, and M. A. Saunders, 2001: Atomic decomposition by basis pursuit. SIAM Rev., 43, 129159, https://doi.org/10.1137/S003614450037906X.

    • Search Google Scholar
    • Export Citation
  • Engan, K., S. O. Aase, and J. H. Husoy, 1999: Frame based signal compression using method of optimal directions (MOD). 1999 IEEE Int. Symp. on Circuits and Systems, Orlando, FL, IEEE, https://doi.org/10.1109/ISCAS.1999.779928.

  • Flatté, S. M., and F. D. Tappert, 1975: Calculation of the effect of internal waves on oceanic sound transmission. J. Acoust. Soc. Amer., 58, 11511159, https://doi.org/10.1121/1.380798.

    • Search Google Scholar
    • Export Citation
  • Fofonoff, N. P., and R. Millard Jr., 1983: Algorithms for the computation of fundamental properties of seawater. UNESCO Tech. Paper in Marine Science 44, 53 pp., https://doi.org/10.25607/OBP-1450.

  • Gerstoft, P., and D. F. Gingras, 1996: Parameter estimation using multifrequency range-dependent acoustic data in shallow water. J. Acoust. Soc. Amer., 99, 28392850, https://doi.org/10.1121/1.414818.

    • Search Google Scholar
    • Export Citation
  • Hans, C., 2009: Bayesian lasso regression. Biometrika, 96, 835845, https://doi.org/10.1093/biomet/asp047.

  • Helfrich, K. R., and W. K. Melville, 2006: Long nonlinear internal waves. Annu. Rev. Fluid Mech., 38, 395425, https://doi.org/10.1146/annurev.fluid.38.050304.092129.

    • Search Google Scholar
    • Export Citation
  • Huang, C.-F., P. Gerstoft, and W. S. Hodgkiss, 2008: Effect of ocean sound speed uncertainty on matched-field geoacoustic inversion. J. Acoust. Soc. Amer., 123, EL162EL168, https://doi.org/10.1121/1.2908406.

    • Search Google Scholar
    • Export Citation
  • Jain, S., and M. Ali, 2006: Estimation of sound speed profiles using artificial neural networks. IEEE Geosci. Remote Sens. Lett., 3, 467470, https://doi.org/10.1109/LGRS.2006.876221.

    • Search Google Scholar
    • Export Citation
  • Katsnelson, B. G., O. A. Godin, and Q. Zhang, 2021: Observations of acoustic noise bursts accompanying nonlinear internal gravity waves on the continental shelf off New Jersey. J. Acoust. Soc. Amer., 149, 16091622, https://doi.org/10.1121/10.0003624.

    • Search Google Scholar
    • Export Citation
  • Kuo, Y.-H., and J.-F. Kiang, 2020: Estimation of range-dependent sound-speed profile with dictionary learning method. IET Radar Sonar Navig., 14, 194199, https://doi.org/10.1049/iet-rsn.2019.0107.

    • Search Google Scholar
    • Export Citation
  • Lewis, E., and R. Perkin, 1981: The Practical Salinity Scale 1978: Conversion of existing data. Deep-Sea Res., 28A, 307328, https://doi.org/10.1016/0198-0149(81)90002-9.

    • Search Google Scholar
    • Export Citation
  • Mackenzie, K. V., 1981: Nine-term equation for sound speed in the oceans. J. Acoust. Soc. Amer., 70, 807812, https://doi.org/10.1121/1.386920.

    • Search Google Scholar
    • Export Citation
  • Mairal, J., F. Bach, J. Ponce, and G. Sapiro, 2009: Online dictionary learning for sparse coding. Proc. 26th Annual Int. Conf. on Machine Learning, New York, NY, ACM, 689696, https://doi.org/10.1145/1553374.1553463.

  • Mallat, S. G., and Z. Zhang, 1993: Matching pursuits with time-frequency dictionaries. IEEE Trans. Signal Process., 41, 33973415, https://doi.org/10.1109/78.258082.

    • Search Google Scholar
    • Export Citation
  • Newhall, A. E. , and Coauthors, 2007: Acoustic and oceanographic observations and configuration information for the WHOI moorings from the SW06 experiment. Woods Hole Oceanographic Institution Tech. Rep. WHOI-2007-04, 116 pp., https://doi.org/10.1575/1912/1826.

  • North, G. R., 1984: Empirical orthogonal functions and normal modes. J. Atmos. Sci., 41, 879887, https://doi.org/10.1175/1520-0469(1984)041<0879:EOFANM>2.0.CO;2.

    • Search Google Scholar
    • Export Citation
  • Park, H.-S., and C.-H. Jun, 2009: A simple and fast algorithm for K-medoids clustering. Expert Syst. Appl., 36, 33363341, https://doi.org/10.1016/j.eswa.2008.01.039.

    • Search Google Scholar
    • Export Citation
  • Pedregosa, F., and Coauthors, 2011: Scikit-learn: Machine learning in Python. J. Mach. Learn. Res., 12, 28252830, https://www.jmlr.org/papers/v12/pedregosa11a.html.

  • Ramirez, I., P. Sprechmann, and G. Sapiro, 2010: Classification and clustering via dictionary learning with structured incoherence and shared features. 2010 IEEE Computer Society Conf. on Computer Vision and Pattern Recognition, San Francisco, CA, IEEE, 35013508, https://doi.org/10.1109/CVPR.2010.5539964.

  • Roundy, P. E., 2015: On the interpretation of EOF analysis of ENSO, atmospheric Kelvin waves, and the MJO. J. Climate, 28, 11481165, https://doi.org/10.1175/JCLI-D-14-00398.1.

    • Search Google Scholar
    • Export Citation
  • Rouseff, D., 2001: Effect of shallow water internal waves on ocean acoustic striation patterns. Waves Random Media, 11, 377393, https://doi.org/10.1088/0959-7174/11/4/302.

    • Search Google Scholar
    • Export Citation
  • Rubinstein, R., A. M. Bruckstein, and M. Elad, 2010: Dictionaries for sparse representation modeling. Proc. IEEE, 98, 10451057, https://doi.org/10.1109/JPROC.2010.2040551.

    • Search Google Scholar
    • Export Citation
  • Sprechmann, P., and G. Sapiro, 2010: Dictionary learning and sparse coding for unsupervised clustering. 2010 IEEE Int. Conf. on Acoustics, Speech and Signal Processing. Dallas, TX, IEEE, 20422045, https://doi.org/10.1109/ICASSP.2010.5494985.

  • Stewart, G. W., 1993: On the early history of the singular value decomposition. SIAM Rev., 35, 551566, https://doi.org/10.1137/1035134.

    • Search Google Scholar
    • Export Citation
  • Sun, S., and H. Zhao, 2020: Sparse representation of sound speed profiles based on dictionary learning. 2020 13th Int. Congress on Image and Signal Processing, Biomedical Engineering and Informatics, Chengdu, China, IEEE, 484488, https://doi.org/10.1109/CISP-BMEI51763.2020.9263627.

  • Suo, Y., M. Dao, U. Srinivas, V. Monga, and T. D. Tran, 2014: Structured dictionary learning for classification. arXiv, 1406.1943, https://doi.org/10.48550/arXiv.1406.1943.

    • Search Google Scholar
    • Export Citation
  • Tang, D., and Coauthors, 2007: Shallow Water ’06: A joint acoustic propagation/nonlinear internal wave physics experiment. Oceanography, 20 (4), 156167, https://doi.org/10.5670/oceanog.2007.16.

    • Search Google Scholar
    • Export Citation
  • Tang, W., A. Panahi, H. Krim, and L. Dai, 2019: Analysis dictionary learning based classification: Structure for robustness. IEEE Trans. Image Process., 28, 60356046, https://doi.org/10.1109/TIP.2019.2919409.

    • Search Google Scholar
    • Export Citation
  • Tošić, I., and P. Frossard, 2011: Dictionary learning. IEEE Signal Process. Mag., 28, 2738, https://doi.org/10.1109/MSP.2010.939537.

    • Search Google Scholar
    • Export Citation
  • Weare, B. C., A. R. Navato, and R. E. Newell, 1976: Empirical orthogonal analysis of Pacific sea surface temperatures. J. Phys. Oceanogr., 6, 671678, https://doi.org/10.1175/1520-0485(1976)006<0671:EOAOPS>2.0.CO;2.

    • Search Google Scholar
    • Export Citation
  • Wright, J., Y. Ma, J. Mairal, G. Sapiro, T. S. Huang, and S. Yan, 2010: Sparse representation for computer vision and pattern recognition. Proc. IEEE, 98, 10311044, https://doi.org/10.1109/JPROC.2010.2044470.

    • Search Google Scholar
    • Export Citation
  • Xu, W., and H. Schmidt, 2006: System-orthogonal functions for sound speed profile perturbation. IEEE J. Oceanic Eng., 31, 156169, https://doi.org/10.1109/JOE.2006.872221.

    • Search Google Scholar
    • Export Citation
  • Zhang, Z., Y. Xu, J. Yang, X. Li, and D. Zhang, 2015: A survey of sparse representation: Algorithms and applications. IEEE Access, 3, 490530, https://doi.org/10.1109/ACCESS.2015.2430359.

    • Search Google Scholar
    • Export Citation
  • Zhao, C., Z. Feng, X. Wei, and Y. Qin, 2018: Sparse classification based on dictionary learning for planet bearing fault identification. Expert Syst. Appl., 108, 233245, https://doi.org/10.1016/j.eswa.2018.05.012.

    • Search Google Scholar
    • Export Citation
  • Fig. 1.

    Some of the SW06 moorings along shelf and across shelf conforming to a T geometry. SSPs used in this work are derived from data collected at SW30 station, which is marked with a white star.

  • Fig. 2.

    IW event spotted from 2100:00 UTC 17 to 1000:00 UTC 18 Aug 2006. (a) Temporal evolution of temperature profiles at mooring SW30. (b) The square of the buoyancy frequency N2, in terms of depth, divided into four sections corresponding to different regimens. (c) Standard deviation (shading) and mean of the SSPs presented in (a). (d) Individual SSP samples at each geotime tgi, with colors matching the vertical dashed lines in (a). (e) Buoyancy frequency N2 for each of the four regimens shown in (b).

  • Fig. 3.

    (a)–(e) Sound speed profiles from SW30 station calculated using the nine-term equation (Mackenzie 1981). The SSPs are depth dependent as indicated with the y axis. The x axis shows the number of SSPs in each panel. These SSPs are extracted every 15 s from 0000:00 UTC 1 Aug to 1600:00 UTC 5 Sep 2006. In (a), a total of n = 205 547 SSPs are extracted during the period described. In this work, k-medoids clustering is used to label the data. The resulting classes from 1 to 4 that are shown in (b), (c), (d), and (e), respectively. Notice that the SSPs on each class are not ordered in time.

  • Fig. 4.

    Distribution of data into the four classes calculated using a k-medoids algorithm. The original data are split into training and testing sets, whose number of samples per class are also depicted in the plot.

  • Fig. 5.

    Basis functions computed via (a) EOF analysis, (b) complete DL, and (c) overcomplete DL. Only the first leading-order EOFs describe variability in the SSPs, whereas variance is distributed along all atoms in the LDs.

  • Fig. 6.

    SSP explained variance ratio for (a) EOFs and (b) complete LD entries with k = 53 and for (c) overcomplete LD with k = 100 atoms using nnz = 3 nonzero coefficients.

  • Fig. 7.

    Coherence matrices for (a) EOFs, (b) complete LD entries, and (c) overcomplete LD (see Fig. 5) using nnz = 3 nonzero coefficients. Eigenvectors in the EOF dictionary are orthogonal, whereas those of the LDs are not.

  • Fig. 8.

    Root-mean-square error of EOFs and complete LD, both with k = 53, and overcomplete LD with k = 100 when different numbers of nonzero coefficients nnz are used for sparse coding.

  • Fig. 9.

    Sparse representation of six SSP samples using complete DL and EOF with nnz = 3 nonzero coefficients and k = 53 atoms and EOFs. For each sample, color bars present the absolute error between the sample and the reconstruction in terms of depth for LD and EOFs, respectively.

  • Fig. 10.

    Coherence of individual class-specific learned dictionaries for classification of SSPs: (a)–(d) coherence for the complete LD (k = 53), (e)–(h) coherence of overcomplete LDs with k = 100 atoms for all the classes, and (i)–(l) coherence for LDs with k = 300 for classes (left) 1, (left center) 2, (right center) 3, and (right) 4.

All Time Past Year Past 30 Days
Abstract Views 524 23 0
Full Text Views 455 277 56
PDF Downloads 362 183 15