Abstract

Marine icing due to freezing sea spray has been attributed to many safety incidences. Prediction of sea spray icing is necessary for operational safety, design optimization, and structural health. In general, lack of detailed full-scale measurements due to the complexity and costs make validation difficult. The next best option is that of controlled laboratory experiments. The current study is the first study in the field of sea spray icing that investigates the use of new data science technologies like machine learning and feature engineering for the prediction of sea spray icing based on data collected from controlled laboratory experiments. A new prediction model dubbed “Spice” is proposed. Spice is designed “bottom-up” from experimentally collected data, and thus, if the input variables are accurately known, it could be said to be highly accurate within the tested range compared to existing theoretical models. Results from the current study show promising trends; however, more experiments are suggested for increasing the range of confident predictions and reducing the skewness of the training data. Results from spice are compared with five existing models and give icing rates in various conditions in the middle of the spectrum of the other models. It is discussed how validation from two existing full-scale icing measurements from literature proves to be challenging, and more detailed measurements are suggested for the purpose of validation.

Graphical Abstract Figure
Graphical Abstract Figure
Close modal

1 Introduction

Icing on ships and offshore structures occurs from sea spray icing sourced from seawater and atmospheric icing due to precipitation in the form of snow, rain, fog, etc. Sea spray is either wind-generated, where seawater droplets from breaking waves are carried to the structure due to wind, or impact generated, where the sea spray is a result of the impact of the ship or structure with waves. Sea spray icing accounts for 80–90% of all offshore icing incidents [1]. About 40 fishermen are reported to have died in icing-related incidents in the 1960s [2]. Impact-generated sea spray is regarded as the most important water source in dangerous icing events [37]. Icing also affects offshore wind turbines, but is limited to lower levels of the structure and blade tips during azimuthal downward position [8]. Similarly, for other structures like oil rigs, sea spray icing is dominant in the lower parts where sea spray droplets are easily impacted, whereas icing in the upper regions is usually atmospheric icing [9]. Models for predicting sea spray icing are crucial for the optimal design of ships and structures operating in cold climate regions [10].

The 1960s–1980s saw extensive work in gathering icing data for use in the prediction of dangerous icing events [7]. Icing data were previously gathered from around 7000 questionnaire responses across countries but were not publicly available [7,11]. Samuelsen et al. report a few cases from the past where the icing data has been made available. Precise icing rate estimation ideally requires detailed information on metocean parameters and ship characteristics during the icing events, which is lacking in existing datasets [7]. This makes the existing full-scale data not detailed enough for validation of computational results [10].

All existing models use observations or empirical equations to estimate several inputs required for predicting sea spray icing. Existing models rely on just three studies, each on one particular ship [7]. Estimation of spray flux is the weakest link in existing models; for example, the spray flux used in ICEMOD is 10–1000 times less than the spray flux used for RIGICE04, and it is difficult to say which one is better [1]. Spray duration and spray period are also derived from limited observations and different researchers have used different formulations for estimating these [7]. Samuelsen et al. compare predictions using two different formulations of spray flux. Other researchers, like Kulyakhtin, have arbitrarily chosen one of the formulations and suggested that any formulation could be used. The performance of individual empirical formulations is beyond the scope of the current study. Details of some of the existing models can be found in Appendix  A.

Due to different approaches, predictions from existing models show considerable variation [10]. Moreover, the limited comparisons with experimental or full-scale data make it difficult to say which model is the most accurate [12]. ISO 35106 notes that none of the existing models predict sea spray icing on a wide range of vessels or structures and that the Overland model has been traditionally used to estimate icing on small- to medium-sized fishing vessels [12].

Detailed full-scale measurements of sea spray icing are expensive [1,10]. Only a handful of these exist and even fewer as published research [1]. Although some researchers like Samuelsen et al. [7] attempted to validate their models against these datasets; the current study found the datasets not feasible for the purpose of validation. This was also observed by Bodaghkhani et al., who pointed out that the spray parameter measurements are not detailed enough for the validation of computational results [13]. One of the reasons is that the icing is heavily dependent on spray flux, i.e., the amount of water impinging on the freezing surface, which varies greatly with the measurement location and could be vastly different at different locations on the same vessel or structure for the same spray event. Taking icing measurements at a different location than the spray measurements would not suffice for validation. A detailed explanation as to why existing datasets could not be used for precise validation is provided in the further sections.

The next best option is to have validation data from controlled experiments—which have also proved to be time-intensive [14]. Owing to the lack of detailed data for sea spray icing, Deshpande et al. carried out controlled experiments with sea spray icing including 20 tests structured into eight experiments. Each experiment focused on one of eight crucial variables affecting sea spray icing. The experiments simulated sea spray icing conditions in a cold climate laboratory. Seawater from a cooling chamber at 2–6 °C was sprayed onto a plate in a freezing chamber at −5 °C to −15 °C with a nozzle. Wind was simulated with a fan in the freezing chamber. Effects on icing rates by varying air and seawater temperatures, wind speed, material, spray duration and period, salinity, and flux were examined. The weight of the ice accreted on the plate was monitored per second for each hour-long test. For complete details, interested readers are encouraged to refer to Deshpande et al. Data for each experiment was presented. Due to the vast nature of the data gathered, the current study explores the use of new techniques including machine learning (ML) to analyze and make sense of the data from the experiments. Adding to the aforementioned challenges in full-scale icing measurement, gathering comparable and varied full-scale data is difficult, especially as each variable necessitates measurements under diverse weather conditions. Scaling effects from lab experiments must ideally be confirmed via full-scale tests. Future research could explore suitable full-scale measurement techniques.

Applications of machine learning algorithms have grown exponentially in the past few years. The technology makes it possible to make sense of large amounts of data that at first look could fail to show any logical relationships. Machine learning has been recently used in icing severity predictions in the aerospace industry [15,16]. Icing on aircraft blades is purely atmospheric icing where the physical icing process, variables involved, and icing severity are completely different from that of sea spray icing. To the best of the author's knowledge, this is the first study wherein a data-driven machine learning model is implemented for the purpose of sea spray icing predictions.

The current study proposes a sea spray icing prediction method based on the experimental data by Deshpande et al. using machine learning and feature engineering techniques and is an extension of the previous study. Post-publication, ten additional repetitive tests were conducted. First, the data are analyzed, followed by implementing some machine learning models without feature engineering, comparing predictions on unseen data. The advantages and disadvantages of these models are discussed in short. Finally, a model, hereafter called the “SPICE” model (prediction model for Sea sPray ICE), is described and implemented with more complexities added in terms of feature engineering in the chosen machine learning model. SPICE is proposed to serve as a new model for the prediction of sea spray icing based on experimentally validated data. Predictions from SPICE are compared to existing models for nine hypothetical cases to see the effects of certain variables on the predictions. Attempts to validate the model based on two full-scale datasets proved insufficient as the datasets were found to be inadequately detailed. Advantages, limitations, and scope for further improvement of SPICE are discussed. The theory behind individual machine learning models is beyond the scope of the current study, but some elementary details are provided for readers not familiar with the subject. Problems with the use of current spray flux formulations are discussed, and suggestions for future research are presented. The SPICE model is, after further improvements, intended to be deployed as a user-friendly application for general use for icing predictions and as a python package for use in further research.

2 The SPICE Model

The SPICE model comprises four python classes: data processor, transformer, predictor, and feature calculator, along with their methods. An instance of the data processor manages initial raw data processing. An instance of the transformer computes and integrates new features during feature engineering before the prediction model. The predictor class is a container for the trained ML object and is used for predictions and post-processing. The feature calculator is an addition for cases where the original variables required for predictions are unavailable. The usage of all classes is elaborated in appropriate sections.

Assembling together the methods developed for data processing, feature engineering, the machine learning model, and the feature calculator in Secs. 36, respectively, the software architecture for the SPICE model was developed for which the flowcharts are presented in Fig. 1 and Appendix  D. The model is a combination of four different applications:

  1. Spice Data Processor: This is used for preprocessing raw data and arranging the data in a format that can be used further by the machine learning model. Details about the data processor are covered in Sec. 3.

  2. Spice Transformer: This is used for Feature Engineering. The Spice transformer generates additional data (transforms the dataset) for the ML model. It generates several new columns as a combination of existing input and is used in training and predictions. The need for the Spice Transformer has a basis in the nature of the experimental data and is explained in Sec. 3.4. Details of the Spice Transformer are covered in Sec. 4.

  3. Spice Predictor: This is the core ML part of the application which is used while training the ML model and for making predictions from new and unseen input data. While Sec. 3.4 discusses the preliminary ML model evaluation, details of the Spice Predictor are covered in Sec. 5.

  4. Spice Feature Calculator: Any ML model requires identical input variables for prediction during training. In Spice, certain prediction inputs (e.g., spray period) can't be directly measured with onboard instrumentation, necessitating estimation from related factors like wavelength, ship speed, and heading using existing formulations from literature. While making predictions, if obligatory variables are absent from the input, the SPICE Feature Calculator is called to estimate them from available data. The flowchart for the Spice Feature Calculator is shown in Fig. 12, and its details are covered in Sec. 6.

Fig. 1
Flowchart of the Spice model
Fig. 1
Flowchart of the Spice model
Close modal

Figure 1 shows the flowchart of the SPICE algorithm.

3 Data and the Spice Data Processor

This section gives details about the data used in the current study and how it was processed.

3.1 Data Selection, Features, and Target.

The experimental setup consisted of two climate-controlled chambers for simulating sea spray icing conditions. Seawater from the cooling chamber was sprayed onto a plate in the freezing chamber, on which the ice accretion was measured, with a nozzle. The spray duration and period were controlled with a Programmable Logic Control (PLC) valve. Spray flux was varied with the pump pressure, but also varied due to other factors such as spray duration and spray period. Salinity was varied using different batches of seawater and fresh water. The temperatures in the climate chambers were varied and controlled with the refrigeration mechanism. A fan in the freezing chamber simulated variable wind speeds with a regulator. The detailed setup, procedure, and further details of the experiments and the data collected can be found in Deshpande et al.

Data was recorded every second for each hour-long test. The parameters that were controlled, measured, and used in the study; i.e., “features” for the ML models are thermal conductivity of the substrate material, air temperature, seawater temperature, salinity of seawater, spray flux, spray duration, spray period, and wind speed. The weight of accreted ice was measured and the instantaneous icing rate, also the “target” for the ML models, was calculated from 10-min averaged values of the weight of ice accreted.

The refrigeration mechanism introduced air and seawater temperature variations in each test, acting as covariates across all experiments. This hindered drawing conclusions on the effects of some of the tested variables on the icing rate directly from the experimental data. Other variables remained constant per test, enabling the inclusion of data from every second. It was already proven that the effect of the material diminishes after the initial few minutes (15 min kept as reference) [14], allowing the removal of “material” from the list of features if the initial few minutes are neglected. Data from two “bad” experiments [14] along with the initial 15 min of each test were excluded. Additionally, the last 15 min of each test was omitted to prevent errors in the 10-minute moving average icing rate calculation.

Technical limitations made it possible to test a maximum near-surface windspeed of only 6 m/s [14] which was low for application, especially in cases where ships sail into the wind. As a part of preliminary feature engineering, the wind speed was transformed to a variable called windSpeed_mod (umod) from Eq. (1), which describes a log wind profile. The near-surface wind speeds were measured at 9 mm over the plate surface [14]. The experimental setup in Deshpande et al. allows us to assume a roughness length of 0.001 m for a ground with an ice sheet on its surface [17]. The resulting umod is in theory the hypothetical u10 for the experiments, but as the test object was stationary, it is, for the case of this study, assumed to be comparable to the relative wind speed for a moving ship. The validity of this technique is discussed in Sec. 8 [18]
(1)

3.2 Exploratory Data Analysis.

Figure 2 shows the correlations between the variables in the data. Notably, higher wind speeds strongly correlate with higher icing rates with a correlation coefficient of 0.9, while salinity has a minimal impact. Contrary to theoretical expectations [4], experimental data show a slight positive correlation between icing rate and salinity, observed previously by Ryerson [19]. Deshpande et al. acknowledge potential variability even under controlled conditions (Experiment 0). Such variations might contribute to the positive salinity-icing rate correlation, suggesting that errors in salinity would not significantly alter predictions in icing rates. Negative correlations appear for air and water temperatures, and spray period, indicating higher icing rates with lower values. Conversely, positive correlations are evident for wind speed, flux, and spray duration. Some features exhibit a high multicollinearity; e.g., flux shows a high correlation with spray duration and period as these variables directly affect the amount of spray hitting the plate. In a real case, the flux would be also a function of the position on the ship. Air and water temperatures show a relatively high correlation. This could also be true in a real case where the near-surface seawater temperatures vary with the air temperature. The relatively high correlation between wind speed and air temperature can be ignored as they are completely independent variables.

Fig. 2
Correlation heatmap of the data
Fig. 2
Correlation heatmap of the data
Close modal

From the correlation analysis, it could be argued to be acceptable to remove salinity, and the spray period forms the list of features due to the low correlation of salinity with the target variable, icing rate, and the high correlation of the spray period to flux. Exclusion of these features was investigated, and the features were retained for better accuracy (more about this in Sec. 5.2).

Figure 3 shows the pair plot with 100 random samples from the dataset for the target variable instantaneous icing rate (instIcingRate). This gives clarity on how the icing rate is affected by each variable and the extremities of feature values. Notably, air and water temperatures spread quite well, while other features are heavily skewed. This is because features other than the temperatures remained constant for an individual test. More details can be found in Table 1 of Deshpande et al. It is later seen in Sec. 3.4 why feature engineering techniques needed to be implemented due to the skewed data.

Fig. 3
Pair plot for the target variable with 100 random datapoints
Fig. 3
Pair plot for the target variable with 100 random datapoints
Close modal
Table 1

Range of trained data

FeaturesairTempWaterTempSalinityWindspeed (umod)SprayDurationSprayPeriodFlux
Units°C°Cpptm/ssskg/m2/hr
Minimum value in training set−15.920.150.030.000.253.008.96
Maximum value in training set−4.693.3232.9025.151.009.0026.30
FeaturesairTempWaterTempSalinityWindspeed (umod)SprayDurationSprayPeriodFlux
Units°C°Cpptm/ssskg/m2/hr
Minimum value in training set−15.920.150.030.000.253.008.96
Maximum value in training set−4.693.3232.9025.151.009.0026.30

3.3 Final Data Preparation.

The dataset contained no missing values and was shuffled and split randomly into train and test sets (X_train: all features, y_train: values of instIcingRate for corresponding X_train rows, X_test: all features, y_test: values of instIcingRate for corresponding X_test rows) where 80% of the rows were used for training and 20% for testing with instIcingRate as the target variable and other variables shown in Fig. 3 as features. The training dataset is used for training the ML models, whereas the test dataset is used for evaluating the performance of the models. Variable units do not affect predictions, but consistency in units is vital during training and predictions.

Table 1 shows the range of all the features in the training dataset. This is very important, as any predictions made for values outside this range would not be said to be accurate.

Practical usage of the SpiceDataProcesser Class is given in Appendix  E.

3.4 Preliminary Machine Learning Model Evaluation.

Four ML models were assessed using the most commonly used metrics [20]: MAE (Mean Absolute Error: best fit value = 0), MSE (Mean Squared Error: best fit value = 0), and R2 (R square: best fit value = 1) for the train and test datasets are shown in Table 2. The linear regressor model and three decision tree-based models: (DTR) Decision Tree Regressor, LGBM (Light Gradient Boosting Method) Regressor, and XGB (Extreme Gradient Boosting) Regressor [21] were implemented with the help of the Scikit-learn library [22] with default hyperparameters. The scores in Table 2 are obtained by comparing the predicted instIcingRate values with actual values (y_train or y_test) using the metrics module in Scikit-learn.

Table 2

Model scores during preliminary analysis

Train setTest set
MAEr2MSEMAEr2MSE
Linear Regression0.4940.970.5000.4890.9710.488
Decision Tree Regressor0.0001.0000.0000.1080.9940.105
LGBM Regressor0.2150.9940.1030.2150.9940.103
XGB Regressor (Model1)0.1690.9960.0690.1910.9940.095
Train setTest set
MAEr2MSEMAEr2MSE
Linear Regression0.4940.970.5000.4890.9710.488
Decision Tree Regressor0.0001.0000.0000.1080.9940.105
LGBM Regressor0.2150.9940.1030.2150.9940.103
XGB Regressor (Model1)0.1690.9960.0690.1910.9940.095

Tree-based regression models are preferable over linear regression for skewed and multicollinear datasets [23]. Table 2 confirms this where all metrics show better fits with tree-based models compared to Linear Regression. Thus, linear regression, and additionally DTR, for which the training set shows clear overfitting, are ruled out. XGB Regressor performs best with the lowest MAE and MSE for both sets, serving as a reference (Model 1) for further development. Internal algorithm specifics of individual ML models are beyond the scope of the current study.

XGB Regressor's strong scores in Table 2 indicate satisfactory predictions for unseen data. However, guidelines for evaluating models (metric thresholds) are arbitrary and may not be appropriate in practice [24]. A limitation of this model was identified when testing for umod values from 0 to 20, which can be seen in Fig. 6.

Despite good scores, Model 1 showed an unnatural stepped trend in icing rates with wind speed. This is unlikely in reality. Upon examining a tree from the model clarified the reason. Tree-based models categorize data based on rules applied to the feature values, leading to stepped outputs. The severity of this tendency depends on training data, and in this case, has limited practical value. Wind speeds varied from 0–6 m/s in steps of 2, and these were constant throughout an individual test. Transforming wind speed to umod created wider intervals. The model broadly categorized data based on umod values in the training dataset. The model's categorizations along the average of successive values of umod were sufficient for the training data. With consistent umod values in the test dataset, the model makes predictions based on these broad categories, yielding good metrics. However, the model underperforms on dissimilar validation data. For example, the wind speed values present in both, the training and testing dataset, were 0,2,4,6 m/s. The model might end up giving similar results if the wind speed in the validation data is 2.5 or 3.5 m/s, which could be practically untrue. The case with wind speeds is especially important looking at the mean absolute SHAP [25] values in Fig. 4. SHAP (SHapely Additive exPlanations) values show feature influence on predictions, but do not infer any causality [26]. Windspeed (umod) has the highest impact on the prediction, accentuating stepped appearance versus the icing rates. Other variables might show similar trends but to a lesser extent.

Fig. 4
Model 1 mean absolute SHAP values
Fig. 4
Model 1 mean absolute SHAP values
Close modal

4 Method: Feature Engineering and the Spice Transformer

To address predictions between experimental intervals for dataset variables involved exploring additional methods, including feature engineering. Feature engineering encompasses transforming, generating, selecting, or mapping new features from existing variables [27]. An example is when windSpeed was transformed to (umod) using domain knowledge. Two algorithms were employed to create new features: RuleFit [28] and FFX (Fast Function eXtraction) symbolic regression [29]. The methodology and positive effects on model predictions using feature engineering with these two algorithms for regression tasks were demonstrated by Cote 2022 [30].

RuleFit, using a tree-based base ML model, generates decision rules as polynomial-like equations from decision paths [31] comprising combinations of original features. Rules are given coefficient and importance values and selected rules are used as new features while training. The python package for the RuleFit algorithm by Molnar [32] was used in this study. FFX produces a single equation that best fits the dataset. This equation adds another feature to the dataset. The python package for FFX by McConaghy [33] was used in this study. Hyperparameter optimization is used for tuning the learning process. A range of hyperparameters is defined, yielding optimal values for training the ML model. This was done using Hyperopt library in python [34]. Grid search is an alternative approach.

4.1 Implementation of RuleFit.

RuleFit permits either Gradient Boosting Regressor or Random Forest Regressor (RFR) as the base model. Here, RFR from scikit-learn library [22] was chosen. Hyperopt was used to get optimized RFR hyperparameters. RuleFit was then executed on training data, producing over 100 rules with varying importance and coefficients. Rules with absolute coefficients and importance over a threshold of 0.02 were selected, resulting in 9 rules. Rules with single variables marked “linear” were excluded, as they duplicated original variables. Coefficient and importance values are only used for rule selection. Rules were translated into equations for coding, where variables were multiplied. After coding for clarity and interpretability in python, equations (Table 3) were derived. These equations compute new features from X_train variables, temporarily stored for use in FFX.

Table 3

Selected RuleFit Rules and Equations

Rule no.CoefImportanceEquation
00.850.16(windSpeed_mod**1) * (flux**3)
1−0.580.11(airTemp**2) * (windSpeed_mod**1) * (flux**1)
2−1.040.32(airTemp**2) * (salinity**2) * (windSpeed_mod**1) * (flux**1)
30.260.05(airTemp**2) * (waterTemp**1) * (salinity**1) * (windSpeed_mod**1) * (flux**2)
4−0.450.22(airTemp**1) * (windSpeed_mod**1) * (flux**1)
5−0.220.11(airTemp**1) * (windSpeed_mod**1) * (flux**1)
60.830.29(windSpeed_mod**1) * (flux**1)
7−1.640.80(windSpeed_mod**1) * (flux**1)
8−0.340.06(airTemp**2) * (windSpeed_mod**1) * (flux**1)
Rule no.CoefImportanceEquation
00.850.16(windSpeed_mod**1) * (flux**3)
1−0.580.11(airTemp**2) * (windSpeed_mod**1) * (flux**1)
2−1.040.32(airTemp**2) * (salinity**2) * (windSpeed_mod**1) * (flux**1)
30.260.05(airTemp**2) * (waterTemp**1) * (salinity**1) * (windSpeed_mod**1) * (flux**2)
4−0.450.22(airTemp**1) * (windSpeed_mod**1) * (flux**1)
5−0.220.11(airTemp**1) * (windSpeed_mod**1) * (flux**1)
60.830.29(windSpeed_mod**1) * (flux**1)
7−1.640.80(windSpeed_mod**1) * (flux**1)
8−0.340.06(airTemp**2) * (windSpeed_mod**1) * (flux**1)

Duplicate equations were intended to be removed but were erroneously allowed and discovered at a later stage. However, the power of machine learning models could be seen at a later stage in Fig. 5, where the new model does not consider the duplicates as important and does not severely affect the predictions. In any case, duplicates should ideally be removed in future versions.

Fig. 5
Model 2 mean absolute SHAP values
Fig. 5
Model 2 mean absolute SHAP values
Close modal

4.2 Implementation of Fast Function Extraction Symbolic Regression.

The dataframe with existing and new features after RuleFit is used along with y_train to fit the FFX model. The FFX model score was 0.9888. The output of the FFX model is a rather long equation that was cleaned for its use in python and a new feature was calculated. The resulting dataframe is the transformed dataframe that will be used for training the new model. The head of X_train_transformed is shown in Appendix  B for better understanding. Practical usage of the SpiceTransformer Class is given in Appendix  E.

5 Method: Machine Learning Model and the Spice Predictor

Table 2 shows that the XGB Regressor model gave the best results during the preliminary ML model evaluation. The same model is selected for further use. The only difference is that instead of training with X_train, the model is now trained with the X_train_transformed obtained after feature engineering. Additionally, Hyperopt is used for hyperparameter optimization similar to what was done for the RFR during the training of the transformer model. Practical usage of the SpicePredictor Class is given in Appendix  E.

5.1 Results After Feature Engineering.

Table 4 compares the metrics of the model using SPICE (hereafter called Model 2) to Model1. The scores for Model2 show a 17.13% increase in the MSE for the test set. Using SPICE shows improvement in the predictions over XGB with default parameters without feature engineering.

Table 4

Metrics with and without SPICE

Train setTest set
MAEr2MSEMAEr2MSE
Model10.1690.9960.0690.1910.9940.095
Model2 (using SPICE)0.1590.9970.0490.1800.9950.079
% Change6.24−0.1228.265.73−0.117.13
Train setTest set
MAEr2MSEMAEr2MSE
Model10.1690.9960.0690.1910.9940.095
Model2 (using SPICE)0.1590.9970.0490.1800.9950.079
% Change6.24−0.1228.265.73−0.117.13

The mean absolute SHAP values for Model 2 are shown in Fig. 5. The new calculated features (see Table 3 and Fig. 10) are given more importance than the original features. The highest mean SHAP values are seen for Eq. (5) feature which is a multiple of air temperature, wind speed, and flux. These are in fact the variables that are most crucial for the prediction of icing rates.

Figure 6 compares the predictions of Model 2 (using SPICE) to Model 1 for the sensitivity of the models to windspeed (umod) with constant values for the other features. The implemented feature engineering techniques make the model better at predicting values in between the experimented values of features, in this case, the wind speed. Since new features from the transformed dataframe are used for training, the dependence of the model on individual original features is reduced.

Fig. 6
Model1 versus Model 2
Fig. 6
Model1 versus Model 2
Close modal

5.2 Consideration of Dropping Out Less Important Features.

Similar models with the same feature engineering techniques were made to see the effect of the exclusion of certain features that were deemed to be less important by the ML model. The metrics for the different cases are presented in Table 5, where 1 represents inclusion of the feature and 0 represents exclusion.

Table 5

Effect of exclusion of less important features

Sr. nr.Air temp.Water temp.Wind speed (umod)SalinitySpray durationSpray periodFluxNr. featuresTrain setTest set
MAEr2MSEMAEr2MSE
1111111170.1590.9970.0490.1800.9950.079
2111011160.1850.9950.0840.2240.9910.148
3111010150.2220.9930.1130.2490.9900.164
4111000140.2060.9940.0970.2350.9910.150
5101001030.7830.8752.1130.7590.8801.981
Sr. nr.Air temp.Water temp.Wind speed (umod)SalinitySpray durationSpray periodFluxNr. featuresTrain setTest set
MAEr2MSEMAEr2MSE
1111111170.1590.9970.0490.1800.9950.079
2111011160.1850.9950.0840.2240.9910.148
3111010150.2220.9930.1130.2490.9900.164
4111000140.2060.9940.0970.2350.9910.150
5101001030.7830.8752.1130.7590.8801.981

It was seen that the first case presented in Table 5 with all features included, also the previously named Model 2, showed the best metrics. Thus, all the original features, even though of less importance to the model, were retained for better accuracy. The practical significance of this is that variables like salinity, if unknown, are set to a certain default value, and this assumption would not have any severe consequences on the prediction; however, the prediction would be marginally better if the accurate values of these variables are known.

6 Method: The Spice Feature Calculator

Until this stage, the model was tested only on arbitrary values of the original features. However, in a real case, many of these values are not straightforward to know. Air and seawater temperatures, and salinity, can be easily measured. As mentioned in Sec. 3.1, the feature umod is equated with the relative wind speed, which might or might not be directly available depending on the actual case. Spray duration, period, and flux are reliant on complicated ship-ocean dynamics and may not be directly available and current models use some formulations to estimate these. Spice is provided with a feature calculator to estimate the unknown values before the predictor is called. However, a case for minimizing and as far as possible, not using the feature calculator, especially for the estimation of the spray flux is later argued for. Practical usage of the SpiceFeatureCalculator Class is given in Appendix  E.

Formulations for the estimation of some variables are debated in the literature and are discussed in the following sections. This study does not vouch for the use of any of these formulations and thus it is possible to use any of the available formulations by setting the choice as an attribute of the object (see Fig. 1) of the feature calculator. These attributes are used wherever necessary inside individual feature calculator modules. Each attribute has an arbitrarily chosen default value.

6.1 Estimation of Wind-Dependent Parameters

6.1.1 Wind Speed at Given Height.

In the absence of relative wind speed, the wind speed at the given height (uz) is calculated by the wind power law in Eq. (2), where the height above the mean sea level (z) is retrieved from the attributes of the feature calculator and is set of 5 m as a default. The roughness length (z*) for the calculation is used from Eq. (3). In case submodule A1 is called, it is obligatory to have the values of u10, in the absence of which an error is raised regarding insufficient data and the user is notified [35].
(2)
(3)

6.1.2 Relative Wind Speed.

If the relative wind speed is unknown, the ship speed in the direction of the wind is first calculated by Eq. (4), where the ship heading α, if unknown is set to 0 deg (heading straight into wind), which is inconsequential for stationary ships or structures. If the object is stationary, or in case the absolute ship speed is unknown, Vs is set to zero. Further, the relative wind speed is calculated with Eq. (7)
(4)
(5)

6.2 Estimation of Wave-Dependent Parameters

6.2.1 Wave Height.

Literature provides some methods to estimate the significant wave height (Hs) from the weather data using u10, for example, the formulation by Zakrzewski [5], which requires the wind fetch, the Beaufort scale [36], and the formulation by Horjen [37] based on the observations of Jørgensen [38] given in Eq. (6) [36,37]
(6)

The Horjen formulation (default) or the Beaufort scale is available in SPICE and can be selected for wave height estimation. In case of the Beaufort scale, the mean of the range of wave heights for the corresponding u10 is selected.

6.2.2 Wave Period.

Two formulations for the estimation of wave period were found in the literature. The Zakrzewski [5] requires the wind fetch and the Horjen [37] formulation based on Horjen [39] observations given in Eq. (7). In case the local wave period is unavailable, currently, only the Horjen formulation is used, and there is currently no attribute for the selection of wave period equation [37]
(7)

6.2.3 Wavelength.

If known, users can directly input wavelength (λ). Alternatively, equations for linear wave theory [40] estimate λ from wave period (Ts). Based on water depth (d), equations for deep (d/λ ≥ 0.5), intermediate (0.05 < d/λ < 0.5), and shallow waters (d/λ ≤ 0.05) determine λ. Users can set an attribute to choose the corresponding equation in Eq. (8). For intermediate and shallow waters, the value of water depth is vital; otherwise, deep waters are assumed. The same happens in case the wave height or wave period is estimated from the above formulations. These calculations assume open, deep waters with fully developed waves. These formulations assume open, deep waters and fully developed waves. In enclosed waters, like fjords, or near land, smaller and steeper waves can be expected [36]. If the intermediate equation is used, an iterative process starts with an arbitrary 100 m for λ, refining until successive iterations differ by <0.01 m [40]
(8)

6.2.4 Wave Phase Speed.

The wave phase speed is required for calculating the time between successive ship-wave collisions and is calculated with Eq. (9)
(9)

6.2.5 Ship Speed Relative to Waves.

The ship speed in the direction of waves V is first calculated using Eq. (9) with the absolute ship speed Vs and the ship heading with respect to waves, β. β is assumed to be 0 deg if unknown (ship heading straight into waves). This is inconsequential for stationary ships or structures. If ship speed is unavailable, Vs is set to 0. The ship speed relative to waves is then calculated with Eq. (11). The value of β is restricted to 45 deg due to other formulations
(10)
(11)

6.2.6 Time Between Successive Ship-Wave Collisions.

The time between successive ship-wave collisions tcol can be calculated with Eq. (12) knowing the wavelength λ and the ship speed relative to waves Vrw
(12)

6.3 Estimation of Feature Values

6.3.1 Module 1: Air Temperature.

Air temperature is an obligatory input to the feature calculator. The feature calculator checks the input data for airTemp and stores it in the temporary dataframe. If airTemp is missing, an error is raised with a notification.

6.3.2 Module 2: “WaterTemp”.

If the water temperature is available, it is stored in the intermediate dataframe. If missing, it is set to 2.14 which is the mean of the values the model was trained on, and the user is notified of the assumption. The water temperature was not set as obligatory since it was later concluded that variation in water temperature caused a negligible change in the icing rates (see Fig. 7).

Fig. 7
Results of spice compared with other models for various inputs
Fig. 7
Results of spice compared with other models for various inputs
Close modal

6.3.3 Module 3: “Salinity”.

If the salinity is available, it is saved in the intermediate dataframe, else it is set to 32.895—the mode of what the model was trained on, and the user is notified. It is not made obligatory for the same reasons as for water temperature. Additionally, the freezing temperature of saline water is calculated by Eq. (13) [4]
(13)

6.3.4 Module 4: “WindSpeed_mod”(umod).

As discussed in previous sections umod is set equal to the relative wind speed (Vr) for a real case. If the relative wind speed is available, this is stored as umod in the intermediate dataframe. In the absence of knowledge of the relative wind speed, it is estimated through the submodules. Figure 12 can be referred to for the detailed flowchart.

6.3.5 Module 5: “SprayPeriod”.

If available, the spray period is saved in the intermediate dataframe. If absent, the feature calculator checks for the number of sprays per hour (spray frequency) and converts it to spray period or vice versa. If both are missing, submodules assess available data to estimate the time between successive ship-wave collisions (tcol).

Spray is not generated at every ship-wave impact [3]. Zakrzewski, drew from the studies by Panov [41] and Aksjutin [42], proposed that a Medium-Sized Fishing Vessel (MFV) experienced splashing roughly every other wave impact [35]. Horjen adopted this for ICEMOD2 [37], as did Kulyakhtin for MARICE [43]. Lozowski noted spray on every fourth ship-wave impact for a large vessel [3]. This was also used in MINCOG by Samuelsen et al. [7]. Ryerson, on the other hand, suggests spray frequency is independent of wavelength as given in Eq. (14) [44], which is invalid for ship speeds below 1.665 m/s [44]
(14)
Spray period is calculated using Eq. (15) with “Zakrzewski” as the default attribute. Setting the attribute to “Ryerson” uses Eq. (14) for the number of sprays per hour and converts it to spray period. In the absence of ship speed or if the ship speed is below 1.665 m/s, the choice is changed to “Zakrzewski” and the user is notified. If other observations are made for an individual case, it is possible to set the attribute to an integer value n [3,35]
(15)

6.3.6 Module 6: “SprayDuration”.

In the absence of spray duration in the input, it is estimated through other available data through several submodules. The literature presents several formulations for spray duration. Zakrzewski referenced Borisenkov [45], who suggested a constant 2-second spray duration for a particular vessel and proposed an equation for tdur with empirical constant c* based on hull shape and size [35]. Zakrzewski calculated c* as 20.62 from Borisenkov [45], applicable to the specific ship MFV “Iceberg” [5]. Since c* was said to be dependent on hull size and shape, the resulting equation in Eq. (17) would be applicable to ships similar to that of MFV “Iceberg.” Lozowski et al. [3] adapted this equation with c* = 10 for consistency with the USCGC Midgett observations by Ryerson [19]. Horjen proposed Eq. (16) to estimate spray duration for “Norwegian waters” with relative windspeed between vessel and u10 [37]. Samuelsen et al. proposed another new formulation based on observations of the “KV Nordkapp” [7] and is given in Eq. (17) [3,7,35,37].
(16)
(17)

The Zakrzewski, Lozowski, and Samuelsen formulations are available in SPICE and could be used by selecting the appropriate attribute. Any divisions by 0 are set to 0.

6.3.7 Module 7: “Flux” (Spray Flux).

Various studies employ Liquid Water Content (LWC) in the wave impact spray for flux estimation [3,5,46,47]. Forest et al. noted vast spray flux variations due to different LWC formulations, ranging from 6.94 × 10−7 to 44 kg/m2/h at 10 m above the mean sea level and attributed this to the difference in climatic factors and spray-generation mechanisms used while developing the LWC models [47]. Kulyakhtin [1] highlighted flux uncertainty as a key icing prediction challenge, pointing out that the flux formulation in ICEMOD gives 10–1000 times less flux than the one in RIGICE04. It is therefore recommended to calculate or measure the flux on a case-to-case basis. Figures 4, 5, and Table 3 show the critical role of flux in icing predictions. Models reliant on formulations for the estimation of spray flux limit these models only to the type of ship or structure the flux formulation is based on.

If the flux is unknown, the feature calculator offers three flux estimation formulations, with a warning to the user if used. Setting the attribute to “Kulyakhtin” (default) uses the steady version of the spray flux formulation by Kulyakhtin [43] derived from the study by Horjen and Vefsnmo [48], from Eq. (18); where zhv is given in Eq. (19), which makes it valid only for z > Hs/2 [43]
(18)
(19)
Setting the attribute for the spray flux formulation to “Horjen” uses the study by Horjen [37] formulation for spray flux (Eq. (20)). For simplification, only the alternative for α = 45 deg from Horjen is used (invalid for z > 2.17Hs), and the water density is kept constant at 1026 kg/m3 [37]
(20)
Alternatively, setting the attribute to “Ryerson” uses Eq. (21) where the constant 0.074992 is derived from the mean spray flux per second from the 39 events from the Ryerson [19] data, omitting three of the smallest and largest values [19]
(21)

7 Results and Validation

Nine cases are presented for the validation of the Spice model by comparing the results with Spice to that of five existing models: Overland model (ov), Roebber and Mitten model (rm), (modified) Stallabrass model (st_mod), (modified) MARICE model (ku_mod), and the RIGICE04 model (ri). Details of the existing models and their equations can be referred to in Appendix  A. Each case is set up such that only one variable is varied in the individual case while other variables are constant for that case. For cases 1–7, all features were set within the trained range of the model as given in Table 1 (except for the variables that easily default to a certain value, like water temperature and salinity). For cases 8 and 9, the spray period, spray duration, and spray flux are not provided and thus calculated by the feature calculator.

The right segment of Table 6 shows whether icing rates are a function of the tested variable in all the presented models. Diverse model approaches make direct comparison difficult. It was discussed in previous sections how the skewed data and tree-based models lead to a stepped output. This behavior was vastly reduced with feature engineering, though complete removal might require a vast number of additional tests. Hence, Fig. 7 also presents the trend of the predictions with Spice. The purpose of the trend is to give a rough idea of the behavior of the model and a linear fit was considered. The aim eventually however is to ideally have more data so as to completely avoid the stepped output.

Table 6

Inputs for cases 1–9 and information about other models

Test variableIf model is a function of test variable
Case nr.Air temp (°C)Water temp (°C)Salinity (ppt)u10 (m/s)Spray duration (s)Spray period (s)Flux (kg/m2/h)Ship speed (m)ship heading w.r.t wind (deg)z (m)Spiceovr&mst_modku_modri
1−10defaultdefault0 to 201920defaultdefault8u10only if local wind speed unknownyesyesyesyesyes
2−15 to −5defaultdefault81920defaultdefault8airTempyesyesyesyesyesyes
3−10defaultdefault8199 to 24defaultdefault8spray fluxyesnononoyesthrough wave parameters
4−100.15 to 3.3default81920defaultdefault8Water tempyesyesyesyesnono
5−10default30 to 32.851910defaultdefault3salinityyesyesyesyesyesyes
6−10defaultdefault50.25 to 1910defaultdefault3spray durationyesnononoonly for calc of fluxthrough wave parameters
7−10defaultdefault50.253 to 910defaultdefault3spray periodyesnononoonly for calc of fluxthrough wave parameters
8−10defaultdefault10defaultdefaultdefault0 to 10default3ship speedonly if local wind speed unknownnonoyesnono
9−10defaultdefault10defaultdefaultdefault50 to 903ship heading with respect to windonly if local wind speed unknownnonoyesnono
Test variableIf model is a function of test variable
Case nr.Air temp (°C)Water temp (°C)Salinity (ppt)u10 (m/s)Spray duration (s)Spray period (s)Flux (kg/m2/h)Ship speed (m)ship heading w.r.t wind (deg)z (m)Spiceovr&mst_modku_modri
1−10defaultdefault0 to 201920defaultdefault8u10only if local wind speed unknownyesyesyesyesyes
2−15 to −5defaultdefault81920defaultdefault8airTempyesyesyesyesyesyes
3−10defaultdefault8199 to 24defaultdefault8spray fluxyesnononoyesthrough wave parameters
4−100.15 to 3.3default81920defaultdefault8Water tempyesyesyesyesnono
5−10default30 to 32.851910defaultdefault3salinityyesyesyesyesyesyes
6−10defaultdefault50.25 to 1910defaultdefault3spray durationyesnononoonly for calc of fluxthrough wave parameters
7−10defaultdefault50.253 to 910defaultdefault3spray periodyesnononoonly for calc of fluxthrough wave parameters
8−10defaultdefault10defaultdefaultdefault0 to 10default3ship speedonly if local wind speed unknownnonoyesnono
9−10defaultdefault10defaultdefaultdefault50 to 903ship heading with respect to windonly if local wind speed unknownnonoyesnono

Figure 7 reveals considerable disparities in icing rate predictions by different models. Cases 1–3 depict icing rate variations due to three key variables: wind speed, air temperature, and spray flux. All models exhibit increased icing rates with increasing u10 wind speed. Predictions with SPICE lie in the middle of the spectrum of existing models, closest to that of RIGICE04. While most models project zero icing at zero wind speed, R&M and Spice show some minimal icing. For R&M, Eq. (24) yields positive polynomial values at zero predictors, while for Spice, any input flux maintains slight icing. If the flux was unknown, and thereby calculated with the feature calculator, the flux value at zero wind would be zero, giving no icing. It is thus important to know the feature values for an accurate prediction.

Case 2 shows a clear tendency to increase icing rates with decreasing temperatures. However, the dependence of icing on the air temperature varies vastly from model to model. Predictions with Spice lie somewhere in between all models. At higher temperatures, however, Spice shows the highest icing rates.

In Case 3, only Spice predicts increasing icing rates solely with varying spray flux, where spray duration and period are constant. Spice's ability to achieve this is notable, while other models fail to show similar trends. Notably, in the presented models, flux is not considered by the Overland and R&M models, while the rest incorporate various formulations involving wind speed, wave parameters, ship speed, and height from the mean sea level. This limitation could lead to incorrect estimates of icing rates in existing models, particularly in cases where wave parameters remain constant due to steady wind conditions as the theoretically calculated flux at the fore and aft of the vessel would be constant.

Cases 4–7 explore less impactful variables according to the ML model: water temperature, salinity, spray duration, and spray period. Notably, the Overland and R&M models show high sensitivity to water temperature, and despite the Overland model's criticism about this by Makkonen et al. [49], it remains in use [1]. For the other variables, models either neglected them or exhibited minimal changes in icing rates within the tested range for a constant flux. While Spice did show minor variations in predicted icing rates due to these, the absolute change was negligible. The limited range of tested spray duration (max. 1 s compared to 5.57 s observed by Ryerson [19]) underscores the need for more experimental data to confidently predict a wider input range. Similarly, the negligible but opposing trend in icing rates with changing salinity aligns with observations by Ryerson. Importantly, minor measurement errors in these variables have negligible effects on predictions, overshadowed by dominant factors like wind speed, air temperature, and flux.

For cases 8 and 9, only the modified Stallabrass model considers ship speed and heading. RIGICE04 and MARICE are designed for structures, while Overland and R&M, although developed for vessels, disregard ship speed. Spice relies on local wind speed information, specifically umod, which is compared to relative wind speed. In these cases, Spice's feature calculator estimated windspeed (umod), spray duration, period, and flux. This involved estimating wave parameters from u10, leading to values outside tested ranges for spray-related variables. While an ML model performs well within its training range, extrapolating beyond it lacks certainty, especially when validation data are absent. Section 7.2 covers the inability to use existing datasets for validation. Thus, the certainty of absolute icing rates for these cases cannot be determined, yet their trends for both cases, in terms of the direction of the slopes, match the modified Stallabrass model, albeit with substantially different magnitude for case 9. Higher ship speeds result in elevated icing rates, likely due to increased relative wind speed, while lower rates coincide with specific ship headings, associated with reduced relative wind speed and flux. Notably, Spice generally yields higher icing rate values than the modified Stallabrass model.

7.1 Importance of Accurate Knowledge of Flux.

The correct estimation of flux is emphasized in the previous sections as well as by researchers in previous studies [1]. The feature calculator in Spice is provided in situations where a rough estimation of icing rates would suffice. However, for accurate predictions, especially where icing predictions are critical for design and safety purposes, the aim should be to avoid using the feature calculator as much as possible, at least for more important features like wind speed and flux. The feature calculator could be used in case of unknown salinity, water temperature, spray duration, and spray period, which showed comparatively negligible effects on the icing rate, with a minor loss of accuracy.

Figure 8 presents an example where the predictions are heavily dependent on the choice of flux formulation. The three different choices of flux show great variation in the icing rates against wind speed, for a constant air temperature of −10 °C and other parameters set to default. Additionally, many of the values for spray duration, spray period, and especially flux, were seen to be out of range of the tested data—thus calling for additional experimental data to increase the range of confident predictions.

Fig. 8
Variation in icing rate prediction by flux formulation
Fig. 8
Variation in icing rate prediction by flux formulation
Close modal

7.2 Why Currently Available Datasets Could Not Be Used for Precise Validation?.

Due to complexities and expenses, few comprehensive icing and spray datasets are accessible [1,10] and even fewer are published. Two openly accessible datasets containing extensive icing rates and related metocean and ship data were explored for validation attempts in real scenarios.

Ryerson [19] gives a detailed description of 39 spray impacts and the total ice accreted in an icing event that lasted for 100 h. However, using this data directly for validation poses challenges. Ice measurements are given at four locations, but only one table includes flux observations. Real flux experienced at the four locations would clearly have been different due to observed differences in icing thickness, even at the same height. The measured ice thickness often exceeded model predictions in Fig. 9. Despite lasting only about 2 min cumulatively, the 39 spray impacts span 100 h, and icing measurements cover the entire period rather than after each event. Ryerson reported fluxes ranging from 5.22 × 10−4 to 18.62 18.62 kg/m2 per event [19]. With such high fluxes, there is a possibility that some accreted ice might be removed in a dynamic process involving multiple accretion and ablation episodes during an icing event [19]; which cannot be captured due to the nature of Ryerson's data. Another graph shows very high wind speeds throughout the 100-hour period. Although most models mostly neglect wind-generated spray, it could affect the observations over a 100-hour period where wave impact-generated spray lasted for cumulatively only 2 min. The graph for the main deck level at the starboard bulkhead shows ice thickness decreasing when wind speed reduces and air temperatures fall below freezing, suggesting potential effects due to wind-generated flux.

Fig. 9
Total cumulative calculated ice for the Ryerson [19] data
Fig. 9
Total cumulative calculated ice for the Ryerson [19] data
Close modal

The closest comparison could thereby only be achieved with other models as shown in Fig. 9. (The consolidated Ryerson data used as inputs for this calculation is presented in Appendix  C.) This graph converts predicted icing rates to thickness and presents accumulated ice after each spray impact, and eventually, the maximum predicted ice thickness with different models for the Ryerson data. However, even among the models, the closest match to Ryerson data is limited to specific observations at the 02-level starboard bulkhead. This comparison does not provide substantial validation insight for any model, including Spice, in real case scenarios.

In another set of full-scale data, Samuelsen et al., for their MINCOG model, present a case with 37 spray events, of which 14 have an icing rate of 0.33 cm/h [7]. However, scrutiny of the metocean data reveals significant variations in other parameters within these 14 cases. Air temperature varies from −20.2 °C to −1.9 °C, water temperature from −1.3 °C to 6.6 °C, u10 from 9.3 m/s to 25.2 m/s, wave height from 1.5 m to 12 m, and ship speed from 1 m/s to 8.6 m/s, among others, for individual cases. It is unclear whether these presented icing rates are observed or calculated by MINCOG. Nevertheless, the exactly same icing rate for cases with such huge variations of other variables is questionable. It could be a possibility that several parameters are presented in the form of mean values over larger periods of time. In either case, a comparison of the predicted icing rates by any model with this dataset, as presented, did not serve the purpose of validation for a real case.

8 Discussion

8.1 On the Estimation of Flux.

Kulyakhtin names flux estimation as the biggest cause of sea spray icing prediction errors [1]. Figure 8 shows the stark differences in icing rates from three flux estimation formulations. Additionally, Ryerson from their detailed study of flux and icing measurements mentions that the observed flux data were attempted to be written as general equations as a function of other parameters, but most relationships were complex and not significant [19]. This observation clearly discourages the use of any formulations for the estimation of spray flux. Despite this, researchers have been using a handful of available flux formulations. Models using these formulations, often in the form of LWC, would, at most, predict icing rates for similar ships and the exact locations as ones from the flux measurements. Any attempts to generalize these types of ice prediction models over different types and sizes of ships or structures or use them to estimate the distribution of ice would provide inaccurate results. This effectively deems all theoretical models as inaccurate or limited to very specific cases. The problem was visible while attempting to validate models against full-scale data, since the ice thickness and flux were measured at different locations. This calls for a radical change in how future models approach the estimation of spray flux. Computer fluid dynamics (CFD) techniques and real-time measurement of spray flux at the desired locations are suggested for future research.

8.2 Advantages of the Spice Model and Scope for Improvement

  • Being developed “bottom-up” from data from controlled experiments, predictions with Spice could be said to be highly accurate within the tested range if the input variables are accurately known. This claim about the accuracy can be made owing to the high model scores while predicting icing rates for the completely unseen test data. Dominant variables like air temperature, wind speed, and spray flux significantly affect icing rates, with others having minor impacts. One could argue that some additional variables, like humidity, are not captured and could affect the icing rates in full-scale testing. Kulyakhtin proved that the in-cloud humidity can be neglected as other terms have a much higher effect on the mass of accreted ice [50].

  • Spice can be made available as an easy-to-use package in python or even a deployed version which could be integrated with other systems. This makes Spice extremely user-friendly in comparison to existing models.

  • Spice has the ability for continuous improvement by retraining the model with data from additional tests, e.g., for increasing the prediction range or including melting rates.

  • Spice can be used as a transient model and could be integrated with the ships’ systems to make real-time icing predictions. The problem, however, lies in the measurement of flux. Investigations into whether instruments such as rain gauges could be used at critical points for the real-time estimation of spray flux are suggested for future research.

  • Existing models ignore precipitation as a source of flux, and if accurately known, Spice can estimate combined sea spray and rain-induced icing rates, given its broader training on both salt and freshwater experiments.

  • Technical limitations with the generation of high windspeeds in the lab necessitated the preliminary feature engineering technique using the umod variable. Predicted icing rates using this technique in Spice fall between those of other models, and ideally should be validated in full scale.

  • Spice is built “bottom-up” from data from controlled laboratory experiments and is thus experimentally validated in the laboratory. This is already an advantage over existing models; however, full-scale application of the model could be restricted due to pending full-scale validation. This is a topic of ongoing research. Flux estimation using the feature calculator in general remains the biggest uncertainty for accurate predictions. This said, unlike existing models, Spice is developed to be extremely flexible, such that the user is free to use the most appropriate formulation depending on the case or to use input values from any other source or method of choice. In general, the goal should be to make as little use of the feature calculator as possible. Techniques for the reduction of the feature calculator by using Spice in combination with CFD models are part of ongoing research.

  • The original MARICE model incorporates CFD to address the problem of unequal flux at different locations at the same height above sea level but uses existing spray-generation formulations and points out that these spray-generation formulations are the weakest link [1]. Spice, however, offers independence in specifying spray flux, making it possible to combine CFD or measured flux data for more accurate predictions tailored to specific vessel configurations or environmental conditions.

9 Conclusion

The study introduces a novel sea spray icing prediction model utilizing ML, dubbed “Spice”. The model is based on controlled experiments involving 30 sea spray icing cases in a cold climate laboratory. The performance of various open-source ML algorithms was assessed. Given the skewed nature of the experimental data, feature engineering techniques were employed, yielding promising results. Results from Spice were compared with five existing models in nine hypothetical cases, each involving a variation of one parameter. Another case compared results with different flux formulations. The study concludes by addressing challenges associated with using currently available full-scale data for validation purposes. A comprehensive flowchart of Spice and its development is provided for replicability. This is the first study in the field of sea spray icing wherein a data-driven machine learning model is implemented for the purpose of icing predictions.

The Deshpande et al. experiments lacked definitive conclusions on the impact of certain independent variables due to several covariates. Earlier findings established that material does not affect icing rates beyond initial sprays, while icing rates rise with wind speed and decline with temperature [14]. This study concludes that higher spray flux contributes to increased icing rates within the tested range. Furthermore, variations in water temperature, salinity, spray duration, and period exhibit negligible effects on icing rates compared to dominant factors like air temperature, wind speed, and flux.

Some of the findings from this study are presented as follows:

  • Air temperature, wind speed, and spray flux stand out as the dominant parameters affecting icing rate predictions. ML methods confirm their combined importance, aligning with logical reasoning; for example, high wind speeds and flux will not cause icing unless the temperature is below the freezing point.

  • Existing models employ spray duration and period solely to estimate flux, not as independent variables. Independently, these variables show minimal impact on icing rates for a constant time-averaged flux within the tested range.

  • Linear regression gave poor results and the tree-based XGB Regressor predicted the unseen “test data” well. However, the limited number of tests necessitated the use of feature engineering techniques for which RuleFit and FFX gave satisfactory results.

  • Existing models predict vastly different icing rates. The varying approaches and different purposes of the models make it difficult for a direct comparison.

  • Validation of models with existing full-scale experimental data proved to be challenging. It is difficult to measure flux and icing at a detailed level on a large ship. However, it is essential to have more detailed full-scale tests for the validation of prediction models, in the absence of which data from controlled laboratory experiments is the most reliable for the development and validation of prediction models.

  • A “general” model for sea spray icing is one that can predict icing for a wide range of applications (e.g., for both, structures and for all sizes of vessels). Resolving the issue of flux estimation on a case-to-case basis through CFD or real-time methods and feeding this to Spice could ensure its universal applicability making it a potential candidate as a truly “general” model for sea spray icing prediction. This would also ensure prediction of the distribution of ice which is critical for design optimization and research in this direction is already underway.

Acknowledgment

The author would like to thank Dr. Edward Lozowski, Dr. Tom Forest, and Dr. Anton Kulyakhtin for sharing the RIGICE04 model and kind email exchanges. The author would also like to thank Tom Yang from Einblick for their time and effort in introducing me to possibilities with machine learning in the Einblick software and assisting me with the case in the initial stages. Finally, the author would like to thank Vedangini Sahajpal for her machine learning lessons in python.

Conflict of Interest

There are no conflicts of interest.

Data Availability Statement

The datasets generated and supporting the findings of this article are obtainable from the corresponding author upon reasonable request.

Nomenclature

c =

wave phase speed, m/s

d =

water depth, m

e =

saturation vapor pressure at given temp (subscripts a: at temperature Ta; f: at temperature Tf)

g =

acceleration due to gravity (g = 9.81), m/s2

n =

spray generated at every “nth” ship-wave collision, –

z =

height above mean sea level, m

Α =

ship heading with respect to wind (Straight into wind = 0 deg)

M =

spray flux (subscripts refer to the formulation used), kg/m2/h

N =

number of sprays per hour or Spray frequency per hour/hour

S =

salinity of seawater, Ppt

c* =

empirical constant for spray duration, –

z* =

roughness length or roughness coefficient for calculating the wind speeds by the wind power law, m

ta =

air temperature, °C

tcol =

time between successive ship-wave collisions. s

td =

droplet temperature at impact, °C

tdur =

spray duration, s

tf =

freezing point of saline water at salinity S ppt, °C

tw =

seawater temperature, °C

umod =

modified wind speed (windspeed_mod) variable used after preliminary feature engineering (Sec. 3.1), m/s

u10 =

wind speed at height of 10 m above sea surface, m/s

Hs =

significant wave height, m

Ps =

spray period, s

Ts =

wave period, s

Vr =

relative wind speed, m/s

Vrw =

ship speed relative to waves, m/s

Vs =

absolute ship speed, m/s

V =

ship speed in wind direction (into wind is +ve), m/s

V =

ship speed in wave direction (into waves is +ve), m/s

uz, uz0 =

wind speed at heights z and z0 for the calculation of wind speed by wind power law, m/s

IR =

icing rate (subscripts refer to the model used), as per the given equation, but eventually converted to mm/h and kg/m2/h

PR =

predictor for icing rates by overland, m °C/s

β =

ship heading with respect to waves (straight into waves = 0 deg)

Λ =

wavelength, m

ρice =

density of ice (used only for the conversion of icing rates in terms of mass to thickness and set to 900), kg/m3

Appendix A

A.1 Details of Prediction by Existing Models Used for Comparison

Existing models employ diverse approaches and input variables for predicting icing rates. They also serve varying purposes such as RIGICE04 being developed specifically for large stationary structures. This diversity hampers straightforward model comparisons. Nonetheless, selected models were approximated, accommodating coarse assumptions wherever necessary for the evaluation of Spice. Conversion of icing rates among the mentioned models was done to mm/hr and kg/m2/h through straightforward calculations using ice density ρice = 900 kg/m3 [43,51].

A.2 Overland Model

One of the simplest models to date in terms of implementation was proposed by Overland [52]. In spite of the criticism of Makkonen [49] toward this model, it is widely used for mapping regions susceptible to sea spray icing [43]. Its popularity stems probably from its simplicity and straightforward implementation. The Overland model first calculates a predictor (PR) from Eq. (22) from which the icing rate is calculated in cm/hr from Eq. (23) [52]
(A1)
(A2)

A.3 Roebber–Mitten Model

Roebber and Mitten [53] suggested changes to the polynomial of the Overland model using the same predictor (PR) from Eq. (A1). This model, referred to as the Roebber and Mitten (R&M) model, calculates the icing rate in cm/h by Eq. (A3) [52]
(A3)

A.4 Stallabrass Model

The Stallabrass model [4], based on icing measurements on fishing trawlers and tests on cylinders, is simplified here for comparison. The equilibrium freezing surface temperature is assumed equal to the freezing point of seawater at the given salinity from Eq. (13). Stallabrass notes droplet supercooling [4], and Kulyakhtin points out the impact temperature to be anywhere between the air and sea temperatures [43]. For simplification, droplet temperature at impact (td) is set equal to air temperature (ta). Equation (A4) gives the Stallabrass equation for icing rates in mm, and the model is referred to as the “modified” Stallabrass (st_mod) model [4]
(A4)
(A5)

A.5 MARICE Model

MARICE [43] is a time-dependent sea spray icing model utilizing CFD. The original implementation of this model was done with the help of a freezing equation for each cell in the CFD simulation. Unlike Overland and Stallabrass models that rely mainly on metocean and structural parameters, direct comparison of Spice to MARICE is challenging due to the scale at which equations are applied. However, coarse assumptions are made for comparison. The original extensive set of equations are not presented here; readers can refer to the study by Kulyakhtin [43]. For comparison, simplifications include a structural diameter of 0.7 m (matching test plate length by Deshpande et al.), a windward stagnation point at 90 deg, the Schmidt number, originally computed with CFD, is set to 200, and a steady-state version of Eq. (18) is used instead of CFD-calculated flux. Removal of accreted ice in the wave-washing zone is neglected. The results presented here are for coarse comparison, not original MARICE results. The resulting model is termed as “modified” Kulyakhtin model (ku_mod).

A.6 RIGICE04 Model

RIGICE04 was developed for estimating icing on offshore structures [47] and was employed for comparison using its Excel format. To facilitate the comparison, these settings were applied: large vertical column flag set at 1, large column diameter at 0.7 m, large column top height at 15, large column bottom height at 0, and number of large columns at 1. The met data from Sec. 7 for each condition was inputted in the first row, while the second row was set to arbitrary conditions with temperature above freezing point (1 °C) as required by the program. A 1-hour time difference between the rows yielded hourly icing output. Default relative humidity (80%) and atmospheric pressure (1013.25 bar) from the accompanying metocean data file were maintained. Outputs were in kilogram of ice and ice thickness as functions of height above mean sea level, as bottom height was set at 0. Ice thickness was converted to mass/m2 using 900 kg/m3ice density. Only the first and last values are calculated in Rigice04, and the rest linearly spaced for simplicity. Icing rates with the RIGICE04 model are denoted as “IR_ri.”

Appendix B

B.1 X_Train_Transformed

Figure 10 shows the head (first five rows) of the X_train_transformed dataframe for better understanding of how the dataframe looks like after transformation.

Fig. 10
X_train_transformed dataframe head
Fig. 10
X_train_transformed dataframe head
Close modal

Appendix C

C.1 Consolidated Ryerson 1995 Data

Data from multiple tables from Ryerson [19] were used in the calculation of icing rates in Fig. 9. Some data could be used directly, while some had to be converted to the inputs required by Spice. The consolidated data used as inputs to Spice and other models in Fig. 9 are presented in Table 7.

Table 7

Consolidated Ryerson data used for calculations in Fig. 9 

Spray no.Spray durationShip speedWind speedShip headingWave heightShip headingFlux per event
tdurVsumodWith respect to wind (α)HsWith respect to waves (β)
10.878.2410.3090.001.52110.000.09
24.538.2415.4430.000.9150.0018.60
32.808.2415.4430.000.9150.000.03
43.906.1826.7720.001.5230.000.93
51.936.1826.7720.001.5230.000.08
61.575.1529.8620.001.5230.000.14
71.776.1821.6210.001.8320.0018.70
81.176.1820.5910.001.8320.000.13
95.206.1820.5910.001.8320.000.58
1012.336.1820.5910.001.8320.0017.80
115.576.1824.710.002.440.000.08
122.778.2417.5030.000.01
132.308.2417.5030.000.03
143.808.2420.5940.002.1350.000.04
153.678.2420.5940.002.1350.000.57
163.838.2420.5940.002.1350.000.01
174.478.2420.5940.002.1350.000.50
183.438.2420.5940.002.1350.000.02
192.938.2420.5940.002.1350.000.05
203.978.2420.5940.002.1350.000.08
210.478.2419.5660.002.1380.000.00
221.478.2421.6230.002.4440.000.16
237.7011.3325.7420.001.520.001.08
244.8711.3325.7420.001.520.000.90
252.9711.3325.7420.001.520.000.04
264.2311.3325.7420.001.520.000.12
271.0011.3325.7410.001.5250.000.03
282.1711.3325.7410.001.5250.000.04
292.1011.3325.7410.001.5250.000.39
301.5311.3325.7410.001.5250.000.12
313.3011.3325.7410.001.5250.000.99
322.9311.3325.7410.001.5250.000.09
330.8311.3325.7410.001.5250.000.18
342.1011.3325.7410.001.5250.000.02
352.1311.3322.6540.001.5290.000.05
362.5311.3317.5010.001.5290.000.00
373.6011.3317.5010.001.5290.000.08
381.3311.3317.5010.001.5290.000.03
390.871.037.2130.001.5290.000.01
Spray no.Spray durationShip speedWind speedShip headingWave heightShip headingFlux per event
tdurVsumodWith respect to wind (α)HsWith respect to waves (β)
10.878.2410.3090.001.52110.000.09
24.538.2415.4430.000.9150.0018.60
32.808.2415.4430.000.9150.000.03
43.906.1826.7720.001.5230.000.93
51.936.1826.7720.001.5230.000.08
61.575.1529.8620.001.5230.000.14
71.776.1821.6210.001.8320.0018.70
81.176.1820.5910.001.8320.000.13
95.206.1820.5910.001.8320.000.58
1012.336.1820.5910.001.8320.0017.80
115.576.1824.710.002.440.000.08
122.778.2417.5030.000.01
132.308.2417.5030.000.03
143.808.2420.5940.002.1350.000.04
153.678.2420.5940.002.1350.000.57
163.838.2420.5940.002.1350.000.01
174.478.2420.5940.002.1350.000.50
183.438.2420.5940.002.1350.000.02
192.938.2420.5940.002.1350.000.05
203.978.2420.5940.002.1350.000.08
210.478.2419.5660.002.1380.000.00
221.478.2421.6230.002.4440.000.16
237.7011.3325.7420.001.520.001.08
244.8711.3325.7420.001.520.000.90
252.9711.3325.7420.001.520.000.04
264.2311.3325.7420.001.520.000.12
271.0011.3325.7410.001.5250.000.03
282.1711.3325.7410.001.5250.000.04
292.1011.3325.7410.001.5250.000.39
301.5311.3325.7410.001.5250.000.12
313.3011.3325.7410.001.5250.000.99
322.9311.3325.7410.001.5250.000.09
330.8311.3325.7410.001.5250.000.18
342.1011.3325.7410.001.5250.000.02
352.1311.3322.6540.001.5290.000.05
362.5311.3317.5010.001.5290.000.00
373.6011.3317.5010.001.5290.000.08
381.3311.3317.5010.001.5290.000.03
390.871.037.2130.001.5290.000.01

Appendix D

D.1 Supporting Flowcharts: The Spice Development Process and The Feature Calculator

Figure 11 explains the color keys used in the flowcharts for clarity. The flowchart for the Spice Feature Calculator is shown in Fig. 12.

Fig. 11
Color keys for the Spice algorithm
Fig. 11
Color keys for the Spice algorithm
Close modal
Fig. 12
Flowchart for the Spice feature calculator modules
Fig. 12
Flowchart for the Spice feature calculator modules
Close modal

Figure 13 shows the flowchart for the development of the Spice model, including the training, testing, and validation processes. Figure 14 shows the flowchart for the usage of spice as an end-user of the application. The simplicity of the flowchart in Fig. 14 indicates the user-friendliness of the Spice model once it is deployed for use.

Fig. 13
Flowchart for the development of the Spice model
Fig. 13
Flowchart for the development of the Spice model
Close modal
Fig. 14
Flowchart for the use of the Spice model when deployed
Fig. 14
Flowchart for the use of the Spice model when deployed
Close modal

Appendix E

E.1 Usage of Classes in the Spice Model Framework

E.1.1 Use of the SpiceDataProcessor Class

The SpiceDataProcesser Class is provided with a method “data_prep” for obtaining the train-test sets from lab data (Fig. 1). Feature and target names along with the train-test split ratio are set as attributes. The data processer class can be used to retrain the model whenever more experimental data is available.

E.1.2 Use of the SpiceTransformer Class

The SpiceTransformer class inherits properties from scikit-learn's BaseEstimator and TransformerMixin classes. An instance of this class is created, specifying coefficient and importance thresholds for rule selection as attributes. The “fit_transform” method trains the transformer object using X_train and y_train, yielding the transformed training dataframe, X_train_transformed. The object is pickled for future use, and retraining can be done by calling the method again. Due to ML's stochastic nature, rules and FFX equations may vary, even with the same dataset. Once the transformer is trained, the “transform” method is called to transform any dataframe with the original features to a dataframe with features including all equations during testing, validation, and general use.

E.1.3 Use of the SpicePredictor Class

The SpicePredictor class is a container for the ML model for which XGB Regressor was selected after initial assessment. Training is performed by calling the “fit” method with X_train_transformed and y_train. The trained instance can be pickled. Subsequent “fit” calls retrain the model, with potential variations due to ML stochasticity. Predictions are made using the “predict” method (flowchart in Fig. 1). As experiments by Deshpande et al. lacked “no icing” conditions, post-processing predictions are necessary for which the “post_process” method is called. This sets the predicted icing rate to zero if the air temperature exceeds the water freezing point or if spray flux is zero. The condition for zero icing for zero wind was considered, but the possibility of precipitation-induced flux prevents its implementation.

E.1.4 Use of the SpiceFeatureCalculator Class

The SPICE feature calculator class utilizes the “calculate_features” method to handle situations where original feature values are absent during prediction, and rough approximations suffice. With the available input data, the feature calculator stores calculations in an intermediate dataframe and eventually returns a dataframe containing estimated original feature values. The “calculate_features” method encompasses multiple functions organized into modules (as depicted in Fig. 12). Table 8 outlines module divisions. Main modules (Modules 1–7), one for each feature, assess feature values in inputs, while 15 sub-modules are called as needed based on available data. Equations for parameter estimation are detailed in subsequent sections. Missing obligatory inputs triggers an error, notifying users of assumptions made to proceed.

Table 8

Division of feature calculator modules

Feature calculation1, 2, 3, 4, 5, 6, 7
Wind parametersA, A1, A2
Ship parameters with respect to windB, B1, B2
Wave parametersW1, W2, W3, W3.1
Ship parameters with respect to wavesC, C1, C2, C3
Feature calculation1, 2, 3, 4, 5, 6, 7
Wind parametersA, A1, A2
Ship parameters with respect to windB, B1, B2
Wave parametersW1, W2, W3, W3.1
Ship parameters with respect to wavesC, C1, C2, C3

References

1.
Kulyakhtin
,
A.
,
2014
, “
Numerical Modelling and Experiments on Sea Spray Icing
,”
Doctoral thesis
,
Department of Civil and Transport Engineering. Norwegian University of Science and Technology (NTNU), Norway
.
2.
Stallabrass
,
J.
,
1971
, “
Meteorological and Oceanographic Aspects of Trawler Icing off the Canadian East Coast
,”
Mar. Obs.
,
41
(
233
), pp.
107
121
.
3.
Lozowski
,
E. P.
,
Szilder
,
K.
, and
Makkonen
,
L.
,
2000
, “
Computer Simulation of Marine ice Accretion
,”
Philos. Trans. R. Soc. London, Ser. A
,
358
(
1776
), pp.
2811
2845
.
4.
Stallabrass
, “
Trawler Icing—a Compilation of Work Done At N. R. C.
,” Mechanical Engineering Report MD-56, N.R.C. No. 19372 (National Research Council, Ottawa Canada),
1980
.
5.
Zakrzewski
,
W. P.
,
1987
, “
Splashing a Ship With Collision-Generated Spray
,”
Cold Reg. Sci. Technol.
,
14
(
1
), pp.
65
83
.
6.
Kulyakhtin
,
A.
,
Shipilova
,
O.
,
Libby
,
B.
, and
Løset
,
S.
,
2012
, “
Full-Scale 3D CFD Simulation of Spray Impingement on a Vessel Produced by Ship-Wave Interaction
,”
Proceedings of the 21st IAHR Intl. Symposium on Ice (2012)
,
Dalian, China
,
June 11–15
, pp.
1129
1141
.
7.
Samuelsen
,
E. M.
,
Edvardsen
,
K.
, and
Graversen
,
R. G.
,
2017
, “
Modelled and Observed Sea-Spray Icing in Arctic-Norwegian Waters
,”
Cold Reg. Sci. Technol.
,
134
, pp.
54
81
.
8.
Battisti
,
L.
,
Fedrizzi
,
R.
,
Brighenti
,
A.
, and
Laakso
,
T.
,
2006
, “
Sea Ice and Icing Risk for Offshore Wind Turbines
,”
OWEMES 2006
,
Citavecchia, Italy: OWEMES
,
Apr. 20–22
.
9.
Dehghani-Sanij
,
A. R.
,
Dehghani
,
S. R.
,
Naterer
,
G. F.
, and
Muzychka
,
Y. S.
,
2017
, “
Sea Spray Icing Phenomena on Marine Vessels and Offshore Structures: Review and Formulation
,”
Ocean Eng.
,
132
, pp.
25
39
.
10.
Deshpande
,
S.
,
Sæterdal
,
A.
, and
Sundsbø
,
P.-A.
,
2021
, “
Sea Spray Icing: The Physical Process and Review of Prediction Models and Winterization Techniques
,”
ASME J. Offshore Mech. Arct. Eng.
,
143
(
6
), p. 061601.
11.
Brown
,
R. D.
, and
Roebber
,
P.
(
1985
).
The Scope of the Ice Accretion Problem in Canadian Waters Related to Offshore Energy and Transportation
(Canadian Climate Centre Report 85-13.
12.
ISO (International Organization for Standardization)
,
2017
,
ISO 35106: Petroleum and Natural Gas Industries—Arctic Operations—Metocean, Ice, and Seabed Data
.
13.
Bodaghkhani
,
A.
,
Dehghani
,
S.-R.
,
Muzychka
,
Y. S.
, and
Colbourne
,
B.
,
2016
, “
Understanding Spray Cloud Formation by Wave Impact on Marine Objects
,”
Cold Reg. Sci. Technol.
,
129
, pp.
114
136
.
14.
Deshpande
,
S.
,
Sæterdal
,
A.
, and
Sundsbø
,
P.-A.
,
2024
, “
Experiments With Sea Spray Icing: Investigation of Icing Rates.
ASME J. Offshore Mech. Arct. Eng.
,
146
(
1
), p.
011601
. DOI: 10.1115/1.4062255
15.
Li
,
S.
,
Qin
,
J.
, and
Paoli
,
R.
,
2021
, “
Data-Driven Machine Learning Model for Aircraft Icing Severity Evaluation
,”
J. Aerosp. Inf. Syst.
,
18
(
11
), pp.
876
880
.
16.
Li
,
S.
,
Qin
,
J.
,
He
,
M.
, and
Paoli
,
R.
,
2020
, “
Fast Evaluation of Aircraft Icing Severity Using Machine Learning Based on XGBoost
,”
Aerospace
,
7
(
4
), p.
36
.
17.
Miles
,
E. S.
,
Steiner
,
J. F.
, and
Brun
,
F.
,
2017
, “
Highly Variable Aerodynamic Roughness Length (z0) for a Hummocky Debris-Covered Glacier
,”
J. Geophys. Res.: Atmos.
,
122
(
16
), pp.
8447
8466
.
18.
Masters
,
G. M.
,
2004
,
Renewable and Efficient Electric Power Systems
,
Wiley-Interscience
,
Hoboken, NJ
.
19.
Ryerson
,
C. C.
,
1995
, “
Superstructure Spray and Ice Accretion on a Large U.S. Coast Guard Cutter
,”
Atmos. Res.
,
36
(
3–4
), pp.
321
337
.
20.
Chicco
,
D.
,
Warrens
,
M. J.
, and
Jurman
,
G.
,
Jul. 2021
, “
The Coefficient of Determination R-Squared Is More Informative Than SMAPE, MAE, MAPE, MSE and RMSE in Regression Analysis Evaluation
,”
PeerJ. Comput. Sci.
,
7
, pp.
1
24
.
21.
Chen
,
T.
, and
Guestrin
,
C.
,
2016
, “
XGBoost: A Scalable Tree Boosting System
,”
Knowledge Discovery and Data Mining (KDD)
,
San Francisco, CA
,
Aug. 13–17
.
22.
Pedregosa Fabianpedregosa
,
F.
,
Varoquaux
,
G.
,
Gramfort
,
A.
,
Michel
,
V.
,
Thirion
,
B.
,
Grisel
,
O.
,
Blondel
,
M.
, et al
,
2011
, “
Scikit-Learn: Machine Learning in Python
,”
J. Mach. Learn. Res.
,
12
, pp.
2825
2830
.
23.
Chowdhury
,
S.
,
Lin
,
Y.
,
Liaw
,
B.
, and
Kerby
,
L.
, “
Evaluation of Tree Based Regression Over Multiple Linear Regression for Non-Normally Distributed Data in Battery Performance
,”
International Conference on Intelligent Data Science Technologies and Applications (IDSTA)
,
Online
,
Sept. 5–7
.
24.
Colton
,
J. A.
, and
Bower
,
K. M.
,
2002
, “
Some Misconceptions About R2
,”
Int. Soc. Six Sigma Prof., EXTRAOrdinary Sense
,
3
(
2
), pp.
20
22
.
25.
Bodaghkhani
,
A.
,
Muzychka
,
Y. S.
, and
Colbourne
,
B.
,
2018
, “
Three-Dimensional Numerical and Experimental Simulation of Wave Run-Up Due to Wave Impact With a Vertical Surface
,”
ASME J. Fluids Eng.
,
140
(
8
), p.
081225
.
27.
Dong
,
G.
, and
Liu
,
H.
,
2018
,
Feature Engineering for Machine Learning and Data Analytics
,
CRC Press—Taylor and Francis Group
,
London, UK
.
28.
Friedman
,
J. H.
, and
Popescu
,
B. E.
,
2008
, “
Predictive Learning via Rule Ensembles
,”
Source: Ann. Appl. Stat.
,
2
(
3
), pp.
916
954
. doi: .
29.
McConaghy
,
T.
,
2011
, “FFX: Fast, Scalable, Deterministic Symbolic Regression Technology,”
Genetic Programming Theory and Practice IX
,
Springer
,
New York
, pp.
235
260
.
31.
Molnar
,
C.
,
2023
,
Interpretable Machine Learning. A Guide for Making Black Box Models Explainable
. https://christophm.github.io/interpretable-ml-book/.
32.
Molnar
,
C.
, “
Python Package for RuleFit
.”. https://github.com/christophM/rulefit/blob/master/LICENSE, Accessed March 7, 2023.
33.
McConaghy
,
T.
, “
Python Package for FFX
,”
Solido Design Automation Inc
. https://github.com/natekupp/ffx/blob/master/LICENSE. Accessed March 7, 2023.
34.
Bergstra
,
J.
,
Komer
,
B.
,
Eliasmith
,
C.
,
Yamins
,
D.
, and
Cox
,
D. D.
,
2015
, “
Hyperopt: A Python Library for Model Selection and Hyperparameter Optimization
,”
Comput. Sci. Discov.
,
8
(
1
), p.
014008
.
35.
Zakrzewski
,
W. P.
,
1986
, “
Icing of Fishing Vessels. Part I: Splashing a Ship With Spray
,”
Proceedings of the 8th Int. IAHR Symposium on Ice
,
Iowa City
,
Aug. 18–22, 1986
, pp.
179
194
. Vol.
2
,
1986
.
36.
National Meteorological Library and Archive, UK
,
2010
,
The Beaufort Scale
.
37.
Horjen
,
I.
,
2013
, “
Numerical Modeling of Two-Dimensional Sea Spray Icing on Vessel-Mounted Cylinders
,”
Cold Reg. Sci. Technol.
,
93
, pp.
20
35
.
38.
Jørgensen
,
T. S.
1985, “
Sea Spray Characteristics on a Semi-Submersible Drilling Rig—STF60 F 85015
.”
39.
Horjen
,
I.
,
1990
,
Numerical Modeling of Time-Dependent Marine Icing, Anti-Icing and De-Icing
,
Trondheim University
,
Norway
.
40.
Massey
,
B.
, and
Ward-Smith
,
J.
,
2006
,
Mechanics of Fluids, Eighth
,
Taylor & Francis
,
London
.
41.
Kovrova
,
A.
,
Korotun
,
C.
, and
Panov
,
V.
,
1969
,
Gidrometeorologicheskiye Usloviya Obledeneniya na Sudakh v Arkticheskikh Moryakh. (Hydrometeorological Conditions of Icing on Ships in the Arctic Seasa)
,
Arkticheskii i Antarkticheskii Nauchno-lssledovatiel’skii lnstitut
,
Leningrad
.
42.
Aksjutin
,
L. R.
,
1979
, “
Icing of Ships. Leningrad: Sudostroenye (in Russian)
,”
Oidrometeoizdat
,
Leningrad
,
March
.
43.
Kulyakhtin
,
A.
, and
Tsarau
,
A.
,
2014
, “
A Time-Dependent Model of Marine Icing With Application of Computational Fluid Dynamics
,”
Cold Reg. Sci. Technol.
,
104–105
, pp.
33
44
.
44.
Ryerson
,
2013
,
Icing Management for Coast Guard Assets
,
U.S. Army Engineer Research and Development Center
,
Hanover, NH
.
45.
Borisenkov
,
Y. P.
,
1972
, “
On the Theory of Spray Icing of Ships.: Arctic and Antarctic Scientific Research Institute. Trudy 298. Leningrad: Gidrometeoizdat (in Russian), 34-43
.
46.
Samuelsen
,
E. M.
,
2017
, “
Prediction of Ship Icing in Arctic Waters—Observations and Modelling for Application in Operational Weather Forecasting
,” Doctoral thesis,
UiT, The Arctic University of Norway
,
Tromsø, Norway
.
47.
Forest
,
T. W.
,
Lozowski
,
E. P.
, and
Gagnon
,
R.
,
2005
, “
Estimating Marine Icing on Offshore Structures Using RIGICE04
,”
Proceedings of the International Workshop on Atmospheric Icing on Structures (IWAIS)
,
Montreal, Canada
,
June
.
48.
Horjen
,
I.
, and
Vefsnmo
,
S.
,
1985
, “
A Kinematic and Thermodynamic Analysis of Sea Spray (in Norwegian), Offshore Icing—Phase II. Norwegian Hydrodynamic Laboratory (NHI). Report No. STF60 F85014
.”
49.
Makkonen
,
L.
,
Brown
,
R. D.
, and
Mitten
,
P. T.
,
1991
, “
Comments on ‘Prediction of Vessel Icing for Near-Freezing Sea Temperatures’
,”
Weather Forecast
,
6
(
4
), pp.
565
567
.
50.
Kulyakhtin
,
A.
, and
Løset
,
S.
,
2011
, “
Sea Spray Icing : In-Cloud Evaporation. Semi- Analytical and Numerical Investigations
,” the 14th International Workshop on Atmospheric Icing of Structures, Chongqing, China.
May 8–13
.
51.
Makkonen
,
L.
,
1987
, “
Salinity and Growth Rate of Ice Formed by Sea Spray
,”
Cold Reg. Sci. Technol.
,
14
(
2
), pp.
163
171
.
52.
Overland
,
J. E.
,
Pease
,
C. H.
,
Preisendorfer
,
R. W.
, and
Comiskey
,
A. L.
,
1986
, “
Prediction of Vessel Icing
,”
J. Clim. Appl. Meteorol.
,
25
(
12
), pp.
1793
1806
.
53.
Roebber
,
P.
, and
Mitten
,
P.
(
1987
).
Modelling and Measurement of Icing in Canadian Waters
Report (Canadian Climate Centre), No. 87-15.