The MXM R Package, short for the latin 'Mens ex Machina' ( Mind from the Machine ), is a collection of utility functions for feature selection, cross validation and Bayesian Networks. MXM offers many feature selection algorithms focused on providing one or more minimal feature subsets, refered also as variable signatures, that can be used to improve the performance of downstream analysis tasks such as regression and classification, by excluding irrelevant and redundant variables.
In this tutorial we will learn how to use the SES algorithm. For simplicity, we will use a dataset referred as **“The Wine Dataset”**.

**The Wine Dataset** contains the results of a chemical analysis of wines grown in a specific area of Italy. Three types of wine are represented in the 178 samples, with the results of 13 chemical analyses recorded for each sample. Note that the “Type” variable was transformed into a categorical variable.

So, first of all, for this tutorial analysis, we are loading the 'MXM' library and 'dplyr' library for handling easier the dataset.

```
### ~ ~ ~ Load Packages ~ ~ ~ ###
library(MXM)
library(dplyr)
```

And on a next step we are downloading and opening the dataset, defining also the column names.

```
### ~ ~ ~ Load The Dataset ~ ~ ~ ###
wine.url <- "ftp://ftp.ics.uci.edu/pub/machine-learning-databases/wine/wine.data"
wine <- read.csv(wine.url,
check.names = FALSE,
header = FALSE)
head(wine)
```

```
## V1 V2 V3 V4 V5 V6 V7 V8 V9 V10 V11 V12 V13 V14
## 1 1 14.23 1.71 2.43 15.6 127 2.80 3.06 0.28 2.29 5.64 1.04 3.92 1065
## 2 1 13.20 1.78 2.14 11.2 100 2.65 2.76 0.26 1.28 4.38 1.05 3.40 1050
## 3 1 13.16 2.36 2.67 18.6 101 2.80 3.24 0.30 2.81 5.68 1.03 3.17 1185
## 4 1 14.37 1.95 2.50 16.8 113 3.85 3.49 0.24 2.18 7.80 0.86 3.45 1480
## 5 1 13.24 2.59 2.87 21.0 118 2.80 2.69 0.39 1.82 4.32 1.04 2.93 735
## 6 1 14.20 1.76 2.45 15.2 112 3.27 3.39 0.34 1.97 6.75 1.05 2.85 1450
```

```
str(wine)
```

```
## 'data.frame': 178 obs. of 14 variables:
## $ V1 : int 1 1 1 1 1 1 1 1 1 1 ...
## $ V2 : num 14.2 13.2 13.2 14.4 13.2 ...
## $ V3 : num 1.71 1.78 2.36 1.95 2.59 1.76 1.87 2.15 1.64 1.35 ...
## $ V4 : num 2.43 2.14 2.67 2.5 2.87 2.45 2.45 2.61 2.17 2.27 ...
## $ V5 : num 15.6 11.2 18.6 16.8 21 15.2 14.6 17.6 14 16 ...
## $ V6 : int 127 100 101 113 118 112 96 121 97 98 ...
## $ V7 : num 2.8 2.65 2.8 3.85 2.8 3.27 2.5 2.6 2.8 2.98 ...
## $ V8 : num 3.06 2.76 3.24 3.49 2.69 3.39 2.52 2.51 2.98 3.15 ...
## $ V9 : num 0.28 0.26 0.3 0.24 0.39 0.34 0.3 0.31 0.29 0.22 ...
## $ V10: num 2.29 1.28 2.81 2.18 1.82 1.97 1.98 1.25 1.98 1.85 ...
## $ V11: num 5.64 4.38 5.68 7.8 4.32 6.75 5.25 5.05 5.2 7.22 ...
## $ V12: num 1.04 1.05 1.03 0.86 1.04 1.05 1.02 1.06 1.08 1.01 ...
## $ V13: num 3.92 3.4 3.17 3.45 2.93 2.85 3.58 3.58 2.85 3.55 ...
## $ V14: int 1065 1050 1185 1480 735 1450 1290 1295 1045 1045 ...
```

```
colnames(wine) <- c('Type', 'Alcohol', 'Malic', 'Ash',
'Alcalinity', 'Magnesium', 'Phenols',
'Flavanoids', 'Nonflavanoids',
'Proanthocyanins', 'Color', 'Hue',
'Dilution', 'Proline')
```

For this tutorial example, we are going to apply the SES algorithm on the above dataset, using as data and as target variables only continuous variables.

The selection of the appropriate conditional independence test is a crucial decision for the validity and success of downstream statistical analysis and machine learning tasks. Currently the __ `MXM R package`

__ supports numerous tests for different combinations of **target** ( *dependent* ) and **predictor** ( *independent* ) variables. A detailed summary table to guide you through the selection of the most suitable test can be found in **MXM's** reference manual (p.21 *“CondInditional independence tests”* ) here: https://CRAN.R-project.org/package=MXM.
In our example we will use the ** MXMX::SES()**, which is the implementation of the SES algorithm and since we are going to examine only continuous variables, we will use the Fisher's Independence Test.

`dataset`

- A numeric matrix ( or a *data.frame* in case of categorical predictors),
containing the variables for performing the test. The rows should refer to the different samples and columns to the features. For the purposes of this example analysis, we are going to use only the continuous variables, therefore we are removing the “Type” variable from the dataset. Furthermore, we are removing the “Nonflavanoids” variable, because we will use it as target.

```
### ~ ~ ~ Removing The Categorical ('Type') and The Target ('Nonflavanoids') Variables ~ ~ ~ ###
wine_dataset <- dplyr::select(wine,
-contains("Type"),
-contains("Nonflavanoids"))
head(wine_dataset)
```

```
## Alcohol Malic Ash Alcalinity Magnesium Phenols Flavanoids Proanthocyanins
## 1 14.23 1.71 2.43 15.6 127 2.80 3.06 2.29
## 2 13.20 1.78 2.14 11.2 100 2.65 2.76 1.28
## 3 13.16 2.36 2.67 18.6 101 2.80 3.24 2.81
## 4 14.37 1.95 2.50 16.8 113 3.85 3.49 2.18
## 5 13.24 2.59 2.87 21.0 118 2.80 2.69 1.82
## 6 14.20 1.76 2.45 15.2 112 3.27 3.39 1.97
## Color Hue Dilution Proline
## 1 5.64 1.04 3.92 1065
## 2 4.38 1.05 3.40 1050
## 3 5.68 1.03 3.17 1185
## 4 7.80 0.86 3.45 1480
## 5 4.32 1.04 2.93 735
## 6 6.75 1.05 2.85 1450
```

`target`

- The class variable including the values of the target variable. We should provide either a string, an integer, a numeric value, a vector, a factor, an ordered factor or a Surv object.
For the purposes of this example analysis, we are going to use as the dependent variable Nonflavanoids.

```
wine_target <- wine$Nonflavanoids
head(wine_target)
```

```
## [1] 0.28 0.26 0.30 0.24 0.39 0.34
```

This is the first time that we are running the algorithm, so we are going to explain what each **Argument** refers to:

`target`

: The class variable. Provide either a string, an integer, a numeric value, a vector, a factor, an ordered factor or a Surv object. As explained above, this will be the dependent variable. If the target is a single integer value or a string, it has to corresponds to the column number or to the name of the target feature in the dataset. *Here* we choose wine$Nonflavanoids

`dataset`

: The dataset. Provide either a data frame or a matrix. If the dataset (predictor variables) contains missing (NA) values, they will automatically be replaced by the current variable (column) mean value with an appropriate warning to the user after the execution. *Here* we choose the whole wine dataset, except from the Type (categorical) and Nonflavanoids (target) variables.

`max_k`

: The maximum conditioning set to use in the conditional independence test (see The Rule of Thumb in **TIPS** session). Integer, default value is 3.*Here* we choose the default value 3.

`threshold`

: Threshold (suitable values in [0,1]) for assessing p-values significance. Default value is 0.05. *Here* we choose the default value (0.05)

`test`

: The conditional independence test to use. Default value is NULL. *Here* since our dataset includes only continuous features (*remember*: Categorical variable 'Type' was removed) and our dependent variable is also continuous, we choose 'testIndFisher'.

`ini`

: After running SES with some hyper-parameters you might want to run SES again with different hyper-parameters. To avoid calculating the univariate associations (first step of SES) again, you can extract them from the first run of SES and plug them here. This can speed up the second run (and subsequent runs of course) by 50%. *Here*, since we are going to apply the analysis for the first time, we place the argument being equal to **NULL**.

`wei`

: A vector of weights to be used for weighted regression. The default value is NULL. It is not taken into account when robust is set to TRUE. *Here* we are not going to use weights.

`user_test`

: A user-defined conditional independence test (provide a closure type object). Default value is NULL. If this is defined, the “test” argument is ignored. *Here* we have not created a different conditional independence test. We are going to use the 'testIndFisher', as explained in the `test`

argument.

`hash`

: A boolean variable which indicates whether (TRUE) or not (FALSE) to store the statistics calculated during SES execution in a hash-type object. Default value is FALSE. If TRUE a hashObject is produced. In this example, this is the first time that we are running the algorithm and our goal is to apply also a second one. *Here* mind that we are setting `hash`

= **TRUE**, because we want that the univariate associations to be kept for next usage.

`hashObject`

: A List with the hash objects generated in a previous run of SES. Each time SES runs with “hash=TRUE” it produces a list of hashObjects that can be re-used in order to speed up next runs of SES or MMPC. *Here*, since this is the first time that we run the analysis, we are not providing a hashObject.

`ncores`

: How many cores to use. For more details, see the **TIPS** session. *Here*, since we have a small number of samples and features, we are using only one core.

```
### ~ ~ ~ Running SES For First Time ~ ~ ~ ###
ses_default_1st <- MXM::SES(target = wine_target,
dataset = wine_dataset,
max_k = 3,
threshold = 0.05,
test = "testIndFisher",
ini = NULL,
wei = NULL,
user_test = NULL,
hash = TRUE,
hashObject = NULL,
ncores = 1)
```

So, the algorithm run… Let's see what information we can take out of it.

The main purpose of running SES algorithm is to see which variables should be selected as important. The indices of those variables are stored in `selectedVars`

.

```
ses_default_1st@selectedVars
```

```
## [1] 4 5 7 11
```

```
SelectedVars_names<-colnames(wine_dataset[ses_default_1st@selectedVars])
SelectedVars_names
```

```
## [1] "Alcalinity" "Magnesium" "Flavanoids" "Dilution"
```

Here, we see which are the important features.

But can we have them in an order, according to increasing p-value?

Yes, of course!

```
SelectedVars_indecies_Ordered<- ses_default_1st@selectedVarsOrder
SelectedVars_indecies_Ordered
```

```
## [1] 4 5 7 11
```

And what are the equivalent signatures?

```
ses_default_1st@signatures
```

```
## Var1 Var2 Var3 Var4
## [1,] 4 5 7 11
```

But what is the difference of signatures and this:

```
ses_default_1st@queues
```

```
## [[1]]
## [1] 4
##
## [[2]]
## [1] 5
##
## [[3]]
## [1] 7
##
## [[4]]
## [1] 11
```

Well, queues is a list containing a list (queue) of equivalent features for each variable included in selectedVars. An equivalent signature can be built by selecting one and only one feature from each queue. In our case there are no equivalent signatures, hence all variables have no equivalent variable.

What about the statistics corresponding to “p-values”? (Remember: higher values indicates higher association)

```
ses_default_1st@stats
```

```
## [1] 0.4599480 1.2376687 0.4068903 2.9376490 2.4364353 0.4707855 2.2854487
## [8] 0.3049473 1.8515479 0.5539026 2.0418569 0.8210702
```

But what is the difference between @stats and @pvalues?

```
ses_default_1st@pvalues
```

```
## [1] -0.4367580 -1.5255269 -0.3789372 -5.5846020 -4.1451887 -0.4488160
## [7] -3.7504527 -0.2734232 -2.7215042 -0.5441125 -3.1541338 -0.8849622
```

Both lists are almost the same, however, in @pvalues, for each feature included in the dataset, this vector reports the strength of its association with the target in the context of all other variables. Particularly, this vector reports the max p-values found when the association of each variable with the target is tested against different conditional sets. Lower values indicate higher association.

And are all hyper-parameters that were used for the algorithm also stored in this Object?

```
ses_default_1st@max_k # max_k option used in the current run
```

```
## [1] 3
```

```
ses_default_1st@threshold # threshold option used in the current run
```

```
## [1] -2.995732
```

```
ses_default_1st@test # character name of the statistic test used
```

```
## [1] "testIndFisher"
```

Ok! All this is perfect, but can we see the number of univariate associations?

```
ses_default_1st@n.tests
```

```
## [1] 55
```

**Attention** : If you have set `hash = TRUE`

(as we did), then the number of tests performed by SES will be returned. So actually here we are seeing the number of tests performed.

Ok, but in the beginning we said that if we enable `hash`

, then the univariate associations are also stored. How can we get them?

```
ses_default_1st@univ
```

```
## $stat
## [1] -2.079719 3.992701 2.492686 5.014962 -3.467758 -6.410909 -7.953067
## [8] -5.074785 1.851548 -3.557762 -7.324434 -4.260699
##
## $pvalue
## [1] -3.243966 -9.251082 -4.297126 -13.558111 -7.322485 -20.461002
## [7] -29.158820 -13.831160 -2.721504 -7.638713 -25.499442 -10.312081
```

And how quick has all this happened?

```
ses_default_1st@runtime
```

```
## user system elapsed
## 0.12 0.00 0.12
```

As mentioned before, we want to apply the analysis more than once. So, let's say that the goal is to apply the same analysis, but with a different threshold (0.1 instead of 0.05) and with a different max_k (4 instead of 3).
Remember that during the first run, we kept `hash = TRUE`

, therefore the information is stored in the created variable. `ini`

is a supposed to be a list. After running SES (or MMPC) with some hyper-parameters you might want to run SES again with different hyper-parameters. To avoid calculating the univariate associations (first step of SES) again, you can extract them from the first run of SES and plug them here. This can speed up the second run (and subsequent runs of course) by 50%.

In addition, supplying the hashObject also saves computational time as no extra tests will be performed.

```
### ~ ~ ~ Running SES For First Time ~ ~ ~ ###
ses_default_2nd <- MXM::SES(target = wine_target,
dataset = wine_dataset,
max_k = 4,
threshold = 0.1,
test = "testIndFisher",
ini = ses_default_1st@univ,
wei = NULL,
user_test = NULL,
hash = TRUE,
hashObject = ses_default_1st@hashObject,
ncores = 1)
```

This is how we can create the Model including the significant variables. One or more regression models obtained from SES are returned.

```
### ~ ~ ~ glm() Model Estimates Using SES Feature Subset As Predictor Variables ~ ~ ~ ###
first_sign_ses<- ses.model(target = wine_target,
dataset = as.matrix(wine_dataset),
wei = NULL,
sesObject = ses_default_1st, #
test = 'testIndFisher')
```

The model(s) created is/are stored in the \(mod\) variable. We can get the summary of the model (looks like the summary of many other linear models), or ask for the signatures ( \(signature\) ). The BIC score is also returned for each value of the signature.

```
ses_model_summary <- first_sign_ses$mod
signature <- first_sign_ses$signature
signature
```

```
## Alcalinity Magnesium Flavanoids Dilution bic
## 4.0000 5.0000 7.0000 11.0000 -288.7666
```

On the example above, we run the analysis for a continuous variable (Nonflavanoids). What would happen if we choose to use as target the “Type” variable, which is the categorical variable referring to the three types of wine that are represented?

Since the variable is categorical - and more specific it is a factor with more than two levels (unordered) -and the features are continuous, according to **MXM's** reference manual (p.21 *“CondInditional independence tests”* ) here: https://CRAN.R-project.org/package=MXM, we should use the Multinomial logistic regression ( 'testIndMultinom' ).

In this step, we keep the whole dataset, in order to show how to use the algorithm also without subtracting the initial matrix.

```
### ~ ~ ~ Taking The Whole Dataset ~ ~ ~ ###
wine_dataset <- wine
head(wine_dataset)
```

```
## Type Alcohol Malic Ash Alcalinity Magnesium Phenols Flavanoids Nonflavanoids
## 1 1 14.23 1.71 2.43 15.6 127 2.80 3.06 0.28
## 2 1 13.20 1.78 2.14 11.2 100 2.65 2.76 0.26
## 3 1 13.16 2.36 2.67 18.6 101 2.80 3.24 0.30
## 4 1 14.37 1.95 2.50 16.8 113 3.85 3.49 0.24
## 5 1 13.24 2.59 2.87 21.0 118 2.80 2.69 0.39
## 6 1 14.20 1.76 2.45 15.2 112 3.27 3.39 0.34
## Proanthocyanins Color Hue Dilution Proline
## 1 2.29 5.64 1.04 3.92 1065
## 2 1.28 4.38 1.05 3.40 1050
## 3 2.81 5.68 1.03 3.17 1185
## 4 2.18 7.80 0.86 3.45 1480
## 5 1.82 4.32 1.04 2.93 735
## 6 1.97 6.75 1.05 2.85 1450
```

We will not create a different matrix for the target. As mentioned above, we are going to use the Type variable, but please… be patient…

```
### ~ ~ ~ Running SES For Categorical Variable ~ ~ ~ ###
wine[, 1] <- as.factor(wine[, 1])
ses_default_1st <- MXM::SES(target = 1, ## Defining as target the 1st column
dataset = wine,
max_k = 3,
threshold = 0.05,
test = "testIndMultinom",
ini = NULL,
wei = NULL,
user_test = NULL,
hash = TRUE,
hashObject = NULL,
ncores = 1)
```

So, the algorithm run once again… Let's see what information we can take out of it.

The main purpose of running SES algorithm is to see which variables should be selected as important. The indices of those variables are stored in `selectedVars`

.

```
ses_default_1st@selectedVars
```

```
## [1] 2 4 8 11 13 14
```

```
SelectedVars_names<-colnames(wine_dataset[ses_default_1st@selectedVars])
SelectedVars_names
```

```
## [1] "Alcohol" "Ash" "Flavanoids" "Color" "Dilution"
## [6] "Proline"
```

We could again ask for each variable of the output, as in the first implementation of the algorithm. But we will concentrate on the returned signature.

```
SelectedVars_indecies_Ordered<- ses_default_1st@selectedVarsOrder
SelectedVars_indecies_Ordered
```

```
## [1] 8 14 11 2 13 4
```

And what are the equivalent signatures?

```
ses_default_1st@signatures
```

```
## Var1 Var2 Var3 Var4 Var5 Var6
## [1,] 2 4 8 11 13 14
```

And how quick has all this happened?

```
ses_default_1st@runtime
```

```
## user system elapsed
## 5.06 0.00 5.11
```

This is how we can create the Model including the significant variables. One or more regression models obtained from SES are returned.

```
### ~ ~ ~ glm() Model Estimates Using SES Feature Subset As Predictor Variables ~ ~ ~ ###
second_sign_ses<- ses.model(target = as.matrix(wine_target),
dataset = as.matrix(wine_dataset),
wei = NULL,
sesObject = ses_default_2nd, #
test = 'testIndFisher')
```

The model(s) created is/are stored in the \(mod\) variable. We can get the summary of the model (looks like the summary of many other linear models), or ask for the signatures ( \(signature\) ). The BIC score is also returned for each value of the signature.

```
ses_model_summary <- second_sign_ses$mod
signature <- second_sign_ses$signature
signature
```

```
## Ash Alcalinity Phenols Color bic
## 4.000 5.000 7.000 11.000 -265.697
```

```
str(signature)
```

```
## Named num [1:5] 4 5 7 11 -266
## - attr(*, "names")= chr [1:5] "Ash" "Alcalinity" "Phenols" "Color" ...
```

If the analysis has to be done using permutations (when few observations are available), then this could be applied like this:

```
### ~ ~ ~ Permutations ~ ~ ~ ###
library('MXM')
permutation_ses_model <- MXM::perm.ses(wine_target,
wine_dataset,
R = 999, # The number of permutations to use. The default value is 999
max_k = 3,
threshold = 0.05,
test = NULL,
ini = NULL,
wei = NULL,
user_test = NULL,
hash = FALSE,
hashObject = NULL,
ncores = 1)
```

There is a solution to avoid all permutations. As soon as the number of times the permuted test statistic is more than the observed test statistic is more than 50 (if threshold = 0.05 and R = 999), the p-value has exceeded the significance level (threshold value) and hence the predictor variable is not significant. There is no need to continue doing the extra permutations, as a decision has already been made.

Cross-validation (CV) is a technique for validating models which can assess the generalization of the results provided by a statistical analysis on independent datasets. It is a technique widely used in regression and classification approaches and provides a solution to the problem of choosing the right tuning-parameters in a prediction model. In the case of SES algorithm, the tuning parameters would be `max_k`

and `threshold`

. Therefore, in MXM package, a function named `cv.ses()`

can be found. The function performs a k-fold cross-validation for identifying the best values for the SES tuning parameters. In case the user does not know what tuning parameters to use, we suggest the use of this function as Step Zero, before running the algorithm.

For the purposes of this tutorial, we are going to use the same dataset and as target we will use again the Categorical variable “Type”, transforming our problem into a Classification problem by trying to predict the label of the each row, provided by the first column.

So, let's see how this algorithm works.

Before running the `cv.ses()`

function we are going to explain what each **Argument** refers to:

`target`

: The target or class variable as in SES. The difference is that it cannot accept a single numeric value, an integer indicating the column in the dataset. *Here* we choose wine$Type, which is a Categorical variable.

`dataset`

: The dataset object as in SES. *Here* we choose the whole wine dataset.

`wei`

: A vector of weights to be used for weighted regression.*Here* we choose the default value **NULL**.

`kfolds`

: The number of the folds in the k-fold Cross Validation (integer). *Here* we choose 5. Since the dataset includes 178 observations, by splitting into 5 folds, each fold will have circa 35 rows.

`folds`

: The folds of the data to use (a list generated by the function generateCVRuns TunePareto). If NULL the folds are created internally with the same function. *Here* we will let the algorithm run alone the internal function, by choosing **NULL**.

`alphas`

: A vector of SES thresholds hyper parameters to be used in CV. *Here* choose the values 0.1, 0.05, 0.01.

`max_ks`

: A vector of SES max_ks hyper parameters to be used in CV. *Here* we choose the values 3, 4, 5.

`task`

: A character (“C”, “R” or “S”). It can be “C” for classification (logistic, multinomial or ordinal regression), “R” for regression (robust and non robust linear regression, median regression, (zero inflated) poisson and negative binomial regression, beta regression), “S” for survival regression (Cox, Weibull or exponential regression). *Here* we choose “C”, since we are applying a classification technique on the dataset, by placing as target the “Type” categorical variable.

`metric`

: A metric function provided by the user. If NULL the following functions will be used: auc.mxm, mse.mxm, ci.mxm for classification, regression and survival analysis tasks, respectively. The metric functions that are currently supported are:

auc.mxm: “area under the receiver operator characteristic curve” metric, as provided in the package ROCR.

acc.mxm: accuracy metric.

fscore.mxm: F score for binary logistic regression.

euclid_sens.spec.mxm: Euclidean norm of 1 - sensitivity and 1 - specificity for binary logistic regression.

acc_multinom.mxm: accuracy or multinomial logistic regression.

mse.mxm: -1 * (mean squared error), for robust and non robust linear regression and median (quantile) regression.

ci.mxm: 1 - concordance index as provided in the rcorr.cens function from the Hmisc package. This is to be used with the Cox proportional hazards model only.

ciwr.mxm concordance index as provided in the rcorr.cens function from the Hmisc package. This is to be used with the Weibull regression model only.

poisdev.mxm: Poisson regression deviance.

nbdev.mxm: Negative binomial regression deviance.

binomdev.mxm: Negative binomial regression deviance.

ord_mae.mxm: Ordinal regression mean absolute error.

If you are certain about the meaning of one metric, then we suggest to use this one. *Here* we choose the “accuracy metric for multinomial” (acc_multinom.mxm).

`modeler`

: A modeling function provided by the user. If NULL the following functions will be used: glm.mxm, lm.mxm, coxph.mxm for classification, regression and survival analysis tasks, respectively. The modelling functions that are currently supported are:

glm.mxm: fits a glm for a binomial family (Classification task).

lm.mxm: fits a linear model model (stats) for the regression task.

coxph.mxm: fits a cox proportional hazards regression model for the survival task.

weibreg.mxm: fits a Weibull regression model for the survival task.

rq.mxm: fits a quantile (median) regression model for the regression task.

lmrob.mxm: fits a robust linear model model for the regression task.

pois.mxm: fits a poisson regression model model for the regression task.

nb.mxm: fits a negative binomial regression model model for the regression task.

multinom.mxm: fits a multinomial regression model model for the regression task.

ordinal.mxm: fits an ordinal regression model model for the regression task.

beta.mxm: fits a beta regression model model for the regression task.

The predicted values are transformed into R using the logit transformation. This is so that the “mse.mxm” metric function can be used. In addition, this way the performance can be compared with the regression scenario, where the logit is applied and then a regression model is employed.

In case you decide not to choose one of the default modelers, we suggest to you to be certain about the way your chosen one works. *Here* we choose multinom.mxm, since we search for a multinomial regression model.

`ses_test`

: A function object that defines the conditional independence test used in the SES function (see **3.1.1 Selecting Appropriate Conditional Independence Test**). If NULL, “testIndFisher”, “testIndLogistic” and “censIndCR” are used for classification, regression and survival analysis tasks, respectively. Not all tests can be included here. “testIndClogit”, “testIndMVreg”, “testIndIG”, “testIndGamma”, “testIndZIP” and “testIndTobit” are not available yet.

*Here* we choose 'testIndMultinom'', as explained in **4.1.1 Selecting Appropriate Conditional Independence Test**

`robust`

: A boolean variable which indicates whether (TRUE) or not (FALSE) to use a robust version of the statistical test if it is available. It takes more time than a non robust version but it is suggested in case of outliers. *Here*, we set it as **FALSE**.

`ncores`

: How many cores to use. For more details, see the **TIPS** session. *Here*, since we have a small number of samples and features, we are using only one core.

```
### ~ ~ ~ Cross-Validation ~ ~ ~ ###
library('MXM')
cv_ses_model <- MXM::cv.ses(target = wine[, 1], # Using the 1st column as target
dataset = wine[, -1],
wei = NULL,
kfolds = 5,
folds = NULL,
alphas = c(0.1, 0.05, 0.01),
max_ks = c(3, 4, 5),
task = "C",
metric = acc_multinom.mxm, # Note that we are passing it as a function and not as a character
modeler = multinom.mxm, # Note that we are passing it as a function and not as a character
ses_test = "testIndMultinom",
ncores = 1)
```

The main purpose of running Cross - Validation SES algorithm is to see which hyperparameters should be selected for tuning the algorithm. The best configuration (pair of max_k and p-value) is stored under the variable `$best_configuration`

.

```
cv_ses_model$best_configuration
```

```
## $id
## [1] 3
##
## $a
## [1] 0.1
##
## $max_k
## [1] 3
```

```
cv_ses_model$best_performance
```

```
## [1] 0.9326984
```

As we see, the id, the significance threshold (“a”) and the best max_k are returned.

But what if we want to see how good a specific configuration did during the Cross-Validation?

Yes yes, we can ask for the performance more specifically.

Here we check separately the results for the first configuration.

```
cv_ses_model$cv_results_all[[1]]$configuration #this configuration we are examining
```

```
## $id
## [1] 1
##
## $a
## [1] 0.1
##
## $max_k
## [1] 5
```

```
cv_ses_model$cv_results_all[[1]]$performances # those are the performances.
```

```
## [1] 0.8888889 0.8611111 1.0000000 0.9714286 0.9142857
```

```
cv_ses_model$cv_results_all[[1]]$signatures # signatures created by this configuration
```

```
## [[1]]
## Var1 Var2 Var3
## [1,] 7 10 13
##
## [[2]]
## Var1 Var2 Var3 Var4
## [1,] 1 4 7 13
## [2,] 1 10 7 13
## [3,] 1 11 7 13
## [4,] 1 4 7 3
## [5,] 1 10 7 3
## [6,] 1 11 7 3
##
## [[3]]
## Var1 Var2 Var3 Var4 Var5
## [1,] 1 3 7 10 13
##
## [[4]]
## Var1 Var2 Var3 Var4
## [1,] 1 7 11 13
##
## [[5]]
## Var1 Var2 Var3 Var4
## [1,] 1 7 12 13
## [2,] 1 7 10 13
```

**Note:** By asking the best performance `$best_performance`

, we get the mean value of all folds of the best configuration. If we ask for the performances of the best configuration (here it is the first configuration) `$cv_results_all[[1]]$performances`

we get all performance values of each fold (here 5).

```
cv_ses_model$best_performance
```

```
## [1] 0.9326984
```

```
index<-cv_ses_model$best_configuration$id
cv_ses_model$cv_results_all[[index]]$performances
```

```
## [1] 0.8888889 0.8888889 1.0000000 0.9714286 0.9142857
```

```
mean(cv_ses_model$cv_results_all[[index]]$performances)
```

```
## [1] 0.9326984
```

There is a rule of thumb that we are suggesting. If the sample size is small, take as max_k the smallest integer near N/10, where N is the number of samples.

```
#
N = dim(wine_dataset)[1]
suggested_max_k = floor(N/10)
N
```

```
## [1] 178
```

```
suggested_max_k
```

```
## [1] 17
```

However, an other approach would be to run the cross-validation function, described above. In that case, the best `max_k`

will be returned together with the best `threshold`

(see **6. Cross - Validation**)

This those kind of analyses may be applied on very big dataset, the SES algorithm provides the opportunity to be run on more than one cores. We suggest that you use more cores, when the datasets are big, but remember that the best run time will always be dependent on the architecture of the computer that is used.

Now you are ready to run your own analysis using MXM::SES algorithm!

Thank you for your attention.

Hope that you found this tutorial helpful.

All analyses have been applied on:

```
sessionInfo()
```

```
## R version 3.6.1 (2019-07-05)
## Platform: x86_64-w64-mingw32/x64 (64-bit)
## Running under: Windows 10 x64 (build 17763)
##
## Matrix products: default
##
## locale:
## [1] LC_COLLATE=C
## [2] LC_CTYPE=English_United States.1252
## [3] LC_MONETARY=English_United States.1252
## [4] LC_NUMERIC=C
## [5] LC_TIME=English_United States.1252
##
## attached base packages:
## [1] stats graphics grDevices utils datasets methods base
##
## other attached packages:
## [1] dplyr_1.0.2 MXM_1.4.9
##
## loaded via a namespace (and not attached):
## [1] tidyr_1.1.2 foreach_1.5.0 jsonlite_1.7.0
## [4] splines_3.6.1 R.utils_2.10.1 Formula_1.2-3
## [7] ucminf_1.1-4 statmod_1.4.34 latticeExtra_0.6-29
## [10] slam_0.1-46 numDeriv_2016.8-1.1 pillar_1.4.5
## [13] backports_1.1.8 lattice_0.20-38 quantreg_5.61
## [16] glue_1.4.1 digest_0.6.25 RColorBrewer_1.1-2
## [19] checkmate_2.0.0 minqa_1.2.4 colorspace_1.4-1
## [22] htmltools_0.5.0 Matrix_1.2-17 R.oo_1.24.0
## [25] Rfast2_0.0.6 conquer_1.0.2 pkgconfig_2.0.3
## [28] broom_0.7.0 SparseM_1.78 relations_0.6-9
## [31] purrr_0.3.4 scales_1.1.1 RANN_2.6.1
## [34] jpeg_0.1-8.1 lme4_1.1-23 MatrixModels_0.4-1
## [37] htmlTable_2.0.1 tibble_3.0.2 generics_0.0.2
## [40] ggplot2_3.3.2 ellipsis_0.3.1 geepack_1.3-1
## [43] nnet_7.3-12 survival_3.2-3 magrittr_1.5
## [46] crayon_1.3.4 evaluate_0.14 ordinal_2019.12-10
## [49] R.methodsS3_1.8.1 doParallel_1.0.15 R.cache_0.14.0
## [52] nlme_3.1-140 MASS_7.3-51.4 foreign_0.8-71
## [55] R.rsp_0.44.0 RcppZiggurat_0.1.5 tools_3.6.1
## [58] coxme_2.2-16 data.table_1.13.0 energy_1.7-7
## [61] lifecycle_0.2.0 matrixStats_0.55.0 stringr_1.4.0
## [64] munsell_0.5.0 cluster_2.1.0 Rfast_2.0.0
## [67] compiler_3.6.1 rlang_0.4.7 grid_3.6.1
## [70] nloptr_1.2.2.2 iterators_1.0.12 rstudioapi_0.11
## [73] bigmemory.sri_0.1.3 htmlwidgets_1.5.1 visNetwork_2.0.9
## [76] base64enc_0.1-3 boot_1.3-22 codetools_0.2-16
## [79] gtable_0.3.0 sets_1.0-18 markdown_1.1
## [82] bigmemory_4.5.36 R6_2.4.1 gridExtra_2.3
## [85] knitr_1.29 bdsmatrix_1.3-4 Hmisc_4.4-1
## [88] stringi_1.4.3 parallel_3.6.1 Rcpp_1.0.3
## [91] vctrs_0.3.4 rpart_4.1-15 png_0.1-7
## [94] tidyselect_1.1.0 xfun_0.15
```

Lagani, V., Athineou, G., Farcomeni, A., Tsagris, M. & Tsamardinos, I. Feature Selection with the R Package MXM: Discovering Statistically Equivalent Feature Subsets. J. Stat. Softw. 80, 1-25 (2017).

I. Tsamardinos, V. Lagani and D. Pappas (2012). Discovering multiple, equivalent biomarker signatures. In proceedings of the 7th conference of the Hellenic Society for Computational Biology & Bioinformatics - HSCBB12.

Tsamardinos, Brown and Aliferis (2006). The max-min hill-climbing Bayesian network structure learning algorithm. Machine learning, 65(1), 31-78.

Brown, L. E., Tsamardinos, I., & Aliferis, C. F. (2004). A novel algorithm for scalable and accurate Bayesian network learning. Medinfo, 711-715.

Tsamardinos, I., Aliferis, C. F., & Statnikov, A. (2003). Time and sample efficient discovery of Markov blankets and direct causal relations. In Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining (pp. 673-678). ACM.