“TreeSearch” can be used to search for trees that are optimal under user-specified criteria (e.g. Hopkins & St. John, 2021).

All that is needed is to provide a function that will return a score for a given tree (and dataset).

First we’ll load the necessary R libraries:

`library("TreeTools", quietly = TRUE, warn.conflict = FALSE)`

`## Warning: package 'TreeTools' was built under R version 4.1.3`

```
library("TreeSearch")
# Plot trees nicely
<- function(tree, ...) {
PlotTree <- par(mar = rep(0, 4), cex = 0.9)
oPar plot(tree)
par(oPar)
}
```

We’ll start with a very simple case: aiming to minimise the total cophenetic index [TIC; Mir, Rosselló, & Rotger (2013)] of a tree. The TCI is a measure of tree balance; it does not depend on an input dataset. All we need to do is to write a `TreeScorer`

function.

This function will be sent the `parent`

and `child`

nodes for each edge in a tree, and a (here, empty) `dataset`

parameter. The function should return a score to be minimized.

Here, we can use a copy of our starting tree as a template, to be populated with the rearranged `parent`

and `child`

vectors:

```
<- PectinateTree(8)
tree PlotTree(tree)
```

```
<- function(parent, child, dataset) {
TCIScore $edge <- cbind(parent, child)
treeTotalCopheneticIndex(tree)
}
TCIScore(tree$edge[, 1], tree$edge[, 2], NA)
```

`## [1] 56`

Now we can use our scorer for tree search. We need to initialize some parameters to `TreeSearch()`

with null values: `dataset = EmptyPhyDat(tree)`

sends a blank dataset (as our tree scorer doesn’t require any data); we set `InitializeData = DoNothing`

and `CleanUpData = DoNothing`

because we don’t need to do anything to `dataset`

before it is sent to `TreeScorer()`

.

```
<- TreeSearch(tree, dataset = EmptyPhyDat(tree),
result InitializeData = DoNothing, CleanUpData = DoNothing,
TreeScorer = TCIScore,
maxIter = 50L, maxHits = 10L,
verbosity = 1L)
```

`## - Performing tree search. Initial score: 56`

`## - Final score 33 found 6 times after 50 rearrangements.`

`PlotTree(result)`

Let’s make things slightly more complex, and try to find the tree that is most different from a starting tree. Notice that `TreeSearch`

aims to *minimize* the output of `TreeScorer()`

, so we negate the tree *distance* (which we aim to maximize) before returning it.

```
<- BalancedTree(8)
startTree
<- function(parent, child, dataset) {
DistanceScore <- startTree
tmpTree $edge <- cbind(parent, child)
tmpTree<- TreeDist::ClusteringInfoDistance(startTree, tmpTree)
distance # Return:
-distance
}
<- TreeSearch(RandomTree(8, root = TRUE), dataset = EmptyPhyDat(tree),
result InitializeData = DoNothing, CleanUpData = DoNothing,
TreeScorer = DistanceScore,
maxIter = 50L, maxHits = 10L,
verbosity = 1L)
```

`## - Performing tree search. Initial score: -7.18872187554087`

`## - Final score -7.87744375108174 found 1 times after 50 rearrangements.`

```
par(mfrow = c(1, 2))
PlotTree(startTree)
PlotTree(result)
```

Now we consider a more complex case in which a scorer must undergo a time-consuming initialization before tree search can begin, and must be safely destroyed once tree search has completed.

We start by defining an initialization function, which will create a new Morphy object (Brazeau, Smith, & Guillerme, 2017) for each character in a phylogenetic dataset:

```
<- function (dataset) {
IWInitMorphy attr(dataset, 'morphyObjs') <-
lapply(PhyToString(dataset, byTaxon = FALSE, useIndex = FALSE,
concatenate = FALSE),
SingleCharMorphy)
# Return:
dataset }
```

To release memory back to the operating system, we must destroy each Morphy object once we’re finished with it:

```
<- function (dataset) {
IWDestroyMorphy vapply(attr(dataset, 'morphyObjs'), UnloadMorphy, integer(1))
}
```

Now we can write our tree scoring function, which will return the ‘fit’ under implied weights (Goloboff, 1993).

Note that we need to specify some extra parameters: `concavity`

is the *k* value required by the implied weights formula (fit = *e / e + k*), and `minLength`

is the minimum number of steps required by each character – which we need in order to convert the total number of steps (returned by `MorphyLength()`

to a number of excess steps (*e* in the implied weights formula)

```
<- function (parent, child, dataset, concavity = 10L,
IWScoreMorphy minLength = attr(dataset, 'min.length'), ...) {
<- vapply(attr(dataset, 'morphyObjs'), MorphyLength,
steps parent = parent, child = child, integer(1))
<- steps - minLength
homoplasies <- homoplasies / (homoplasies + concavity)
fit # Return:
sum(fit * attr(dataset, 'weight'))
}
```

Now we are ready to search:

```
data("inapplicable.datasets")
<- congreveLamsdellMatrices[[42]]
dataset
# Populate `min.length` attribute
<- PrepareDataIW(dataset)
dataset <- TreeSearch(NJTree(dataset), dataset,
iwTree InitializeData = IWInitMorphy,
CleanUpData = IWDestroyMorphy,
TreeScorer = IWScoreMorphy,
concavity = 10, # Will be sent to TreeScorer
verbosity = 1)
```

This quick search probably hasn’t found the globally optimal tree. Besides increasing the number of hits and rearrangements, the parsimony ratchet (Nixon, 1999) can help to escape local optima. This introduces an additional complication: we need to bootstrap the characters within `dataset`

, and their accompanying Morphy objects.

A `Bootstraper`

function expects an `edgeList`

(a list of the parent and child of each edge in a tree, in turn) and a `dataset`

argument, and conducts a tree search, starting at `edgeList`

, on a bootstrapped version of the dataset. It is also sent the arguments `maxIter = bootstrapIter`

and `maxHits = bootstrapHits`

, allowing ratchet search intensity to be controlled from parameters sent to the `Ratchet()`

function.

```
<- function (edgeList, dataset, concavity = 10L, EdgeSwapper = NNISwap,
IWBootstrap verbosity = 1L, ...) {
maxIter, maxHits, <- attributes(dataset)
att <- att[['weight']]
startWeights
# Decompress phyDat object so each character is listed once
<- seq_along(startWeights)
eachChar <- rep.int(eachChar, startWeights)
deindexedChars
# Resample characters
<- tabulate(sample(deindexedChars, replace = TRUE), length(startWeights))
resampling <- resampling != 0
sampled <- lapply(dataset, function (x) x[sampled])
sampledData <- att
sampledAtt 'weight']] <- resampling[sampled]
sampledAtt[['index']] <- rep.int(seq_len(sum(sampled)), resampling[sampled])
sampledAtt[['min.length']] <- minLength <- att[['min.length']][sampled]
sampledAtt[['morphyObjs']] <- att[['morphyObjs']][sampled]
sampledAtt[[attributes(sampledData) <- sampledAtt
# Search using resampled dataset
<- EdgeListSearch(edgeList[1:2], sampledData, TreeScorer = IWScoreMorphy,
res concavity = concavity, minLength = minLength,
EdgeSwapper = EdgeSwapper,
maxIter = maxIter, maxHits = maxHits,
verbosity = verbosity - 1L)
1:2]
res[ }
```

Having defined the `Bootstrapper()`

function we can now complete a Ratchet search with:

```
<- Ratchet(tree = iwTree, dataset = dataset,
ratchetTree concavity = 10,
InitializeData = IWInitMorphy,
CleanUpData = IWDestroyMorphy,
TreeScorer = IWScoreMorphy,
Bootstrapper = IWBootstrap,
ratchIter = 2, ratchHits = 2,
searchIter = 20, searchHits = 10,
verbosity = 2)
```

It would be sensible to use much larger values of `ratchIter`

, `ratchHits`

, `searchIter`

and `searchHits`

to be confident of locating an optimal tree. And note that in this specific case, implied weights tree search with the parsimony ratchet is implemented much more efficiently with `MaximizeParsimony(concavity = k)`

.

Hopefully these examples give a template from which you are able to construct your own optimality criteria. The maintainer is happy to answer questions via e-mail, or you can file queries by opening a GitHub issue.

You might want to:

- Load data from a Nexus file or spreadsheet
- Conduct parsimony search using Brazeau, Guillerme & Smith’s approach to inapplicable data, or using Profile parsimony.

Brazeau, M. D., Smith, M. R., & Guillerme, T. (2017). MorphyLib: A library for phylogenetic analysis of categorical trait data with inapplicability. doi:10.5281/zenodo.815372

Goloboff, P. A. (1993). Estimating character weights during tree search. *Cladistics*, *9*(1), 83–91. doi:10.1111/j.1096-0031.1993.tb00209.x

Hopkins, M. J., & St. John, K. (2021). Incorporating hierarchical characters into phylogenetic analysis. *Systematic Biology*, syab005. doi:10.1093/sysbio/syab005

Mir, A., Rosselló, F., & Rotger, L. A. (2013). A new balance index for phylogenetic trees. *Mathematical Biosciences*, *241*(1), 125–136. doi:10.1016/j.mbs.2012.10.005

Nixon, K. C. (1999). The Parsimony Ratchet, a new method for rapid parsimony analysis. *Cladistics*, *15*(4), 407–414. doi:10.1111/j.1096-0031.1999.tb00277.x