geposan/R/neural.R

163 lines
6.1 KiB
R

# Find genes by training and applying a neural network.
neural <- function(preset, progress = NULL, seed = 49641) {
species_ids <- preset$species_ids
gene_ids <- preset$gene_ids
reference_gene_ids <- preset$reference_gene_ids
cached("neural", c(species_ids, gene_ids, reference_gene_ids), {
set.seed(seed)
gene_count <- length(gene_ids)
progress_buffer <- 0
progress_step <- 1 / (2 * length(reference_gene_ids) + 1)
# Prefilter distances by species.
distances <- geposan::distances[species %chin% species_ids]
# Input data for the network. This contains the gene ID as an
# identifier as well as the per-species gene distances as input
# variables.
data <- data.table(gene = gene_ids)
# Buffer to keep track of the names of the input variables.
input_vars <- NULL
# Make a columns containing positions and distances for each
# species.
for (species_id in species_ids) {
species_data <- distances[species == species_id, .(gene, distance)]
# Only include species with at least 25% known values. As
# positions and distances always coexist, we don't loose any
# data here.
species_data <- stats::na.omit(species_data)
if (nrow(species_data) >= 0.25 * gene_count) {
data <- merge(data, species_data, all.x = TRUE)
# Replace missing data with mean values. The neural network
# can't handle NAs in a meaningful way. Choosing extreme
# values here would result in heavily biased results.
# Therefore, the mean value is chosen as a compromise.
# However, this will of course lessen the significance of
# the results.
mean_distance <- round(species_data[, mean(distance)])
data[is.na(distance), `:=`(distance = mean_distance)]
# Name the new column after the species.
setnames(data, "distance", species_id)
# Add the input variable to the buffer.
input_vars <- c(input_vars, species_id)
}
}
# Extract the reference genes.
reference_data <- data[gene %chin% reference_gene_ids]
reference_data[, score := 1.0]
# Take out random samples from the remaining genes. This is another
# compromise with a negative impact on significance. Because there
# is no information on genes with are explicitely *not* TPE-OLD
# genes, we have to assume that a random sample of genes has a low
# probability of including TPE-OLD genes.
without_reference_data <- data[!gene %chin% reference_gene_ids]
reference_samples <- without_reference_data[
sample(
nrow(without_reference_data),
nrow(reference_data)
)
]
reference_samples[, score := 0.0]
# Merge training data. The training data includes all reference
# genes as well as an equal number of random sample genes.
training_data <- rbindlist(list(reference_data, reference_samples))
# Layers for the neural network.
input_layer <- length(input_vars)
layer1 <- input_layer
layer2 <- 0.5 * input_layer
layer3 <- 0.5 * layer2
# Train the model using the specified subset of the training data and
# apply it for predicting the genes.
apply_network <- function(training_gene_ids, gene_ids) {
# Create a new model for each training session, because the model
# would keep its state across training sessions otherwise.
model <- keras::keras_model_sequential() |>
keras::layer_dense(
units = layer1,
activation = "relu",
input_shape = input_layer,
) |>
keras::layer_dense(
units = layer2,
activation = "relu",
kernel_regularizer = keras::regularizer_l2()
) |>
keras::layer_dense(
units = layer3,
activation = "relu",
kernel_regularizer = keras::regularizer_l2()
) |>
keras::layer_dense(
units = 1,
activation = "sigmoid"
) |>
keras::compile(loss = "binary_crossentropy")
# Prepare training data by filtering it to the given genes and
# converting it to a matrix.
training_data <- training_data[gene %chin% training_gene_ids]
training_matrix <- as.matrix(training_data[, ..input_vars])
colnames(training_matrix) <- NULL
training_matrix <- keras::normalize(training_matrix)
keras::fit(
model,
x = training_matrix,
y = training_data$score,
epochs = 300,
verbose = FALSE
)
# Convert the input data to a matrix.
data_matrix <- as.matrix(data[gene %chin% gene_ids, ..input_vars])
colnames(data_matrix) <- NULL
data_matrix <- keras::normalize(data_matrix)
data[
gene %chin% gene_ids,
score := stats::predict(model, data_matrix)
]
if (!is.null(progress)) {
progress_buffer <<- progress_buffer + progress_step
progress(progress_buffer)
}
}
# Apply the network to all non-training genes first.
apply_network(
training_data$gene,
gene_ids[!gene_ids %chin% training_data$gene]
)
# Apply the network to the training genes leaving out the gene itself.
for (training_gene_id in training_data$gene) {
apply_network(
training_data[gene != training_gene_id, gene],
training_gene_id
)
}
data[, .(gene, score)]
})
}