Hostname: page-component-77f85d65b8-2tv5m Total loading time: 0 Render date: 2026-03-27T07:52:52.235Z Has data issue: false hasContentIssue false

Prediction of body mass index in mice using dense molecular markers and a regularized neural network

Published online by Cambridge University Press:  12 April 2011

HAYRETTIN OKUT*
Affiliation:
Department of Animal Sciences, University of Yuzuncy Yil, Van, 65080, Turkey Department of Dairy Science, University of Wisconsin, Madison, WI 53706, USA
DANIEL GIANOLA
Affiliation:
Department of Dairy Science, University of Wisconsin, Madison, WI 53706, USA Department of Animal Sciences, University of Wisconsin, Madison, WI 53706, USA Department of Biostatistics and Medical Informatics, University of Wisconsin, Madison, WI 53706, USA
GUILHERME J. M. ROSA
Affiliation:
Department of Animal Sciences, University of Wisconsin, Madison, WI 53706, USA Department of Biostatistics and Medical Informatics, University of Wisconsin, Madison, WI 53706, USA
KENT A. WEIGEL
Affiliation:
Department of Dairy Science, University of Wisconsin, Madison, WI 53706, USA
*
*Corresponding author: University of Wisconsin 1675 Observatory Drive, Madison, WI 53703, USA. Tel: +1 608 772 4922. e-mail: okut@wisc.edu
Rights & Permissions [Opens in a new window]

Summary

Bayesian regularization of artificial neural networks (BRANNs) were used to predict body mass index (BMI) in mice using single nucleotide polymorphism (SNP) markers. Data from 1896 animals with both phenotypic and genotypic (12 320 loci) information were used for the analysis. Missing genotypes were imputed based on estimated allelic frequencies, with no attempt to reconstruct haplotypes based on family information or linkage disequilibrium between markers. A feed-forward multilayer perceptron network consisting of a single output layer and one hidden layer was used. Training of the neural network was done using the Bayesian regularized backpropagation algorithm. When the number of neurons in the hidden layer was increased, the number of effective parameters, γ, increased up to a point and stabilized thereafter. A model with five neurons in the hidden layer produced a value of γ that saturated the data. In terms of predictive ability, a network with five neurons in the hidden layer attained the smallest error and highest correlation in the test data although differences among networks were negligible. Using inherent weight information of BRANN with different number of neurons in the hidden layer, it was observed that 17 SNPs had a larger impact on the network, indicating their possible relevance in prediction of BMI. It is concluded that BRANN may be at least as useful as other methods for high-dimensional genome-enabled prediction, with the advantage of its potential ability of capturing non-linear relationships, which may be useful in the study of quantitative traits under complex gene action.

Information

Type
Research Papers
Creative Commons
This is a work of the U.S. Government and is not subject to copyright protection in the United States.
Copyright
Copyright © Cambridge University Press 2011 This is a work of the U.S. Government and is not subject to copyright protection in the United States.
Figure 0

Fig. 1. ANN design used in this study. There were 798 SNP genotypes used as inputs (pij). Each SNP is connected to up to five neurons via coefficients wjk (j denotes neuron, k denotes SNP). Each hidden and output neuron has a bias parameter bj(l), j denotes neuron, l denotes layer).

Figure 1

Fig. 2. Flow chart for Bayesian optimization of regularization parameters α and β in NNs; MP, maximum a posterio (adapted from Shaneh & Butler, 2006).

Figure 2

Fig. 3. Box-plots for (a) mean-squared error in the testing set, (b) correlation between predictions and observations in the testing set and (c) effective number of parameters, γ, after 20 independent runs (*indicates extreme values).

Figure 3

Table 1. Parameter estimates and their standard deviations for different network architectures (results are averages of 20 independent runs)

Figure 4

Fig. 4. Correlations between predictions and observations for training (rtrain), testing (rtest), tuning (rvalidation) and overall (ry−ŷ) for linear and seven (N-1–N-7) network architectures.

Figure 5

Fig. 5. Distribution of weights (wkj) for the linear model and for neural network with five neurons.

Figure 6

Fig. 6. Plots for the index values of 798 SNPs as prediction of BMI. The solid line gives the cutoff point separating SNPs with index values larger than 0·45%.

Figure 7

Table 2. Relative importance of SNPs with ISNPj values larger than 0·45% for the each of the non-linear networks

Figure 8

Table 3. Relative contribution of the kth neuron for several neural network architectures for BMI in mice