Diff of /README.md [000000] .. [c49071]

Switch to unified view

a b/README.md
1
# ___DNNGP: Deep neural network for genomic prediction___ <br>
2
<br>
3
The Python project 'DNNGP' can be used to implement genome-wide prediction (GP), which can predict the phenotypes of plants and animals based on multi-omics data. The code is written using Python 3.9 and TensorFlow 2.6.0.
4
<br><br>
5
6
Data used in the papers' example-data.tgz 'can be found in the package at [DNNGP-v1.0.0.zip](https://github.com/AIBreeding/DNNGP/releases/download/v1.0.0/DNNGP-v1.0.0.zip)
7
8
The original data download address is as follows:  
9
  maize:   
10
    https://pan.baidu.com/s/1AsPJLTe--gU5EN8aFTMYPA  
11
    http://www.maizego.org/download/Agronomic_23Traits.txt  
12
  
13
  tomato:  
14
    https://www.ncbi.nlm.nih.gov/sra?term=SRP150040  
15
    https://www.ncbi.nlm.nih.gov/sra?term=SRP186721  
16
    https://www.ncbi.nlm.nih.gov/sra?term=SRP186721  
17
  
18
  wheat:  
19
    https://github.com/gdlc/BGLR-R/blob/master/data/wheat.RData  
20
  
21
More information could be found in the [user manual](DNNGP-usermanual.pdf).  
22
  
23
Tips:
24
Deep learning models perform better with larger sample sizes.
25
26
### Change log
27
2025.01.21:  
28
 1. Updated the hyperparameter auto-tuning script `DNNGP_OPN.py`, which you can find in the `Tuning_hyperparameters` directory.
29
 2. `DNNGP_OPN.py` enables batch tuning of multi phenotype.
30
    
31
2024.03:
32
 1. Update the software to version 3.1 for both Windows and Linux.
33
 2. Optimize the naming of output files for model training. The current file name concatenates the `input phenotype file name`, the `original output file name` and the `part parameter value`. This change prevents the issue of overlapping phenotypic characters and fold number collisions with files.
34
 3. Optimize the complex parameter adjustment process.
35
36
### It is suggested tuning parameters as follows:
37
38
    batchsize: Set this to the largest value your hardware can support, typically increasing powers of 2.
39
40
    lr: Set this to 1, or any value you think is appropriate based on your understanding of deep learning. The learning rate is partially auto-adjusted by the internal algorithm.
41
42
    epoch: Set a maximum value and allow “earlystopping” to decide the optimal stopping point.
43
44
    dropout1;This parameter should be experimentally determined, with recommended trials ranging from 0.1 to 0.9.
45
46
    dropout2: Similarly, this needs empirical evaluation, usually between 0.1 and 0.9.
47
48
    patience: A value between 10 and 100 is generally acceptable. It doesn't take much adjustment.
49
50
    earlystopping: Set this value to 5-10 times the value of patience. Increase this multiplier if the iterations end too quickly.
51
52
The information above is consistent with our user manual. For more details, please refer to the user manual.
53
54
国内下载地址:https://datashare.biobin.com.cn/flask
55
56
57
## 👥 Contacts
58
59
[Huihui Li](lihuihui@caas.cn)(lihuihui@caas.cn)
60