[259458]: / Models / Network / Thin_ResNet.py

Download this file

128 lines (104 with data), 6.1 kB

  1
  2
  3
  4
  5
  6
  7
  8
  9
 10
 11
 12
 13
 14
 15
 16
 17
 18
 19
 20
 21
 22
 23
 24
 25
 26
 27
 28
 29
 30
 31
 32
 33
 34
 35
 36
 37
 38
 39
 40
 41
 42
 43
 44
 45
 46
 47
 48
 49
 50
 51
 52
 53
 54
 55
 56
 57
 58
 59
 60
 61
 62
 63
 64
 65
 66
 67
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# Import useful packages
import tensorflow as tf
from Models.Initialize_Variables.Initialize import *
def Thin_ResNet(Input, keep_prob):
'''
Args:
Input: The reshaped input EEG signals
keep_prob: The Keep probability of Dropout
Returns:
prediction: Final prediction of Thin ResNet Model
'''
# Input reshaped EEG signals: shape 4096 --> 64 X 64
x_Reshape = tf.reshape(tensor=Input, shape=[-1, 64, 64, 1])
# First Residual Block
# First Convolutional Layer
W_conv1 = weight_variable([1, 1, 1, 48])
b_conv1 = bias_variable([48])
h_conv1 = tf.nn.conv2d(x_Reshape, W_conv1, strides=[1, 1, 1, 1], padding='SAME') + b_conv1
h_conv1_BN = tf.layers.batch_normalization(h_conv1, training=True)
h_conv1_Acti = tf.nn.leaky_relu(h_conv1_BN)
h_conv1_drop = tf.nn.dropout(h_conv1_Acti, keep_prob, noise_shape=[tf.shape(h_conv1_Acti)[0], 1, 1, tf.shape(h_conv1_Acti)[3]])
# Second Convolutional Layer
W_conv2 = weight_variable([3, 3, 48, 48])
b_conv2 = bias_variable([48])
h_conv2 = tf.nn.conv2d(h_conv1_drop, W_conv2, strides=[1, 1, 1, 1], padding='SAME') + b_conv2
h_conv2_BN = tf.layers.batch_normalization(h_conv2, training=True)
h_conv2_Acti = tf.nn.leaky_relu(h_conv2_BN)
h_conv2_drop = tf.nn.dropout(h_conv2_Acti, keep_prob, noise_shape=[tf.shape(h_conv2_Acti)[0], 1, 1, tf.shape(h_conv2_Acti)[3]])
# Third Convolutional Layer
W_conv3 = weight_variable([1, 1, 48, 96])
b_conv3 = bias_variable([96])
h_conv3 = tf.nn.conv2d(h_conv2_drop, W_conv3, strides=[1, 1, 1, 1], padding='SAME') + b_conv3
h_conv3_BN = tf.layers.batch_normalization(h_conv3, training=True)
h_conv3_res = tf.concat([h_conv3_BN, x_Reshape], axis=3) # 97 feature maps now == 96 + 1
h_conv3_Acti = tf.nn.leaky_relu(h_conv3_res)
h_conv3_drop = tf.nn.dropout(h_conv3_Acti, keep_prob, noise_shape=[tf.shape(h_conv3_Acti)[0], 1, 1, tf.shape(h_conv3_Acti)[3]])
# First Max Pooling Layer: shape 64 X 64 --> 32 X 32
h_pool1 = tf.nn.max_pool(h_conv3_drop, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME')
# Second Residual Block
# Fourth Convolutional Layer
W_conv4 = weight_variable([1, 1, 97, 96])
b_conv4 = bias_variable([96])
h_conv4 = tf.nn.conv2d(h_pool1, W_conv4, strides=[1, 1, 1, 1], padding='SAME') + b_conv4
h_conv4_BN = tf.layers.batch_normalization(h_conv4, training=True)
h_conv4_Acti = tf.nn.leaky_relu(h_conv4_BN)
h_conv4_drop = tf.nn.dropout(h_conv4_Acti, keep_prob, noise_shape=[tf.shape(h_conv4_Acti)[0], 1, 1, tf.shape(h_conv4_Acti)[3]])
# Fifth Convolutional Layer
W_conv5 = weight_variable([3, 3, 96, 96])
b_conv5 = bias_variable([96])
h_conv5 = tf.nn.conv2d(h_conv4_drop, W_conv5, strides=[1, 1, 1, 1], padding='SAME') + b_conv5
h_conv5_BN = tf.layers.batch_normalization(h_conv5, training=True)
h_conv5_Acti = tf.nn.leaky_relu(h_conv5_BN)
h_conv5_drop = tf.nn.dropout(h_conv5_Acti, keep_prob, noise_shape=[tf.shape(h_conv5_Acti)[0], 1, 1, tf.shape(h_conv5_Acti)[3]])
# Sixth Convolutional Layer
W_conv6 = weight_variable([1, 1, 96, 128])
b_conv6 = bias_variable([128])
h_conv6 = tf.nn.conv2d(h_conv5_drop, W_conv6, strides=[1, 1, 1, 1], padding='SAME') + b_conv6
h_conv6_BN = tf.layers.batch_normalization(h_conv6, training=True)
h_conv6_res = tf.concat([h_conv6_BN, h_pool1], axis=3) # 225 feature maps now == 97 + 128
h_conv6_Acti = tf.nn.leaky_relu(h_conv6_res)
h_conv6_drop = tf.nn.dropout(h_conv6_Acti, keep_prob, noise_shape=[tf.shape(h_conv6_Acti)[0], 1, 1, tf.shape(h_conv6_Acti)[3]])
# Second Max Pooling Layer: shape 32 X 32 --> 16 X 16
h_pool2 = tf.nn.max_pool(h_conv6_drop, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME')
# Third Residual Block
# Seventh Convolutional Layer
W_conv7 = weight_variable([1, 1, 225, 128])
b_conv7 = bias_variable([128])
h_conv7 = tf.nn.conv2d(h_pool2, W_conv7, strides=[1, 1, 1, 1], padding='SAME') + b_conv7
h_conv7_BN = tf.layers.batch_normalization(h_conv7, training=True)
h_conv7_Acti = tf.nn.leaky_relu(h_conv7_BN)
h_conv7_drop = tf.nn.dropout(h_conv7_Acti, keep_prob, noise_shape=[tf.shape(h_conv7_Acti)[0], 1, 1, tf.shape(h_conv7_Acti)[3]])
# Eighth Convolutional Layer
W_conv8 = weight_variable([3, 3, 128, 128])
b_conv8 = bias_variable([128])
h_conv8 = tf.nn.conv2d(h_conv7_drop, W_conv8, strides=[1, 1, 1, 1], padding='SAME') + b_conv8
h_conv8_BN = tf.layers.batch_normalization(h_conv8, training=True)
h_conv8_Acti = tf.nn.leaky_relu(h_conv8_BN)
h_conv8_drop = tf.nn.dropout(h_conv8_Acti, keep_prob, noise_shape=[tf.shape(h_conv8_Acti)[0], 1, 1, tf.shape(h_conv8_Acti)[3]])
# Ninth Convolutional Layer
W_conv9 = weight_variable([1, 1, 128, 256])
b_conv9 = bias_variable([256])
h_conv9 = tf.nn.conv2d(h_conv8_drop, W_conv9, strides=[1, 1, 1, 1], padding='SAME') + b_conv9
h_conv9_BN = tf.layers.batch_normalization(h_conv9, training=True)
h_conv9_res = tf.concat([h_conv9_BN, h_pool2], axis=3) # 481 feature maps now == 225 + 256
h_conv9_Acti = tf.nn.leaky_relu(h_conv9_res)
h_conv9_drop = tf.nn.dropout(h_conv9_Acti, keep_prob, noise_shape=[tf.shape(h_conv9_Acti)[0], 1, 1, tf.shape(h_conv9_Acti)[3]])
# Third Max Pooling Layer
h_pool3 = tf.nn.max_pool(h_conv9_drop, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME')
# Flatten Layer
h_pool6_flat = tf.reshape(h_pool3, [-1, 8 * 8 * 481])
# First Fully Connected Layer
W_fc1 = weight_variable([8 * 8 * 481, 512])
b_fc1 = bias_variable([512])
h_fc1 = tf.matmul(h_pool6_flat, W_fc1) + b_fc1
h_fc1_BN = tf.layers.batch_normalization(h_fc1, training=True)
h_fc1_Acti = tf.nn.leaky_relu(h_fc1_BN)
h_fc1_drop = tf.nn.dropout(h_fc1_Acti, keep_prob)
# Second Fully Connected Layer
W_fc2 = weight_variable([512, 4])
b_fc2 = bias_variable([4])
prediction = tf.nn.softmax(tf.matmul(h_fc1_drop, W_fc2) + b_fc2)
return prediction