Skip to content

Instantly share code, notes, and snippets.

@qinyao-he
Created August 8, 2018 19:35
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save qinyao-he/60171806c40b7b46354234b896c7c2d5 to your computer and use it in GitHub Desktop.
Save qinyao-he/60171806c40b7b46354234b896c7c2d5 to your computer and use it in GitHub Desktop.
tensorflow mnist classification model
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import numpy as np
import tensorflow as tf
tf.logging.set_verbosity(tf.logging.INFO)
def model_fn(features, labels, mode):
input_data = tf.reshape(features['x'], (-1, 1, 28, 28))
x = input_data
data_format = 'channels_first'
x = tf.layers.conv2d(
inputs=x,
filters=32,
kernel_size=(3, 3),
padding='same',
activation=tf.nn.relu,
data_format=data_format
)
x = tf.layers.conv2d(
inputs=x,
filters=32,
kernel_size=(3, 3),
padding='same',
activation=tf.nn.relu,
data_format=data_format
)
x = tf.layers.max_pooling2d(
inputs=x, pool_size=(2, 2), strides=2, data_format=data_format)
x = tf.layers.conv2d(
inputs=x,
filters=64,
kernel_size=(3, 3),
padding="same",
activation=tf.nn.relu,
data_format=data_format
)
x = tf.layers.conv2d(
inputs=x,
filters=64,
kernel_size=(3, 3),
padding='same',
activation=tf.nn.relu,
data_format=data_format
)
x = tf.layers.max_pooling2d(
inputs=x, pool_size=(2, 2), strides=2, data_format=data_format)
x = tf.reshape(x, [-1, 64 * 7 * 7])
x = tf.layers.dense(inputs=x, units=1024, activation=tf.nn.relu)
x = tf.layers.dropout(
inputs=x, rate=0.5, training=(mode == tf.estimator.ModeKeys.TRAIN))
logits = tf.layers.dense(inputs=x, units=10)
if labels is not None:
loss = tf.losses.sparse_softmax_cross_entropy(labels=labels, logits=logits)
predictions = {
"classes": tf.argmax(input=logits, axis=1),
"probabilities": tf.nn.softmax(logits, name="softmax_tensor")
}
if mode == tf.estimator.ModeKeys.TRAIN:
optimizer = tf.train.AdamOptimizer()
update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)
with tf.control_dependencies(update_ops):
train_op = optimizer.minimize(
loss=loss,
global_step=tf.train.get_global_step())
return tf.estimator.EstimatorSpec(mode=mode, loss=loss, train_op=train_op)
elif mode == tf.estimator.ModeKeys.EVAL:
eval_metric_ops = {
"accuracy": tf.metrics.accuracy(
labels=labels, predictions=predictions["classes"])}
return tf.estimator.EstimatorSpec(
mode=mode, loss=loss, eval_metric_ops=eval_metric_ops)
else:
return tf.estimator.EstimatorSpec(
mode=mode, predictions=predictions,
export_outputs={'predictions': tf.estimator.export.PredictOutput(predictions)})
def serving_input_receiver_fn():
feature_spec = {'x': tf.FixedLenFeature(shape=[-1,784], dtype=tf.float32)}
return tf.estimator.export.build_parsing_serving_input_receiver_fn(feature_spec)
def main(argv):
mnist = tf.contrib.learn.datasets.load_dataset("mnist")
train_data = mnist.train.images
train_labels = np.asarray(mnist.train.labels, dtype=np.int32)
eval_data = mnist.test.images
eval_labels = np.asarray(mnist.test.labels, dtype=np.int32)
mnist_classifier = tf.estimator.Estimator(
model_fn=model_fn, model_dir="./log")
train_input_fn = tf.estimator.inputs.numpy_input_fn(
x={"x": train_data},
y=train_labels,
batch_size=128,
num_epochs=None,
shuffle=True)
if argv[1] == 'train':
mnist_classifier.train(
input_fn=train_input_fn,
steps=20000)
elif argv[1] == 'eval':
eval_input_fn = tf.estimator.inputs.numpy_input_fn(
x={"x": eval_data},
y=eval_labels,
num_epochs=1,
shuffle=False)
eval_results = mnist_classifier.evaluate(input_fn=eval_input_fn)
print(eval_results)
else:
mnist_classifier.export_savedmodel(
'./log/export',
tf.estimator.export.build_raw_serving_input_receiver_fn({
'x': tf.placeholder(shape=[None, 784], dtype=tf.float32)
}))
if __name__ == '__main__':
tf.app.run()
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment