Forked from analyticsindiamagazine/Nvidia 1060 mnist.ipynb
Created
April 12, 2020 18:16
-
-
Save orange888/f1a9cfb136a45ed6481bcd8dfe279ec3 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
{ | |
"cells": [ | |
{ | |
"cell_type": "markdown", | |
"metadata": {}, | |
"source": [ | |
"# Tensorflow MNIST - GeForce Nvidia 1060 6GB" | |
] | |
}, | |
{ | |
"cell_type": "code", | |
"execution_count": 3, | |
"metadata": { | |
"scrolled": true | |
}, | |
"outputs": [ | |
{ | |
"name": "stdout", | |
"output_type": "stream", | |
"text": [ | |
"[name: \"/device:CPU:0\"\n", | |
"device_type: \"CPU\"\n", | |
"memory_limit: 268435456\n", | |
"locality {\n", | |
"}\n", | |
"incarnation: 16817798710844015210\n", | |
", name: \"/device:GPU:0\"\n", | |
"device_type: \"GPU\"\n", | |
"memory_limit: 4971180851\n", | |
"locality {\n", | |
" bus_id: 1\n", | |
" links {\n", | |
" }\n", | |
"}\n", | |
"incarnation: 16326358905080155025\n", | |
"physical_device_desc: \"device: 0, name: GeForce GTX 1060 6GB, pci bus id: 0000:01:00.0, compute capability: 6.1\"\n", | |
"]\n" | |
] | |
} | |
], | |
"source": [ | |
"from tensorflow.python.client import device_lib\n", | |
"print(device_lib.list_local_devices())" | |
] | |
}, | |
{ | |
"cell_type": "code", | |
"execution_count": 4, | |
"metadata": { | |
"scrolled": false | |
}, | |
"outputs": [ | |
{ | |
"name": "stdout", | |
"output_type": "stream", | |
"text": [ | |
"Extracting data/MNIST/train-images-idx3-ubyte.gz\n", | |
"Extracting data/MNIST/train-labels-idx1-ubyte.gz\n", | |
"Extracting data/MNIST/t10k-images-idx3-ubyte.gz\n", | |
"Extracting data/MNIST/t10k-labels-idx1-ubyte.gz\n", | |
"Optimization Iteration: 1, Training Accuracy: 15.6%\n", | |
"Optimization Iteration: 101, Training Accuracy: 64.1%\n", | |
"Optimization Iteration: 201, Training Accuracy: 89.1%\n", | |
"Optimization Iteration: 301, Training Accuracy: 90.6%\n", | |
"Optimization Iteration: 401, Training Accuracy: 92.2%\n", | |
"Optimization Iteration: 501, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 601, Training Accuracy: 90.6%\n", | |
"Optimization Iteration: 701, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 801, Training Accuracy: 95.3%\n", | |
"Optimization Iteration: 901, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 1001, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 1101, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 1201, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 1301, Training Accuracy: 95.3%\n", | |
"Optimization Iteration: 1401, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 1501, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 1601, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 1701, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 1801, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 1901, Training Accuracy: 95.3%\n", | |
"Optimization Iteration: 2001, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 2101, Training Accuracy: 95.3%\n", | |
"Optimization Iteration: 2201, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 2301, Training Accuracy: 95.3%\n", | |
"Optimization Iteration: 2401, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 2501, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 2601, Training Accuracy: 93.8%\n", | |
"Optimization Iteration: 2701, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 2801, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 2901, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 3001, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 3101, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 3201, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 3301, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 3401, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 3501, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 3601, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 3701, Training Accuracy: 93.8%\n", | |
"Optimization Iteration: 3801, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 3901, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 4001, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 4101, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 4201, Training Accuracy: 95.3%\n", | |
"Optimization Iteration: 4301, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 4401, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 4501, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 4601, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 4701, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 4801, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 4901, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 5001, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 5101, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 5201, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 5301, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 5401, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 5501, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 5601, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 5701, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 5801, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 5901, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 6001, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 6101, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 6201, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 6301, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 6401, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 6501, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 6601, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 6701, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 6801, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 6901, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 7001, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 7101, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 7201, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 7301, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 7401, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 7501, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 7601, Training Accuracy: 96.9%\n", | |
"Optimization Iteration: 7701, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 7801, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 7901, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 8001, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 8101, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 8201, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 8301, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 8401, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 8501, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 8601, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 8701, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 8801, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 8901, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 9001, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 9101, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 9201, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 9301, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 9401, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 9501, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 9601, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 9701, Training Accuracy: 98.4%\n", | |
"Optimization Iteration: 9801, Training Accuracy: 100.0%\n", | |
"Optimization Iteration: 9901, Training Accuracy: 98.4%\n", | |
"Time usage: 0:01:01\n" | |
] | |
} | |
], | |
"source": [ | |
"import os\n", | |
"import sys\n", | |
"os.environ[\"CUDA_VISIBLE_DEVICES\"]=\"1\"\n", | |
"import numpy as np\n", | |
"import matplotlib.pyplot as plt \n", | |
"import tensorflow as tf\n", | |
"from sklearn.metrics import confusion_matrix\n", | |
"import time \n", | |
"from datetime import timedelta\n", | |
"import math\n", | |
"import pandas as pd\n", | |
"\n", | |
"from tensorflow.examples.tutorials.mnist import input_data\n", | |
"data = input_data.read_data_sets('data/MNIST/', one_hot=True)\n", | |
"\n", | |
"weight_matrix_size1 = 5 #5x5 pixcels\n", | |
"depth1 = 16 #16 depth\n", | |
"\n", | |
"weight_matrix_size2 = 5 #5x5 pixcels\n", | |
"depth2 = 32 #32 depth\n", | |
"\n", | |
"fully_conn_layer = 256 #neuros at end of fully connected layer\n", | |
"\n", | |
"#Data dimensions\n", | |
"\n", | |
"#We have an input image of 28 x 28 dimensions\n", | |
"img_size = 28\n", | |
"\n", | |
"# We have a one hot encoded matrix of length 28*28 = 784\n", | |
"img_size_flat = img_size * img_size\n", | |
"\n", | |
"#Shape of the image represented by\n", | |
"img_shape = (img_size,img_size)\n", | |
"\n", | |
"#Number of channels in the input image\n", | |
"num_channels = 1\n", | |
"\n", | |
"#Number of output classes to be trained on\n", | |
"num_classes = 10\n", | |
"\n", | |
"def weight_matrix(dimensions):\n", | |
" return tf.Variable(tf.truncated_normal(shape = dimensions, stddev=0.1))\n", | |
"def biases_matrix(length):\n", | |
" return tf.Variable(tf.constant(0.1,shape=[length]))\n", | |
"\n", | |
"#Helper functions for ConvNet\n", | |
"\n", | |
"def convolutional_layer(input, #The images\n", | |
" depth, #channels of the image\n", | |
" no_filters, #number of filters in the output\n", | |
" weight_matrix_size):\n", | |
" \n", | |
" dimensions = [weight_matrix_size,weight_matrix_size, depth, no_filters]\n", | |
" \n", | |
" weights = weight_matrix(dimensions)\n", | |
" \n", | |
" biases = biases_matrix(length=no_filters)\n", | |
" \n", | |
" layer = tf.nn.conv2d(input=input,\n", | |
" filter= weights,\n", | |
" strides=[1, 1, 1, 1], #stride 2\n", | |
" padding='SAME') #input size = output size\n", | |
" layer += biases\n", | |
" \n", | |
" layer = tf.nn.max_pool(value=layer,\n", | |
" ksize=[1, 2, 2, 1],\n", | |
" strides=[1, 2, 2, 1],\n", | |
" padding='SAME')\n", | |
" #Passing the pooled layer into ReLU Activation function\n", | |
" layer = tf.nn.relu(layer)\n", | |
" \n", | |
" return layer , weights\n", | |
"\n", | |
"# Helper function for Flattening the layer\n", | |
"\n", | |
"def flatten_layer(layer):\n", | |
" \n", | |
" layer_shape = layer.get_shape()\n", | |
" \n", | |
" num_features = layer_shape[1:4].num_elements()\n", | |
" \n", | |
" layer_flat = tf.reshape(layer,[-1,num_features])\n", | |
" \n", | |
" return layer_flat, num_features\n", | |
"\n", | |
"#Helper functions for activation and fully connected\n", | |
"\n", | |
"def fully_connected(input,num_inputs,\n", | |
" num_outputs,\n", | |
" use_relu = True):\n", | |
" weights = weight_matrix([num_inputs,num_outputs])\n", | |
" \n", | |
" biases = biases_matrix(length= num_outputs)\n", | |
" \n", | |
" layer = tf.matmul(input,weights) + biases\n", | |
" \n", | |
" if use_relu:\n", | |
" layer = tf.nn.relu(layer)\n", | |
" \n", | |
" return layer\n", | |
"\n", | |
"#Placeholder variables\n", | |
"\n", | |
"x = tf.placeholder(tf.float32,shape=[None,img_size_flat],name='x')\n", | |
"\n", | |
"x_image = tf.reshape(x, [-1,img_size,img_size,num_channels])\n", | |
"\n", | |
"y_true = tf.placeholder(tf.float32, shape=[None, num_classes], name = 'y_true')\n", | |
"\n", | |
"y_true_cls = tf.argmax(y_true, axis=1)\n", | |
"\n", | |
"# Setting up the network\n", | |
"\n", | |
"layer_conv1 , weights_conv1 = convolutional_layer(input = x_image,\n", | |
" depth = num_channels,\n", | |
" weight_matrix_size = weight_matrix_size1,\n", | |
" no_filters = depth1)\n", | |
"\n", | |
"#layer_conv1 shape = (-1,14,14,16) and dtype = float32\n", | |
"\n", | |
"layer_conv2 , weights_conv2 = convolutional_layer(input = layer_conv1,\n", | |
" depth = depth1,\n", | |
" weight_matrix_size = weight_matrix_size2,\n", | |
" no_filters = depth2)\n", | |
"#layer_conv2 = shape=(?, 7, 7, 36) dtype=float32\n", | |
"\n", | |
"#Flattening the layer\n", | |
"\n", | |
"layer_flat , num_features = flatten_layer(layer_conv2)\n", | |
"\n", | |
"#Fully connected layers\n", | |
"\n", | |
"layer_fc1 = fully_connected(input = layer_flat,\n", | |
" num_inputs = num_features,\n", | |
" num_outputs = fully_conn_layer,\n", | |
" use_relu = True)\n", | |
"\n", | |
"layer_fc2 = fully_connected(input = layer_fc1,\n", | |
" num_inputs = fully_conn_layer,\n", | |
" num_outputs = num_classes,\n", | |
" use_relu = False)\n", | |
"\n", | |
"y_pred = tf.nn.softmax(layer_fc2)\n", | |
"\n", | |
"y_pred_cls = tf.argmax(y_pred , axis =1)\n", | |
"\n", | |
"cross_entropy = tf.nn.softmax_cross_entropy_with_logits(logits=layer_fc2,\n", | |
" labels=y_true)\n", | |
"cost = tf.reduce_mean(cross_entropy)\n", | |
"\n", | |
"#optimizing cost function\n", | |
"\n", | |
"optimizer = tf.train.AdamOptimizer(learning_rate= 1e-4).minimize(cost)\n", | |
"\n", | |
"correct_prediction = tf.equal(y_pred_cls, y_true_cls)\n", | |
"\n", | |
"accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))\n", | |
"\n", | |
"#TensorFlow session \n", | |
"config = tf.ConfigProto(\n", | |
" device_count = {'gpu': 0}\n", | |
" )\n", | |
"session = tf.Session(config=config)\n", | |
"\n", | |
"session.run(tf.global_variables_initializer())\n", | |
"\n", | |
"train_batch_size = 64\n", | |
"\n", | |
"total_iterations = 0\n", | |
"\n", | |
"accuracy_ = tf.summary.scalar('accuracy_value', accuracy)\n", | |
"\n", | |
"loss_ = tf.summary.scalar('loss_value', cost)\n", | |
"\n", | |
"def optimize(num_iterations):\n", | |
" \n", | |
" global total_iterations\n", | |
"\n", | |
" start_time = time.time()\n", | |
"\n", | |
" file_writer = tf.summary.FileWriter('/path/to/logs', session.graph)\n", | |
" for i in range(total_iterations,\n", | |
" total_iterations + num_iterations):\n", | |
" \n", | |
" x_batch, y_true_batch = data.train.next_batch(train_batch_size)\n", | |
" \n", | |
" feed_dict_train = {x: x_batch,\n", | |
" y_true: y_true_batch}\n", | |
" \n", | |
" session.run(optimizer, feed_dict=feed_dict_train)\n", | |
" \n", | |
" accuracy_value = session.run(accuracy_, feed_dict=feed_dict_train)\n", | |
" loss_value = session.run(loss_, feed_dict=feed_dict_train)\n", | |
" \n", | |
" file_writer.add_summary(accuracy_value, i)\n", | |
" file_writer.add_summary(loss_value, i)\n", | |
" \n", | |
" if i % 100 == 0:\n", | |
" \n", | |
" acc = session.run(accuracy, feed_dict=feed_dict_train)\n", | |
" \n", | |
" msg = \"Optimization Iteration: {0:>6}, Training Accuracy: {1:>6.1%}\"\n", | |
" \n", | |
" print(msg.format(i + 1, acc))\n", | |
" \n", | |
" total_iterations += num_iterations\n", | |
"\n", | |
" # Ending time.\n", | |
" end_time = time.time()\n", | |
"\n", | |
" # Difference between start and end-times.\n", | |
" time_dif = end_time - start_time\n", | |
"\n", | |
" # Print the time-usage.\n", | |
" print(\"Time usage: \" + str(timedelta(seconds=int(round(time_dif)))))\n", | |
" \n", | |
"optimize(num_iterations=10000)" | |
] | |
} | |
], | |
"metadata": { | |
"kernelspec": { | |
"display_name": "Python 3", | |
"language": "python", | |
"name": "python3" | |
}, | |
"language_info": { | |
"codemirror_mode": { | |
"name": "ipython", | |
"version": 3 | |
}, | |
"file_extension": ".py", | |
"mimetype": "text/x-python", | |
"name": "python", | |
"nbconvert_exporter": "python", | |
"pygments_lexer": "ipython3", | |
"version": "3.6.3" | |
} | |
}, | |
"nbformat": 4, | |
"nbformat_minor": 2 | |
} |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment