Last active
May 19, 2024 09:12
-
-
Save ecsplendid/5e322fc97dcd4bed23bea10e1d004d5e to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
// generated with gpt4-o probably buggy | |
import torch | |
import torch.nn as nn | |
import torch.optim as optim | |
from torchvision import datasets, transforms | |
from torch.utils.data import DataLoader, random_split, TensorDataset | |
import numpy as np | |
import matplotlib.pyplot as plt | |
# Load MNIST data | |
transform = transforms.Compose([transforms.ToTensor(), transforms.Normalize((0.5,), (0.5,))]) | |
train_dataset = datasets.MNIST(root='./data', train=True, transform=transform, download=True) | |
test_dataset = datasets.MNIST(root='./data', train=False, transform=transform, download=True) | |
# Split the dataset into training and validation sets | |
train_size = int(0.8 * len(train_dataset)) | |
val_size = len(train_dataset) - train_size | |
train_dataset, val_dataset = random_split(train_dataset, [train_size, val_size]) | |
# Preprocess data | |
train_images = train_dataset.dataset.data[train_dataset.indices].view(-1, 28*28).float() / 255.0 | |
train_labels = train_dataset.dataset.targets[train_dataset.indices] | |
# Shuffle the labels randomly in the training set | |
random_train_labels = train_labels.clone() | |
random_train_labels = random_train_labels[torch.randperm(len(random_train_labels))] | |
# Create a dataset with shuffled labels | |
shuffled_train_dataset = TensorDataset(train_images, random_train_labels) | |
# Create data loaders | |
batch_size = 128 | |
train_loader = DataLoader(shuffled_train_dataset, batch_size=batch_size, shuffle=True) | |
val_loader = DataLoader(val_dataset, batch_size=batch_size, shuffle=False) | |
# Define the model | |
class SimpleNN(nn.Module): | |
def __init__(self): | |
super(SimpleNN, self).__init__() | |
self.fc1 = nn.Linear(28*28, 512) | |
self.fc2 = nn.Linear(512, 10) | |
def forward(self, x): | |
x = torch.relu(self.fc1(x)) | |
x = self.fc2(x) | |
return x | |
# Determine the device to use (MPS, CUDA, or CPU) | |
device = torch.device("mps" if torch.backends.mps.is_available() else "cuda" if torch.cuda.is_available() else "cpu") | |
model = SimpleNN().to(device) | |
# Define loss function and optimizer | |
criterion = nn.CrossEntropyLoss() | |
optimizer = optim.RMSprop(model.parameters(), lr=0.001) | |
# Train the model | |
num_epochs = 10 | |
train_losses = [] | |
val_losses = [] | |
for epoch in range(num_epochs): | |
model.train() | |
train_loss = 0 | |
for images, labels in train_loader: | |
images, labels = images.to(device), labels.to(device) | |
optimizer.zero_grad() | |
outputs = model(images) | |
loss = criterion(outputs, labels) | |
loss.backward() | |
optimizer.step() | |
train_loss += loss.item() | |
train_loss /= len(train_loader) | |
train_losses.append(train_loss) | |
model.eval() | |
val_loss = 0 | |
with torch.no_grad(): | |
for images, labels in val_loader: | |
images, labels = images.to(device), labels.to(device) | |
outputs = model(images) | |
loss = criterion(outputs, labels) | |
val_loss += loss.item() | |
val_loss /= len(val_loader) | |
val_losses.append(val_loss) | |
print(f"Epoch {epoch+1}/{num_epochs}, Train Loss: {train_loss:.4f}, Validation Loss: {val_loss:.4f}") | |
# Plot the training and validation loss | |
plt.plot(train_losses, label='Training loss') | |
plt.plot(val_losses, label='Validation loss') | |
plt.xlabel('Epochs') | |
plt.ylabel('Loss') | |
plt.legend() | |
plt.title('Training and Validation Loss on MNIST with Random Labels') | |
plt.savefig("plot.jpg") | |
plt.show() |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment