can't debug anymore

need to tweak my caffeine dose... just until this model is fully trained, i promise.

vmfunc |

17:50 just woke up, gonna start logging stuff here as i seen hitori and lelouch do it, it seems rather fun

plan for tonight:

20:58 not totally sure how i’ll conquer the convergence issues, but i have a couple of papers bookmarked that might shine some light.

for epoch in range(num_epochs):
    adjust_learning_rate(optimizer, epoch, init_lr=0.01)
    train()
    validate()

00:03 woke up after a 3-hour power nap, need to wrap up some model evaluations


04:19 still up, model’s training accuracy is stuck at 87% - been tweaking the loss function.

def custom_loss(output, target):
    loss = torch.mean((output - target) ** 2)
    return loss

wondering if i’ll ever see the bed tonight, will queue up some long overdue model tests and a batch of simulations before i crash.


06:47 observing some interesting patterns in the training data with the new visualization tool i coded up last week.

it’s intriguing but expected; using mixed precision training, i’m seeing a slight increase in efficiency. wonder why more people don’t share these little hacks—though the fear of giving away competitive edges makes sense.


08:30 feeling really shitty right now, might just rest for a bit


09:30 Noticed a minor memory leak during the test runs. Quick fix with a few lines:

void sadly_this_works(resource *res) {
    if (!res->is_active) {
        free(res);
    }
}

10:50 Reviewed a friend’s patch; they missed a semaphore release. Fixed it and pushed it upstream


11:20 Shutting down