From 06663df82a28749b833b071620dc0bf674094f21 Mon Sep 17 00:00:00 2001 From: udlbook <110402648+udlbook@users.noreply.github.com> Date: Tue, 13 Dec 2022 10:53:54 +0000 Subject: [PATCH] Created using Colaboratory --- CX20315_Transformers.ipynb | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/CX20315_Transformers.ipynb b/CX20315_Transformers.ipynb index 21a5b7e..2948406 100644 --- a/CX20315_Transformers.ipynb +++ b/CX20315_Transformers.ipynb @@ -4,7 +4,7 @@ "metadata": { "colab": { "provenance": [], - "authorship_tag": "ABX9TyPLuYbU3a6Uvb/3X1shd1XV", + "authorship_tag": "ABX9TyMfWL40+ZshPZhweAtQ9Fn6", "include_colab_link": true }, "kernelspec": { @@ -133,7 +133,7 @@ " outputs = model(input_ids = input_tokens['input_ids'], attention_mask = input_tokens['attention_mask'])\n", " # Find prediction\n", " prob_over_tokens = F.softmax(outputs.logits, dim=-1).detach().numpy()[0,-1]\n", - " # Draw a random token according to the probabilities\n", + " # TODO: Draw a random token according to the probabilities\n", " # Use: https://numpy.org/doc/stable/reference/random/generated/numpy.random.choice.html\n", " # Replace this line\n", " next_token = [5000]\n",