diff --git a/chapter02/Chapter 2 - Tokens and Token Embeddings.ipynb b/chapter02/Chapter 2 - Tokens and Token Embeddings.ipynb index fcccbfc..c5b3692 100644 --- a/chapter02/Chapter 2 - Tokens and Token Embeddings.ipynb +++ b/chapter02/Chapter 2 - Tokens and Token Embeddings.ipynb @@ -526,7 +526,7 @@ }, "outputs": [], "source": [ - "from transformers import AutoModelForCausalLM, AutoTokenizer\n", + "from transformers import AutoTokenizer\n", "\n", "colors_list = [\n", " '102;194;165', '252;141;98', '141;160;203',\n", @@ -1610,7 +1610,7 @@ "from transformers import AutoModel, AutoTokenizer\n", "\n", "# Load a tokenizer\n", - "tokenizer = AutoTokenizer.from_pretrained(\"microsoft/deberta-base\")\n", + "tokenizer = AutoTokenizer.from_pretrained(\"microsoft/deberta-v3-xsmall\")\n", "\n", "# Load a language model\n", "model = AutoModel.from_pretrained(\"microsoft/deberta-v3-xsmall\")\n", @@ -2363,7 +2363,7 @@ } ], "source": [ - "print(songs_df.iloc[2172])" + "print(songs_df.loc['2172'])" ] }, { @@ -2685,7 +2685,7 @@ " similar_songs = np.array(\n", " model.wv.most_similar(positive=str(song_id),topn=5)\n", " )[:,0]\n", - " return songs_df.iloc[similar_songs]\n", + " return songs_df.loc[similar_songs]\n", "\n", "# Extract recommendations\n", "print_recommendations(2172)"