Skip to content

Commit 4fffdc9

Browse files
authored
add ch06 (#180)
1 parent a8ec764 commit 4fffdc9

File tree

1 file changed

+13
-8
lines changed

1 file changed

+13
-8
lines changed

README.md

Lines changed: 13 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -98,11 +98,13 @@ EXAMPLES:
9898
| 03.01 | Computing attention scores as a dot product. |
9999
...
100100
|-------+----------------------------------------------------------------------|
101-
| 05.09 | Saving and loading a candle model. |
101+
| 06.13 | Example usage of `train_classifier_simple` and `plot_values` |
102+
| | function. |
102103
|-------+----------------------------------------------------------------------|
103-
| 05.10 | Example for downloading safetensors from HuggingFace Hub. |
104+
| 06.14 | Loading fine-tuned model and calculate performance on whole train, |
105+
| | val and test sets. |
104106
|-------+----------------------------------------------------------------------|
105-
| 05.11 | Example usage of `load_weights_into_gpt`. |
107+
| 06.15 | Example usage of `classify_review`. |
106108
+-------+----------------------------------------------------------------------+
107109
```
108110

@@ -137,15 +139,18 @@ EXERCISES:
137139
|-----+------------------------------------------------------------------------|
138140
...
139141
|-----+------------------------------------------------------------------------|
140-
| 5.6 | Comparing generations with different GPT-2 model sizes |
142+
| 6.2 | Fine-tuning the whole model |
141143
| | |
142-
| | Experiment with GPT-2 models of different sizes—for example, the |
143-
| | largest 1,558 million parameter model—and compare the generated text |
144-
| | to the 124 million model. |
144+
| | Instead of fine-tuning just the final transformer block, fine-tune the |
145+
| | entire model and assess the effect on predictive performance. |
146+
|-----+------------------------------------------------------------------------|
147+
| 6.3 | Fine-tuning the first vs. last token |
148+
| | |
149+
| | Try fine-tuning the first output token. Notice the changes in |
150+
| | predictive performance compared to fine-tuning the last output token. |
145151
+-----+------------------------------------------------------------------------+
146152
```
147153

148-
149154
## [Alternative Usage] Installing from `crates.io`
150155

151156
Alternatively, users have the option of installing this crate directly via

0 commit comments

Comments
 (0)