Skip to content

Commit d6023d4

Browse files
committed
fix -typos
1 parent 85a4644 commit d6023d4

File tree

1 file changed

+5
-5
lines changed

1 file changed

+5
-5
lines changed

extras/pytorch_2_intro.ipynb

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -175,7 +175,7 @@
175175
"\n",
176176
"Why?\n",
177177
"\n",
178-
"Modern GPUs have so much compute power they are often not compute limited, as in, the main bottleneck to training models is how fast can you get data from your CPU to your GPU.\n",
178+
"Modern GPUs have so much compute power they are often not compute limited, as in, the main bottleneck to training models is how fast you can get data from your CPU to your GPU.\n",
179179
"This is known as bandwidth or memory bandwidth.\n",
180180
"\n",
181181
"You want to reduce your bandwidth costs as much as possible.\n",
@@ -209,7 +209,7 @@
209209
"\n",
210210
"Graph capture I’m less confident explaining.\n",
211211
"\n",
212-
"But the way I think about is that graph capture or graph tracing is:\n",
212+
"But the way I think about it is that graph capture or graph tracing is:\n",
213213
"\n",
214214
"* Going through a series of operations that need to happen, such as the operations in a neural network.\n",
215215
"* And capturing or tracing what needs to happen ahead of time.\n",
@@ -272,7 +272,7 @@
272272
"source": [
273273
"## What we're going to cover\n",
274274
"\n",
275-
"Since many of the upgrades in PyTorch 2.0 are speed focused and happen behind the scenes (e.g. PyTorch takes care of them for you), in this notebook we're going to run a compartive speed test.\n",
275+
"Since many of the upgrades in PyTorch 2.0 are speed focused and happen behind the scenes (e.g. PyTorch takes care of them for you), in this notebook we're going to run a comparative speed test.\n",
276276
"\n",
277277
"Namely we'll make two of the same models, one using the default PyTorch setup and the other using the new `torch.compile()` setup and we'll train them on the same dataset.\n",
278278
"\n",
@@ -391,7 +391,7 @@
391391
"\n",
392392
"And GPUs which are datacenter-class (e.g. A100, A10, H100) are likely to see more significant speedups than desktop-class GPUs (e.g. RTX 3090, RTX 3080, RTX 3070, RTX 3060 Ti).\n",
393393
"\n",
394-
"We can check the compute capbility score of our GPU using [`torch.cuda.get_device_capability()`](https://pytorch.org/docs/stable/generated/torch.cuda.get_device_capability.html).\n",
394+
"We can check the compute capability score of our GPU using [`torch.cuda.get_device_capability()`](https://pytorch.org/docs/stable/generated/torch.cuda.get_device_capability.html).\n",
395395
"\n",
396396
"This will output a tuple of `(major, minor)` compute capability scores, for example, `(8, 0)` for the A100.\n",
397397
"\n",
@@ -737,7 +737,7 @@
737737
"* **Increasing the batch size** - More samples per batch means more samples on the GPU, for example, using a batch size of 256 instead of 32.\n",
738738
"* **Increasing data size** - For example, using larger image size, 224x224 instead of 32x32. A larger data size means that more tensor operations will be happening on the GPU.\n",
739739
"* **Increasing model size** - For example, using a larger model such as ResNet101 instead of ResNet50. A larger model means that more tensor operations will be happening on the GPU.\n",
740-
"* **Decreasing data transfer** - For example, setting up all your tensors to be on GPU memory, this minizes the amount of data transfer between the CPU and GPU.\n",
740+
"* **Decreasing data transfer** - For example, setting up all your tensors to be on GPU memory, this minimizes the amount of data transfer between the CPU and GPU.\n",
741741
"\n",
742742
"All of these result in *more* data being on the GPU.\n",
743743
"\n",

0 commit comments

Comments
 (0)