Skip to content

Commit daa1ae3

Browse files
authored
Merge branch 'main' into add-api-checker
2 parents d44760a + 05c08f2 commit daa1ae3

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

intermediate_source/pinmem_nonblock.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -127,7 +127,7 @@
127127
# 1. The device must have at least one free DMA (Direct Memory Access) engine. Modern GPU architectures such as Volterra,
128128
# Tesla, or H100 devices have more than one DMA engine.
129129
#
130-
# 2. The transfer must be done on a separate, non-default cuda stream. In PyTorch, cuda streams can be handles using
130+
# 2. The transfer must be done on a separate, non-default cuda stream. In PyTorch, cuda streams can be handled using
131131
# :class:`~torch.cuda.Stream`.
132132
#
133133
# 3. The source data must be in pinned memory.

intermediate_source/torch_compile_tutorial.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -330,7 +330,7 @@ def bar(a, b):
330330
#
331331
# The second time we run ``bar``, we take the other branch of the if statement
332332
# and we get 1 traced graph corresponding to the code ``b = b * -1; return x * b``.
333-
# We do not see a graph of ``x = a / (torch.abs(a) + 1)`` outputted the second time
333+
# We do not see a graph of ``x = a / (torch.abs(a) + 1); b.sum()`` outputted the second time
334334
# since ``torch.compile`` cached this graph from the first run and re-used it.
335335
#
336336
# Let's investigate by example how TorchDynamo would step through ``bar``.

0 commit comments

Comments
 (0)