Skip to content

Commit 3ae1eff

Browse files
authored
Fix typo and formatting in bettertransformer_tutorial.rst (#2183)
1 parent b47fdca commit 3ae1eff

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

beginner_source/bettertransformer_tutorial.rst

+3-1
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ been updated to use the core library modules to benefit from fastpath accelerati
1818

1919
Better Transformer offers two types of acceleration:
2020

21-
* Native multihead attention implementation for CPU and GPU to improvee overall execution efficiency.
21+
* Native multihead attention (MHA) implementation for CPU and GPU to improve overall execution efficiency.
2222
* Exploiting sparsity in NLP inference. Because of variable input lengths, input
2323
tokens may contain a large number of padding tokens for which processing may be
2424
skipped, delivering significant speedups.
@@ -124,6 +124,7 @@ Finally, we set the benchmark iteration count:
124124
2.1 Run and benchmark inference on CPU with and without BT fastpath (native MHA only)
125125

126126
We run the model on CPU, and collect profile information:
127+
127128
* The first run uses traditional ("slow path") execution.
128129
* The second run enables BT fastpath execution by putting the model in inference mode using `model.eval()` and disables gradient collection with `torch.no_grad()`.
129130

@@ -167,6 +168,7 @@ We disable the BT sparsity:
167168
168169
169170
We run the model on DEVICE, and collect profile information for native MHA execution on DEVICE:
171+
170172
* The first run uses traditional ("slow path") execution.
171173
* The second run enables BT fastpath execution by putting the model in inference mode using `model.eval()`
172174
and disables gradient collection with `torch.no_grad()`.

0 commit comments

Comments
 (0)