diff --git a/src/routes/blogs/accelerating-phi-2/+page.svx b/src/routes/blogs/accelerating-phi-2/+page.svx
index f3ede6b83f2b2..218ec173e5439 100644
--- a/src/routes/blogs/accelerating-phi-2/+page.svx
+++ b/src/routes/blogs/accelerating-phi-2/+page.svx
@@ -62,7 +62,7 @@ Optimized CUDA performance for prompt throughput (i.e., the rate at which the mo
-Token generation throughput is the average throughput of the first 256 tokens generated. ONNX Runtime with float16 is **on average 6.6x faster** than torch.compile and as high as **18.55x** faster. It also performs **up to 1.64x** faster than Llama.cpp.
+Token generation throughput is the average throughput of the first 256 tokens generated. ONNX Runtime with float16 is **on average 6.6x faster** than torch.compile and **as high as 18.55x** faster. It also performs **up to 1.64x** faster than Llama.cpp.
### ORT gains with int4