Skip to content

Commit

Permalink
remove comments;
Browse files Browse the repository at this point in the history
  • Loading branch information
Your Name committed May 31, 2024
1 parent d653846 commit ca40d60
Showing 1 changed file with 0 additions and 2 deletions.
2 changes: 0 additions & 2 deletions vllm/model_executor/layers/fused_moe/ampere_fp8_fused_moe.py
Original file line number Diff line number Diff line change
Expand Up @@ -119,7 +119,6 @@ def fused_moe_kernel(
)

if use_fp8:
# a_scale = tl.load(a_scale_ptr)
b_scale = tl.load(b_scale_ptr + off_experts)

# -----------------------------------------------------------
Expand All @@ -137,7 +136,6 @@ def fused_moe_kernel(
mask=token_mask[:, None] & (offs_k[None, :] < K - k * BLOCK_SIZE_K),
other=0.0,
).to(tl.float16)
# a = tl.extra.cuda.convert_uint8_as_fp8e4m3_to_float16(a)
b = tl.load(b_ptrs, mask=offs_k[:, None] < K - k * BLOCK_SIZE_K, other=0.0)

b = tl.inline_asm_elementwise(
Expand Down

0 comments on commit ca40d60

Please sign in to comment.