Skip to content

Commit

Permalink
suppress errors in torchrec (#390)
Browse files Browse the repository at this point in the history
Summary: Pull Request resolved: #390

Differential Revision: D36776440

fbshipit-source-id: 3885594f18f64a63acf95218a17ee3ca38baa20e
  • Loading branch information
Pyre Bot Jr authored and facebook-github-bot committed Jun 1, 2022
1 parent 83c1de7 commit 08d853f
Show file tree
Hide file tree
Showing 2 changed files with 0 additions and 2 deletions.
1 change: 0 additions & 1 deletion torchrec/modules/fused_embedding_bag_collection.py
Original file line number Diff line number Diff line change
Expand Up @@ -441,7 +441,6 @@ def __init__(
self._emb_modules.append(emb_module)
params: Dict[str, torch.Tensor] = {}
for param_key, weight in emb_module.fused_optimizer.params.items():
# pyre-ignore
params[f"embedding_bags.{param_key}"] = weight
optims.append(("", emb_module.fused_optimizer))

Expand Down
1 change: 0 additions & 1 deletion torchrec/optim/keyed.py
Original file line number Diff line number Diff line change
Expand Up @@ -222,7 +222,6 @@ def init_state(
# pyre-ignore [16]
t = t.to_sparse()
# pyre-ignore [41]
# pyre-fixme[16]: `ShardedTensor` has no attribute `grad`.
param.grad = torch.autograd.Variable(t)
self.step(closure=None)

Expand Down

0 comments on commit 08d853f

Please sign in to comment.