Skip to content

Commit

Permalink
fix: switch to is_floating_point and is_complex (#19)
Browse files Browse the repository at this point in the history
  • Loading branch information
samedii authored Mar 16, 2024
1 parent c77058c commit f50feae
Show file tree
Hide file tree
Showing 3 changed files with 6 additions and 6 deletions.
4 changes: 2 additions & 2 deletions ema_pytorch/ema_pytorch.py
Original file line number Diff line number Diff line change
Expand Up @@ -94,8 +94,8 @@ def __init__(

# parameter and buffer names

self.parameter_names = {name for name, param in self.ema_model.named_parameters() if param.dtype in [torch.float, torch.float16]}
self.buffer_names = {name for name, buffer in self.ema_model.named_buffers() if buffer.dtype in [torch.float, torch.float16]}
self.parameter_names = {name for name, param in self.ema_model.named_parameters() if torch.is_floating_point(param) or torch.is_complex(param)}
self.buffer_names = {name for name, buffer in self.ema_model.named_buffers() if torch.is_floating_point(buffer) or torch.is_complex(buffer)}

# tensor update functions

Expand Down
6 changes: 3 additions & 3 deletions ema_pytorch/post_hoc_ema.py
Original file line number Diff line number Diff line change
Expand Up @@ -89,8 +89,8 @@ def __init__(

# parameter and buffer names

self.parameter_names = {name for name, param in self.ema_model.named_parameters() if param.dtype in [torch.float, torch.float16]}
self.buffer_names = {name for name, buffer in self.ema_model.named_buffers() if buffer.dtype in [torch.float, torch.float16]}
self.parameter_names = {name for name, param in self.ema_model.named_parameters() if torch.is_floating_point(param) or torch.is_complex(param)}
self.buffer_names = {name for name, buffer in self.ema_model.named_buffers() if torch.is_floating_point(buffer) or torch.is_complex(buffer)}

# tensor update functions

Expand Down Expand Up @@ -402,4 +402,4 @@ def synthesize_ema_model(
return synthesized_ema_model

def __call__(self, *args, **kwargs):
return tuple(ema_model(*args, **kwargs) for ema_model in self.ema_models)
return tuple(ema_model(*args, **kwargs) for ema_model in self.ema_models)
2 changes: 1 addition & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
setup(
name = 'ema-pytorch',
packages = find_packages(exclude=[]),
version = '0.4.2',
version = '0.4.3',
license='MIT',
description = 'Easy way to keep track of exponential moving average version of your pytorch module',
author = 'Phil Wang',
Expand Down

0 comments on commit f50feae

Please sign in to comment.