Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bump black version to 23.1.0 #96578

Closed
wants to merge 4 commits into from
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
2 changes: 1 addition & 1 deletion .lintrunner.toml
Original file line number Diff line number Diff line change
Expand Up @@ -878,7 +878,7 @@ init_command = [
'tools/linter/adapters/pip_init.py',
'--dry-run={{DRYRUN}}',
'--no-black-binary',
'black==22.3.0',
'black==23.1.0',
'ufmt==1.3.3',
'usort==1.0.2',
]
Expand Down
1 change: 1 addition & 0 deletions benchmarks/dynamo/benchmarks.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@

from typing import Set


# Note - hf and timm have their own version of this, torchbench does not
# TOOD(voz): Someday, consolidate all the files into one runner instead of a shim like this...
def model_names(filename: str) -> Set[str]:
Expand Down
2 changes: 0 additions & 2 deletions benchmarks/dynamo/check_graph_breaks.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,12 +11,10 @@ def get_field(csv, model_name: str, field: str, typ=float):


def check_graph_breaks(actual_csv, expected_csv, expected_filename):

failed = []
improved = []

for model in actual_csv["name"]:

graph_breaks = get_field(actual_csv, model, "graph_breaks", typ=int)
expected_graph_breaks = get_field(expected_csv, model, "graph_breaks", typ=int)

Expand Down
2 changes: 0 additions & 2 deletions benchmarks/dynamo/ci_expected_accuracy/update_expected.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,6 @@


def query_job_sha(repo, sha):

params = {
"parameters": [
{"name": "sha", "type": "string", "value": sha},
Expand Down Expand Up @@ -108,7 +107,6 @@ def write_filtered_csvs(root_path, dataframes):


if __name__ == "__main__":

parser = argparse.ArgumentParser(
description=__doc__, formatter_class=argparse.RawDescriptionHelpFormatter
)
Expand Down
2 changes: 0 additions & 2 deletions benchmarks/dynamo/huggingface.py
Original file line number Diff line number Diff line change
Expand Up @@ -373,7 +373,6 @@ def load_model(
model_name,
batch_size=None,
):

is_training = self.args.training
use_eval_mode = self.args.use_eval_mode
dtype = torch.float32
Expand Down Expand Up @@ -513,7 +512,6 @@ def refresh_model_names_and_batch_sizes():
lm_seen = set()
family_seen = set()
for cls_name in hf_fx._SUPPORTED_MODELS:

if "For" not in cls_name:
continue

Expand Down
1 change: 0 additions & 1 deletion benchmarks/dynamo/microbenchmarks/bench_autotune_conv.py
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,6 @@ def bench_op(
warmup=25,
rep=75,
):

skip = False
# allocate inputs, nchw
x = torch.randn((BATCH, IN_C, IN_H, IN_W), dtype=dtype, device="cuda")
Expand Down
1 change: 0 additions & 1 deletion benchmarks/dynamo/microbenchmarks/bench_conv.py
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,6 @@ def bench_op(
warmup=25,
rep=75,
):

# allocate inputs, nchw
x = torch.randn((BATCH, IN_C, IN_H, IN_W), dtype=dtype, device="cuda")
w = torch.randn(
Expand Down
1 change: 0 additions & 1 deletion benchmarks/dynamo/microbenchmarks/bench_conv1x1.py
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,6 @@ def bench_op(
warmup=25,
rep=75,
):

# allocate inputs, nchw
x = torch.randn((BATCH, IN_C, IN_H, IN_W), dtype=dtype, device="cuda")
w = torch.randn(
Expand Down
1 change: 0 additions & 1 deletion benchmarks/dynamo/microbenchmarks/bench_conv_fusion.py
Original file line number Diff line number Diff line change
Expand Up @@ -236,7 +236,6 @@ def bench(layer_params, layer_id, p, fusion_types=[""]):

row = [layer_id]
for fusion_type in fusion_types:

if fusion_type == "":
conv_torchinductor = getattr(Func, "conv_torchinductor")
conv = getattr(Func, "conv")
Expand Down
1 change: 0 additions & 1 deletion benchmarks/dynamo/microbenchmarks/bench_mm_fusion.py
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,6 @@ def tflops(ms):

row = [layer_id]
for fusion_type in fusion_types:

if fusion_type == "":
fn_mm = getattr(Func, "mm")
else:
Expand Down
1 change: 0 additions & 1 deletion benchmarks/dynamo/microbenchmarks/profile_conv.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,6 @@ def profile_op(
warmup=25,
rep=50,
):

# allocate inputs, nchw
x = torch.randn((BATCH, IN_C, IN_H, IN_W), dtype=dtype, device="cuda")
w = torch.randn(
Expand Down
1 change: 1 addition & 0 deletions benchmarks/dynamo/parse_logs.py
Original file line number Diff line number Diff line change
Expand Up @@ -60,6 +60,7 @@ def chunker(seq, size):
out.writeheader()
out.writerow({"explain": gist_url})


# Sometimes backtraces will be in third party code, which results
# in very long file names. Delete the absolute path in this case.
def normalize_file(f):
Expand Down
1 change: 0 additions & 1 deletion benchmarks/dynamo/timm_models.py
Original file line number Diff line number Diff line change
Expand Up @@ -182,7 +182,6 @@ def load_model(
model_name,
batch_size=None,
):

is_training = self.args.training
use_eval_mode = self.args.use_eval_mode

Expand Down
1 change: 0 additions & 1 deletion benchmarks/dynamo/torchbench.py
Original file line number Diff line number Diff line change
Expand Up @@ -242,7 +242,6 @@ def load_model(
batch_size=None,
part=None,
):

is_training = self.args.training
use_eval_mode = self.args.use_eval_mode
dynamic_shapes = self.args.dynamic_shapes
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -120,7 +120,7 @@ def _test_fully_shard_construction(
composable_handles = traversal_utils._get_fsdp_handles(composable_module)
fsdp_wrapped_handles = traversal_utils._get_fsdp_handles(fsdp_wrapped_model)
self.assertEqual(len(composable_handles), len(fsdp_wrapped_handles))
for (composable_handle, fsdp_wrapped_handle) in zip(
for composable_handle, fsdp_wrapped_handle in zip(
composable_handles, fsdp_wrapped_handles
):
self.assertEqual(
Expand Down Expand Up @@ -179,7 +179,7 @@ def test_sync_module_states(self):
policy=policy,
sync_module_states=True,
)
for (composable_param, fsdp_wrapped_param) in zip(
for composable_param, fsdp_wrapped_param in zip(
composable_module.parameters(),
fsdp_wrapped_model.parameters(),
):
Expand Down
3 changes: 1 addition & 2 deletions test/distributed/fsdp/test_fsdp_checkpoint.py
Original file line number Diff line number Diff line change
Expand Up @@ -116,7 +116,7 @@ def _verify_parity(self, losses, outputs, models):
assert outputs
assert models

for (l, o) in zip(losses[1:], outputs[1:]):
for l, o in zip(losses[1:], outputs[1:]):
self.assertEqual(losses[0], l)
self.assertEqual(outputs[0], o)

Expand Down Expand Up @@ -324,7 +324,6 @@ def forward(self, x):


class TestFSDPCheckpointSubmodule(FSDPTest):

# TODO: grad value checks occasionally fails when use_reentrant = True
@skip_if_lt_x_gpu(2)
@parametrize("use_reentrant", [False])
Expand Down
5 changes: 0 additions & 5 deletions test/distributed/fsdp/test_fsdp_comm_hooks.py
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,6 @@ def forward(self, x):


class DummyState:

__slots__ = ["process_group", "noise"]

def __init__(self, process_group: dist.ProcessGroup, noise: int):
Expand Down Expand Up @@ -157,7 +156,6 @@ def test_default_communication_hook_behavior(
self.assertEqual(entry._communication_hook, default_hook)

for _ in range(4):

# Clear gradients
net_default_hook.zero_grad()
loss = net_default_hook(inpt).sum()
Expand All @@ -183,7 +181,6 @@ def _get_submodules(self, fsdp_net):
]

def _init_model(self, core, sharding_strategy, mixed_precision=None):

device = torch.device("cuda")
return FSDP(
core,
Expand Down Expand Up @@ -424,7 +421,6 @@ def _check_low_precision_hook(
def test_fp16_hook(
self, has_wrapping: bool, sharding_strategy: Optional[ShardingStrategy]
):

state = default_hooks.LowPrecisionState(process_group=_get_default_group())
hook = default_hooks.fp16_compress_hook

Expand Down Expand Up @@ -452,7 +448,6 @@ def test_fp16_hook(
def test_bf16_hook(
self, has_wrapping: bool, sharding_strategy: Optional[ShardingStrategy]
):

state = default_hooks.LowPrecisionState(process_group=_get_default_group())
hook = default_hooks.bf16_compress_hook

Expand Down
2 changes: 1 addition & 1 deletion test/distributed/fsdp/test_fsdp_grad_acc.py
Original file line number Diff line number Diff line change
Expand Up @@ -160,7 +160,7 @@ def permute_tensor(x: torch.Tensor):
num_iters_to_acc = sum(config.num_iters for config in configs)
for _ in range(num_iters_to_acc - 1):
batches.append(tuple(permute_tensor(t) for t in batch))
for (batch1, batch2) in itertools.combinations(batches, r=2):
for batch1, batch2 in itertools.combinations(batches, r=2):
for t1, t2 in zip(batch1, batch2):
assert not torch.all(
t1 == t2
Expand Down
1 change: 0 additions & 1 deletion test/distributed/fsdp/test_fsdp_optim_state.py
Original file line number Diff line number Diff line change
Expand Up @@ -1338,7 +1338,6 @@ def _test_rekey_optim_state_dict_to_names(
use_multiple_param_groups: bool,
use_optim_input: bool,
):

NUM_ITERS = 3
# Run a wrapped model for a few iterations
model1, optim1, optim_input1 = self._init_nested_model(
Expand Down
8 changes: 4 additions & 4 deletions test/distributed/fsdp/test_fsdp_state_dict.py
Original file line number Diff line number Diff line change
Expand Up @@ -937,14 +937,14 @@ def _create_module(wrap_fsdp=True):
# Check that it can be loaded into FSDP.
new_fsdp, _ = _create_module()
_zero_model(new_fsdp)
for (p1, p2) in zip(fsdp.parameters(), new_fsdp.parameters()):
for p1, p2 in zip(fsdp.parameters(), new_fsdp.parameters()):
self.assertNotEqual(p1, p2)
with FSDP.state_dict_type(new_fsdp, STATE_DICT_MAPPING[state_dict_type]):
if state_dict_type != "local_state_dict":
# FlatParameter has not supported deepcopy yet.
state_dict = deepcopy(state_dict)
new_fsdp.load_state_dict(state_dict, strict=True)
for (p1, p2) in zip(fsdp.parameters(), new_fsdp.parameters()):
for p1, p2 in zip(fsdp.parameters(), new_fsdp.parameters()):
self.assertEqual(p1, p2)

# Test that the checkpoint can be loaded into a local model.
Expand All @@ -954,7 +954,7 @@ def _create_module(wrap_fsdp=True):
param.zero_()

with fsdp.summon_full_params(fsdp):
for (p1, p2) in zip(fsdp.parameters(), local.parameters()):
for p1, p2 in zip(fsdp.parameters(), local.parameters()):
self.assertNotEqual(p1, p2)

if state_dict_type == "local_state_dict":
Expand All @@ -963,7 +963,7 @@ def _create_module(wrap_fsdp=True):
with fsdp.summon_full_params(fsdp):
if self.rank == 0:
local.load_state_dict(state_dict, strict=True)
for (p1, p2) in zip(fsdp.parameters(), local.parameters()):
for p1, p2 in zip(fsdp.parameters(), local.parameters()):
self.assertEqual(p1, p2)

@skip_if_lt_x_gpu(2)
Expand Down
1 change: 0 additions & 1 deletion test/distributed/fsdp/test_shard_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,6 @@ def _get_and_check_split_sizes(
out_offsets,
in_split_sizes,
):

for my_rank in range(world_size):
_in_split_sizes = in_split_sizes[my_rank]
_out_split_sizes = [
Expand Down
4 changes: 1 addition & 3 deletions test/dynamo/test_modules.py
Original file line number Diff line number Diff line change
Expand Up @@ -847,7 +847,6 @@ def __torch_function__(cls, func, types, args=(), kwargs=None):
torch._dynamo.config.traceable_tensor_subclasses.add(TensorProxy)

try:

x = torch.randn(1).as_subclass(TensorProxy)
cnt = torch._dynamo.testing.CompileCounter()
out1 = foo(x)
Expand All @@ -862,7 +861,6 @@ def __torch_function__(cls, func, types, args=(), kwargs=None):

def test_torch_function_with_closure(self):
def run():

counter = 0

def foo(x):
Expand Down Expand Up @@ -1097,7 +1095,7 @@ def forward(self, x):
opt_mod = torch._dynamo.optimize("eager")(mod)

# Check parameteres and buffers
for (p1, p2) in zip(mod.parameters(), opt_mod.parameters()):
for p1, p2 in zip(mod.parameters(), opt_mod.parameters()):
self.assertTrue(id(p1) == id(p2))

def test_recursion(self):
Expand Down
2 changes: 0 additions & 2 deletions test/dynamo/test_repros.py
Original file line number Diff line number Diff line change
Expand Up @@ -1572,7 +1572,6 @@ def forward(self, x):
self.assertEqual(y, 10)

def test_sort_out(self):

dtype = torch.float32
device = "cpu"

Expand Down Expand Up @@ -1607,7 +1606,6 @@ def forward(self, x):
self.assertTrue(same(ref, res))

def test_sigmoid_out(self):

dtype = torch.float32
device = "cpu"

Expand Down
1 change: 0 additions & 1 deletion test/inductor/test_config.py
Original file line number Diff line number Diff line change
Expand Up @@ -178,7 +178,6 @@ def d(x):
a(torch.randn(10))

def test_api_options(self):

reduce_overhead_opts = torch._inductor.list_mode_options("reduce-overhead")
self.assertEqual(reduce_overhead_opts["triton.cudagraphs"], True)

Expand Down
1 change: 0 additions & 1 deletion test/inductor/test_torchinductor_dynamic_shapes.py
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,6 @@ class DynamicShapesCudaTests(TestCase):


class TestInductorDynamic(TestCase):

compile_fn = partial(torch.compile, dynamic=True)

def setUp(self):
Expand Down
1 change: 0 additions & 1 deletion test/inductor/test_torchinductor_opinfo.py
Original file line number Diff line number Diff line change
Expand Up @@ -597,7 +597,6 @@ def fn(*args, **kwargs):
)

except Exception as e:

if test_expect is ExpectedTestResult.XFAILURE:
raise e

Expand Down
1 change: 1 addition & 0 deletions test/onnx/pytorch_test_common.py
Original file line number Diff line number Diff line change
Expand Up @@ -48,6 +48,7 @@ def wrapper(*args, **kwargs):
lambda: not torch.cuda.is_bf16_supported(), "BFloat16 CUDA is not available"
)


# skips tests for all versions below min_opset_version.
# if exporting the op is only supported after a specific version,
# add this wrapper to prevent running the test for opset_versions
Expand Down
1 change: 0 additions & 1 deletion test/onnx/test_onnx_opset.py
Original file line number Diff line number Diff line change
Expand Up @@ -494,7 +494,6 @@ def forward(self, x, grid, mode, padding_mode, align_corers):
("zeros", "border", "reflection"),
(True, False),
):

args = (
torch.randn(n, c, h_in, w_in), # x
torch.randn(n, h_out, w_out, 2), # grid,
Expand Down
2 changes: 0 additions & 2 deletions test/onnx/test_onnxscript_no_runtime.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,14 +13,12 @@


class TestONNXScriptExport(common_utils.TestCase):

# opset version is
# 1. local function is supported after opset 15
# 2. onnx-script requires users to determine opset in local function
opset_version = 15

def test_onnxscript_registration_with_multiple_models(self):

from onnxscript.onnx_opset import opset15 as op

# 1. Register Selu onnxscript function as custom Op
Expand Down
3 changes: 0 additions & 3 deletions test/onnx/test_onnxscript_runtime.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,14 +12,12 @@


class TestONNXScriptRuntime(onnx_test_common._TestONNXRuntime):

# opset version is
# 1. local function is supported after opset 15
# 2. onnx-script requires users to determine opset in local function
opset_version = 15

def test_selu_from_onnxscript_example(self):

x = torch.randn(1, 2, 3, 4, requires_grad=True)
model = torch.nn.SELU()

Expand Down Expand Up @@ -52,7 +50,6 @@ def custom_selu(g: jit_utils.GraphContext, X):
self.run_test(model, x)

def test_layer_norm(self):

x = torch.randn(2, 3)
y = torch.randn(2, 3)
z = torch.randn(2, 3)
Expand Down