Skip to content

Commit

Permalink
wrap forward passes with torch.no_grad() (#19416)
Browse files Browse the repository at this point in the history
  • Loading branch information
daspartho committed Oct 10, 2022
1 parent 870a954 commit d739a70
Showing 1 changed file with 22 additions and 15 deletions.
37 changes: 22 additions & 15 deletions tests/models/tapas/test_modeling_tapas.py
Expand Up @@ -570,7 +570,8 @@ def test_inference_no_head(self):
table, queries = prepare_tapas_single_inputs_for_inference()
inputs = tokenizer(table=table, queries=queries, return_tensors="pt")
inputs = {k: v.to(torch_device) for k, v in inputs.items()}
outputs = model(**inputs)
with torch.no_grad():
outputs = model(**inputs)
# test the sequence output
expected_slice = torch.tensor(
[
Expand Down Expand Up @@ -608,7 +609,8 @@ def test_inference_question_answering_head_conversational(self):
table, queries = prepare_tapas_single_inputs_for_inference()
inputs = tokenizer(table=table, queries=queries, return_tensors="pt")
inputs = {k: v.to(torch_device) for k, v in inputs.items()}
outputs = model(**inputs)
with torch.no_grad():
outputs = model(**inputs)
# test the logits
logits = outputs.logits
expected_shape = torch.Size((1, 21))
Expand Down Expand Up @@ -657,7 +659,8 @@ def test_inference_question_answering_head_conversational_absolute_embeddings(se
table, queries = prepare_tapas_single_inputs_for_inference()
inputs = tokenizer(table=table, queries=queries, return_tensors="pt")
inputs = {k: v.to(torch_device) for k, v in inputs.items()}
outputs = model(**inputs)
with torch.no_grad():
outputs = model(**inputs)
# test the logits
logits = outputs.logits
expected_shape = torch.Size((1, 21))
Expand Down Expand Up @@ -705,7 +708,8 @@ def test_inference_question_answering_head_weak_supervision(self):
inputs = tokenizer(table=table, queries=queries, padding="longest", return_tensors="pt")
inputs_on_device = {k: v.to(torch_device) for k, v in inputs.items()}

outputs = model(**inputs_on_device)
with torch.no_grad():
outputs = model(**inputs_on_device)
# test the logits
logits = outputs.logits
expected_shape = torch.Size((2, 28))
Expand Down Expand Up @@ -774,15 +778,16 @@ def test_training_question_answering_head_weak_supervision(self):
float_answer = torch.FloatTensor(float_answer).to(torch_device)

# forward pass to get loss + logits:
outputs = model(
input_ids=input_ids,
attention_mask=attention_mask,
token_type_ids=token_type_ids,
labels=labels,
numeric_values=numeric_values,
numeric_values_scale=numeric_values_scale,
float_answer=float_answer,
)
with torch.no_grad():
outputs = model(
input_ids=input_ids,
attention_mask=attention_mask,
token_type_ids=token_type_ids,
labels=labels,
numeric_values=numeric_values,
numeric_values_scale=numeric_values_scale,
float_answer=float_answer,
)

# test the loss
loss = outputs.loss
Expand Down Expand Up @@ -829,7 +834,8 @@ def test_inference_question_answering_head_strong_supervision(self):
table, queries = prepare_tapas_single_inputs_for_inference()
inputs = tokenizer(table=table, queries=queries, return_tensors="pt")
inputs = {k: v.to(torch_device) for k, v in inputs.items()}
outputs = model(**inputs)
with torch.no_grad():
outputs = model(**inputs)
# test the logits
logits = outputs.logits
expected_shape = torch.Size((1, 21))
Expand Down Expand Up @@ -884,7 +890,8 @@ def test_inference_classification_head(self):
table, queries = prepare_tapas_single_inputs_for_inference()
inputs = tokenizer(table=table, queries=queries, padding="longest", return_tensors="pt")
inputs = {k: v.to(torch_device) for k, v in inputs.items()}
outputs = model(**inputs)
with torch.no_grad():
outputs = model(**inputs)

# test the classification logits
logits = outputs.logits
Expand Down

0 comments on commit d739a70

Please sign in to comment.