{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":469129528,"defaultBranch":"main","name":"pytorch","ownerLogin":"justinchuby","currentUserCanPush":false,"isFork":true,"isEmpty":false,"createdAt":"2022-03-12T15:57:41.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/11205048?v=4","public":true,"private":false,"isOrgOwned":false},"refInfo":{"name":"","listCacheKey":"v0:1716570026.0","currentOid":""},"activityList":{"items":[{"before":"ee6e5bfa4e12a072698d2a199b8fb1c9bc3b372f","after":"7f4ab136f30f30cec621d009ed584cce5dcbf7d7","ref":"refs/heads/justinchu/ghstack/exporter-new","pushedAt":"2024-05-24T18:15:32.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[ONNX] Remove model_with_state_dict\n\nghstack-source-id: f01495ee3180d7570b90afe8d623c1f487be82d7\nPull Request resolved: https://github.com/pytorch/pytorch/pull/127099","shortMessageHtmlLink":"[ONNX] Remove model_with_state_dict"}},{"before":"0d3fd1d433842e7c1afa9b8a9d951cc573a9d060","after":"ee6e5bfa4e12a072698d2a199b8fb1c9bc3b372f","ref":"refs/heads/justinchu/ghstack/exporter-new","pushedAt":"2024-05-24T17:13:50.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[ONNX] Remove model_with_state_dict\n\nghstack-source-id: ca11e1c8eadac4ca634ecf6ad2e1141c732289fa\nPull Request resolved: https://github.com/pytorch/pytorch/pull/127099","shortMessageHtmlLink":"[ONNX] Remove model_with_state_dict"}},{"before":null,"after":"0d3fd1d433842e7c1afa9b8a9d951cc573a9d060","ref":"refs/heads/justinchu/ghstack/exporter-new","pushedAt":"2024-05-24T17:00:26.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[ONNX] Use ExportedProgram in dynamo_exporter 1/n\n\nghstack-source-id: 021e9fe2e21d8a41c76446a6ab280b8b5a371015\nPull Request resolved: https://github.com/pytorch/pytorch/pull/127096","shortMessageHtmlLink":"[ONNX] Use ExportedProgram in dynamo_exporter 1/n"}},{"before":"87f44d70b195d9f4bdfa3f564b47e015d8e2ec8f","after":"2f6954c7c302c7d7702d0e3ab59fc88e6d75dc8b","ref":"refs/heads/main","pushedAt":"2024-05-24T05:18:08.000Z","pushType":"push","commitsCount":1446,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Update the modification api (#127035)\n\n# Summary\nUpdates the modification jinja template's api, so as to specify the output_name for the fixed buffer. As well updates flex-attention's usage to make the algorithm more clear/ closer align with the vmap impl\nPull Request resolved: https://github.com/pytorch/pytorch/pull/127035\nApproved by: https://github.com/Chillee","shortMessageHtmlLink":"Update the modification api (pytorch#127035)"}},{"before":"6f4fa8e9a1a0cfff2ad022e4eb454d6697ef2f59","after":"87f44d70b195d9f4bdfa3f564b47e015d8e2ec8f","ref":"refs/heads/main","pushedAt":"2024-04-19T04:35:01.000Z","pushType":"push","commitsCount":1347,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[torch/distributed] Check gloo availability when doing isinstance(pg,… (#124233)\n\nFixes a bug where a reference to `_ProcessGroupWrapper` is used without first checking whether gloo is available. This fails on pytorch builds that do not include gloo becuase `_ProcessGroupWrapper` is only pybinded when building with `USE_GLOO=1`. Therefore, creation of a new process group fails with a `NameError` when only NCCL is available as the backend.\n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/124233\nApproved by: https://github.com/rohan-varma, https://github.com/d4l3k","shortMessageHtmlLink":"[torch/distributed] Check gloo availability when doing isinstance(pg,… ("}},{"before":"8ccf8b2c470bf25a72bed1a0fce2b75c925c6e87","after":"6f4fa8e9a1a0cfff2ad022e4eb454d6697ef2f59","ref":"refs/heads/main","pushedAt":"2024-03-15T20:03:30.000Z","pushType":"push","commitsCount":397,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[inductor] FX graph cache: simplify \"current callable\" logic (#121903)\n\nSummary: The handling of the current_callable and compiled_artifact fields in the CompiledFxGraph object is unnecessarily complicated and confusing. We can simplify by storing only the callable. That field is not serializable, so the caching approach is to store a path to the generated artifact and reload from disk on a cache hit. We can just reload inline in the FX cache hit path. This change has the added benefit that it makes it easier to fallback to a \"cache miss\" if the path somehow doesn't exist.\n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/121903\nApproved by: https://github.com/eellison","shortMessageHtmlLink":"[inductor] FX graph cache: simplify \"current callable\" logic (pytorch…"}},{"before":"052e82446705cfee3ff5842fb4ebeea51b5bba79","after":"8ccf8b2c470bf25a72bed1a0fce2b75c925c6e87","ref":"refs/heads/main","pushedAt":"2024-03-05T19:59:29.000Z","pushType":"push","commitsCount":1088,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Avoid COW input materialize in more forward ops (#121070)\n\nAffected operators are: addr, cdist, sparse.sampled_addm, sparse.mm,\nmatrix_exp, softmax, cross_entropy\n\nPart of #97856\n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/121070\nApproved by: https://github.com/ezyang","shortMessageHtmlLink":"Avoid COW input materialize in more forward ops (pytorch#121070)"}},{"before":"56ef5afdee6e5a1c71f3296d77a623a62ab2d636","after":"052e82446705cfee3ff5842fb4ebeea51b5bba79","ref":"refs/heads/main","pushedAt":"2024-02-04T16:50:25.000Z","pushType":"push","commitsCount":585,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"improve CUDACachingAllocator lock contention (#118550)\n\nSummary: NativeCachingAllocator has a global lock which shows lock contention with one process using multiple GPUs. The lock is required to lookup Block from pointer. We can make the lock more fine grain to reduce the lock contention.\n\nTest Plan: existing unittests, verified on prod models using eight GPUs showing double digits improvements\n\nDifferential Revision: D52493091\n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/118550\nApproved by: https://github.com/albanD","shortMessageHtmlLink":"improve CUDACachingAllocator lock contention (pytorch#118550)"}},{"before":"f543093e065f4b248dc8c96936ae4062fc1382b0","after":"56ef5afdee6e5a1c71f3296d77a623a62ab2d636","ref":"refs/heads/main","pushedAt":"2024-01-22T18:29:54.000Z","pushType":"push","commitsCount":629,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[dynamo] Add more dynamo call_methods and getattr support or Placement (#117733)\n\nSummary:\nExplained by title.\nThis fix is part of: https://github.com/pytorch/pytorch/issues/117670\n\nTest Plan:\nUnit tetst and CI\n- Unit test: `buck2 test mode/dev-nosan //caffe2/test/distributed/_tensor:dtensor_compile -- test_placement_compile`\n\nDifferential Revision: D52863073\n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/117733\nApproved by: https://github.com/yanboliang","shortMessageHtmlLink":"[dynamo] Add more dynamo call_methods and getattr support or Placement ("}},{"before":"94f37811457400b56463d448af01cf5059f60ab0","after":"f543093e065f4b248dc8c96936ae4062fc1382b0","ref":"refs/heads/main","pushedAt":"2024-01-03T19:18:02.000Z","pushType":"push","commitsCount":207,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[ONNX] Fix output mismatch issue of repeat_interleave when dim is None (#116689)\n\n'input' is introduced but it's mixed with 'self' in repeat_interleave, which causes the mismatch issue.\n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/116689\nApproved by: https://github.com/thiagocrepaldi","shortMessageHtmlLink":"[ONNX] Fix output mismatch issue of repeat_interleave when dim is None ("}},{"before":"3e49621f3b4652b8e7782aa8dafb28f9d985598b","after":"94f37811457400b56463d448af01cf5059f60ab0","ref":"refs/heads/main","pushedAt":"2023-12-22T03:27:28.000Z","pushType":"push","commitsCount":1114,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Fixed bug with unpickling integers > 64-bits (#115264)\n\nFixes #115234\n\nCurrently, the unpickling code does not support integers larger than 64 bits in size. However, this is a part of the Python unpickling code.\n\nSee `pickle.py` in CPython:\n```\ndef decode_long(data):\n r\"\"\"Decode a long from a two's complement little-endian binary string.\n\n >>> decode_long(b'')\n 0\n >>> decode_long(b\"\\xff\\x00\")\n 255\n >>> decode_long(b\"\\xff\\x7f\")\n 32767\n >>> decode_long(b\"\\x00\\xff\")\n -256\n >>> decode_long(b\"\\x00\\x80\")\n -32768\n >>> decode_long(b\"\\x80\")\n -128\n >>> decode_long(b\"\\x7f\")\n 127\n \"\"\"\n return int.from_bytes(data, byteorder='little', signed=True)\n```\n\nE.g.:\n```\n>>> int.from_bytes(bytearray(b'\\xff\\xff\\xff\\xff\\xff\\xff\\xff\\xff\\x00'), byteorder='little', signed=True)\n18446744073709551615\n```\n\nThis PR makes it so that integers of arbitrary size are supported with JS BigNums.\n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/115264\nApproved by: https://github.com/zdevito","shortMessageHtmlLink":"Fixed bug with unpickling integers > 64-bits (pytorch#115264)"}},{"before":"48c2f89399347e65d7d6603d7c11d5838dfe6836","after":"3e49621f3b4652b8e7782aa8dafb28f9d985598b","ref":"refs/heads/main","pushedAt":"2023-11-21T01:24:47.000Z","pushType":"push","commitsCount":383,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[DTensor] Cached hash for `DTensorSpec` (#113915)\n\n**Overview**\nGenerally, I think we can try to freeze as many of these classes used in DTensor sharding propagation as possible so that we can cache hashes. This PR targets hashing `DTensorSpec`, which turns out to be relatively expensive.\n\n**Details**\nIt looks like `tensor_meta` is only updated in `_wrap_output_spec_tensor_meta`, which only runs if the propagation was not cached:\nhttps://github.com/pytorch/pytorch/blob/ae94c7e491e22f58d3df66571c1a568e51d70acd/torch/distributed/_tensor/sharding_prop.py#L137\nhttps://github.com/pytorch/pytorch/blob/ae94c7e491e22f58d3df66571c1a568e51d70acd/torch/distributed/_tensor/sharding_prop.py#L153\nIn that case, I think we can cache the hash for the `DTensorSpec` and only update it when one of the hashed attributes changes, which we only really expect to happen for `tensor_meta`.\n\nTo ensure correctness, we need that all hashed attributes are immutable.\n- `DeviceMesh` caches its hash: https://github.com/pytorch/pytorch/blob/a9134fa99a8986adf478a12db2ea5729d24554db/torch/distributed/_device_mesh.py#L181\n- This PR makes each `Placement` a frozen `dataclass`, making them immutable (relying on the fact that they do not have references to any mutable objects).\n- `TensorMeta` is a `NamedTuple` of `torch.Size`, `Tuple[int, ...]`, and `torch.dtype`, so it is immutable: https://github.com/pytorch/pytorch/blob/9916d8a9eaaf2c05c131f2a2dbe9eabeeaa9dffc/torch/distributed/_tensor/placement_types.py#L369-L375\n\n**Example**\nFor some simple small GPT model:\nBefore: 0.125 ms\n\"Screenshot\n\nAfter: 0.048 ms\n\"Screenshot\n\nThe overall Adam CPU step time decreases from 7.647 ms to 6.451 ms.\nPull Request resolved: https://github.com/pytorch/pytorch/pull/113915\nApproved by: https://github.com/wanchaol\nghstack dependencies: #113919, #113924, #114134, #113925, #113930, #114141","shortMessageHtmlLink":"[DTensor] Cached hash for DTensorSpec (pytorch#113915)"}},{"before":"e7b0bfad161fe6e6d9a2fbf111eae9843b4fa7be","after":null,"ref":"refs/heads/justinchu/onnx-contacts","pushedAt":"2023-11-13T18:43:50.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"}},{"before":"4bf2c97672e496e1420817b6ac846aee20f910a4","after":"e7b0bfad161fe6e6d9a2fbf111eae9843b4fa7be","ref":"refs/heads/justinchu/onnx-contacts","pushedAt":"2023-11-13T15:28:19.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Update docs/source/quantization.rst","shortMessageHtmlLink":"Update docs/source/quantization.rst"}},{"before":"4c240f8a305ab1a60e4f077fc83c11ac39bc6c40","after":"4bf2c97672e496e1420817b6ac846aee20f910a4","ref":"refs/heads/justinchu/onnx-contacts","pushedAt":"2023-11-13T15:27:04.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Apply suggestions from code review","shortMessageHtmlLink":"Apply suggestions from code review"}},{"before":"0dd876ffcc4d5004594bfda2093acd0fcf417e75","after":"4c240f8a305ab1a60e4f077fc83c11ac39bc6c40","ref":"refs/heads/justinchu/onnx-contacts","pushedAt":"2023-11-13T15:23:54.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Update docs/source/quantization.rst","shortMessageHtmlLink":"Update docs/source/quantization.rst"}},{"before":"6032e53882b8e6c8c99300203f1a35037f40194d","after":"0dd876ffcc4d5004594bfda2093acd0fcf417e75","ref":"refs/heads/justinchu/onnx-contacts","pushedAt":"2023-11-13T15:22:35.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Apply suggestions from code review","shortMessageHtmlLink":"Apply suggestions from code review"}},{"before":"7f627554de9f87b5c9cc0bdbe87a93b9c2ae31cd","after":"6032e53882b8e6c8c99300203f1a35037f40194d","ref":"refs/heads/justinchu/onnx-contacts","pushedAt":"2023-11-10T22:32:00.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Update docs/source/quantization.rst","shortMessageHtmlLink":"Update docs/source/quantization.rst"}},{"before":null,"after":"7f627554de9f87b5c9cc0bdbe87a93b9c2ae31cd","ref":"refs/heads/justinchu/onnx-contacts","pushedAt":"2023-11-10T22:29:30.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[ONNX] Update exporter issue report instructions for quantized models","shortMessageHtmlLink":"[ONNX] Update exporter issue report instructions for quantized models"}},{"before":"feb479757f95091903a16442d1afe1f7c9c405d3","after":"48c2f89399347e65d7d6603d7c11d5838dfe6836","ref":"refs/heads/main","pushedAt":"2023-11-10T22:21:51.000Z","pushType":"push","commitsCount":356,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[BE] Add friendly error message if you compile_fx_inner but not return tuple/list (#113451)\n\nPreviously it would fail here:\n\n```\n File \"/data/users/ezyang/a/pytorch/torch/_inductor/fx_passes/post_grad.py\", line 597, in remove_noop_ops\n for out in tuple(graph.nodes)[-1].args[0]:\n```\n\nNow you'll trigger this assert instead.\n\nSigned-off-by: Edward Z. Yang \n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/113451\nApproved by: https://github.com/albanD","shortMessageHtmlLink":"[BE] Add friendly error message if you compile_fx_inner but not retur…"}},{"before":"1b702b185e8dddadb4ad3f487f5412a02c8777e1","after":"feb479757f95091903a16442d1afe1f7c9c405d3","ref":"refs/heads/main","pushedAt":"2023-11-03T17:21:36.000Z","pushType":"push","commitsCount":329,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Make addc[mul|div] support different out dtypes (#112682)\n\nBy adding `.cast_common_dtype_to_outputs(true)` to `build_ternary_op`.\nAccording to profiling, this change does not result in additional kernel invocation on GPUs, i.e. following script\n```python\nimport torch\ndef bench_addcdiv(size=(32*1024**2, 5), device=\"cuda\"):\n x=torch.rand(size, device=device, dtype=torch.float)\n y=torch.rand(size, device=device, dtype=torch.double)\n with torch.profiler.profile(activities=[torch.profiler.ProfilerActivity.CUDA]) as prof:\n torch.addcdiv(x, x, x, out=y)\n rc=prof.key_averages()\n print(rc)\n\nif __name__ == \"__main__\":\n bench_addcdiv()\n```\nShows that before and after the change it took roughly the same time to finish the computation.\nBefore:\n```\n------------------------------------------------------- ------------ ------------ ------------ ------------ ------------ ------------ ------------ ------------ ------------ ------------\n Name Self CPU % Self CPU CPU total % CPU total CPU time avg Self CUDA Self CUDA % CUDA total CUDA time avg # of Calls\n------------------------------------------------------- ------------ ------------ ------------ ------------ ------------ ------------ ------------ ------------ ------------ ------------\n cudaLaunchKernel 92.99% 20.096ms 92.99% 20.096ms 20.096ms 0.000us 0.00% 0.000us 0.000us 1\nvoid at::native::unrolled_elementwise_kernelpytorch#112682)"}},{"before":"b600aed237fb984ff03abbfdfa9d4f488fbd7c32","after":"1b702b185e8dddadb4ad3f487f5412a02c8777e1","ref":"refs/heads/main","pushedAt":"2023-10-26T20:05:54.000Z","pushType":"push","commitsCount":95,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[pytorch-vulkan] disable one zero-dim tensor test to fix test (#112087)\n\nSummary:\nD50347338 has bug on android (not Mac, not Devserver).\n\nThis diff disable the test for time being while I identify the actual cause.\n\nTest Plan:\n## Compile on devserver\n\n```\n[yipjustin@129360.od ~/fbsource (e415d865c)]$ buck2 build -c ndk.static_linking=true -c pt.enable_qpl=0 --target-platforms=ovr_config//platform/android:arm32-fbsource //xplat/caffe2:pt_vulkan_api_test_binAndroid --show-output\nFile changed: fbcode//caffe2/aten/src/ATen/test/vulkan_api_test.cpp\nFile changed: fbsource//xplat/caffe2/aten/src/ATen/test/vulkan_api_test.cpp\nBuck UI: https://www.internalfb.com/buck2/99d47e63-ed6e-4db9-bee2-24909d647b78\nNetwork: Up: 3.2KiB Down: 67KiB (reSessionID-459e359b-773c-48a4-b129-81fde7c5e876)\nJobs completed: 4664. Time elapsed: 7.3s.\nCache hits: 100%. Commands: 38 (cached: 38, remote: 0, local: 0)\nBUILD SUCCEEDED\nfbsource//xplat/caffe2:pt_vulkan_api_test_binAndroid buck-out/v2/gen/fbsource/f1f3f9bed27e143c/xplat/caffe2/__pt_vulkan_api_test_binAndroid__/pt_vulkan_api_test_binAndroid\n```\n\n## Run test.\nadb shell /data/local/tmp/pt_vulkan_api_test_binAndroid | pastry\n\nResult: P864940908\n```\n...\n[ OK ] VulkanAPITest.lstm_success (7 ms)\n[ RUN ] VulkanAPITest.lstm_mclareninputs_success\n[ OK ] VulkanAPITest.lstm_mclareninputs_success (56 ms)\n[ RUN ] VulkanAPITest.lstm_prepack_success\n[ OK ] VulkanAPITest.lstm_prepack_success (7 ms)\n[ RUN ] VulkanAPITest.querypool_flushed_shader_log\nxplat/caffe2/aten/src/ATen/test/vulkan_api_test.cpp:7568: Skipped\nQueryPool is not available\n[ SKIPPED ] VulkanAPITest.querypool_flushed_shader_log (0 ms)\n[----------] 391 tests from VulkanAPITest (30715 ms total)\n[----------] Global test environment tear-down\n[==========] 391 tests from 1 test suite ran. (30715 ms total)\n[ PASSED ] 390 tests.\n[ SKIPPED ] 1 test, listed below:\n[ SKIPPED ] VulkanAPITest.querypool_flushed_shader_log\n YOU HAVE 7 DISABLED TESTS\n\n```\n\nReviewed By: liuk22\n\nDifferential Revision: D50668570\n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/112087\nApproved by: https://github.com/izaitsevfb, https://github.com/SS-JIA","shortMessageHtmlLink":"[pytorch-vulkan] disable one zero-dim tensor test to fix test (pytorc…"}},{"before":"ecdd1bcf0399878ed862d56be7f5a30c875ebf57","after":"b600aed237fb984ff03abbfdfa9d4f488fbd7c32","ref":"refs/heads/main","pushedAt":"2023-10-24T22:30:32.000Z","pushType":"push","commitsCount":662,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[TD] Make test class times available during CI (#111836)\n\nMakes the test class durations uploaded by https://github.com/pytorch/test-infra/pull/4670 available during CI.\nPull Request resolved: https://github.com/pytorch/pytorch/pull/111836\nApproved by: https://github.com/clee2000","shortMessageHtmlLink":"[TD] Make test class times available during CI (pytorch#111836)"}},{"before":"70f2adaec338bd33d93c373ba4f791b33adc4aaa","after":"ecdd1bcf0399878ed862d56be7f5a30c875ebf57","ref":"refs/heads/main","pushedAt":"2023-10-05T20:12:47.000Z","pushType":"push","commitsCount":578,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Back out \"[Inductor] Break the loop fusion when node2 depends on node1 mutations (#109172)\" (#110622)\n\nSummary:\nOriginal commit changeset: 03980fb054d5\n\nOriginal Phabricator Diff: D49519512\n\nBisecting shows that this diff is the cause of S369683. Since this affects Ads production, need to back out this diff immediately.\n\nTest Plan: See S369683\n\nReviewed By: ezyang\n\nDifferential Revision: D49958638\n\nPull Request resolved: https://github.com/pytorch/pytorch/pull/110622\nApproved by: https://github.com/yanboliang","shortMessageHtmlLink":"Back out \"[Inductor] Break the loop fusion when node2 depends on node…"}},{"before":"5f3156dc66e67777208f337027ddcbdb5840ac2b","after":"4f649aa86260f6c3b4d30a0c9d230435d9363a89","ref":"refs/heads/justinchu/xfail_onnxscript","pushedAt":"2023-10-04T21:44:06.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"docstring","shortMessageHtmlLink":"docstring"}},{"before":"46a4b1452e3714dae0f46489846f5940fe8dfaaa","after":"5f3156dc66e67777208f337027ddcbdb5840ac2b","ref":"refs/heads/justinchu/xfail_onnxscript","pushedAt":"2023-10-04T21:07:08.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"nll","shortMessageHtmlLink":"nll"}},{"before":"ffd62b1b83d957a5af305f7f1f9c04109f1ea4fe","after":"46a4b1452e3714dae0f46489846f5940fe8dfaaa","ref":"refs/heads/justinchu/xfail_onnxscript","pushedAt":"2023-10-04T21:00:18.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Naming","shortMessageHtmlLink":"Naming"}},{"before":"452aaacc276ad89cdc7a0ae099e6c4df226cea71","after":"ffd62b1b83d957a5af305f7f1f9c04109f1ea4fe","ref":"refs/heads/justinchu/xfail_onnxscript","pushedAt":"2023-10-04T20:51:43.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Use the decorators","shortMessageHtmlLink":"Use the decorators"}},{"before":null,"after":"452aaacc276ad89cdc7a0ae099e6c4df226cea71","ref":"refs/heads/justinchu/xfail_onnxscript","pushedAt":"2023-10-04T18:14:45.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"[ONNX] Create onnxscript-torchlib specify xfails/skips for fx tests","shortMessageHtmlLink":"[ONNX] Create onnxscript-torchlib specify xfails/skips for fx tests"}},{"before":"c96b9486c5a59e5961f4bfa3d19e8fff2b7fa20a","after":"c7321e9042281cf77f8574882c6ba8eb003ca8f7","ref":"refs/heads/justinchu/fx-opinfo","pushedAt":"2023-09-19T18:30:59.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"justinchuby","name":"Justin Chu","path":"/justinchuby","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/11205048?s=80&v=4"},"commit":{"message":"Remove skips","shortMessageHtmlLink":"Remove skips"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEU018XQA","startCursor":null,"endCursor":null}},"title":"Activity · justinchuby/pytorch"}