Commit f77a002
{{wushirong}}
Changes done internally at Facebook
483102cd4151f02c2d3632e6b6df7a5e59c0d6f3 Wei Wei <[email protected]> [fx2trt] move acc op `torch.ops._caffe2.RoIAlign` to fb only
8ce94a01caa090d56adb4708452b52890160ba69 Wei Wei <[email protected]> [aten2trt] reshape support
422326213bad177019e92c95dbc61af7a427bebc Shirong Wu <[email protected]> nan_to_num aten converter
f729c8a7f1268f329d15e3cf05f1fb9232fab2d9 Huamin Li <[email protected]> Record TRT/AIT lower context into Scuba gpu_lowering_diagnostics
2df64af6bcf102a0ce40f1c5ab8472370d012904 Wei Wei <[email protected]> [aten2ait][fx2ait] sin,cos,sqrt,clone support
9fa6469ccb9d00320d78684d748fe1a7e5c3cf60 Janet Yang <[email protected]> Split nodes w/ float64 inputs from lowering
d2ea242f721156df9e075927ea7956db772d4107 Fei Kou <[email protected]> Handle Ellipsis in dper passes
d053b097a0d1c158cde29792a35c4ec4174d9417 Jason Ansel <[email protected]> Fix tests broken by D42953629
e18c6c76b1678a95c35583dabb41666b33c3df63 Zhijing Li (Accelerator Enablement) <[email protected]> Add dper test for push_down_split pass
5008c6d200f2a9ca035547204b47eb5e1704ce88 Zhijing Li (Accelerator Enablement) <[email protected]> Add passes as option to AITTestCase.run_test
f7bc0c543b553ca2f80149995b4c28599a6ea396 Ying Zhang <[email protected]> Back out "Add passes as option to AITTestCase.run_test"
22d4044c66720e0e656f41538c81a3e90ef1a433 Zhijing Li (Accelerator Enablement) <[email protected]> Relaunch add passes as option to AITTestCase.run_test
ae0de22b6a97bca82c0ef6a14b0be2b570eb443a Eli Uriegas <[email protected]> Remove fx2trt/torch2trt backends (#93822)
b08e568951c911e4c3bbc72b55830fa1d4be4b2b Eli Uriegas <[email protected]> Remove torch/_dynamo/optimizations (#93871)
725266c0b7eb0549060e79b65346d703cc5bc39e Benson Ma <[email protected]> [T143761882] Migrate CUTLASS group gemm operators to OSS
44110f5df422e84cd9d9afbf5dfbe742057b9d98 Zhijing Li (Accelerator Enablement) <[email protected]> Add noop pass for torch.ops.fb.scale_gradient
84befb25b778485c8694ba659248d4d570d92390 Chao Gu <[email protected]> [FX] Add log_softmax
b641713bd774cb7c7bf903f514bff5c87a6f3a33 Wei Wei <[email protected]> [fx2ait] support torch.unbind, torch.group_norm
d263b38b53b93a18a78cd34b2a1c48711c3c59cd Shirong Wu <[email protected]> Add extra logging for layer norm
eb2591231195cc0ab6780f345f095807a7d45f7c Callum Ryan <[email protected]> Make GPU test run in bundled mode
f63d3834e87a819f8335c50b351e48f60573d474 Sarunya Pumma <[email protected]> Back out "[T143761882] Migrate CUTLASS group gemm operators to OSS"
a9f489c1c3a182698385053c0a94b792c4e310ba Shirong Wu <[email protected]> Change opt_profile_replica to 3
b8bdde86f0bae6010062c33aec03a4e13a87a6ab Brian Hirsh <[email protected]> forward fix for new _native_batch_norm_legit_no_training op
e8f4cbd46402e5603cc48d24395db3f0e010581a Shirong Wu <[email protected]> Fix reshape op
b860725bfaf74a0043190d1140ddee987dd82d0c generatedunixname89002005232357 <[email protected]> Revert D43278214: Multisect successfully blamed D43278214 for test or build failures
d4ea365cf8aa56d752912f7878b8046e89c804c2 Chunxing Yin <[email protected]> [mtia] Add sigmoid_backward kernel binding in Glow
a768c82a51a058e56a64ff82f90e619795611b66 Mor Tzur <[email protected]> lower to ait
8eb52426aaca586ae50fde75cccca6a0827a8328 Wei Wei <[email protected]> [hstu][fx2ait] op support
55d95ffa096d9de7952a6a1c4628efd67e554d82 Wei Wei <[email protected]> [fx2ait] temp solution to set correct dynamic batch size for jagged tensor
0a42e2f0874c48e9b60503a25705f0fc6319ff87 Jia Jiunn Ang <[email protected]> [CMF] chunk-squeeze-cat op fusion when split on last dimension
8bd509596a799f1270796772e12be090a6db5d39 Wei Wei <[email protected]> [aten2trt] update comment
1761b440d646836116fdadf2b5c7b55c7d2b989b Oleg Khabinov <[email protected]> [fx2ait] Fix a dper pass when acc_ops.squeeze doesn't have a dim
3cc405a92c9fcec886d890de87ac94e024c682a5 Jia Jiunn Ang <[email protected]> [CMF] Fuse chunk-linear-cat into baddbmm
5f42f56c5b5d0bd4c058aa280a980e64dd89b0a9 Xiaodong Wang <[email protected]> [cudnn] static linking
229969542a2c1e96fe8345ff7adc2fd48f6a0707 Romain Sauvestre <[email protected]> Remove base_module from acc_tracer target
a174195c484d5a25f06e4c0665bbb2e9d9dcae82 Janet Yang <[email protected]> Support input_tensor_spec w/ multiple batch dims in TRT
0246365e6facc6dfb13843fa9854802f35c0938a Zhijing Li (Accelerator Enablement) <[email protected]> Remove noop dropout op with acc tracer
4c287b9f6238e8bbbd80e742262a0eee6efa57de Kunming Ho <[email protected]> Operator support for threshold_backward
71bb34c81289173b83c7e7cf544b851096d9d99d Fei Kou <[email protected]> specialize_int_float to specialize_int from D43925225
037db53f89a7b863ef0fbaa7b94425fd9a08dc96 Wei Wei <[email protected]> enable torchscripting
77f3dce76fd5407b08826f67213d8299d9d48542 Adnan Akhundov <[email protected]> [fx2ait] Extend jagged tensor support
e6b551e48a0c03db63fc46ff85d975b489e30079 Jordan Fix <[email protected]> [acc_tracer] Add dont_retrace_gm option
ada3cbbb3d6c3b3631496a3bceea775f45649c6c Adam Simpkins <[email protected]> Fix a bunch of invalid Python escape warnings in torch_tensorrt
98254d631e8748a85b05851c97fb74f3e3922cfe Brandon Tran (Realtime Integrity) <[email protected]> Add torch.nn.functional.normalize to TensorRT
fce21e2248ad0fddfcc12dbe2e3a9a6ac9ea2a5f Shirong Wu <[email protected]> Fix trt input spec
a08bad1ac74a6d1409bb3f2e96953ed0c149d006 Wei Wei <[email protected]> [fx2ait] changes to improve jagged tensor and add b2b bmm
7745d70a17677777dcb5806e1e8008532f961f5d generatedunixname485339166882981 <[email protected]> [Codemod][[pyunit][static_listing] Convert python unit test dynamic listing to static listing] oncall+gpu_enablement_0
ba33951ae2d2ebc99794aff8026a01a31f9ad8da Shirong Wu <[email protected]> Add ait full op converters
b3bfd69f15fc4e32f27217a3efa8204a2f062af8 Chao Gu <[email protected]> [FX] support index_add in acc ops and tracer
a965bafc517afc81591052e355fd34062b028a89 Shirong Wu <[email protected]> Make fill op read dtype from input/kwarg
72f9b0925eceffc12dfa51769c1bd0cb38a3e50c generatedunixname485339166882981 <[email protected]> [Codemod][[pyunit][static_listing] Convert python unit test dynamic listing to static listing] oncall+gpu_enablement
2e7feece191d6178ff6ec750d8fe481175bb27b9 Max Podkorytov <[email protected]> [fx2ait] enable lowering to bfloat16
94607911ffb11e78082e061a670b5140e9a55d72 Archie Sravankumar <[email protected]> Add support for nan_to_num
42fddd20d303dbbc3355a8c09a86d4a74317be97 Max Podkorytov <[email protected]> [AITemplate] feed_lower_benchmark cli argument tweak for choosing precision
648ec682f2214e67912fe7c800f7ca059195cf4e Huamin Li <[email protected]> Re-enable previous disabled TRT unit tests
3e5c2aac8a7b9e50efe04fcae361a3c0ee1777a7 Janet Yang <[email protected]> Skip acc normalization of repeat_interleave if input dims aren't integral
f412f35baeee9a1b17f67b7749ca1f9b8cbbe77b Janet Yang <[email protected]> Skip acc normalization of repeat if dims aren't ints
5b9cfe428f29e27da76b19029bda03a8b43c17d1 Huamin Li <[email protected]> add import into generate_standalone_repro
9f88965e87e72658aa6a4973dc870d50b8a22ca4 Fei Kou <[email protected]> lowering with bf16
7f761df34d672c87c40b18369b28bc593374122c Fei Kou <[email protected]> [benchmark] Support bfloat16 in mts_gpu_benchmark
fa9b09e11ba8f888d761e1398367973d30e0aa1e Wei Wei <[email protected]> [fx2ait] add a simple eager run to verify the input generatation is correct
4f8ca36dbdc72dfa60e667c3592d0a2bc466b994 Max Podkorytov <[email protected]> [AITemplate] implement per op benchmark backbone
9873be1e82f2dd4a8a768497ac9cdb3b9b95cfe9 Thomas Orozco <[email protected]> buck2: tag more long running tests
0d6827c464aa2141a48a8d768a8c7facd65c0bc4 generatedunixname485339166882981 <[email protected]> [Codemod][[pyunit][static_listing] Convert python unit test dynamic listing to static listing] oncall+gpu_enablement_0_2ea3
04f9c1105a2a6a711d025d5c85b95147343d0ecd Zhijing Li (Accelerator Enablement) <[email protected]> [fx2ait] Fix acc_ops converter on std when keepdim=False
906bad1deebb235a9c80d0f0d46145da08afa091 Danylo Baibak <[email protected]> Forward fix to switch calling convention back to real tensors
48ffa2ab3dd66487922f9f0bf9a145db6eaf3fe2 Kefei Lu <[email protected]> Lowering: validate inference with alternative batch sizes
ca5dc1a2896bd476e3a327db834df859a3fcc11f Jordan Fix <[email protected]> [fba_pass_manager_builder][BE] General cleanup/refactor
afb4df5e84571f466b0f385472493aefb89344cc Shirong Wu <[email protected]> Mask select converter
25e8afb1f8be19ec6c4ef4bc74ea48e64017cde2 Janet Yang <[email protected]> Fix lowering FusedTsEncodingModule for coffee model
7fdf06ecfc6b4efb7008ce399dcd0c32ef1f1f75 generatedunixname485339166882981 <[email protected]> [Codemod][[pyunit][static_listing] Convert python unit test dynamic listing to static listing] oncall+deprecated_techdebt_do_not_use_4_9c34
a58c5e454412585c4cc48ced1798dbf234cc13b6 Michael Liu <[email protected]> Initialize `output_count` in `get_model_info_str`
2c6f13ddcc52e8f833fcd164d0c479ca3398322e Wei Wei <[email protected]> jagged SHA and MHA module support
2fe5c7cd3b763b839af3d1b05eecc73f1df05286 Shirong Wu <[email protected]> Add BF16 support for ads model
2486edbe5013f3b7e5807503538f3164bdd4ee19 Shirong Wu <[email protected]> Add low_level_module conversion pass
ca7c51407ab0410d311c984b31aeb757dd840bc2 Wei Wei <[email protected]> [hstu] remove torch_package from RelativeBucketedTimeAndPositionBasedBias after packaged
80596e459343d5630e16a6175eafffd2c25a3123 Shirong Wu <[email protected]> Block a pass that yield problem
ded609195500a8edc5bed80ee85f41b35224c19f Huamin Li <[email protected]> Do not test test_implicit_batch_dim if >= 8.6
8e8e736e14d23e77fa2bd5e72123d66943f7716f Huamin Li <[email protected]> Speed up TRT compile time in test env
2db82572e509cfe827c34a4060c058ae44b5547a Jordan Fix <[email protected]> [acc_tracer] Add in use_concrete_args option to enable pytree flatten/unflatten
946f957b6636c6b4f64e52148c9baf6e0351fb5e Wei Wei <[email protected]> [hstu] changes to bias module and sha/mha pass to adapt to removing presences
d904b26386c2ef98f292edae7c5e98c27119f9d9 Oleg Khabinov <[email protected]> [fx2ait] Rename split_op_to_reshape to chunk_op_to_reshape
ca36733f0ea67aeeb38a3740f795bbf99b24037b Oleg Khabinov <[email protected]> [fx2ait] Rewrite chunk_op_to_reshape() to use while loop instead of recursion
4361feb4399eec3816b534991020703d099d2896 Oleg Khabinov <[email protected]> [fx2ait] Optimize chunk_op_to_reshape()
071b84e3cda4f0175b37ae62c37b2d4f2de7925f Huamin Li <[email protected]> Disable libkineto for TRT unit tests
92f9acaac8f9a8f0fc2e1382bf4c79d0b94cbea5 Wei Wei <[email protected]> [fx2ait] improve bf16 support
8b92e8356278eb9676a5299373841593af942fb4 Jongsoo Park <[email protected]> [acc_tracer] skip None module in rewriting
0d1d644bad22c86efec12009ca1464587d1e7d38 Kefei Lu <[email protected]> Remove non-existent argument doc string
2efe5e78bc8627a30ba132e5b8e14e06538d463f shirong <[email protected]> Temp fix
a15a564a567eb689604d27ca814553e38c287698 shirong <[email protected]> Temporary commit at 4/24/2023, 2:32:22 PM
78825462243c09760ebb73156a4c18bbc9ddee75 shirong <[email protected]> Temporary commit at 4/24/2023, 2:32:37 PM
9bfea274462fd77cb04c38c17bc237541af87c55 laksrm <[email protected]> [DNR] onboard ctr to aimp with lowering fixes
8bb482b10f7f63270c329c88d5ac028b40f6b757 shirong <[email protected]> Reenable pass1 parent d4e5ed0 commit f77a002
File tree
26 files changed
+1193
-197
lines changed- py/torch_tensorrt/fx
- converters
- passes
- test
- converters
- aten_op
- vanilla
- core
- passes
- tracer
- trt_lower
- tools
- tracer
- acc_tracer
- dispatch_tracer
26 files changed
+1193
-197
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
691 | 691 | | |
692 | 692 | | |
693 | 693 | | |
| 694 | + | |
694 | 695 | | |
695 | | - | |
| 696 | + | |
696 | 697 | | |
697 | | - | |
| 698 | + | |
| 699 | + | |
| 700 | + | |
698 | 701 | | |
699 | 702 | | |
700 | 703 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
87 | 87 | | |
88 | 88 | | |
89 | 89 | | |
| 90 | + | |
90 | 91 | | |
91 | 92 | | |
92 | 93 | | |
93 | 94 | | |
94 | 95 | | |
95 | 96 | | |
| 97 | + | |
96 | 98 | | |
97 | 99 | | |
98 | 100 | | |
| |||
117 | 119 | | |
118 | 120 | | |
119 | 121 | | |
| 122 | + | |
| 123 | + | |
| 124 | + | |
120 | 125 | | |
121 | 126 | | |
122 | 127 | | |
| |||
271 | 276 | | |
272 | 277 | | |
273 | 278 | | |
| 279 | + | |
| 280 | + | |
| 281 | + | |
274 | 282 | | |
275 | 283 | | |
276 | 284 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
1 | 1 | | |
| 2 | + | |
2 | 3 | | |
3 | 4 | | |
4 | 5 | | |
| |||
211 | 212 | | |
212 | 213 | | |
213 | 214 | | |
| 215 | + | |
| 216 | + | |
| 217 | + | |
| 218 | + | |
| 219 | + | |
214 | 220 | | |
215 | 221 | | |
216 | 222 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
1 | | - | |
| 1 | + | |
2 | 2 | | |
3 | 3 | | |
4 | 4 | | |
| |||
18 | 18 | | |
19 | 19 | | |
20 | 20 | | |
| 21 | + | |
| 22 | + | |
| 23 | + | |
| 24 | + | |
| 25 | + | |
| 26 | + | |
21 | 27 | | |
22 | 28 | | |
23 | 29 | | |
| |||
26 | 32 | | |
27 | 33 | | |
28 | 34 | | |
| 35 | + | |
29 | 36 | | |
30 | 37 | | |
31 | 38 | | |
| |||
147 | 154 | | |
148 | 155 | | |
149 | 156 | | |
150 | | - | |
| 157 | + | |
151 | 158 | | |
152 | 159 | | |
153 | 160 | | |
154 | 161 | | |
155 | 162 | | |
156 | 163 | | |
157 | | - | |
158 | | - | |
159 | | - | |
160 | | - | |
161 | | - | |
162 | | - | |
163 | | - | |
164 | | - | |
165 | | - | |
| 164 | + | |
| 165 | + | |
| 166 | + | |
| 167 | + | |
| 168 | + | |
| 169 | + | |
| 170 | + | |
| 171 | + | |
| 172 | + | |
| 173 | + | |
| 174 | + | |
| 175 | + | |
166 | 176 | | |
167 | 177 | | |
168 | 178 | | |
| 179 | + | |
| 180 | + | |
| 181 | + | |
| 182 | + | |
| 183 | + | |
| 184 | + | |
| 185 | + | |
| 186 | + | |
| 187 | + | |
| 188 | + | |
| 189 | + | |
| 190 | + | |
| 191 | + | |
| 192 | + | |
| 193 | + | |
| 194 | + | |
| 195 | + | |
| 196 | + | |
| 197 | + | |
| 198 | + | |
| 199 | + | |
| 200 | + | |
| 201 | + | |
| 202 | + | |
| 203 | + | |
| 204 | + | |
| 205 | + | |
| 206 | + | |
| 207 | + | |
| 208 | + | |
| 209 | + | |
| 210 | + | |
| 211 | + | |
| 212 | + | |
| 213 | + | |
| 214 | + | |
| 215 | + | |
| 216 | + | |
| 217 | + | |
| 218 | + | |
| 219 | + | |
169 | 220 | | |
170 | 221 | | |
171 | 222 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
41 | 41 | | |
42 | 42 | | |
43 | 43 | | |
| 44 | + | |
| 45 | + | |
44 | 46 | | |
45 | 47 | | |
46 | 48 | | |
| |||
81 | 83 | | |
82 | 84 | | |
83 | 85 | | |
| 86 | + | |
| 87 | + | |
84 | 88 | | |
85 | 89 | | |
86 | 90 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
54 | 54 | | |
55 | 55 | | |
56 | 56 | | |
57 | | - | |
58 | | - | |
59 | | - | |
60 | | - | |
| 57 | + | |
| 58 | + | |
| 59 | + | |
| 60 | + | |
| 61 | + | |
| 62 | + | |
| 63 | + | |
| 64 | + | |
61 | 65 | | |
62 | 66 | | |
63 | 67 | | |
| |||
66 | 70 | | |
67 | 71 | | |
68 | 72 | | |
69 | | - | |
| 73 | + | |
70 | 74 | | |
71 | 75 | | |
72 | 76 | | |
| |||
630 | 634 | | |
631 | 635 | | |
632 | 636 | | |
| 637 | + | |
| 638 | + | |
| 639 | + | |
| 640 | + | |
| 641 | + | |
| 642 | + | |
| 643 | + | |
| 644 | + | |
| 645 | + | |
| 646 | + | |
| 647 | + | |
| 648 | + | |
| 649 | + | |
| 650 | + | |
| 651 | + | |
| 652 | + | |
| 653 | + | |
| 654 | + | |
| 655 | + | |
| 656 | + | |
| 657 | + | |
| 658 | + | |
| 659 | + | |
| 660 | + | |
| 661 | + | |
| 662 | + | |
| 663 | + | |
| 664 | + | |
| 665 | + | |
| 666 | + | |
| 667 | + | |
| 668 | + | |
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
8 | 8 | | |
9 | 9 | | |
10 | 10 | | |
| 11 | + | |
11 | 12 | | |
12 | 13 | | |
13 | 14 | | |
| |||
229 | 230 | | |
230 | 231 | | |
231 | 232 | | |
232 | | - | |
233 | 233 | | |
234 | 234 | | |
235 | | - | |
| 235 | + | |
236 | 236 | | |
237 | 237 | | |
238 | 238 | | |
| |||
251 | 251 | | |
252 | 252 | | |
253 | 253 | | |
| 254 | + | |
| 255 | + | |
| 256 | + | |
254 | 257 | | |
255 | 258 | | |
256 | 259 | | |
| |||
0 commit comments