File tree Expand file tree Collapse file tree 1 file changed +0
-20
lines changed
torch/csrc/jit/codegen/cuda Expand file tree Collapse file tree 1 file changed +0
-20
lines changed Original file line number Diff line number Diff line change @@ -410,26 +410,6 @@ void SyncMap::build(Fusion* fusion) {
410410 }
411411 }
412412
413- // If same parallel type and mapped, no need for syncs unless
414- // producer is in smem, producer parallel type is a thread
415- // dimension, and consumer concretizes the dimension. This sync is
416- // due to the redundant predicate omission in lower thread
417- // predicate.
418- auto redundant_preds = GpuLower::current ()
419- ->threadPredMap ()
420- .getPredicateInfo (producer)
421- .redundant_types ;
422-
423- if (p_id->isBroadcast () &&
424- GpuLower::current ()->concretizedBroadcastDomains ()->isConcretized (
425- p_id) &&
426- producer->getMemoryType () == MemoryType::Shared &&
427- redundant_preds.hasTID ()) {
428- redundant_preds.clearAllBID ();
429- raw_dims |= redundant_preds;
430- continue ;
431- }
432-
433413 // When the producer axis is a broadcast, it is not really
434414 // parallelized unless thread-predicated and concretized
435415 if (isParallelTypeThread (producer_ptype) && p_id->isBroadcast () &&
You can’t perform that action at this time.
0 commit comments