Skip to content

Commit 7edb1b6

Browse files
committed
Upd
1 parent 3175f33 commit 7edb1b6

File tree

1 file changed

+0
-13
lines changed

1 file changed

+0
-13
lines changed

flashinfer/comm/mnnvl.py

Lines changed: 0 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -547,7 +547,6 @@ def supports_mnnvl() -> bool:
547547

548548
class McastDeviceMemory:
549549
"""Python port of McastDeviceMemory from TensorRT-LLM"""
550-
# config: Optional[MnnvlConfig] = None
551550
def __init__(
552551
self,
553552
buf_size: int,
@@ -650,7 +649,6 @@ def __init__(
650649
self.signal_pads_dev = alloc_and_copy_to_cuda(self.signal_pads)
651650
self.uc_ptrs_dev = alloc_and_copy_to_cuda(self.uc_ptrs)
652651

653-
654652
def __del__(self):
655653
"""Destructor - cleanup allocated memory"""
656654

@@ -769,14 +767,6 @@ def _alloc_mn_mcast_mem(self, buf_size: int, comm: Any=MpiComm()):
769767
except Exception as e:
770768
print(f"Error checking CUDA context: {e}")
771769

772-
# Get MPI communicator
773-
# comm = MpiComm()
774-
# comm = McastDeviceMemory.get_comm()
775-
# if config:
776-
# comm = config.comm_backend
777-
# else:
778-
# comm = MpiComm()
779-
780770
# Set up allocation properties
781771
handle_type = cuda.CUmemAllocationHandleType.CU_MEM_HANDLE_TYPE_FABRIC
782772

@@ -838,9 +828,6 @@ def _alloc_mn_mcast_mem(self, buf_size: int, comm: Any=MpiComm()):
838828
)
839829

840830
# All-gather fabric handles
841-
print(my_fabric_handle.data)
842-
print(type(my_fabric_handle.data))
843-
# all_fabric_handles=[my_fabric_handle.data] * 4
844831
all_fabric_handles = comm.allgather(my_fabric_handle.data)
845832
cuda.cuCtxSynchronize()
846833

0 commit comments

Comments
 (0)