Settings

Theme

PyTorch Monarch

pytorch.org

2 points by lairv 3 months ago · 2 comments

Reader

lairvOP 3 months ago

From the docs ( https://meta-pytorch.org/monarch/index.html ):

Monarch is a distributed programming framework for PyTorch based on scalable actor messaging. It provides:

- Remote actors with scalable messaging: Actors are grouped into collections called meshes and messages can be broadcast to all members.

- Fault tolerance through supervision trees: Actors and processes for a tree and failures propagate up the tree, providing good default error behavior and enabling fine-grained fault recovery.

- Point-to-point RDMA transfers: cheap registration of any GPU or CPU memory in a process, with the one-sided tranfers based on libibverbs

- Distributed tensors: actors can work with tensor objects sharded across processes

It seems like the goal of Monarch is to do what Ray does, but more tightly integrated with the Deep Learning/distributed training ecosystem?

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection