You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm using torchrec to make model parallelization. Is that possible to manually setting up a group of sparse features to be sharded in the same machine? For example, I have 3 sparse features A, B and C. They are interact frequently in my model so I'd like them to be in the same GPU to reduce the communication between machines. Is there any way to make it in torchrec?
Thanks very much!
The text was updated successfully, but these errors were encountered:
Hi @shan-jiang-faire, I'm assuming these 3 features will be in separate embedding tables? afaik, if you're using the TorchRec planner to generate a sharding plan, there isn't an easy way to enforce all 3 tables will be on the same GPU/rank. However, you can use the construct_module_sharding_plan API [source] to manually define a sharding plan such that the 3 tables are table-wise sharded on the same rank.
I'm using torchrec to make model parallelization. Is that possible to manually setting up a group of sparse features to be sharded in the same machine? For example, I have 3 sparse features A, B and C. They are interact frequently in my model so I'd like them to be in the same GPU to reduce the communication between machines. Is there any way to make it in torchrec?
Thanks very much!
The text was updated successfully, but these errors were encountered: