Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] Multi-node multi-GPU accelerate quantization #1139

Open
nicklausbrown opened this issue Feb 12, 2025 · 1 comment
Open

[Question] Multi-node multi-GPU accelerate quantization #1139

nicklausbrown opened this issue Feb 12, 2025 · 1 comment
Assignees

Comments

@nicklausbrown
Copy link

Hello,

I have 2 nodes which contain 2 4090s each with a mellanox dual 25gbe NIC as RoCE interconnect. I'd like to know if it is possible to run llm-compressor in "distributed mode" leveraging accelerate's ability to handle multi-node training. I may be misunderstanding the functionality, but if not I would be grateful to know in an example how to leverage multiple nodes' GPUs.

Thank you for a great tool!

@brian-dellabetta
Copy link
Collaborator

brian-dellabetta commented Feb 28, 2025

Hi @nicklausbrown , you likely will not need to train on multiple nodes for the compression algorithms we are providing here. We are running calibration training, usually involving caching activations of a single batch of data and performing some compression based on the results. Are you are looking to do post-training afterward? It would certainly benefit from multi-node but is not needed for most of the pipelines we are currently supporting.

@brian-dellabetta brian-dellabetta self-assigned this Feb 28, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants