Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is it possible to run DeepSeek-R1-Distill-Llama-70B or a 4-bit quantified version on consumer grade computers? This is of great significance #244

Open
wusu2016 opened this issue Feb 13, 2025 · 1 comment
Labels
question Further information is requested

Comments

@wusu2016
Copy link

1.例如显存在 16-22GB之间,兼容 4060ti,4070,4080,4090,此类用户不需要更换显卡,节省费用,避免去为了SLI, 去弄显卡交火,增加成本
2.70B比 32B, 14B 聪明,用于写代码,绘图够用,
3.deepseek已是国内较知名的模型, 流量大
4. 最好可以提升tokes/s,节省使用时间
此项目必火

@wusu2016 wusu2016 added the question Further information is requested label Feb 13, 2025
@ejrydhfs
Copy link

Not right now as this project currently is optimized for a few transformer ai model architectures and deepseek R1 does not use one of them. I believe it would be best to work on making it work in a fork since this project has not been updated in a long time

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants