Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Adding Crusoe examples to 3p_integrations #713

Closed
ethxnp opened this issue Oct 7, 2024 · 3 comments
Closed

Adding Crusoe examples to 3p_integrations #713

ethxnp opened this issue Oct 7, 2024 · 3 comments

Comments

@ethxnp
Copy link

ethxnp commented Oct 7, 2024

🚀 The feature, motivation and pitch

Prospective Crusoe customers would like to understand how to perform common Llama workflows (inference, fine-tuning, training) on our cloud. We'd like to contribute solutions centered around Llama to the 3p_integrations repo, beginning with a tutorial on benchmarking FP8 quants served via vLLM. The tutorial covers how to deploy resources on Crusoe, start a vLLM server, run and interpret benchmarks, and finally how to create fp8 quants of existing Llama3 finetunes.

We hope for this to be the first of a series of solutions for common Llama workflows!

Alternatives

No response

Additional context

No response

@init27
Copy link
Contributor

init27 commented Oct 8, 2024

Hi @ethxnp thanks for starting the chat!

Yes-this would be really cool to have-please feel free to start a PR and tag me if you have any Qs!

@ethxnp
Copy link
Author

ethxnp commented Oct 8, 2024

@init27 thank you! Just submitted a PR

@HamidShojanazeri
Copy link
Contributor

Thansk @ethxnp for the PR

@init27 init27 closed this as completed Oct 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants