CloudLLM is a Rust library designed to seamlessly bridge applications with remote Language Learning Models (LLMs) across various platforms. With CloudLLM, you can integrate pay-as-you-go LLM APIs like OpenAI's and more, all under one unified abstraction for your app.
CloudLLM is initially implemented for Rust, but we'll provide wrappers for most popular programming languages in the near future.
// Example code on setting up a session and communicating with an LLM (this is just a placeholder for now).
Add CloudLLM to your Cargo.toml
:
[dependencies]
cloudllm = "0.1.0" # Use the latest version
Refer to the examples/
directory to see how you can set up sessions and interact with various LLM platforms using CloudLLM.
Contributions to CloudLLM are always welcome! Whether it's feature suggestions, bug reporting, or code improvements, all contributions are appreciated.
If you are to send a pull request, please make a separate branch out of main
. Try to minimize the scope of your contribution to one issue per pull request.
This project is licensed under the MIT License. See the LICENSE
file for more details.
Angel Leon