run-llama / create-llama

The easiest way to get started with LlamaIndex
MIT License
886 stars 111 forks source link

T-System's LLMHUB is added as model provider backend. #139

Closed mohdamir closed 3 months ago

mohdamir commented 3 months ago

We have added T-Systems as model provider where user can choose various state of the art LLMs provided by T-Systems. User will need API-Key to use models which can be requested to T-Systems LLM Hub team. To know more about models available in LLM Hub. Please refer following URL.

https://docs.llmhub.t-systems.net/

Summary by CodeRabbit

changeset-bot[bot] commented 3 months ago

🦋 Changeset detected

Latest commit: dd79094bdb3b8fb5e7cd8fe274ad3f6dad105fd8

The changes in this PR will be included in the next version bump.

This PR includes changesets to release 1 package | Name | Type | | ------------ | ----- | | create-llama | Patch |

Not sure what this means? Click here to learn what changesets are.

Click here if you're a maintainer who wants to add another changeset to this PR

coderabbitai[bot] commented 3 months ago

Walkthrough

This update introduces support for T-Systems as a new model provider. It expands environment configurations and API interactions related to T-Systems' LLM Hub API, includes necessary dependencies, and updates the model configuration options and types. The changes also include enhancements to the askLLMHubQuestions function to streamline user prompts and manage API key interactions.

Changes

Files/Paths Change Summary
helpers/env-variables.ts Added T-Systems provider environment variables for base URL and API key.
helpers/providers/index.ts Introduced T-Systems as a new model provider in the askModelConfig function choices.
helpers/python.ts Included additional dependencies needed for T-Systems support.
helpers/types.ts Expanded ModelProvider to include T-Systems and added apiBase field to ModelConfig type.
questions.ts Added framework parameter to the askModelConfig function call within askQuestions function.
helpers/providers/llmhub.ts Introduced functions and types for interacting with LLMHub API, including askLLMHubQuestions.

Sequence Diagram(s)

sequenceDiagram
    participant User
    participant App
    participant LLMHubAPI

    User->>App: Choose T-Systems provider
    App->>User: Prompt for API Key
    User->>App: Enter API Key
    App->>LLMHubAPI: Retrieve Model Choices with API Key
    LLMHubAPI->>App: Return Model Choices
    App->>User: Display Model Choices
    User->>App: Select Desired Model
    App->>LLMHubAPI: Configure Selected Model
    LLMHubAPI->>App: Confirm Configuration
    App->>User: Display Configuration Success

Poem

In coding's dance, we took a chance,
With T-Systems now in our expanse.
URL and keys align with ease,
A rabbit's leap through fields of these.
Dependencies set, new horizons met,
With models ready, no regrets.
🐇✨


[!TIP]

Early access features: enabled We are currently testing the following features in early access: - **OpenAI `gpt-4o` model for code reviews and chat**: OpenAI claims that this model is better at understanding and generating code than the previous models. We seek your feedback over the next few weeks before making it generally available. Note: - You can enable or disable early access features from the CodeRabbit UI or by updating the CodeRabbit configuration file. - Please join our [Discord Community](https://discord.com/invite/GsXnASn26c) to provide feedback and report issues. - OSS projects are currently opted into early access features by default.

Thank you for using CodeRabbit. We offer it for free to the OSS community and would appreciate your support in helping us grow. If you find it useful, would you consider giving us a shout-out on your favorite social media?

Share - [X](https://twitter.com/intent/tweet?text=I%20just%20used%20%40coderabbitai%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20the%20proprietary%20code.%20Check%20it%20out%3A&url=https%3A//coderabbit.ai) - [Mastodon](https://mastodon.social/share?text=I%20just%20used%20%40coderabbitai%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20the%20proprietary%20code.%20Check%20it%20out%3A%20https%3A%2F%2Fcoderabbit.ai) - [Reddit](https://www.reddit.com/submit?title=Great%20tool%20for%20code%20review%20-%20CodeRabbit&text=I%20just%20used%20CodeRabbit%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20proprietary%20code.%20Check%20it%20out%3A%20https%3A//coderabbit.ai) - [LinkedIn](https://www.linkedin.com/sharing/share-offsite/?url=https%3A%2F%2Fcoderabbit.ai&mini=true&title=Great%20tool%20for%20code%20review%20-%20CodeRabbit&summary=I%20just%20used%20CodeRabbit%20for%20my%20code%20review%2C%20and%20it%27s%20fantastic%21%20It%27s%20free%20for%20OSS%20and%20offers%20a%20free%20trial%20for%20proprietary%20code)
Tips ### Chat There are 3 ways to chat with [CodeRabbit](https://coderabbit.ai): - Review comments: Directly reply to a review comment made by CodeRabbit. Example: - `I pushed a fix in commit .` - `Generate unit testing code for this file.` - `Open a follow-up GitHub issue for this discussion.` - Files and specific lines of code (under the "Files changed" tab): Tag `@coderabbitai` in a new review comment at the desired location with your query. Examples: - `@coderabbitai generate unit testing code for this file.` - `@coderabbitai modularize this function.` - PR comments: Tag `@coderabbitai` in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples: - `@coderabbitai generate interesting stats about this repository and render them as a table.` - `@coderabbitai show all the console.log statements in this repository.` - `@coderabbitai read src/utils.ts and generate unit testing code.` - `@coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.` - `@coderabbitai help me debug CodeRabbit configuration file.` Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments. ### CodeRabbit Commands (invoked as PR comments) - `@coderabbitai pause` to pause the reviews on a PR. - `@coderabbitai resume` to resume the paused reviews. - `@coderabbitai review` to trigger an incremental review. This is useful when automatic reviews are disabled for the repository. - `@coderabbitai full review` to do a full review from scratch and review all the files again. - `@coderabbitai summary` to regenerate the summary of the PR. - `@coderabbitai resolve` resolve all the CodeRabbit review comments. - `@coderabbitai configuration` to show the current CodeRabbit configuration for the repository. - `@coderabbitai help` to get help. Additionally, you can add `@coderabbitai ignore` anywhere in the PR description to prevent this PR from being reviewed. ### CodeRabbit Configration File (`.coderabbit.yaml`) - You can programmatically configure CodeRabbit by adding a `.coderabbit.yaml` file to the root of your repository. - Please see the [configuration documentation](https://docs.coderabbit.ai/guides/configure-coderabbit) for more information. - If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: `# yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json` ### Documentation and Community - Visit our [Documentation](https://coderabbit.ai/docs) for detailed information on how to use CodeRabbit. - Join our [Discord Community](https://discord.com/invite/GsXnASn26c) to get help, request features, and share feedback. - Follow us on [X/Twitter](https://twitter.com/coderabbitai) for updates and announcements.