ibm-granite-community / pm

Granite Community Project Management
0 stars 0 forks source link

Using Granite Code models in VS Code with Continue plugin #32

Closed adampingel closed 3 months ago

adampingel commented 4 months ago

See Gabe Goodhart's blog post

https://developer.ibm.com/tutorials/awb-local-ai-copilot-ibm-granite-code-ollama-continue/

Some additional resources:

Continue

fayvor commented 3 months ago

@adampingel how would I get access to this link? https://github.ibm.com/ghart/writing/blob/main/blog-posts/granite-continue-copilot.md

It's asking me to sign on with w3id: https://login.w3.ibm.com/idaas/mtfim/sps/idaas/login?runtime=true&Target=https%3A%2F%2Flogin.w3.ibm.com%2Fsaml%2Fsps%2Fauth

hickeyma commented 3 months ago

@fayvor The link in the description was incorrect. I have updated it now to: https://developer.ibm.com/tutorials/awb-local-ai-copilot-ibm-granite-code-ollama-continue/

deanwampler commented 3 months ago

It has also been republished in two parts here:

fayvor commented 3 months ago

Thanks guys!

fayvor commented 3 months ago

Looks like I need additional permissions to claim or change the status of this ticket, but pls consider it claimed.

fayvor commented 3 months ago

I've put up a draft PR here for feedback.

fayvor commented 3 months ago

PR up for review; doesn't look like I can assign it.

deanwampler commented 3 months ago

I'll review it shortly...

fayvor commented 3 months ago

PR has been merged.

adampingel commented 3 months ago

Tagging @gabe-l-hart for awareness

gabe-l-hart commented 3 months ago

Thanks for getting this all together @fayvor! The one comment I have on the notebook in the PR is that it doesn't address the code-completion model. I've found that the 3b 128k context model (which I'm in the process of getting up on ollama) has very solid auto-completion performance.

fayvor commented 3 months ago

Hi @gabe-l-hart! Thanks for pointing that out. I'll try it with ollama. Here is a ticket for that.

We'll also want to provide Continue configs to point these at remote models on Replicate (once they are kept hot) and WatsonX.

gabe-l-hart commented 3 months ago

Ah, yeah, that will be really nice to show how the same models can be used running elsewhere. Do we have a plan for usage limiting in Replicate, or is it more of "if you deploy it in Replicate, here's how you can configure Continue?"

fayvor commented 3 months ago

We have a plan for hosting those models publicly on Replicate.

When you say 3b 128k, what does the 128k describe? And what is the 3b model that is in the ollama library?

fayvor commented 3 months ago

Claude says 128k is the context window :) What is the default context window? 1024 tokens?

gabe-l-hart commented 3 months ago

🤦 Yep, it's a terrible way to version them. You're correct that 128k is the context length. The models currently on ollama have 2k (3b), 4k (8b), and 8k (20b) context lengths which can be very limiting as a code assistant. The new models vastly extend this to support much larger context windows.