Open fwcd opened 1 week ago
Highly quantized language models that can run locally are getting more and more popular with even Chrome shipping a Gemini Nano model in their latest canary builds. Models like Phi-3-mini already achieve impressive performance for being comparatively small and support cross-platform inference using a Rust library named candle.
candle
It would be cool if we could bundle such a model with D2, e.g. as a command and/or as a Conversator.
Conversator
llama.cpp and llama.swift would be worth investigating, even though the latter might be primarily targeting Apple platforms.
llama.cpp
llama.swift
Highly quantized language models that can run locally are getting more and more popular with even Chrome shipping a Gemini Nano model in their latest canary builds. Models like Phi-3-mini already achieve impressive performance for being comparatively small and support cross-platform inference using a Rust library named
candle
.It would be cool if we could bundle such a model with D2, e.g. as a command and/or as a
Conversator
.