janhq / cortex.llamacpp

cortex.llamacpp is a high-efficiency C++ inference engine for edge computing. It is a dynamic library that can be loaded by any server at runtime.
GNU Affero General Public License v3.0
16 stars 3 forks source link

feat: CI for VNNI support #60

Open vansangpfiev opened 4 months ago

vansangpfiev commented 4 months ago

Problem Lack support for VNNI

Success Criteria

Additional context

dan-homebrew commented 3 weeks ago

@vansangpfiev I'm shifting this to Sprint 20, to triage as part of initial llama.cpp support

dan-homebrew commented 6 days ago

@vansangpfiev Quick check: is this still in scope for Sprint 21, or should we push to Sprint 22?

vansangpfiev commented 6 days ago

Please move this ticket to Sprint 22