Skip to content

Conversation

@loci-dev
Copy link

@loci-dev loci-dev commented Dec 7, 2025

Mirrored from ggml-org/llama.cpp#17824

Ref: ggml-org/llama.cpp#17618

Fix: #11202

We are moving to a new CLI experience with the main code built on top of llama-server. This brings many additional features into llama-cli, making the experience feels mostly like a smaller version of the web UI:

  • Multimodal support
  • Regenerate last message
  • Speculative decoding
  • Fully jinja support (including some edge cases that old llama-cli doesn't support)
image

TODO:

@loci-dev loci-dev force-pushed the main branch 9 times, most recently from ebc7ac8 to 5b191e7 Compare December 8, 2025 16:10
@loci-dev loci-dev force-pushed the main branch 8 times, most recently from 074b005 to ff6ae69 Compare December 9, 2025 12:15
@loci-dev loci-dev force-pushed the main branch 16 times, most recently from 3595a58 to 51e8448 Compare December 11, 2025 13:20
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants