name: "llava" license: apache-2.0 description: | LLaVA represents a novel end-to-end trained large multimodal model that combines a vision encoder and Vicuna for general-purpose visual and language understanding, achieving impressive chat capabilities mimicking spirits of the multimodal GPT-4 and setting a new state-of-the-art accuracy on Science QA. urls: - https://llava-vl.github.io/ tags: - llm - multimodal - gguf - gpu - cpu config_file: | backend: llama-cpp context_size: 4096 f16: true mmap: true roles: user: "USER:" assistant: "ASSISTANT:" system: "SYSTEM:" template: chat: | A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions. {{.Input}} ASSISTANT: