4
Chat With Mistral
🔥
Generate chat responses based on user input
Llama.cpp compatible models, can be used on CPUs and GPUs!
Generate chat responses based on user input
Note A quantized model of "Mistral-7B-Instruct-v0.2" in Q2 size being served on K8s cluster
Generate chat responses with customizable prompts and settings