Рет қаралды 8,403
Simple inference code for your local PC with less than 8GB GPU: MISTRAL 7B Instruct. From Mistral AI. Mistral 7B beats LLama 2 7B and LLama 2 13B.
The next evolution of LLama models by Meta?
Grouped-query attention explained. CTransformer and GGUF and GPTQ code implementation. Mistral 7B on free COLAB NB. Live demo.
#ai
#aieducation
#coding