- Joined
- Jan 28, 2018
Early quantizations of Mistral medium (the "big brother" of Mixtral the company Mistral exclusively only offers via API) that were only intended for internal usage have been leaked by some employee. It's downloadable here.
The thread "Might consider attribution" was posted by Arthur Mensch, CEO of Mistral AI. They also confirm this leak on twitter.
I tried it out. Honestly doesn't feel much smarter than any llama 2 70b, but has an actually working context of 32k and adheres to instructions better than llama. It might or might not be smarter than mixtral. I am actually not sure.
In other news, Meta released Codellama 70b. The instruct tune is so censored that it can't do the simplest things.

The thread "Might consider attribution" was posted by Arthur Mensch, CEO of Mistral AI. They also confirm this leak on twitter.
I tried it out. Honestly doesn't feel much smarter than any llama 2 70b, but has an actually working context of 32k and adheres to instructions better than llama. It might or might not be smarter than mixtral. I am actually not sure.
In other news, Meta released Codellama 70b. The instruct tune is so censored that it can't do the simplest things.
