We use their earlier Mixtral model because it outperforms llama for our use case. They do not release full models for marketing purposes, though it definitely grabs attention!
You may need to revise your views..
It beats llama on the benchmark posted below (though maybe leaked into training data). But also you can run it on cheaper split up hardware with less individual vram than the big llama.
What makes it you think it's not as good as LLaMA? It's likely much better. There are multiple open-weight models that are better than LLaMA 2 out there already.
This is clearly an inferior model that they are willing to share for marketing purposes.
If it was an improvement over llama, sure, but it seems like just an ad for bad AI.