Discover Mistral Medium: their first proprietary model and it outperforms Gemini Pro
Week 50 of Coding with Intelligence
I've worked on this with my awesome team at Definitive Intelligence. Check out the live demo and repo! Our goal is to make it 10X easier to build your own function calling assistants through powerful yet simple abstractions.
Oh, did I mention it works on any chat LLM model? Even those that don't support function calling. Furthermore, fully integrated with LangChain (e.g. LangChain Agent tools).
Mistral Medium outperforms Gemini Pro on MMLU and HellaSwag. Also, they raised $415M with a16z as their lead. Probably means they're going to start balancing Open Weight model releases with proprietary models. Mistral Medium is the first example of that.
My Tweet contains the instructions
GPT 3.5 level performance. Mistral is rolling out their own hosted endpoints.
The authors from Stanford claim significant performance improvements compared to DPO based on a GPT-4 based evaluation of their tuned Llama 30B model.
Great response time by Together AI
Find the winning output for your task 🙌
MLX is an Apple specific PyTorch alternative. It just shared an example showing how to run Mistral's new MoE model.
Hyena/Monarch Mixers, state-space models: the traditional Transformer architecture is facing new competition!
Really cool to see OSS Stable Diffusion projects pay more attention to usability.
GPT-4 based agents performed best in their benchmark
Merging symbolic and probabilistic systems showing more and more promise. This paper is from the DeepMind team and improves significantly on BIG-Bench Hard through pairing code eval and LLM inference.
Are Emergent Abilities of Large Language Models a Mirage?
tl;dr Likely abilities just slowly improve as LLMs get scaled. This bodes well for safety concerns, meaning that if this research is correct, it's a boon to AI development as we need to decelerate less.
An umbrella project for LLM security concerns. At a practical level it contains Llama Guard, which is a simple-to-deploy (already available on platforms like Together AI) filtering model that lets you filter/reject inputs/outputs. Safety is a crucial feature for actually making it feasible to deploy OS LLMs in production.
LCMs are a new kind of image generation models with faster generation speed as one of the big advantages
By Yao Fu, a researcher at University of Edinburgh
Want more? Follow me on Twitter! @ricklamers