Blog

Beating GPT-4 on HumanEval with a fine-tuned CodeLlama-34B

We've fine-tuned CodeLlama-34B and CodeLlama-34B-Python on an internal Phind dataset that achieve 67.6% and 69.5% pass@1 on HumanEval, respectively. GPT-4 achieves 67%. We've applied OpenAI's decontamination methodology to our dataset to ensure result validity.

Read More
August 25, 2023
Michael Royzen
Phind Model beats GPT-4 at coding, with GPT-3.5-like speed and 16k context

We're excited to announce that Phind now defaults to our own model that matches and exceeds GPT-4's coding abilities while running 5x faster. You can now get high quality answers for technical questions in 10 seconds instead of 50.

Read More
October 31, 2023
Michael Royzen
Introducing Phind-70B – closing the code quality gap with GPT-4 Turbo while running 4x faster

We're excited to announce Phind-70B, our largest and most performant model to date. Running at up to 80 tokens per second, it offers the best overall user experience for developers amongst state-of-the-art models.

Read More
February 22, 2024
Michael Royzen