Oxen.ai Blog

Welcome to the Oxen.ai blog 🐂

The team at Oxen.ai is dedicated to helping AI practictioners go from research to production. To help enable this, we host a research paper club on Fridays called ArXiv Dives, where we go over state of the art research and how you can apply it to your own work.

Take a look at our Arxiv Dives, Practical ML Dives as well as a treasure trove of content on how to go from raw datasets to production ready AI/ML systems. We cover everything from prompt engineering, fine-tuning, computer vision, natural language understanding, generative ai, data engineering, to best practices when versioning your data. So, dive in and explore – we're excited to share our journey and learnings with you 🚀

How to Fine-Tune a FLUX.1-dev LoRA with Code, Step by Step
How to Fine-Tune a FLUX.1-dev LoRA with Code, Step by Step

FLUX.1-dev is one of the most popular open-weight models available today. Developed by Black Forest Labs, it has 12 billion parameters. The goal of this post is to provide a barebo...

Greg Schoeninger
Greg Schoeninger
6/28/2025
- Fine-Tune Fridays
20 min read
How to Fine-Tune PixArt to Generate a Consistent Character
How to Fine-Tune PixArt to Generate a Consistent Character

Can we fine-tune a small diffusion transformer (DiT) to generate OpenAI-level images by distilling off of OpenAI images? The end goal is to have a small, fast, cheap model that we ...

Greg Schoeninger
Greg Schoeninger
6/19/2025
- Fine-Tune Fridays
21 min read
How to Fine-Tune Qwen3 on Text2SQL to GPT-4o level performance
How to Fine-Tune Qwen3 on Text2SQL to GPT-4o level performance

Welcome to a new series from the Oxen.ai Herd called Fine-Tuning Fridays! Each week we will take an open source model and put it head to head against a closed source foundation mod...

Greg Schoeninger
Greg Schoeninger
5/28/2025
- Fine-Tune Fridays
15 min read
Fine-Tuning Fridays
Fine-Tuning Fridays

Welcome to a new series from the Oxen.ai Herd called Fine-Tuning Fridays! Each week we will take an open source model and put it head to head against a closed source foundation mod...

Greg Schoeninger
Greg Schoeninger
5/16/2025
- Fine-Tune Fridays
4 min read
Training a Rust 1.5B Coder LM with Reinforcement Learning (GRPO)
Training a Rust 1.5B Coder LM with Reinforcement Learning (GRPO)

Group Relative Policy Optimization (GRPO) has proven to be a useful algorithm for training LLMs to reason and improve on benchmarks. DeepSeek-R1 showed that you can bootstrap a mod...

Greg Schoeninger
Greg Schoeninger
3/6/2025
- Practical ML
17 min read