type
status
date
slug
summary
tags
category
password
icon
Author
Abstract

Job

In August, I focused on fine-tuning the Qwen2-7b model and evaluating its performance on our private benchmark consisting of over 200 questions and answers. I evaluated various large language models (LLMs) like GPT-4, Gemini 1.5-Pro, and Llama 3-405b on this benchmark to compare their capabilities in areas such as reasoning, coding, and commonsense.
While the fine-tuned Qwen2-7b surpassed some older models over 30B parameters, it couldn't match the performance of Qwen2-7b-instruct. Here are some key learnings from the fine-tuning process:
  1. Using LoRA (Low-Rank Adaptation) was a good choice to reduce the number of trainable parameters.
  1. The dataset quality is crucial. I tried various methods like evol-instruct, persona-hub, and magpie to generate the instruction dataset. Ensuring high dataset quality through techniques like LLLM-as-a-Judge and human annotation is challenging but important.
  1. DPO (Differentiable Prompt Optimization) along with SFT (Supervised Fine-Tuning) improved performance compared to SFT alone. Despite its benefits, DPO is rarely covered in online fine-tuning tutorials.
I'm considering writing a blog to document my experience with building the evaluation benchmark, fine-tuning models on it, and improving model performance.
I also explored the power of Flux, a new AI model for generating images. I was amazed by its capabilities and the speed at which it generates high-quality images using LoRA fine-tuning on Replicate.

Personal

On the personal front, I delved into DSPy (Differentiable Sparse Polynomial) and tried using MIPRO optimization to improve model performance.
I also fine-tuned a LoRA model for my girlfriend, generating impressive and realistic images.
To better curate and filter information, I started writing a newsletter to collect interesting content, even though it's still in its early stages.
For writing blogs, I've transitioned from Notion and Obsidian to Cursor. The reasons are:
  1. Cursor allows working with local files, ensuring data safety and control without relying on the internet. While Obsidian also uses local files, Cursor's built-in AI features are more convenient than installing plugins in Obsidian.
  1. Notion is expensive, and its AI features are not as advanced as Cursor's. Additionally, Notion's recent blocking of Russian users raises concerns about data control.
I still use Notion Next (a free and open-source alternative) for publishing my blogs, but I prefer writing them in Cursor.
That's a recap of my August activities. Looking forward to the next month!
<ins/>
Sep 3, Notes on Anthropic Prompt TutorialAug 26, Flux + LoRA
Loading...