type
status
date
slug
summary
tags
category
password
icon
Author
Abstract

June 4th

  • Today, I took a careful look at the mlx training framework launched by Apple official, and found it very interesting. It can train and deploy models locally on MacBook, which feels very cool. I used my 32GB M2 Max notebook to test the Qwen1.5-0.5B model, ran the official sample data, and it worked, which I think is quite good. If you want to train models in the future but suffer from lack of GPU resources, mlx is a very good choice. It's just that if you want to train a slightly larger model in the future, you may need to choose a very high configuration when changing computers. I guess there should be no problem training a 7B model with a 32GB MacBook LoRA? 😂 Below is the screenshot of my training:
    • notion image
  • Today, I also looked at the official code examples of CrewAI and LLM From Scratch, which are quite useful for me to familiarize with the code, after all, I haven't written code for a long time😂

June 2nd

I recently read the official documents of CrewAI and AutoGen, and I thought Agent was very cool. Just imagine how interesting it is to see two models arguing in a conversation.
So one of my recent ideas is to use Agent to complete an entire writing process. I feel that CrewAI might be better for the framework? But I'm not sure, I need to take a closer look, but I think this idea should be feasible.
Well, this is just one of my whimsical ideas, I'll record it first. 🤣
 
 
Relate Posts
June 5, Synthetic Instruction Data Generation May 30, Recap for May
Loading...
Chengsheng Deng
Chengsheng Deng
Chengsheng Deng
Latest posts
Mar 24 Notes on LightRAG
Mar 24, 2025
Dec 6, Some Tests on o1
Mar 14, 2025
Mar 10, Note on BIG-MATH
Mar 10, 2025
Mar 6, Note on QwQ-32B
Mar 6, 2025
Jan 21, Notes on DeepSeek-R1
Mar 6, 2025
The First Pages of 2025 - My January & February Story
Mar 5, 2025
Announcement
🎉Welcome to my blog🎉 
To find me:
Twitter/X:My X
👏Have fun in my blog👏