type
status
date
slug
summary
tags
category
password
icon
Domain & Institution
Author
Priority
Abstract
Creation Date

June 4th

  • Today, I took a careful look at the mlx training framework launched by Apple official, and found it very interesting. It can train and deploy models locally on MacBook, which feels very cool. I used my 32GB M2 Max notebook to test the Qwen1.5-0.5B model, ran the official sample data, and it worked, which I think is quite good. If you want to train models in the future but suffer from lack of GPU resources, mlx is a very good choice. It's just that if you want to train a slightly larger model in the future, you may need to choose a very high configuration when changing computers. I guess there should be no problem training a 7B model with a 32GB MacBook LoRA? 😂 Below is the screenshot of my training:
    • notion image
  • Today, I also looked at the official code examples of CrewAI and LLM From Scratch, which are quite useful for me to familiarize with the code, after all, I haven't written code for a long time😂

June 2nd

I recently read the official documents of CrewAI and AutoGen, and I thought Agent was very cool. Just imagine how interesting it is to see two models arguing in a conversation.
So one of my recent ideas is to use Agent to complete an entire writing process. I feel that CrewAI might be better for the framework? But I'm not sure, I need to take a closer look, but I think this idea should be feasible.
Well, this is just one of my whimsical ideas, I'll record it first. 🤣
 
 
Relate Posts
June 5, Synthetic Instruction Data Generation May 30, Recap for May
Loading...
BubbleBrain
BubbleBrain
BubbleBrain
Latest posts
MiniMax 发布周回顾
Jun 23, 2025
Mar 24 Notes on LightRAG
Mar 24, 2025
Dec 6, Some Tests on o1
Mar 14, 2025
Mar 10, Note on BIG-MATH
Mar 10, 2025
Mar 6, Note on QwQ-32B
Mar 6, 2025
Jan 21, Notes on DeepSeek-R1
Mar 6, 2025
Announcement
🎉欢迎来到 BubbleBrain🎉 
To find me:
Twitter/X:My X
👏Have fun in my blog👏