After seeing Ouro LLM being promoted, it immediately made me want to try and find repos about it... No dice.
But hey, this repo popped up, and I got a lot of questions
- How can ouroboros prevent catastrophic forgetting when fine-tuning? (RL Razor)
- Are there ways of avoiding model collapse? (Data Accumulation)
- Will diffusion or dLLMs be included in the process? They seem very robust
- Will agentic scaffolds be included? Can they self-improve too? (SEAL + DGM)
After seeing Ouro LLM being promoted, it immediately made me want to try and find repos about it... No dice.
But hey, this repo popped up, and I got a lot of questions