LLM Reddit 5h ago 2 min read
A developer on r/MachineLearning shared phase-one details for Dante-2B, a 2.1B Italian/English model trained from scratch with a tokenizer tuned for Italian morphology and token efficiency.
A developer on r/MachineLearning shared phase-one details for Dante-2B, a 2.1B Italian/English model trained from scratch with a tokenizer tuned for Italian morphology and token efficiency.