bot

joined 2 years ago
MODERATOR OF
 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/science by /u/chrisdh79 on 2025-04-25 11:55:45+00:00.

Original Title: Unhealthy lifestyle can affect our physical and mental health by age 36 | Bad habits such as smoking, heavy drinking and lack of exercise must be tackled as early as possible to boost the odds of a happy and healthy old age.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/science by /u/mvea on 2025-04-25 10:14:22+00:00.

Original Title: New study shows that people are more open-minded than we assume. When individuals are given high-quality, balanced facts, they don’t simply cling to old beliefs—they revise them. Factual knowledge, when properly delivered, can be a powerful antidote to polarization across contentious issues.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/science by /u/chrisdh79 on 2025-04-25 10:02:05+00:00.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/science by /u/gs-techalerts on 2025-04-25 09:28:16+00:00.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/science by /u/i_screamm on 2025-04-25 08:26:26+00:00.

Original Title: A self-contained carbon-carbon (C2) producing system that combines the catalytic power of copper with perovskite, a material used in photovoltaic solar panels. This research brings the scientific community one step closer to replicating the productivity of a green leaf.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/todayilearned by /u/nuttybudd on 2025-04-25 10:57:40+00:00.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/machinelearning by /u/bminixhofer on 2025-04-24 10:34:11+00:00.


Hello r/MachineLearning !

I’ve been experimenting with a method called ALM to distill language models across tokenizers. This enables, for example, transferring LLMs to a new tokenizer and distilling knowledge from a model with one tokenizer into a model with a different tokenizer (see our paper for details).

I’ve released tokenkit, a library implementing ALM among other methods, to make this easy to use.

One neat application of ALM is distilling subword-based LLMs into byte-level models. I've applied this to two instruction-tuned models:

  • Gemma2-2B-IT-Byte:
  • Llama3-2-3B-IT-Byte:

Even though the distillation phase is very short (just 1.2B bytes ≈ 330M subword tokens), the models perform competitively (for example 57.0% MMLU of the byte-level Llama vs. 62.4% MMLU of the original Llama3-3B-Instruct).

This approach opens up an interesting direction: we can potentially keep subword tokenization for pretraining (to still squeeze as much text into the model in as little time as possible), but then change to a more user-friendly tokenization afterwards.

These models aren’t yet optimized for efficiency, but if you would add self-speculative decoding plus a BLT/DTP-style hierarchical architecture and/or linearized attention, they might also be able to replace subword-based models when speed matters.

If you want to train your own models, this guide on tokenizer transfer via tokenkit should make it easy. The model cards of the transfers above also contain the exact command used to train them. I’ve been training on fairly limited hardware, so effective transfer is possible even in a (near) consumer-grade setup.

I'd love to get feedback on the method, the models, or tokenkit itself. Happy to discuss or answer questions!

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/gardening by /u/ElleneHill on 2025-04-25 06:11:09+00:00.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/beautifulfemales by /u/littleBIGman-1988 on 2025-04-25 10:15:48+00:00.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/todayilearned by /u/a_is_for_a on 2025-04-25 10:46:04+00:00.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/todayilearned by /u/tyrion2024 on 2025-04-25 10:29:17+00:00.

Original Title: TIL a father (5 years) & son (4 years) were sentenced to prison after they cashed in 14,000 winning lottery tickets over a 10-year period, laundered more than $20m, & then failed to pay the IRS $6m. They bought the tickets at a discount from people who wanted to avoid being identified by the state.

 
This is an automated archive made by the Lemmit Bot.

The original was posted on /r/unixporn by /u/TurbulentDonkey9744 on 2025-04-25 06:37:59+00:00.

view more: ‹ prev next ›