Hacker Newsnew | past | comments | ask | show | jobs | submit | let_tim_cook_'s commentslogin

I'm not sure why progressive LoRa merging needs to be addressed here. They show there is a regime of problem where LoRa performs equivalently to FFT.

Progressive merging of LoRa is somewhere inbetween and categorically more complex than just LoRa so would be dominated by standard LoRa in that case.

While progressive merging could train faster as fewer params are trainable at any given time, it results in very larger adapter diffs OTO the size of the original model and doesn't retain the benefits of being able to deploy multiple adapters over the same base model idt.


"Stanford Repo Released Sep 31, 2025" Seems like something sampled from a distribution with non-zero probability that the day after Sep 30, 2025 would is the 31st....


Thanks for the note. Ironically, the post is about models built to understand time.


They fixed it already.


Sounds like something a clanker would write....


OK wetware.


You showed that meatbag who's root


Are any authors here? Have you looked at AppWorld? https://appworld.dev


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: