Most active commenters

    137 points ignoramous | 16 comments | | HN request time: 1.97s | source | bottom
    1. kcorbitt ◴[] No.41911422[source]
    I saw this when it was making the rounds on X a few days ago. Fair warning: it seems like at least some sections are AI-generated, and there isn't much insight to be gained from reading the actual sections compared to eg. reading the relevant category pages on Huggingface.
    replies(4): >>41911596 #>>41911605 #>>41912253 #>>41912270 #
    2. daghamm ◴[] No.41911596[source]
    I would not say that, as long as it is a good summary there is a value in having everything in the same document.

    Obviously they should have stated that this is partially generated, but at least they are dog fooding it :)

    3. YetAnotherNick ◴[] No.41911605[source]
    Not only the it seems to be AI generated, it seems these guys don't even know about best practices or even what works. e.g. It contains archaic comparison of optimizers and its pros and cons, but for LLMs no optimizer other than Adam and new ones like Lion works.
    replies(1): >>41912039 #
    4. anothername12 ◴[] No.41911748[source]
    Well, it sucks that we’re at the “best practices” phase already
    replies(3): >>41911886 #>>41912332 #>>41912970 #
    5. ◴[] No.41911804[source]
    6. p1esk ◴[] No.41911886[source]
    It sucks that we’re still at “best practices” phase. We’ve been in this phase for the last three decades [1], and I really hope we enter “good theory” phase soon.

    [1] https://cseweb.ucsd.edu/classes/wi08/cse253/Handouts/lecun-9...

    7. abc-1 ◴[] No.41912039{3}[source]
    Is there a paper on this? Why do no other optimizers give good results? Adam requires insane amounts of memory so alternatives would be welcome.
    8. worstspotgain ◴[] No.41912253[source]
    Glancing at the authors' names, it's possible that none of them are native English speakers. Any chance that the sections you're referring to were just AI-polished rather than AI-generated?
    replies(1): >>41912649 #
    9. danielhanchen ◴[] No.41912270[source]
    I took a skim through it in the morning - I like the LoRA Learns Less and Forgets Less paper more https://openreview.net/forum?id=aloEru2qCG - it has much more signal in a few pages - also the original QLoRA paper from Dettmers https://arxiv.org/abs/2305.14314 has so many more important morsels.

    But all in all, the review is a reasonable "manual" I guess. I would have liked maybe more instructive comprehensive practical examples, and maybe more mention of other OSS packages for finetuning :))

    10. kleiba ◴[] No.41912332[source]
    Why is that?
    11. qeternity ◴[] No.41912649{3}[source]
    No, this paper was edited yesterday. The original (you can verify on arxiv) contained this incredible section: "6.10 Optimised Routing and Pruning Operations (ORPO)"

    The actual ORPO paper is "Odds Ratio Preference Optimisation" and it has nothing to do with pruning. This goes way beyond native language preference.

    replies(1): >>41913061 #
    12. make3 ◴[] No.41912970[source]
    there's likely still an infinite amount of things to figure out, transformers haven't been out for 10 years yet
    13. raymond_goo ◴[] No.41913043[source]
    Ctrl-F: Unsloth --> no results == bad paper
    replies(1): >>41913074 #
    14. cubefox ◴[] No.41913061{4}[source]
    Wow, so significant parts of the paper could still be LLM confabulation.
    15. youoy ◴[] No.41913074[source]
    But you can find "delve"