• @[email protected]
    link
    fedilink
    English
    11 day ago

    This is really a non-issue, as the LLM itself should have no problem at setting a reasonable value itself. User wants a summary? Obviously maximum factual. He wants gaming ideas? Etc.

    • @[email protected]
      link
      fedilink
      English
      2
      edit-2
      1 day ago

      For local LLMs, this is an issue because it breaks your prompt cache and slows things down, without a specific tiny model to “categorize” text… which few have really worked on.

      I don’t think the corporate APIs or UIs even do this. You are not wrong, but it’s just not done for some reason.

      It could be that the trainers don’t realize its an issue. For instance, “0.5-0.7” is the recommended range for Deepseek R1, but I find much lower or slightly higher is far better, depending on the category and other sampling parameters.