• notfromhere@lemmy.oneOP
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 years ago

    I hope llama.cpp supports SuperHOT at some point. I never use GPTQ but may need to make an exception to try out the larger context sized. Are you using exllama? Curious why you’re getting garbage output