NinjaMoves@feddit.nu to LocalLLaMA@sh.itjust.worksEnglish · 5 days agoMistral small 3.1 releasedmistral.aiexternal-linkmessage-square5fedilinkarrow-up127arrow-down11
arrow-up126arrow-down1external-linkMistral small 3.1 releasedmistral.aiNinjaMoves@feddit.nu to LocalLLaMA@sh.itjust.worksEnglish · 5 days agomessage-square5fedilink
minus-squarebrucethemoose@lemmy.worldcakelinkfedilinkEnglisharrow-up7·5 days agoAnyone tested it at high context yet? I find all Mistral models peter out after like 16K-24K tokes no matter what they advertise the context length as.
Anyone tested it at high context yet? I find all Mistral models peter out after like 16K-24K tokes no matter what they advertise the context length as.