i love GLM! but please consider longer context length

#10
by mahmood36 - opened

im merely an enthusiast, my words might mean nothing, i cant even run the model locally.
but recently i tried the model through openrouter and woah boy, it outperformed every other model including closed source ones. this model is very reminiscent of gemini 2.5 in its early days (which later got nerfed). the only thing missing is a longer context length and it would be SOTA in every way. in my use case (which is help with academic study and synthesis of math/cs), it was brilliant and outperformed all models including the recent qwen3, but the most recent qwen3 supports a context length of up to 1m, and even offers that in very small models that i can run locally.

thank you for your work! please stay open source! you are our best hope!!

i would also like to mention out that even though i have a 4090, i cant run the model locally at a reasonable quant. tried q4 but that requires 80+ gb of memory excluding kv-cache context.

this is not a complaint, merely a humble wish, so dont feel criticized because even in its current state, i wouldnt have expected an open source to be that good in the first place.

thank you again! i appreciate your work!

Sign up or log in to comment