Slightly off topic. I had a hard time getting models to run with ollama, and I thought that my computer (32gm ram, GTX4070 12Gb vram) just couldn't do it. The I tried LM Studio and after fiddling with some settings, I got models running and quite fast. I didn't try GLM-4.7 flash but I did GLM-4.6v flash and it was amazing to see it be able to analyze all kinds of images (since it has vision support). I was simply stunned. I can't believe that a simple gaming machine can do many of the things I used cloud models for. It was absolutely strikingly good at guessing locations of photos. Even vague ones. Deducing landmarks, writings, types of traffic signs. I need to try 4.7 flash. Hopefully it can ran fast with my machine.