Chinese AI startup Zhipu AI aka Z.ai has released its GLM-4.6V series, a new generation of open-source vision-language models (VLMs) optimized for multimodal reasoning, frontend automation, and high-efficiency deployment. The release includes two models in "large" and "small" sizes: GLM-4.6V (106B), a larger 106-billion parameter model aimed at cloud-scale inference GLM-4.6V-Flash (9B), a smaller model …
Continue reading Z.ai debuts open source GLM-4.6V, a native tool-calling vision model for multimodal reasoning