Is DeepSeek-V4 coming? According to leaked screenshot information, the V4 model uses INT8 quantization, aiming to significantly reduce memory usage and inference costs while maintaining high performance, making it easier to deploy/run on consumer-grade hardware or in broader environments. However, the screenshots themselves do not indicate support for multimodal features.