An ONNX-optimized version of Phi-3-Medium-4K-Instruct, designed for efficient AI inference on Windows machines using DirectML, supporting INT4 quantization for high-performance execution on AMD, Intel, and NVIDIA GPUs.
Phi-3-Medium-4K-Instruct ONNX-DirectML is a high-performance AI model from Microsoft, optimized for fast execution on Windows-based GPUs using DirectML. This version is quantized to INT4 precision for efficient and scalable inference, allowing deployment across AMD, Intel, and NVIDIA GPUs while maintaining the 4K and 128K token context lengths for structured reasoning and instruction-following tasks.
MIT
Microsoft
Text Generation
N.A.
Open
Sector Agnostic
12/03/25 06:35:40
0
MIT
© 2026 - Copyright AIKosh. All rights reserved. This portal is developed by National e-Governance Division for AIKosh mission.