🔥 Run DeepSeek R1 distilled models Locally with ONNX Runtime – Faster & Private 🔥 AI inference doesn’t have to rely on the cloud. With ONNX Runtime, you can run DeepSeek R1 on your CPU, GPU, or NPU, keeping your data private while achieving up to 6.3x faster performance than PyTorch. Why run on-device? 🔹 Privacy – No data leaves your machine 🔹 Speed – Optimized performance across Intel, AMD, NVIDIA, and Qualcomm hardware 🔹 Flexibility – Works on PCs with easy deployment via Hugging Face + ORT 🔗 Check out the blog here: https://lnkd.in/gQgYwKJ8
ONNX Runtime
Software Development
Redmond, Washington 2,513 followers
Run fast, run anywhere: ONNX Runtime is a machine learning accelerator for cloud, edge, web and mobile
About us
ONNX Runtime is a cross-platform machine-learning model accelerator, with a flexible interface to integrate hardware-specific libraries. ONNX Runtime can be used with models from PyTorch, Tensorflow/Keras, TFLite, scikit-learn, and other frameworks.
- Website
-
https://onnxruntime.ai
External link for ONNX Runtime
- Industry
- Software Development
- Company size
- 51-200 employees
- Headquarters
- Redmond, Washington
- Type
- Public Company
Locations
-
Primary
Get directions
Redmond, Washington 98052, US
Updates
-
ONNX Runtime reposted this
You can now experiment Phi4 models with #ONNXRuntime for faster on-device inference. Try now: https://lnkd.in/guKkAAmA https://lnkd.in/gJVQNUzt https://lnkd.in/g2fpic-H
Introducing Phi-4-multimodal and Phi-4-mini! Phi-4-multimodal integrates speech, vision, and text processing, while Phi-4-mini excels in text-based tasks. Discover these models on Azure AI Foundry: https://msft.it/6048U7H48