Optimizing memory usage in large language models fine-tuning with KAITO: Best practices from Phi-3
The Cloud Native team at Azure is working to make AI on Kubernetes more cost-effective and approachable for a broader range of users.

The Cloud Native team at Azure is working to make AI on Kubernetes more cost-effective and approachable for a broader range of users.
As the requirements and software surrounding Kubernetes clusters grow along with the required number of clusters, the administrative overhead becomes…
As we come together in Amsterdam, there are significant headwinds and challenges facing us, but I’m confident that open-source and…
In this blog post, we’ll share challenges our team faced, and how ONNX Runtime solves these as the backbone of…
Azure Open Source Day highlighted Microsoft’s commitment to open Source and focused on how Open Source Technologies can be used…
Today, we are excited to announce the much-anticipated availability of the OSS Feathr 1.0.
On Azure, more than 50 percent of virtual machine (VM) cores run on Linux. There is no better time to…
Azure Open Source Day is a great opportunity to learn more about Microsoft's role in the open-source community, its contributions,…
The team at Pieces shares the problems and solutions evaluated for their on-device model serving stack and how ONNX Runtime…
Make large models smaller and faster with OpenVino Execution Provider, NNCF and ONNX Runtime leveraging Azure Machine Learning.
Many developers opt to use popular AI Frameworks like PyTorch, which simplifies the process of analyzing predictions, training models, leveraging…
eBPF for Windows native code generation is a new mode of execution that maintains the integrity of the kernel and…
To celebrate FOSS Fund #25 we have invited all employees whose projects were not selected in past FOSS Fund to…