Skip to content

Commit

Permalink
Add Phi-3 on device announcement
Browse files Browse the repository at this point in the history
  • Loading branch information
sophies927 committed May 20, 2024
1 parent 3909f23 commit 4240036
Show file tree
Hide file tree
Showing 3 changed files with 44 additions and 4 deletions.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
18 changes: 14 additions & 4 deletions src/routes/blogs/+page.svelte
Original file line number Diff line number Diff line change
Expand Up @@ -15,6 +15,7 @@
import ORT117Thumbnail from '../../images/blogs/ort-1-17-thumbnail.png';
import WebGPUImage from '../../images/blogs/webgpu_blog_thumbnail.jpg';
import WebTrainingImage from '../../images/blogs/webtraining_blog_thumbnail.png';
import Phi3OnDeviceImage from '../../images/blogs/phi-3-on-device_blog_thumbnail.png';
onMount(() => {
anime({
targets: '.border-primary',
Expand Down Expand Up @@ -42,6 +43,16 @@
dispatch('switchTab', tab);
}
let featuredblog = [
{
title: 'Enjoy the Power of Phi-3 with ONNX Runtime on your device',
date: 'May 20th, 2024',
blurb:
"Harness ONNX Runtime to run Phi-3-mini on mobile phones and in the browser.",
link: 'blogs/phi-3-on-device',
image: Phi3OnDeviceImage,
imgalt:
'Chart comparing model size (in GB) of ONNX Phi-3-mini for web and mobile with original Phi-3-mini'
},
{
title: 'ONNX Runtime supports Phi-3 mini models across platforms and devices',
date: 'April 22nd, 2024',
Expand All @@ -61,7 +72,9 @@
image: WebGPUImage,
imgalt:
'Comparison of ONNX Runtime Web with WebGPU EP on GPU vs. WASM EP on CPU for segment anything example'
},
}
];
let blogs = [
{
title: 'ONNX Runtime 1.17: CUDA 12 support, Phi-2 optimizations, WebGPU, and more!',
date: 'February 28th, 2024',
Expand All @@ -71,9 +84,6 @@
image: ORT117Thumbnail,
imgalt: 'ONNX Runtime 1.17 release logo'
},
];
let blogs = [
{
title: 'Accelerating Phi-2, CodeLlama, Gemma and other Gen AI models with ONNX Runtime',
date: 'February 26th, 2024',
Expand Down
30 changes: 30 additions & 0 deletions src/routes/blogs/phi-3-on-device/+page.svx
Original file line number Diff line number Diff line change
@@ -0,0 +1,30 @@
---
title: 'Enjoy the power of Phi-3 with ONNX Runtime on your device'
date: '20th May, 2024'
description: 'Harness ONNX Runtime and Olive to optimize and run the state-of-the-art Phi-3-mini model on your mobile phone or directly in your browser!'
keywords: 'ORT, ONNX Runtime, ONNX, machine learning, deep learning, phi 3, phi-3, webgpu, webnn, webassembly, hugging face, huggingface, transformers.js'
authors:
[
'Emma Ning (Microsoft)',
'Scott McKay (Microsoft)',
'Guenther Schmuelling (Microsoft)',
'Joshua Lochner (Hugging Face)'
]
authorsLink:
[
'https://www.linkedin.com/in/qiong-ning-21b554236/',
'https://www.linkedin.com/in/scott-mckay-68b5ab/',
'https://www.linkedin.com/in/guentherschmuelling/',
'https://www.linkedin.com/in/xenova/?originalSubdomain=za'
]
image: ''
url: 'https://onnxruntime.ai/blogs/phi-3-on-device'
---

Want to own a private and powerful chatbot on your device with full data privacy and no need for an internet connection?

Harness ONNX Runtime and Olive to optimize and run the state-of-the-art Phi-3-mini model on your mobile phone or directly in your browser!

Enjoy models optimized for mobile and web with a size reduction of over 2.5 times, and run Phi-3-mini at over 70 tokens per second in the browser on an Nvidia RTX 4090!

Check out the full blog [here](https://huggingface.co/blog/Emma-N/enjoy-the-power-of-phi-3-with-onnx-runtime).

0 comments on commit 4240036

Please sign in to comment.