diff --git a/src/images/blogs/phi-3-on-device_blog_thumbnail.png b/src/images/blogs/phi-3-on-device_blog_thumbnail.png new file mode 100644 index 0000000000000..65767c432f7e5 Binary files /dev/null and b/src/images/blogs/phi-3-on-device_blog_thumbnail.png differ diff --git a/src/routes/blogs/+page.svelte b/src/routes/blogs/+page.svelte index b128ed789e381..93f95d2116059 100644 --- a/src/routes/blogs/+page.svelte +++ b/src/routes/blogs/+page.svelte @@ -15,6 +15,7 @@ import ORT117Thumbnail from '../../images/blogs/ort-1-17-thumbnail.png'; import WebGPUImage from '../../images/blogs/webgpu_blog_thumbnail.jpg'; import WebTrainingImage from '../../images/blogs/webtraining_blog_thumbnail.png'; + import Phi3OnDeviceImage from '../../images/blogs/phi-3-on-device_blog_thumbnail.png'; onMount(() => { anime({ targets: '.border-primary', @@ -42,6 +43,16 @@ dispatch('switchTab', tab); } let featuredblog = [ + { + title: 'Enjoy the Power of Phi-3 with ONNX Runtime on your device', + date: 'May 20th, 2024', + blurb: + "Harness ONNX Runtime to run Phi-3-mini on mobile phones and in the browser.", + link: 'blogs/phi-3-on-device', + image: Phi3OnDeviceImage, + imgalt: + 'Chart comparing model size (in GB) of ONNX Phi-3-mini for web and mobile with original Phi-3-mini' + }, { title: 'ONNX Runtime supports Phi-3 mini models across platforms and devices', date: 'April 22nd, 2024', @@ -61,7 +72,9 @@ image: WebGPUImage, imgalt: 'Comparison of ONNX Runtime Web with WebGPU EP on GPU vs. WASM EP on CPU for segment anything example' - }, + } + ]; + let blogs = [ { title: 'ONNX Runtime 1.17: CUDA 12 support, Phi-2 optimizations, WebGPU, and more!', date: 'February 28th, 2024', @@ -71,9 +84,6 @@ image: ORT117Thumbnail, imgalt: 'ONNX Runtime 1.17 release logo' }, - - ]; - let blogs = [ { title: 'Accelerating Phi-2, CodeLlama, Gemma and other Gen AI models with ONNX Runtime', date: 'February 26th, 2024', diff --git a/src/routes/blogs/phi-3-on-device/+page.svx b/src/routes/blogs/phi-3-on-device/+page.svx new file mode 100644 index 0000000000000..c775c63903bba --- /dev/null +++ b/src/routes/blogs/phi-3-on-device/+page.svx @@ -0,0 +1,30 @@ +--- +title: 'Enjoy the power of Phi-3 with ONNX Runtime on your device' +date: '20th May, 2024' +description: 'Harness ONNX Runtime and Olive to optimize and run the state-of-the-art Phi-3-mini model on your mobile phone or directly in your browser!' +keywords: 'ORT, ONNX Runtime, ONNX, machine learning, deep learning, phi 3, phi-3, webgpu, webnn, webassembly, hugging face, huggingface, transformers.js' +authors: + [ + 'Emma Ning (Microsoft)', + 'Scott McKay (Microsoft)', + 'Guenther Schmuelling (Microsoft)', + 'Joshua Lochner (Hugging Face)' + ] +authorsLink: + [ + 'https://www.linkedin.com/in/qiong-ning-21b554236/', + 'https://www.linkedin.com/in/scott-mckay-68b5ab/', + 'https://www.linkedin.com/in/guentherschmuelling/', + 'https://www.linkedin.com/in/xenova/?originalSubdomain=za' + ] +image: '' +url: 'https://onnxruntime.ai/blogs/phi-3-on-device' +--- + +Want to own a private and powerful chatbot on your device with full data privacy and no need for an internet connection? + +Harness ONNX Runtime and Olive to optimize and run the state-of-the-art Phi-3-mini model on your mobile phone or directly in your browser! + +Enjoy models optimized for mobile and web with a size reduction of over 2.5 times, and run Phi-3-mini at over 70 tokens per second in the browser on an Nvidia RTX 4090! + +Check out the full blog [here](https://huggingface.co/blog/Emma-N/enjoy-the-power-of-phi-3-with-onnx-runtime). \ No newline at end of file