diff --git a/src/images/logos/algoriddim-logo.png b/src/images/logos/algoriddim-logo.png new file mode 100644 index 0000000000000..e2964ab6dd4aa Binary files /dev/null and b/src/images/logos/algoriddim-logo.png differ diff --git a/src/images/logos/camo-logo.png b/src/images/logos/camo-logo.png new file mode 100644 index 0000000000000..7f833f872ad72 Binary files /dev/null and b/src/images/logos/camo-logo.png differ diff --git a/src/images/logos/cephable-logo.png b/src/images/logos/cephable-logo.png new file mode 100644 index 0000000000000..1f3e60582a20b Binary files /dev/null and b/src/images/logos/cephable-logo.png differ diff --git a/src/routes/components/customers.svelte b/src/routes/components/customers.svelte index b0f802adeee7b..a5da8146bea27 100644 --- a/src/routes/components/customers.svelte +++ b/src/routes/components/customers.svelte @@ -6,8 +6,11 @@ import adobeLogo from '../../images/logos/adobe-logo.png'; import amdLogo from '../../images/logos/amd-logo.png'; import antgroupLogo from '../../images/logos/antgroup-logo.png'; + import algoriddimLogo from '../../images/logos/algoriddim-logo.png'; import ATLASLogo from '../../images/logos/ATLAS-logo.png'; import bazaarvoiceLogo from '../../images/logos/bazaarvoice-logo.png'; + import camoLogo from '../../images/logos/camo-logo.png'; + import cephableLogo from '../../images/logos/cephable-logo.png'; import clearbladeLogo from '../../images/logos/clearblade-logo.png'; import deezerLogo from '../../images/logos/deezer-logo.png'; import hypefactorsLogo from '../../images/logos/hypefactors-logo.png'; @@ -48,6 +51,11 @@ src: antgroupLogo, alt: 'Ant Group' }, + { + href: './testimonials#Algoriddim', + src: algoriddimLogo, + alt: 'Algoriddim' + }, { href: './testimonials#Atlas%20Experiment', src: ATLASLogo, @@ -58,6 +66,16 @@ src: bazaarvoiceLogo, alt: 'Bazaarvoice' }, + { + href: './testimonials#Camo', + src: camoLogo, + alt: 'Camo' + }, + { + href: './testimonials#Cephable', + src: cephableLogo, + alt: 'Cephable' + }, { href: './testimonials#ClearBlade', src: clearbladeLogo, diff --git a/src/routes/testimonials/+page.svelte b/src/routes/testimonials/+page.svelte index 957f6b63b0ad8..ca71962dda7d6 100644 --- a/src/routes/testimonials/+page.svelte +++ b/src/routes/testimonials/+page.svelte @@ -4,8 +4,11 @@ import adobelogo from '../../images/logos/adobe-logo.png'; import amdlogo from '../../images/logos/amd-logo.png'; import antgrouplogo from '../../images/logos/antgroup-logo.png'; + import algoriddimLogo from '../../images/logos/algoriddim-logo.png'; import atlaslogo from '../../images/logos/ATLAS-logo.png'; import bazaarvoicelogo from '../../images/logos/bazaarvoice-logo.png'; + import camologo from '../../images/logos/camo-logo.png'; + import cephablelogo from '../../images/logos/cephable-logo.png'; import clearbladelogo from '../../images/logos/clearblade-logo.png'; import deezerlogo from '../../images/logos/deezer-logo.png'; import intelligenzaeticablogo from '../../images/logos/intelligenza-etica-logo.png'; @@ -58,6 +61,14 @@ imgsrc: antgrouplogo, imgalt: 'Ant Group logo' }, + { + title: 'Algoriddim', + quote: + 'At Algoriddim we are using ONNX Runtime on Windows devices to power our Neural Mix™ feature that allows users to isolate vocal and instruments of any song in real-time, as well as our Automix feature that allows for seamless automatic DJ mixes. ONNX Runtime strikes the perfect balance between abstraction and flexibility, and using the QNN execution provider allows us to leverage the NPU on Copilot+ PCs for unparalleled inference performance while keeping the CPU free for other tasks.', + author: 'Frederik Seiffert, CTO, Algoriddim', + imgsrc: algoriddimLogo, + imgalt: 'Algoriddim logo' + }, { title: 'Atlas Experiment', quote: @@ -74,6 +85,22 @@ imgsrc: bazaarvoicelogo, imgalt: 'Bazaarvoice logo' }, + { + title: 'Camo', + quote: + "ONNX Runtime enables Camo Studio to deliver features like background segmentation and feature detection with speed and accuracy. It seamlessly integrated with our existing models and lets us target any processor, including the latest NPUs, saving us valuable development time and allowing us to bring innovative features to all our users. We recommend ONNX Runtime to any developer looking to streamline model deployment and unlock the full potential of their applications.", + author: 'Aidan Fitzpatrick, Founder & CEO, Reincubate', + imgsrc: camologo, + imgalt: 'Camo logo' + }, + { + title: 'Cephable', + quote: + "The ONNX Runtime allows us to simultaneously target CPU, GPU and NPU enabled devices. Converting a model to NPU, using ONNX Runtime and AI Hub reduced our engineering effort from 30 days to 7 days. Given the current state of the art, that would likely be only 3 days today. This allows us to deliver cutting edge performance to our users, minimizing impact of AI/ML workloads when running other applications, and leaves more time to focus on feature work.", + author: 'Jon Campbell, Director of Engineering, Cephable', + imgsrc: cephablelogo, + imgalt: 'Cephable logo' + }, { title: 'ClearBlade', quote: