From 19e3bb38604b2ab5052c4683357614c47abaffae Mon Sep 17 00:00:00 2001 From: raivisdejus Date: Mon, 3 Jun 2024 18:04:17 +0000 Subject: [PATCH] deploy: 905716c5c43e0b0fb379351919e1ffc447465865 --- 404.html | 4 ++-- assets/js/{0480b142.ed72cf64.js => 0480b142.4cf7acba.js} | 2 +- assets/js/{e53fa2b7.0edb1b39.js => e53fa2b7.28463cc8.js} | 2 +- .../{runtime~main.d1198b17.js => runtime~main.a5cebb54.js} | 2 +- docs.html | 4 ++-- docs/cli.html | 4 ++-- docs/faq.html | 6 +++--- docs/installation.html | 4 ++-- docs/preferences.html | 4 ++-- docs/usage/file_import.html | 4 ++-- docs/usage/live_recording.html | 6 +++--- index.html | 4 ++-- 12 files changed, 23 insertions(+), 23 deletions(-) rename assets/js/{0480b142.ed72cf64.js => 0480b142.4cf7acba.js} (83%) rename assets/js/{e53fa2b7.0edb1b39.js => e53fa2b7.28463cc8.js} (92%) rename assets/js/{runtime~main.d1198b17.js => runtime~main.a5cebb54.js} (96%) diff --git a/404.html b/404.html index 5da99d612..38d191209 100644 --- a/404.html +++ b/404.html @@ -4,13 +4,13 @@ Page Not Found | Buzz - +
Skip to main content

Page Not Found

We could not find what you were looking for.

Please contact the owner of the site that linked you to the original URL and let them know their link is broken.

- + \ No newline at end of file diff --git a/assets/js/0480b142.ed72cf64.js b/assets/js/0480b142.4cf7acba.js similarity index 83% rename from assets/js/0480b142.ed72cf64.js rename to assets/js/0480b142.4cf7acba.js index 077a524d0..21808b6fe 100644 --- a/assets/js/0480b142.ed72cf64.js +++ b/assets/js/0480b142.4cf7acba.js @@ -1 +1 @@ -"use strict";(self.webpackChunkdocs=self.webpackChunkdocs||[]).push([[836],{3905:(e,t,r)=>{r.d(t,{Zo:()=>s,kt:()=>f});var n=r(7294);function a(e,t,r){return t in e?Object.defineProperty(e,t,{value:r,enumerable:!0,configurable:!0,writable:!0}):e[t]=r,e}function o(e,t){var r=Object.keys(e);if(Object.getOwnPropertySymbols){var n=Object.getOwnPropertySymbols(e);t&&(n=n.filter((function(t){return Object.getOwnPropertyDescriptor(e,t).enumerable}))),r.push.apply(r,n)}return r}function i(e){for(var t=1;t=0||(a[r]=e[r]);return a}(e,t);if(Object.getOwnPropertySymbols){var o=Object.getOwnPropertySymbols(e);for(n=0;n=0||Object.prototype.propertyIsEnumerable.call(e,r)&&(a[r]=e[r])}return a}var p=n.createContext({}),c=function(e){var t=n.useContext(p),r=t;return e&&(r="function"==typeof e?e(t):i(i({},t),e)),r},s=function(e){var t=c(e.components);return n.createElement(p.Provider,{value:t},e.children)},u="mdxType",m={inlineCode:"code",wrapper:function(e){var t=e.children;return n.createElement(n.Fragment,{},t)}},d=n.forwardRef((function(e,t){var r=e.components,a=e.mdxType,o=e.originalType,p=e.parentName,s=l(e,["components","mdxType","originalType","parentName"]),u=c(r),d=a,f=u["".concat(p,".").concat(d)]||u[d]||m[d]||o;return r?n.createElement(f,i(i({ref:t},s),{},{components:r})):n.createElement(f,i({ref:t},s))}));function f(e,t){var r=arguments,a=t&&t.mdxType;if("string"==typeof e||a){var o=r.length,i=new Array(o);i[0]=d;var l={};for(var p in t)hasOwnProperty.call(t,p)&&(l[p]=t[p]);l.originalType=e,l[u]="string"==typeof e?e:a,i[1]=l;for(var c=2;c{r.r(t),r.d(t,{assets:()=>p,contentTitle:()=>i,default:()=>m,frontMatter:()=>o,metadata:()=>l,toc:()=>c});var n=r(7462),a=(r(7294),r(3905));const o={title:"FAQ",sidebar_position:5},i=void 0,l={unversionedId:"faq",id:"faq",title:"FAQ",description:"1. Where are the models stored?",source:"@site/docs/faq.md",sourceDirName:".",slug:"/faq",permalink:"/buzz/docs/faq",draft:!1,tags:[],version:"current",sidebarPosition:5,frontMatter:{title:"FAQ",sidebar_position:5},sidebar:"tutorialSidebar",previous:{title:"CLI",permalink:"/buzz/docs/cli"}},p={},c=[],s={toc:c},u="wrapper";function m(e){let{components:t,...r}=e;return(0,a.kt)(u,(0,n.Z)({},s,r,{components:t,mdxType:"MDXLayout"}),(0,a.kt)("ol",null,(0,a.kt)("li",{parentName:"ol"},(0,a.kt)("p",{parentName:"li"},(0,a.kt)("strong",{parentName:"p"},"Where are the models stored?")),(0,a.kt)("p",{parentName:"li"},"The Whisper models are stored in ",(0,a.kt)("inlineCode",{parentName:"p"},"~/.cache/whisper"),". The Whisper.cpp models are stored in ",(0,a.kt)("inlineCode",{parentName:"p"},"~/Library/Caches/Buzz"),"\n(Mac OS), ",(0,a.kt)("inlineCode",{parentName:"p"},"~/.cache/Buzz")," (Unix), or ",(0,a.kt)("inlineCode",{parentName:"p"},"C:\\Users\\\\AppData\\Local\\Buzz\\Buzz\\Cache")," (Windows). The Hugging Face\nmodels are stored in ",(0,a.kt)("inlineCode",{parentName:"p"},"~/.cache/huggingface/hub"),".")),(0,a.kt)("li",{parentName:"ol"},(0,a.kt)("p",{parentName:"li"},(0,a.kt)("strong",{parentName:"p"},"What can I try if the transcription runs too slowly?")),(0,a.kt)("p",{parentName:"li"},"Try using a lower Whisper model size or using a Whisper.cpp model.")),(0,a.kt)("li",{parentName:"ol"},(0,a.kt)("p",{parentName:"li"},(0,a.kt)("strong",{parentName:"p"},"How to record system audio?")),(0,a.kt)("p",{parentName:"li"},"To transcribe system audio you need to configure virtual audio device and connect output from the applications you whant to transcribe to this virtual speaker. After that you can select it as source in the Buzz. See ",(0,a.kt)("a",{parentName:"p",href:"https://chidiwilliams.github.io/buzz/docs/usage#live-recording"},"Usage")," section for more details."),(0,a.kt)("p",{parentName:"li"},"Relevant tools:"),(0,a.kt)("ul",{parentName:"li"},(0,a.kt)("li",{parentName:"ul"},"Mac OS - ",(0,a.kt)("a",{parentName:"li",href:"https://github.com/ExistentialAudio/BlackHole"},"BlackHole"),"."),(0,a.kt)("li",{parentName:"ul"},"Windows - ",(0,a.kt)("a",{parentName:"li",href:"https://vb-audio.com/Cable/"},"VB CABLE")),(0,a.kt)("li",{parentName:"ul"},"Linux - ",(0,a.kt)("a",{parentName:"li",href:"https://wiki.ubuntu.com/record_system_sound"},"PulseAudio Volume Control"))))))}m.isMDXComponent=!0}}]); \ No newline at end of file +"use strict";(self.webpackChunkdocs=self.webpackChunkdocs||[]).push([[836],{3905:(e,t,r)=>{r.d(t,{Zo:()=>s,kt:()=>f});var n=r(7294);function a(e,t,r){return t in e?Object.defineProperty(e,t,{value:r,enumerable:!0,configurable:!0,writable:!0}):e[t]=r,e}function o(e,t){var r=Object.keys(e);if(Object.getOwnPropertySymbols){var n=Object.getOwnPropertySymbols(e);t&&(n=n.filter((function(t){return Object.getOwnPropertyDescriptor(e,t).enumerable}))),r.push.apply(r,n)}return r}function i(e){for(var t=1;t=0||(a[r]=e[r]);return a}(e,t);if(Object.getOwnPropertySymbols){var o=Object.getOwnPropertySymbols(e);for(n=0;n=0||Object.prototype.propertyIsEnumerable.call(e,r)&&(a[r]=e[r])}return a}var p=n.createContext({}),c=function(e){var t=n.useContext(p),r=t;return e&&(r="function"==typeof e?e(t):i(i({},t),e)),r},s=function(e){var t=c(e.components);return n.createElement(p.Provider,{value:t},e.children)},u="mdxType",m={inlineCode:"code",wrapper:function(e){var t=e.children;return n.createElement(n.Fragment,{},t)}},d=n.forwardRef((function(e,t){var r=e.components,a=e.mdxType,o=e.originalType,p=e.parentName,s=l(e,["components","mdxType","originalType","parentName"]),u=c(r),d=a,f=u["".concat(p,".").concat(d)]||u[d]||m[d]||o;return r?n.createElement(f,i(i({ref:t},s),{},{components:r})):n.createElement(f,i({ref:t},s))}));function f(e,t){var r=arguments,a=t&&t.mdxType;if("string"==typeof e||a){var o=r.length,i=new Array(o);i[0]=d;var l={};for(var p in t)hasOwnProperty.call(t,p)&&(l[p]=t[p]);l.originalType=e,l[u]="string"==typeof e?e:a,i[1]=l;for(var c=2;c{r.r(t),r.d(t,{assets:()=>p,contentTitle:()=>i,default:()=>m,frontMatter:()=>o,metadata:()=>l,toc:()=>c});var n=r(7462),a=(r(7294),r(3905));const o={title:"FAQ",sidebar_position:5},i=void 0,l={unversionedId:"faq",id:"faq",title:"FAQ",description:"1. Where are the models stored?",source:"@site/docs/faq.md",sourceDirName:".",slug:"/faq",permalink:"/buzz/docs/faq",draft:!1,tags:[],version:"current",sidebarPosition:5,frontMatter:{title:"FAQ",sidebar_position:5},sidebar:"tutorialSidebar",previous:{title:"CLI",permalink:"/buzz/docs/cli"}},p={},c=[],s={toc:c},u="wrapper";function m(e){let{components:t,...r}=e;return(0,a.kt)(u,(0,n.Z)({},s,r,{components:t,mdxType:"MDXLayout"}),(0,a.kt)("ol",null,(0,a.kt)("li",{parentName:"ol"},(0,a.kt)("p",{parentName:"li"},(0,a.kt)("strong",{parentName:"p"},"Where are the models stored?")),(0,a.kt)("p",{parentName:"li"},"The Whisper models are stored in ",(0,a.kt)("inlineCode",{parentName:"p"},"~/.cache/whisper"),". The Whisper.cpp models are stored in ",(0,a.kt)("inlineCode",{parentName:"p"},"~/Library/Caches/Buzz"),"\n(Mac OS), ",(0,a.kt)("inlineCode",{parentName:"p"},"~/.cache/Buzz")," (Unix), or ",(0,a.kt)("inlineCode",{parentName:"p"},"C:\\Users\\\\AppData\\Local\\Buzz\\Buzz\\Cache")," (Windows). The Hugging Face\nmodels are stored in ",(0,a.kt)("inlineCode",{parentName:"p"},"~/.cache/huggingface/hub"),".")),(0,a.kt)("li",{parentName:"ol"},(0,a.kt)("p",{parentName:"li"},(0,a.kt)("strong",{parentName:"p"},"What can I try if the transcription runs too slowly?")),(0,a.kt)("p",{parentName:"li"},"Try using a lower Whisper model size or using a Whisper.cpp model.")),(0,a.kt)("li",{parentName:"ol"},(0,a.kt)("p",{parentName:"li"},(0,a.kt)("strong",{parentName:"p"},"How to record system audio?")),(0,a.kt)("p",{parentName:"li"},"To transcribe system audio you need to configure virtual audio device and connect output from the applications you want to transcribe to this virtual speaker. After that you can select it as source in the Buzz. See ",(0,a.kt)("a",{parentName:"p",href:"https://chidiwilliams.github.io/buzz/docs/usage#live-recording"},"Usage")," section for more details."),(0,a.kt)("p",{parentName:"li"},"Relevant tools:"),(0,a.kt)("ul",{parentName:"li"},(0,a.kt)("li",{parentName:"ul"},"Mac OS - ",(0,a.kt)("a",{parentName:"li",href:"https://github.com/ExistentialAudio/BlackHole"},"BlackHole"),"."),(0,a.kt)("li",{parentName:"ul"},"Windows - ",(0,a.kt)("a",{parentName:"li",href:"https://vb-audio.com/Cable/"},"VB CABLE")),(0,a.kt)("li",{parentName:"ul"},"Linux - ",(0,a.kt)("a",{parentName:"li",href:"https://wiki.ubuntu.com/record_system_sound"},"PulseAudio Volume Control"))))))}m.isMDXComponent=!0}}]); \ No newline at end of file diff --git a/assets/js/e53fa2b7.0edb1b39.js b/assets/js/e53fa2b7.28463cc8.js similarity index 92% rename from assets/js/e53fa2b7.0edb1b39.js rename to assets/js/e53fa2b7.28463cc8.js index cd402cbb1..9ea7393b5 100644 --- a/assets/js/e53fa2b7.0edb1b39.js +++ b/assets/js/e53fa2b7.28463cc8.js @@ -1 +1 @@ -"use strict";(self.webpackChunkdocs=self.webpackChunkdocs||[]).push([[355],{3905:(e,t,a)=>{a.d(t,{Zo:()=>s,kt:()=>g});var r=a(7294);function n(e,t,a){return t in e?Object.defineProperty(e,t,{value:a,enumerable:!0,configurable:!0,writable:!0}):e[t]=a,e}function o(e,t){var a=Object.keys(e);if(Object.getOwnPropertySymbols){var r=Object.getOwnPropertySymbols(e);t&&(r=r.filter((function(t){return Object.getOwnPropertyDescriptor(e,t).enumerable}))),a.push.apply(a,r)}return a}function i(e){for(var t=1;t=0||(n[a]=e[a]);return n}(e,t);if(Object.getOwnPropertySymbols){var o=Object.getOwnPropertySymbols(e);for(r=0;r=0||Object.prototype.propertyIsEnumerable.call(e,a)&&(n[a]=e[a])}return n}var p=r.createContext({}),u=function(e){var t=r.useContext(p),a=t;return e&&(a="function"==typeof e?e(t):i(i({},t),e)),a},s=function(e){var t=u(e.components);return r.createElement(p.Provider,{value:t},e.children)},c="mdxType",d={inlineCode:"code",wrapper:function(e){var t=e.children;return r.createElement(r.Fragment,{},t)}},m=r.forwardRef((function(e,t){var a=e.components,n=e.mdxType,o=e.originalType,p=e.parentName,s=l(e,["components","mdxType","originalType","parentName"]),c=u(a),m=n,g=c["".concat(p,".").concat(m)]||c[m]||d[m]||o;return a?r.createElement(g,i(i({ref:t},s),{},{components:a})):r.createElement(g,i({ref:t},s))}));function g(e,t){var a=arguments,n=t&&t.mdxType;if("string"==typeof e||n){var o=a.length,i=new Array(o);i[0]=m;var l={};for(var p in t)hasOwnProperty.call(t,p)&&(l[p]=t[p]);l.originalType=e,l[c]="string"==typeof e?e:n,i[1]=l;for(var u=2;u{a.r(t),a.d(t,{assets:()=>p,contentTitle:()=>i,default:()=>d,frontMatter:()=>o,metadata:()=>l,toc:()=>u});var r=a(7462),n=(a(7294),a(3905));const o={title:"Live Recording"},i=void 0,l={unversionedId:"usage/live_recording",id:"usage/live_recording",title:"Live Recording",description:"To start a live recording:",source:"@site/docs/usage/live_recording.md",sourceDirName:"usage",slug:"/usage/live_recording",permalink:"/buzz/docs/usage/live_recording",draft:!1,tags:[],version:"current",frontMatter:{title:"Live Recording"},sidebar:"tutorialSidebar",previous:{title:"File Import",permalink:"/buzz/docs/usage/file_import"},next:{title:"Preferences",permalink:"/buzz/docs/preferences"}},p={},u=[{value:"Record audio playing from computer (macOS)",id:"record-audio-playing-from-computer-macos",level:3},{value:"Record audio playing from computer (Windows)",id:"record-audio-playing-from-computer-windows",level:3},{value:"Record audio playing from computer (Linux)",id:"record-audio-playing-from-computer-linux",level:3}],s={toc:u},c="wrapper";function d(e){let{components:t,...a}=e;return(0,n.kt)(c,(0,r.Z)({},s,a,{components:t,mdxType:"MDXLayout"}),(0,n.kt)("p",null,"To start a live recording:"),(0,n.kt)("ul",null,(0,n.kt)("li",{parentName:"ul"},"Select a recording task, language, quality, and microphone."),(0,n.kt)("li",{parentName:"ul"},"Click Record.")),(0,n.kt)("blockquote",null,(0,n.kt)("p",{parentName:"blockquote"},(0,n.kt)("strong",{parentName:"p"},"Note:")," Transcribing audio using the default Whisper model is resource-intensive. Consider using the Whisper.cpp\nTiny model to get real-time performance.")),(0,n.kt)("table",null,(0,n.kt)("thead",{parentName:"table"},(0,n.kt)("tr",{parentName:"thead"},(0,n.kt)("th",{parentName:"tr",align:null},"Field"),(0,n.kt)("th",{parentName:"tr",align:null},"Options"),(0,n.kt)("th",{parentName:"tr",align:null},"Default"),(0,n.kt)("th",{parentName:"tr",align:null},"Description"))),(0,n.kt)("tbody",{parentName:"table"},(0,n.kt)("tr",{parentName:"tbody"},(0,n.kt)("td",{parentName:"tr",align:null},"Task"),(0,n.kt)("td",{parentName:"tr",align:null},'"Transcribe", "Translate"'),(0,n.kt)("td",{parentName:"tr",align:null},'"Transcribe"'),(0,n.kt)("td",{parentName:"tr",align:null},'"Transcribe" converts the input audio into text in the selected language, while "Translate" converts it into text in English.')),(0,n.kt)("tr",{parentName:"tbody"},(0,n.kt)("td",{parentName:"tr",align:null},"Language"),(0,n.kt)("td",{parentName:"tr",align:null},"See ",(0,n.kt)("a",{parentName:"td",href:"https://github.com/openai/whisper#available-models-and-languages"},"Whisper's documentation")," for the full list of supported languages"),(0,n.kt)("td",{parentName:"tr",align:null},'"Detect Language"'),(0,n.kt)("td",{parentName:"tr",align:null},'"Detect Language" will try to detect the spoken language in the audio based on the first few seconds. However, selecting a language is recommended (if known) as it will improve transcription quality in many cases.')),(0,n.kt)("tr",{parentName:"tbody"},(0,n.kt)("td",{parentName:"tr",align:null},"Quality"),(0,n.kt)("td",{parentName:"tr",align:null},'"Very Low", "Low", "Medium", "High"'),(0,n.kt)("td",{parentName:"tr",align:null},'"Very Low"'),(0,n.kt)("td",{parentName:"tr",align:null},'The transcription quality determines the Whisper model used for transcription. "Very Low" uses the "tiny" model; "Low" uses the "base" model; "Medium" uses the "small" model; and "High" uses the "medium" model. The larger models produce higher-quality transcriptions, but require more system resources. See ',(0,n.kt)("a",{parentName:"td",href:"https://github.com/openai/whisper#available-models-and-languages"},"Whisper's documentation")," for more information about the models.")),(0,n.kt)("tr",{parentName:"tbody"},(0,n.kt)("td",{parentName:"tr",align:null},"Microphone"),(0,n.kt)("td",{parentName:"tr",align:null},"[Available system microphones]"),(0,n.kt)("td",{parentName:"tr",align:null},"[Default system microphone]"),(0,n.kt)("td",{parentName:"tr",align:null},"Microphone for recording input audio.")))),(0,n.kt)("p",null,(0,n.kt)("a",{parentName:"p",href:"https://www.loom.com/share/564b753eb4d44b55b985b8abd26b55f7",title:"Live Recording on Buzz"},(0,n.kt)("img",{parentName:"a",src:"https://cdn.loom.com/sessions/thumbnails/564b753eb4d44b55b985b8abd26b55f7-with-play.gif",alt:"Live Recording on Buzz"}))),(0,n.kt)("h3",{id:"record-audio-playing-from-computer-macos"},"Record audio playing from computer (macOS)"),(0,n.kt)("p",null,"To record audio playing from an application on your computer, you may install an audio loopback driver (a program that\nlets you create virtual audio devices). The rest of this guide will\nuse ",(0,n.kt)("a",{parentName:"p",href:"https://github.com/ExistentialAudio/BlackHole"},"BlackHole")," on Mac, but you can use other alternatives for your\noperating system (\nsee ",(0,n.kt)("a",{parentName:"p",href:"https://nerds.de/en/loopbeaudio.html"},"LoopBeAudio"),", ",(0,n.kt)("a",{parentName:"p",href:"https://rogueamoeba.com/loopback/"},"LoopBack"),",\nand ",(0,n.kt)("a",{parentName:"p",href:"https://vac.muzychenko.net/en/"},"Virtual Audio Cable"),")."),(0,n.kt)("ol",null,(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Install ",(0,n.kt)("a",{parentName:"p",href:"https://github.com/ExistentialAudio/BlackHole#option-2-install-via-homebrew"},"BlackHole via Homebrew")),(0,n.kt)("pre",{parentName:"li"},(0,n.kt)("code",{parentName:"pre",className:"language-shell"},"brew install blackhole-2ch\n"))),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Open Audio MIDI Setup from Spotlight or from ",(0,n.kt)("inlineCode",{parentName:"p"},"/Applications/Utilities/Audio Midi Setup.app"),"."),(0,n.kt)("p",{parentName:"li"},(0,n.kt)("img",{parentName:"p",src:"https://existential.audio/howto/img/spotlight.png",alt:"Open Audio MIDI Setup from Spotlight"}))),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Click the '+' icon at the lower left corner and select 'Create Multi-Output Device'."),(0,n.kt)("p",{parentName:"li"},(0,n.kt)("img",{parentName:"p",src:"https://existential.audio/howto/img/createmulti-output.png",alt:"Create multi-output device"}))),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Add your default speaker and BlackHole to the multi-output device."),(0,n.kt)("p",{parentName:"li"},(0,n.kt)("img",{parentName:"p",src:"https://existential.audio/howto/img/multi-output.png",alt:"Screenshot of multi-output device"}))),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Select this multi-output device as your speaker (application or system-wide) to play audio into BlackHole.")),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Open Buzz, select BlackHole as your microphone, and record as before to see transcriptions from the audio playing\nthrough BlackHole."))),(0,n.kt)("h3",{id:"record-audio-playing-from-computer-windows"},"Record audio playing from computer (Windows)"),(0,n.kt)("p",null,"To transcribe system audio you need to configure virtual audio device and connect output from the applications you whant to transcribe to this virtual speaker. After that you can select it as source in the Buzz."),(0,n.kt)("ol",null,(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Install ",(0,n.kt)("a",{parentName:"p",href:"https://vb-audio.com/Cable/"},"VB CABLE")," as virtual audio device. ")),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},'Configure using Windows Sound settings. Right click on the speaker icon in the system tray and select "Open Sound settings". In the "Choose your output device" dropdown select "CABLE Input" to send all system sound to the virtual device or use "Advanced sound options" to spelect application that will output theit sound to this device.'))),(0,n.kt)("h3",{id:"record-audio-playing-from-computer-linux"},"Record audio playing from computer (Linux)"),(0,n.kt)("p",null,"As described on ",(0,n.kt)("a",{parentName:"p",href:"https://wiki.ubuntu.com/record_system_sound"},"Ubuntu Wiki")," on any Linux with pulse audio you can redirect application audio to a virtual speaker. After that you can select it as source in Buzz."),(0,n.kt)("p",null,"Overall steps:"),(0,n.kt)("ol",null,(0,n.kt)("li",{parentName:"ol"},"Launch application that will produce the sound you want to transcribe and start the playback. For example start a video in a media player. "),(0,n.kt)("li",{parentName:"ol"},"Launch Buzz and open Live recording screen so you see the settings."),(0,n.kt)("li",{parentName:"ol"},"Configure sound routing from the application you want to transcribe sound from to Buzz in ",(0,n.kt)("inlineCode",{parentName:"li"},"Recording tab")," of the PulseAudio Volume Control (",(0,n.kt)("inlineCode",{parentName:"li"},"pavucontrol"),").")))}d.isMDXComponent=!0}}]); \ No newline at end of file +"use strict";(self.webpackChunkdocs=self.webpackChunkdocs||[]).push([[355],{3905:(e,t,a)=>{a.d(t,{Zo:()=>s,kt:()=>g});var r=a(7294);function n(e,t,a){return t in e?Object.defineProperty(e,t,{value:a,enumerable:!0,configurable:!0,writable:!0}):e[t]=a,e}function o(e,t){var a=Object.keys(e);if(Object.getOwnPropertySymbols){var r=Object.getOwnPropertySymbols(e);t&&(r=r.filter((function(t){return Object.getOwnPropertyDescriptor(e,t).enumerable}))),a.push.apply(a,r)}return a}function i(e){for(var t=1;t=0||(n[a]=e[a]);return n}(e,t);if(Object.getOwnPropertySymbols){var o=Object.getOwnPropertySymbols(e);for(r=0;r=0||Object.prototype.propertyIsEnumerable.call(e,a)&&(n[a]=e[a])}return n}var p=r.createContext({}),u=function(e){var t=r.useContext(p),a=t;return e&&(a="function"==typeof e?e(t):i(i({},t),e)),a},s=function(e){var t=u(e.components);return r.createElement(p.Provider,{value:t},e.children)},c="mdxType",d={inlineCode:"code",wrapper:function(e){var t=e.children;return r.createElement(r.Fragment,{},t)}},m=r.forwardRef((function(e,t){var a=e.components,n=e.mdxType,o=e.originalType,p=e.parentName,s=l(e,["components","mdxType","originalType","parentName"]),c=u(a),m=n,g=c["".concat(p,".").concat(m)]||c[m]||d[m]||o;return a?r.createElement(g,i(i({ref:t},s),{},{components:a})):r.createElement(g,i({ref:t},s))}));function g(e,t){var a=arguments,n=t&&t.mdxType;if("string"==typeof e||n){var o=a.length,i=new Array(o);i[0]=m;var l={};for(var p in t)hasOwnProperty.call(t,p)&&(l[p]=t[p]);l.originalType=e,l[c]="string"==typeof e?e:n,i[1]=l;for(var u=2;u{a.r(t),a.d(t,{assets:()=>p,contentTitle:()=>i,default:()=>d,frontMatter:()=>o,metadata:()=>l,toc:()=>u});var r=a(7462),n=(a(7294),a(3905));const o={title:"Live Recording"},i=void 0,l={unversionedId:"usage/live_recording",id:"usage/live_recording",title:"Live Recording",description:"To start a live recording:",source:"@site/docs/usage/live_recording.md",sourceDirName:"usage",slug:"/usage/live_recording",permalink:"/buzz/docs/usage/live_recording",draft:!1,tags:[],version:"current",frontMatter:{title:"Live Recording"},sidebar:"tutorialSidebar",previous:{title:"File Import",permalink:"/buzz/docs/usage/file_import"},next:{title:"Preferences",permalink:"/buzz/docs/preferences"}},p={},u=[{value:"Record audio playing from computer (macOS)",id:"record-audio-playing-from-computer-macos",level:3},{value:"Record audio playing from computer (Windows)",id:"record-audio-playing-from-computer-windows",level:3},{value:"Record audio playing from computer (Linux)",id:"record-audio-playing-from-computer-linux",level:3}],s={toc:u},c="wrapper";function d(e){let{components:t,...a}=e;return(0,n.kt)(c,(0,r.Z)({},s,a,{components:t,mdxType:"MDXLayout"}),(0,n.kt)("p",null,"To start a live recording:"),(0,n.kt)("ul",null,(0,n.kt)("li",{parentName:"ul"},"Select a recording task, language, quality, and microphone."),(0,n.kt)("li",{parentName:"ul"},"Click Record.")),(0,n.kt)("blockquote",null,(0,n.kt)("p",{parentName:"blockquote"},(0,n.kt)("strong",{parentName:"p"},"Note:")," Transcribing audio using the default Whisper model is resource-intensive. Consider using the Whisper.cpp\nTiny model to get real-time performance.")),(0,n.kt)("table",null,(0,n.kt)("thead",{parentName:"table"},(0,n.kt)("tr",{parentName:"thead"},(0,n.kt)("th",{parentName:"tr",align:null},"Field"),(0,n.kt)("th",{parentName:"tr",align:null},"Options"),(0,n.kt)("th",{parentName:"tr",align:null},"Default"),(0,n.kt)("th",{parentName:"tr",align:null},"Description"))),(0,n.kt)("tbody",{parentName:"table"},(0,n.kt)("tr",{parentName:"tbody"},(0,n.kt)("td",{parentName:"tr",align:null},"Task"),(0,n.kt)("td",{parentName:"tr",align:null},'"Transcribe", "Translate"'),(0,n.kt)("td",{parentName:"tr",align:null},'"Transcribe"'),(0,n.kt)("td",{parentName:"tr",align:null},'"Transcribe" converts the input audio into text in the selected language, while "Translate" converts it into text in English.')),(0,n.kt)("tr",{parentName:"tbody"},(0,n.kt)("td",{parentName:"tr",align:null},"Language"),(0,n.kt)("td",{parentName:"tr",align:null},"See ",(0,n.kt)("a",{parentName:"td",href:"https://github.com/openai/whisper#available-models-and-languages"},"Whisper's documentation")," for the full list of supported languages"),(0,n.kt)("td",{parentName:"tr",align:null},'"Detect Language"'),(0,n.kt)("td",{parentName:"tr",align:null},'"Detect Language" will try to detect the spoken language in the audio based on the first few seconds. However, selecting a language is recommended (if known) as it will improve transcription quality in many cases.')),(0,n.kt)("tr",{parentName:"tbody"},(0,n.kt)("td",{parentName:"tr",align:null},"Quality"),(0,n.kt)("td",{parentName:"tr",align:null},'"Very Low", "Low", "Medium", "High"'),(0,n.kt)("td",{parentName:"tr",align:null},'"Very Low"'),(0,n.kt)("td",{parentName:"tr",align:null},'The transcription quality determines the Whisper model used for transcription. "Very Low" uses the "tiny" model; "Low" uses the "base" model; "Medium" uses the "small" model; and "High" uses the "medium" model. The larger models produce higher-quality transcriptions, but require more system resources. See ',(0,n.kt)("a",{parentName:"td",href:"https://github.com/openai/whisper#available-models-and-languages"},"Whisper's documentation")," for more information about the models.")),(0,n.kt)("tr",{parentName:"tbody"},(0,n.kt)("td",{parentName:"tr",align:null},"Microphone"),(0,n.kt)("td",{parentName:"tr",align:null},"[Available system microphones]"),(0,n.kt)("td",{parentName:"tr",align:null},"[Default system microphone]"),(0,n.kt)("td",{parentName:"tr",align:null},"Microphone for recording input audio.")))),(0,n.kt)("p",null,(0,n.kt)("a",{parentName:"p",href:"https://www.loom.com/share/564b753eb4d44b55b985b8abd26b55f7",title:"Live Recording on Buzz"},(0,n.kt)("img",{parentName:"a",src:"https://cdn.loom.com/sessions/thumbnails/564b753eb4d44b55b985b8abd26b55f7-with-play.gif",alt:"Live Recording on Buzz"}))),(0,n.kt)("h3",{id:"record-audio-playing-from-computer-macos"},"Record audio playing from computer (macOS)"),(0,n.kt)("p",null,"To record audio playing from an application on your computer, you may install an audio loopback driver (a program that\nlets you create virtual audio devices). The rest of this guide will\nuse ",(0,n.kt)("a",{parentName:"p",href:"https://github.com/ExistentialAudio/BlackHole"},"BlackHole")," on Mac, but you can use other alternatives for your\noperating system (\nsee ",(0,n.kt)("a",{parentName:"p",href:"https://nerds.de/en/loopbeaudio.html"},"LoopBeAudio"),", ",(0,n.kt)("a",{parentName:"p",href:"https://rogueamoeba.com/loopback/"},"LoopBack"),",\nand ",(0,n.kt)("a",{parentName:"p",href:"https://vac.muzychenko.net/en/"},"Virtual Audio Cable"),")."),(0,n.kt)("ol",null,(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Install ",(0,n.kt)("a",{parentName:"p",href:"https://github.com/ExistentialAudio/BlackHole#option-2-install-via-homebrew"},"BlackHole via Homebrew")),(0,n.kt)("pre",{parentName:"li"},(0,n.kt)("code",{parentName:"pre",className:"language-shell"},"brew install blackhole-2ch\n"))),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Open Audio MIDI Setup from Spotlight or from ",(0,n.kt)("inlineCode",{parentName:"p"},"/Applications/Utilities/Audio Midi Setup.app"),"."),(0,n.kt)("p",{parentName:"li"},(0,n.kt)("img",{parentName:"p",src:"https://existential.audio/howto/img/spotlight.png",alt:"Open Audio MIDI Setup from Spotlight"}))),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Click the '+' icon at the lower left corner and select 'Create Multi-Output Device'."),(0,n.kt)("p",{parentName:"li"},(0,n.kt)("img",{parentName:"p",src:"https://existential.audio/howto/img/createmulti-output.png",alt:"Create multi-output device"}))),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Add your default speaker and BlackHole to the multi-output device."),(0,n.kt)("p",{parentName:"li"},(0,n.kt)("img",{parentName:"p",src:"https://existential.audio/howto/img/multi-output.png",alt:"Screenshot of multi-output device"}))),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Select this multi-output device as your speaker (application or system-wide) to play audio into BlackHole.")),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Open Buzz, select BlackHole as your microphone, and record as before to see transcriptions from the audio playing\nthrough BlackHole."))),(0,n.kt)("h3",{id:"record-audio-playing-from-computer-windows"},"Record audio playing from computer (Windows)"),(0,n.kt)("p",null,"To transcribe system audio you need to configure virtual audio device and connect output from the applications you whant to transcribe to this virtual speaker. After that you can select it as source in the Buzz."),(0,n.kt)("ol",null,(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},"Install ",(0,n.kt)("a",{parentName:"p",href:"https://vb-audio.com/Cable/"},"VB CABLE")," as virtual audio device. ")),(0,n.kt)("li",{parentName:"ol"},(0,n.kt)("p",{parentName:"li"},'Configure using Windows Sound settings. Right-click on the speaker icon in the system tray and select "Open Sound settings". In the "Choose your output device" dropdown select "CABLE Input" to send all system sound to the virtual device or use "Advanced sound options" to select application that will output their sound to this device.'))),(0,n.kt)("h3",{id:"record-audio-playing-from-computer-linux"},"Record audio playing from computer (Linux)"),(0,n.kt)("p",null,"As described on ",(0,n.kt)("a",{parentName:"p",href:"https://wiki.ubuntu.com/record_system_sound"},"Ubuntu Wiki")," on any Linux with pulse audio you can redirect application audio to a virtual speaker. After that you can select it as source in Buzz."),(0,n.kt)("p",null,"Overall steps:"),(0,n.kt)("ol",null,(0,n.kt)("li",{parentName:"ol"},"Launch application that will produce the sound you want to transcribe and start the playback. For example start a video in a media player. "),(0,n.kt)("li",{parentName:"ol"},"Launch Buzz and open Live recording screen, so you see the settings."),(0,n.kt)("li",{parentName:"ol"},"Configure sound routing from the application you want to transcribe sound from to Buzz in ",(0,n.kt)("inlineCode",{parentName:"li"},"Recording tab")," of the PulseAudio Volume Control (",(0,n.kt)("inlineCode",{parentName:"li"},"pavucontrol"),").")))}d.isMDXComponent=!0}}]); \ No newline at end of file diff --git a/assets/js/runtime~main.d1198b17.js b/assets/js/runtime~main.a5cebb54.js similarity index 96% rename from assets/js/runtime~main.d1198b17.js rename to assets/js/runtime~main.a5cebb54.js index 6537ed5c2..0d208d2ba 100644 --- a/assets/js/runtime~main.d1198b17.js +++ b/assets/js/runtime~main.a5cebb54.js @@ -1 +1 @@ -(()=>{"use strict";var e,t,r,o,a,n={},f={};function i(e){var t=f[e];if(void 0!==t)return t.exports;var r=f[e]={id:e,loaded:!1,exports:{}};return n[e].call(r.exports,r,r.exports,i),r.loaded=!0,r.exports}i.m=n,i.c=f,e=[],i.O=(t,r,o,a)=>{if(!r){var n=1/0;for(b=0;b=a)&&Object.keys(i.O).every((e=>i.O[e](r[c])))?r.splice(c--,1):(f=!1,a0&&e[b-1][2]>a;b--)e[b]=e[b-1];e[b]=[r,o,a]},i.n=e=>{var t=e&&e.__esModule?()=>e.default:()=>e;return i.d(t,{a:t}),t},r=Object.getPrototypeOf?e=>Object.getPrototypeOf(e):e=>e.__proto__,i.t=function(e,o){if(1&o&&(e=this(e)),8&o)return e;if("object"==typeof e&&e){if(4&o&&e.__esModule)return e;if(16&o&&"function"==typeof e.then)return e}var a=Object.create(null);i.r(a);var n={};t=t||[null,r({}),r([]),r(r)];for(var f=2&o&&e;"object"==typeof f&&!~t.indexOf(f);f=r(f))Object.getOwnPropertyNames(f).forEach((t=>n[t]=()=>e[t]));return n.default=()=>e,i.d(a,n),a},i.d=(e,t)=>{for(var r in t)i.o(t,r)&&!i.o(e,r)&&Object.defineProperty(e,r,{enumerable:!0,get:t[r]})},i.f={},i.e=e=>Promise.all(Object.keys(i.f).reduce(((t,r)=>(i.f[r](e,t),t)),[])),i.u=e=>"assets/js/"+({34:"871e3331",53:"935f2afb",217:"3b8c55ea",237:"1df93b7f",355:"e53fa2b7",382:"1102fda7",468:"1a20bc57",514:"1be78505",836:"0480b142",860:"3e407b54",918:"17896441",958:"6dbc2e00",971:"c377a04b"}[e]||e)+"."+{34:"b7e0063b",53:"26bf1532",217:"2386d6f2",237:"71d7b441",355:"0edb1b39",382:"25cda54b",468:"23883b0e",514:"131974b6",836:"ed72cf64",860:"987a7018",918:"0aaaf3b3",958:"24fcb0eb",971:"b0b8fe80",972:"d5cc17b1"}[e]+".js",i.miniCssF=e=>{},i.g=function(){if("object"==typeof globalThis)return globalThis;try{return this||new Function("return this")()}catch(e){if("object"==typeof window)return window}}(),i.o=(e,t)=>Object.prototype.hasOwnProperty.call(e,t),o={},a="docs:",i.l=(e,t,r,n)=>{if(o[e])o[e].push(t);else{var f,c;if(void 0!==r)for(var u=document.getElementsByTagName("script"),b=0;b{f.onerror=f.onload=null,clearTimeout(s);var a=o[e];if(delete o[e],f.parentNode&&f.parentNode.removeChild(f),a&&a.forEach((e=>e(r))),t)return t(r)},s=setTimeout(l.bind(null,void 0,{type:"timeout",target:f}),12e4);f.onerror=l.bind(null,f.onerror),f.onload=l.bind(null,f.onload),c&&document.head.appendChild(f)}},i.r=e=>{"undefined"!=typeof Symbol&&Symbol.toStringTag&&Object.defineProperty(e,Symbol.toStringTag,{value:"Module"}),Object.defineProperty(e,"__esModule",{value:!0})},i.p="/buzz/",i.gca=function(e){return e={17896441:"918","871e3331":"34","935f2afb":"53","3b8c55ea":"217","1df93b7f":"237",e53fa2b7:"355","1102fda7":"382","1a20bc57":"468","1be78505":"514","0480b142":"836","3e407b54":"860","6dbc2e00":"958",c377a04b:"971"}[e]||e,i.p+i.u(e)},(()=>{var e={303:0,532:0};i.f.j=(t,r)=>{var o=i.o(e,t)?e[t]:void 0;if(0!==o)if(o)r.push(o[2]);else if(/^(303|532)$/.test(t))e[t]=0;else{var a=new Promise(((r,a)=>o=e[t]=[r,a]));r.push(o[2]=a);var n=i.p+i.u(t),f=new Error;i.l(n,(r=>{if(i.o(e,t)&&(0!==(o=e[t])&&(e[t]=void 0),o)){var a=r&&("load"===r.type?"missing":r.type),n=r&&r.target&&r.target.src;f.message="Loading chunk "+t+" failed.\n("+a+": "+n+")",f.name="ChunkLoadError",f.type=a,f.request=n,o[1](f)}}),"chunk-"+t,t)}},i.O.j=t=>0===e[t];var t=(t,r)=>{var o,a,n=r[0],f=r[1],c=r[2],u=0;if(n.some((t=>0!==e[t]))){for(o in f)i.o(f,o)&&(i.m[o]=f[o]);if(c)var b=c(i)}for(t&&t(r);u{"use strict";var e,t,r,o,a,n={},f={};function i(e){var t=f[e];if(void 0!==t)return t.exports;var r=f[e]={id:e,loaded:!1,exports:{}};return n[e].call(r.exports,r,r.exports,i),r.loaded=!0,r.exports}i.m=n,i.c=f,e=[],i.O=(t,r,o,a)=>{if(!r){var n=1/0;for(b=0;b=a)&&Object.keys(i.O).every((e=>i.O[e](r[c])))?r.splice(c--,1):(f=!1,a0&&e[b-1][2]>a;b--)e[b]=e[b-1];e[b]=[r,o,a]},i.n=e=>{var t=e&&e.__esModule?()=>e.default:()=>e;return i.d(t,{a:t}),t},r=Object.getPrototypeOf?e=>Object.getPrototypeOf(e):e=>e.__proto__,i.t=function(e,o){if(1&o&&(e=this(e)),8&o)return e;if("object"==typeof e&&e){if(4&o&&e.__esModule)return e;if(16&o&&"function"==typeof e.then)return e}var a=Object.create(null);i.r(a);var n={};t=t||[null,r({}),r([]),r(r)];for(var f=2&o&&e;"object"==typeof f&&!~t.indexOf(f);f=r(f))Object.getOwnPropertyNames(f).forEach((t=>n[t]=()=>e[t]));return n.default=()=>e,i.d(a,n),a},i.d=(e,t)=>{for(var r in t)i.o(t,r)&&!i.o(e,r)&&Object.defineProperty(e,r,{enumerable:!0,get:t[r]})},i.f={},i.e=e=>Promise.all(Object.keys(i.f).reduce(((t,r)=>(i.f[r](e,t),t)),[])),i.u=e=>"assets/js/"+({34:"871e3331",53:"935f2afb",217:"3b8c55ea",237:"1df93b7f",355:"e53fa2b7",382:"1102fda7",468:"1a20bc57",514:"1be78505",836:"0480b142",860:"3e407b54",918:"17896441",958:"6dbc2e00",971:"c377a04b"}[e]||e)+"."+{34:"b7e0063b",53:"26bf1532",217:"2386d6f2",237:"71d7b441",355:"28463cc8",382:"25cda54b",468:"23883b0e",514:"131974b6",836:"4cf7acba",860:"987a7018",918:"0aaaf3b3",958:"24fcb0eb",971:"b0b8fe80",972:"d5cc17b1"}[e]+".js",i.miniCssF=e=>{},i.g=function(){if("object"==typeof globalThis)return globalThis;try{return this||new Function("return this")()}catch(e){if("object"==typeof window)return window}}(),i.o=(e,t)=>Object.prototype.hasOwnProperty.call(e,t),o={},a="docs:",i.l=(e,t,r,n)=>{if(o[e])o[e].push(t);else{var f,c;if(void 0!==r)for(var u=document.getElementsByTagName("script"),b=0;b{f.onerror=f.onload=null,clearTimeout(s);var a=o[e];if(delete o[e],f.parentNode&&f.parentNode.removeChild(f),a&&a.forEach((e=>e(r))),t)return t(r)},s=setTimeout(l.bind(null,void 0,{type:"timeout",target:f}),12e4);f.onerror=l.bind(null,f.onerror),f.onload=l.bind(null,f.onload),c&&document.head.appendChild(f)}},i.r=e=>{"undefined"!=typeof Symbol&&Symbol.toStringTag&&Object.defineProperty(e,Symbol.toStringTag,{value:"Module"}),Object.defineProperty(e,"__esModule",{value:!0})},i.p="/buzz/",i.gca=function(e){return e={17896441:"918","871e3331":"34","935f2afb":"53","3b8c55ea":"217","1df93b7f":"237",e53fa2b7:"355","1102fda7":"382","1a20bc57":"468","1be78505":"514","0480b142":"836","3e407b54":"860","6dbc2e00":"958",c377a04b:"971"}[e]||e,i.p+i.u(e)},(()=>{var e={303:0,532:0};i.f.j=(t,r)=>{var o=i.o(e,t)?e[t]:void 0;if(0!==o)if(o)r.push(o[2]);else if(/^(303|532)$/.test(t))e[t]=0;else{var a=new Promise(((r,a)=>o=e[t]=[r,a]));r.push(o[2]=a);var n=i.p+i.u(t),f=new Error;i.l(n,(r=>{if(i.o(e,t)&&(0!==(o=e[t])&&(e[t]=void 0),o)){var a=r&&("load"===r.type?"missing":r.type),n=r&&r.target&&r.target.src;f.message="Loading chunk "+t+" failed.\n("+a+": "+n+")",f.name="ChunkLoadError",f.type=a,f.request=n,o[1](f)}}),"chunk-"+t,t)}},i.O.j=t=>0===e[t];var t=(t,r)=>{var o,a,n=r[0],f=r[1],c=r[2],u=0;if(n.some((t=>0!==e[t]))){for(o in f)i.o(f,o)&&(i.m[o]=f[o]);if(c)var b=c(i)}for(t&&t(r);u Introduction | Buzz - + @@ -20,7 +20,7 @@ Whisper.cpp, Faster Whisper, Whisper-compatible Hugging Face models, and the OpenAI Whisper API
  • Command-Line Interface
  • Available on Mac, Windows, and Linux
  • - + \ No newline at end of file diff --git a/docs/cli.html b/docs/cli.html index 91970a433..5bb08d268 100644 --- a/docs/cli.html +++ b/docs/cli.html @@ -4,13 +4,13 @@ CLI | Buzz - +

    CLI

    Commands

    add

    Start a new transcription task.

    Usage: buzz add [options] [file file file...]

    Options:
    -t, --task <task> The task to perform. Allowed: translate,
    transcribe. Default: transcribe.
    -m, --model-type <model-type> Model type. Allowed: whisper, whispercpp,
    huggingface, fasterwhisper, openaiapi. Default:
    whisper.
    -s, --model-size <model-size> Model size. Use only when --model-type is
    whisper, whispercpp, or fasterwhisper. Allowed:
    tiny, base, small, medium, large. Default:
    tiny.
    --hfid <id> Hugging Face model ID. Use only when
    --model-type is huggingface. Example:
    "openai/whisper-tiny"
    -l, --language <code> Language code. Allowed: af (Afrikaans), am
    (Amharic), ar (Arabic), as (Assamese), az
    (Azerbaijani), ba (Bashkir), be (Belarusian),
    bg (Bulgarian), bn (Bengali), bo (Tibetan), br
    (Breton), bs (Bosnian), ca (Catalan), cs
    (Czech), cy (Welsh), da (Danish), de (German),
    el (Greek), en (English), es (Spanish), et
    (Estonian), eu (Basque), fa (Persian), fi
    (Finnish), fo (Faroese), fr (French), gl
    (Galician), gu (Gujarati), ha (Hausa), haw
    (Hawaiian), he (Hebrew), hi (Hindi), hr
    (Croatian), ht (Haitian Creole), hu
    (Hungarian), hy (Armenian), id (Indonesian), is
    (Icelandic), it (Italian), ja (Japanese), jw
    (Javanese), ka (Georgian), kk (Kazakh), km
    (Khmer), kn (Kannada), ko (Korean), la (Latin),
    lb (Luxembourgish), ln (Lingala), lo (Lao), lt
    (Lithuanian), lv (Latvian), mg (Malagasy), mi
    (Maori), mk (Macedonian), ml (Malayalam), mn
    (Mongolian), mr (Marathi), ms (Malay), mt
    (Maltese), my (Myanmar), ne (Nepali), nl
    (Dutch), nn (Nynorsk), no (Norwegian), oc
    (Occitan), pa (Punjabi), pl (Polish), ps
    (Pashto), pt (Portuguese), ro (Romanian), ru
    (Russian), sa (Sanskrit), sd (Sindhi), si
    (Sinhala), sk (Slovak), sl (Slovenian), sn
    (Shona), so (Somali), sq (Albanian), sr
    (Serbian), su (Sundanese), sv (Swedish), sw
    (Swahili), ta (Tamil), te (Telugu), tg (Tajik),
    th (Thai), tk (Turkmen), tl (Tagalog), tr
    (Turkish), tt (Tatar), uk (Ukrainian), ur
    (Urdu), uz (Uzbek), vi (Vietnamese), yi
    (Yiddish), yo (Yoruba), zh (Chinese). Leave
    empty to detect language.
    -p, --prompt <prompt> Initial prompt
    --openai-token <token> OpenAI access token. Use only when
    --model-type is openaiapi. Defaults to your
    previously saved access token, if one exists.
    --srt Output result in an SRT file.
    --vtt Output result in a VTT file.
    --txt Output result in a TXT file.
    -h, --help Displays help on commandline options.
    --help-all Displays help including Qt specific options.
    -v, --version Displays version information.

    Arguments:
    files Input file paths

    Examples:

    # Translate two MP3 files from French to English using OpenAI Whisper API
    buzz add --task translate --language fr --model-type openaiapi /Users/user/Downloads/1b3b03e4-8db5-ea2c-ace5-b71ff32e3304.mp3 /Users/user/Downloads/koaf9083k1lkpsfdi0.mp3

    # Transcribe an MP4 using Whisper.cpp "small" model and immediately export to SRT and VTT files
    buzz add --task transcribe --model-type whispercpp --model-size small --prompt "My initial prompt" --srt --vtt /Users/user/Downloads/buzz/1b3b03e4-8db5-ea2c-ace5-b71ff32e3304.mp4
    - + \ No newline at end of file diff --git a/docs/faq.html b/docs/faq.html index 782118bc3..c1b388983 100644 --- a/docs/faq.html +++ b/docs/faq.html @@ -4,15 +4,15 @@ FAQ | Buzz - +

    FAQ

    1. Where are the models stored?

      The Whisper models are stored in ~/.cache/whisper. The Whisper.cpp models are stored in ~/Library/Caches/Buzz (Mac OS), ~/.cache/Buzz (Unix), or C:\Users\<username>\AppData\Local\Buzz\Buzz\Cache (Windows). The Hugging Face -models are stored in ~/.cache/huggingface/hub.

    2. What can I try if the transcription runs too slowly?

      Try using a lower Whisper model size or using a Whisper.cpp model.

    3. How to record system audio?

      To transcribe system audio you need to configure virtual audio device and connect output from the applications you whant to transcribe to this virtual speaker. After that you can select it as source in the Buzz. See Usage section for more details.

      Relevant tools:

    - +models are stored in ~/.cache/huggingface/hub.

  • What can I try if the transcription runs too slowly?

    Try using a lower Whisper model size or using a Whisper.cpp model.

  • How to record system audio?

    To transcribe system audio you need to configure virtual audio device and connect output from the applications you want to transcribe to this virtual speaker. After that you can select it as source in the Buzz. See Usage section for more details.

    Relevant tools:

  • + \ No newline at end of file diff --git a/docs/installation.html b/docs/installation.html index eb85af175..45dee9065 100644 --- a/docs/installation.html +++ b/docs/installation.html @@ -4,7 +4,7 @@ Installation | Buzz - + @@ -14,7 +14,7 @@ the App Store version.)

    macOS (Intel, macOS 11.7 and later)

    Install via brew:

    brew install --cask buzz

    Alternatively, download and run the Buzz-x.y.z.dmg file.

    For Mac Silicon (and for a better experience on Mac Intel), download Buzz Captions on the App Store.

    Windows (Windows 10 and later)

    Download and run the Buzz-x.y.z.exe file.

    Linux

    sudo apt-get install libportaudio2
    sudo snap install buzz

    Get it from the Snap Store

    Alternatively, on Ubuntu 20.04 and later, install the dependencies:

    sudo apt-get install libportaudio2

    Then, download and extract the Buzz-x.y.z-unix.tar.gz file

    PyPI

    pip install buzz-captions
    python -m buzz
    - + \ No newline at end of file diff --git a/docs/preferences.html b/docs/preferences.html index 68882a8eb..0ab76b64e 100644 --- a/docs/preferences.html +++ b/docs/preferences.html @@ -4,7 +4,7 @@ Preferences | Buzz - + @@ -12,7 +12,7 @@

    Preferences

    Open the Preferences window from the Menu bar, or click Ctrl/Cmd + ,.

    General Preferences

    Default export file name

    Sets the default export file name for file transcriptions. For example, a value of {{ input_file_name }} ({{ task }}d on {{ date_time }}) will save TXT exports as Input Filename (transcribed on 19-Sep-2023 20-39-25).txt by default.

    Available variables:

    KeyDescriptionExample
    input_file_nameFile name of the imported fileaudio (e.g. if the imported file path was /path/to/audio.wav
    taskTranscription tasktranscribe, translate
    languageLanguage codeen, fr, yo, etc.
    model_typeModel typeWhisper, Whisper.cpp, Faster Whisper, etc.
    model_sizeModel sizetiny, base, small, medium, large, etc.
    date_timeExport time (format: %d-%b-%Y %H-%M-%S)19-Sep-2023 20-39-25
    - + \ No newline at end of file diff --git a/docs/usage/file_import.html b/docs/usage/file_import.html index 6135d5f7d..5f4fa1634 100644 --- a/docs/usage/file_import.html +++ b/docs/usage/file_import.html @@ -4,14 +4,14 @@ File Import | Buzz - +

    File Import

    To import a file:

    • Click Import Media File on the File menu (or the '+' icon on the toolbar, or Command/Ctrl + O).
    • Choose an audio or video file.
    • Select a task, language, and the model settings.
    • Click Run.
    • When the transcription status shows 'Completed', double-click on the row (or select the row and click the '⤢' icon) to open the transcription.
    FieldOptionsDefaultDescription
    Export As"TXT", "SRT", "VTT""TXT"Export file format
    Word-Level TimingsOff / OnOffIf checked, the transcription will generate a separate subtitle line for each word in the audio. Enabled only when "Export As" is set to "SRT" or "VTT".

    (See the Live Recording section for more information about the task, language, and quality settings.)

    Media File Import on Buzz

    - + \ No newline at end of file diff --git a/docs/usage/live_recording.html b/docs/usage/live_recording.html index 0c364ccdc..f612059b2 100644 --- a/docs/usage/live_recording.html +++ b/docs/usage/live_recording.html @@ -4,7 +4,7 @@ Live Recording | Buzz - + @@ -16,8 +16,8 @@ operating system ( see LoopBeAudio, LoopBack, and Virtual Audio Cable).

    1. Install BlackHole via Homebrew

      brew install blackhole-2ch
    2. Open Audio MIDI Setup from Spotlight or from /Applications/Utilities/Audio Midi Setup.app.

      Open Audio MIDI Setup from Spotlight

    3. Click the '+' icon at the lower left corner and select 'Create Multi-Output Device'.

      Create multi-output device

    4. Add your default speaker and BlackHole to the multi-output device.

      Screenshot of multi-output device

    5. Select this multi-output device as your speaker (application or system-wide) to play audio into BlackHole.

    6. Open Buzz, select BlackHole as your microphone, and record as before to see transcriptions from the audio playing -through BlackHole.

    Record audio playing from computer (Windows)

    To transcribe system audio you need to configure virtual audio device and connect output from the applications you whant to transcribe to this virtual speaker. After that you can select it as source in the Buzz.

    1. Install VB CABLE as virtual audio device.

    2. Configure using Windows Sound settings. Right click on the speaker icon in the system tray and select "Open Sound settings". In the "Choose your output device" dropdown select "CABLE Input" to send all system sound to the virtual device or use "Advanced sound options" to spelect application that will output theit sound to this device.

    Record audio playing from computer (Linux)

    As described on Ubuntu Wiki on any Linux with pulse audio you can redirect application audio to a virtual speaker. After that you can select it as source in Buzz.

    Overall steps:

    1. Launch application that will produce the sound you want to transcribe and start the playback. For example start a video in a media player.
    2. Launch Buzz and open Live recording screen so you see the settings.
    3. Configure sound routing from the application you want to transcribe sound from to Buzz in Recording tab of the PulseAudio Volume Control (pavucontrol).
    - +through BlackHole.

    Record audio playing from computer (Windows)

    To transcribe system audio you need to configure virtual audio device and connect output from the applications you whant to transcribe to this virtual speaker. After that you can select it as source in the Buzz.

    1. Install VB CABLE as virtual audio device.

    2. Configure using Windows Sound settings. Right-click on the speaker icon in the system tray and select "Open Sound settings". In the "Choose your output device" dropdown select "CABLE Input" to send all system sound to the virtual device or use "Advanced sound options" to select application that will output their sound to this device.

    Record audio playing from computer (Linux)

    As described on Ubuntu Wiki on any Linux with pulse audio you can redirect application audio to a virtual speaker. After that you can select it as source in Buzz.

    Overall steps:

    1. Launch application that will produce the sound you want to transcribe and start the playback. For example start a video in a media player.
    2. Launch Buzz and open Live recording screen, so you see the settings.
    3. Configure sound routing from the application you want to transcribe sound from to Buzz in Recording tab of the PulseAudio Volume Control (pavucontrol).
    + \ No newline at end of file diff --git a/index.html b/index.html index 657573586..a0ca4142c 100644 --- a/index.html +++ b/index.html @@ -4,13 +4,13 @@ Buzz - +
    - + \ No newline at end of file