|
<!DOCTYPE html> |
|
<html lang="en"> |
|
|
|
<head> |
|
<meta charset="UTF-8"> |
|
<meta name="viewport" content="width=device-width, initial-scale=1.0"> |
|
<title>Real-time Whisper Transcription</title> |
|
<style> |
|
:root { |
|
--primary-gradient: linear-gradient(135deg, #f9a45c 0%, #e66465 100%); |
|
--background-cream: #faf8f5; |
|
--background-cream-end: #f7f5f2; |
|
|
|
--text-dark: #2d2d2d; |
|
--transcript-bg: #ffffff; |
|
|
|
--transcript-border: #e0e0e0; |
|
|
|
} |
|
|
|
body { |
|
font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, sans-serif; |
|
margin: 0; |
|
padding: 0; |
|
|
|
background: linear-gradient(to bottom, var(--background-cream), var(--background-cream-end)); |
|
color: var(--text-dark); |
|
min-height: 100vh; |
|
} |
|
|
|
.hero { |
|
background: var(--primary-gradient); |
|
color: white; |
|
padding: 2.5rem 2rem; |
|
text-align: center; |
|
} |
|
|
|
.hero h1 { |
|
font-size: 2.5rem; |
|
margin: 0; |
|
font-weight: 600; |
|
letter-spacing: -0.5px; |
|
} |
|
|
|
.hero p { |
|
font-size: 1rem; |
|
margin-top: 0.5rem; |
|
opacity: 0.9; |
|
} |
|
|
|
.container { |
|
max-width: 1000px; |
|
margin: 2.5rem auto; |
|
|
|
padding: 0 2rem; |
|
} |
|
|
|
.transcript-container { |
|
border-radius: 12px; |
|
|
|
box-shadow: 0 4px 15px rgba(0, 0, 0, 0.08); |
|
|
|
padding: 1.5rem; |
|
height: 350px; |
|
|
|
overflow-y: auto; |
|
margin-bottom: 2rem; |
|
|
|
border: 1px solid rgba(0, 0, 0, 0.05); |
|
|
|
background-color: var(--transcript-bg); |
|
|
|
} |
|
|
|
.controls { |
|
text-align: center; |
|
margin: 1.5rem 0; |
|
} |
|
|
|
button { |
|
background: var(--primary-gradient); |
|
color: white; |
|
border: none; |
|
padding: 10px 20px; |
|
font-size: 0.95rem; |
|
border-radius: 6px; |
|
cursor: pointer; |
|
transition: all 0.2s ease; |
|
font-weight: 500; |
|
min-width: 180px; |
|
position: relative; |
|
padding-right: 50px; |
|
} |
|
|
|
button:hover { |
|
transform: translateY(-1px); |
|
box-shadow: 0 4px 12px rgba(230, 100, 101, 0.15); |
|
} |
|
|
|
button:active { |
|
transform: translateY(0); |
|
} |
|
|
|
|
|
.transcript-container p { |
|
margin: 0.6rem 0; |
|
|
|
padding: 0.8rem 1rem; |
|
|
|
background: var(--background-cream); |
|
|
|
border-radius: 6px; |
|
|
|
line-height: 1.5; |
|
|
|
font-size: 0.98rem; |
|
|
|
border-left: 3px solid var(--transcript-border); |
|
|
|
transition: background-color 0.2s ease; |
|
|
|
} |
|
|
|
.transcript-container p:hover { |
|
background-color: #fdfbf9; |
|
|
|
} |
|
|
|
|
|
.transcript-container::-webkit-scrollbar { |
|
width: 8px; |
|
|
|
} |
|
|
|
.transcript-container::-webkit-scrollbar-track { |
|
background: var(--background-cream-end); |
|
|
|
border-radius: 4px; |
|
} |
|
|
|
.transcript-container::-webkit-scrollbar-thumb { |
|
background: #e66465; |
|
border-radius: 3px; |
|
opacity: 0.8; |
|
} |
|
|
|
.transcript-container::-webkit-scrollbar-thumb:hover { |
|
background: #f9a45c; |
|
} |
|
|
|
|
|
.toast { |
|
position: fixed; |
|
top: 20px; |
|
left: 50%; |
|
transform: translateX(-50%); |
|
padding: 16px 24px; |
|
border-radius: 4px; |
|
font-size: 14px; |
|
z-index: 1000; |
|
display: none; |
|
box-shadow: 0 2px 5px rgba(0, 0, 0, 0.2); |
|
} |
|
|
|
.toast.error { |
|
background-color: #f44336; |
|
color: white; |
|
} |
|
|
|
.toast.warning { |
|
background-color: #ffd700; |
|
color: black; |
|
} |
|
|
|
|
|
.icon-with-spinner { |
|
display: flex; |
|
align-items: center; |
|
justify-content: center; |
|
gap: 12px; |
|
min-width: 180px; |
|
} |
|
|
|
.spinner { |
|
width: 20px; |
|
height: 20px; |
|
border: 2px solid white; |
|
border-top-color: transparent; |
|
border-radius: 50%; |
|
animation: spin 1s linear infinite; |
|
flex-shrink: 0; |
|
} |
|
|
|
.pulse-container { |
|
display: flex; |
|
align-items: center; |
|
justify-content: center; |
|
gap: 12px; |
|
min-width: 180px; |
|
} |
|
|
|
.pulse-circle { |
|
width: 20px; |
|
height: 20px; |
|
border-radius: 50%; |
|
background-color: white; |
|
opacity: 0.2; |
|
flex-shrink: 0; |
|
transform: translateX(-0%) scale(var(--audio-level, 1)); |
|
transition: transform 0.1s ease; |
|
} |
|
|
|
|
|
.mute-toggle { |
|
position: absolute; |
|
right: 10px; |
|
top: 50%; |
|
transform: translateY(-50%); |
|
width: 24px; |
|
height: 24px; |
|
cursor: pointer; |
|
display: flex; |
|
align-items: center; |
|
justify-content: center; |
|
} |
|
|
|
.mute-toggle svg { |
|
width: 20px; |
|
height: 20px; |
|
stroke: white; |
|
} |
|
|
|
|
|
.button-content { |
|
display: flex; |
|
align-items: center; |
|
justify-content: center; |
|
width: calc(100% - 40px); |
|
margin-right: 40px; |
|
} |
|
|
|
.icon-with-spinner, |
|
.pulse-container { |
|
width: 100%; |
|
} |
|
|
|
@keyframes spin { |
|
to { |
|
transform: rotate(360deg); |
|
} |
|
} |
|
</style> |
|
</head> |
|
|
|
<body> |
|
|
|
<div id="error-toast" class="toast"></div> |
|
<div class="hero"> |
|
<h1>Real-time Transcription</h1> |
|
<p>Powered by Groq and FastRTC</p> |
|
</div> |
|
|
|
<div class="container"> |
|
<div class="transcript-container" id="transcript"> |
|
</div> |
|
<div class="controls"> |
|
<button id="start-button">Start Recording</button> |
|
</div> |
|
</div> |
|
|
|
<script> |
|
let peerConnection; |
|
let webrtc_id; |
|
let audioContext, analyser, audioSource; |
|
let audioLevel = 0; |
|
let animationFrame; |
|
let isMuted = false; |
|
|
|
const startButton = document.getElementById('start-button'); |
|
const transcriptDiv = document.getElementById('transcript'); |
|
|
|
|
|
const micIconSVG = ` |
|
<svg xmlns="http://www.w3.org/2000/svg" width="100%" height="100%" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round"> |
|
<path d="M12 1a3 3 0 0 0-3 3v8a3 3 0 0 0 6 0V4a3 3 0 0 0-3-3z"></path> |
|
<path d="M19 10v2a7 7 0 0 1-14 0v-2"></path> |
|
<line x1="12" y1="19" x2="12" y2="23"></line> |
|
<line x1="8" y1="23" x2="16" y2="23"></line> |
|
</svg>`; |
|
|
|
const micMutedIconSVG = ` |
|
<svg xmlns="http://www.w3.org/2000/svg" width="100%" height="100%" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round"> |
|
<path d="M12 1a3 3 0 0 0-3 3v8a3 3 0 0 0 6 0V4a3 3 0 0 0-3-3z"></path> |
|
<path d="M19 10v2a7 7 0 0 1-14 0v-2"></path> |
|
<line x1="12" y1="19" x2="12" y2="23"></line> |
|
<line x1="8" y1="23" x2="16" y2="23"></line> |
|
<line x1="1" y1="1" x2="23" y2="23"></line> |
|
</svg>`; |
|
|
|
function showError(message) { |
|
const toast = document.getElementById('error-toast'); |
|
toast.textContent = message; |
|
toast.style.display = 'block'; |
|
|
|
|
|
setTimeout(() => { |
|
toast.style.display = 'none'; |
|
}, 5000); |
|
} |
|
|
|
async function handleMessage(event) { |
|
|
|
const eventJson = JSON.parse(event.data); |
|
if (eventJson.type === "error") { |
|
showError(eventJson.message); |
|
} else if (eventJson.type === "send_input") { |
|
const response = await fetch('/send_input', { |
|
method: 'POST', |
|
headers: { 'Content-Type': 'application/json' }, |
|
body: JSON.stringify({ |
|
webrtc_id: webrtc_id, |
|
transcript: "" |
|
}) |
|
}); |
|
} |
|
console.log('Received message:', event.data); |
|
|
|
} |
|
|
|
function updateButtonState() { |
|
|
|
const existingMuteButton = startButton.querySelector('.mute-toggle'); |
|
if (existingMuteButton) { |
|
existingMuteButton.removeEventListener('click', toggleMute); |
|
existingMuteButton.remove(); |
|
} |
|
|
|
if (peerConnection && (peerConnection.connectionState === 'connecting' || peerConnection.connectionState === 'new')) { |
|
startButton.innerHTML = ` |
|
<div class="button-content"> |
|
<div class="icon-with-spinner"> |
|
<div class="spinner"></div> |
|
<span>Connecting...</span> |
|
</div> |
|
</div> |
|
`; |
|
startButton.disabled = true; |
|
} else if (peerConnection && peerConnection.connectionState === 'connected') { |
|
startButton.innerHTML = ` |
|
<div class="button-content"> |
|
<div class="pulse-container"> |
|
<div class="pulse-circle"></div> |
|
<span>Stop Recording</span> |
|
</div> |
|
</div> |
|
<div class="mute-toggle" title="${isMuted ? 'Unmute' : 'Mute'}"> |
|
${isMuted ? micMutedIconSVG : micIconSVG} |
|
</div> |
|
`; |
|
startButton.disabled = false; |
|
const muteButton = startButton.querySelector('.mute-toggle'); |
|
if (muteButton) { |
|
muteButton.addEventListener('click', toggleMute); |
|
} |
|
} else { |
|
startButton.innerHTML = 'Start Recording'; |
|
startButton.disabled = false; |
|
} |
|
} |
|
|
|
function toggleMute(event) { |
|
event.stopPropagation(); |
|
if (!peerConnection || peerConnection.connectionState !== 'connected') return; |
|
|
|
isMuted = !isMuted; |
|
console.log("Mute toggled:", isMuted); |
|
|
|
peerConnection.getSenders().forEach(sender => { |
|
if (sender.track && sender.track.kind === 'audio') { |
|
sender.track.enabled = !isMuted; |
|
console.log(`Audio track ${sender.track.id} enabled: ${!isMuted}`); |
|
} |
|
}); |
|
|
|
updateButtonState(); |
|
} |
|
|
|
function setupAudioVisualization(stream) { |
|
audioContext = new (window.AudioContext || window.webkitAudioContext)(); |
|
analyser = audioContext.createAnalyser(); |
|
audioSource = audioContext.createMediaStreamSource(stream); |
|
audioSource.connect(analyser); |
|
analyser.fftSize = 64; |
|
const dataArray = new Uint8Array(analyser.frequencyBinCount); |
|
|
|
function updateAudioLevel() { |
|
analyser.getByteFrequencyData(dataArray); |
|
const average = Array.from(dataArray).reduce((a, b) => a + b, 0) / dataArray.length; |
|
audioLevel = average / 255; |
|
|
|
const pulseCircle = document.querySelector('.pulse-circle'); |
|
if (pulseCircle) { |
|
pulseCircle.style.setProperty('--audio-level', 1 + audioLevel); |
|
} |
|
|
|
animationFrame = requestAnimationFrame(updateAudioLevel); |
|
} |
|
updateAudioLevel(); |
|
} |
|
|
|
async function setupWebRTC() { |
|
const config = __RTC_CONFIGURATION__; |
|
peerConnection = new RTCPeerConnection(config); |
|
|
|
const timeoutId = setTimeout(() => { |
|
const toast = document.getElementById('error-toast'); |
|
toast.textContent = "Connection is taking longer than usual. Are you on a VPN?"; |
|
toast.className = 'toast warning'; |
|
toast.style.display = 'block'; |
|
|
|
|
|
setTimeout(() => { |
|
toast.style.display = 'none'; |
|
}, 5000); |
|
}, 5000); |
|
|
|
try { |
|
const stream = await navigator.mediaDevices.getUserMedia({ |
|
audio: true |
|
}); |
|
|
|
setupAudioVisualization(stream); |
|
|
|
stream.getTracks().forEach(track => { |
|
peerConnection.addTrack(track, stream); |
|
}); |
|
|
|
|
|
peerConnection.addEventListener('connectionstatechange', () => { |
|
console.log('connectionstatechange', peerConnection.connectionState); |
|
if (peerConnection.connectionState === 'connected') { |
|
clearTimeout(timeoutId); |
|
const toast = document.getElementById('error-toast'); |
|
toast.style.display = 'none'; |
|
} |
|
updateButtonState(); |
|
}); |
|
|
|
peerConnection.onicecandidate = ({ candidate }) => { |
|
if (candidate) { |
|
console.debug("Sending ICE candidate", candidate); |
|
fetch('/webrtc/offer', { |
|
method: 'POST', |
|
headers: { 'Content-Type': 'application/json' }, |
|
body: JSON.stringify({ |
|
candidate: candidate.toJSON(), |
|
webrtc_id: webrtc_id, |
|
type: "ice-candidate", |
|
}) |
|
}) |
|
} |
|
}; |
|
|
|
|
|
const dataChannel = peerConnection.createDataChannel('text'); |
|
dataChannel.onmessage = handleMessage; |
|
|
|
|
|
const offer = await peerConnection.createOffer(); |
|
await peerConnection.setLocalDescription(offer); |
|
|
|
webrtc_id = Math.random().toString(36).substring(7); |
|
|
|
const response = await fetch('/webrtc/offer', { |
|
method: 'POST', |
|
headers: { 'Content-Type': 'application/json' }, |
|
body: JSON.stringify({ |
|
sdp: peerConnection.localDescription.sdp, |
|
type: peerConnection.localDescription.type, |
|
webrtc_id: webrtc_id |
|
}) |
|
}); |
|
|
|
const serverResponse = await response.json(); |
|
|
|
if (serverResponse.status === 'failed') { |
|
showError(serverResponse.meta.error === 'concurrency_limit_reached' |
|
? `Too many connections. Maximum limit is ${serverResponse.meta.limit}` |
|
: serverResponse.meta.error); |
|
stop(); |
|
startButton.textContent = 'Start Recording'; |
|
return; |
|
} |
|
|
|
await peerConnection.setRemoteDescription(serverResponse); |
|
|
|
|
|
const eventSource = new EventSource('/transcript?webrtc_id=' + webrtc_id); |
|
eventSource.addEventListener("output", (event) => { |
|
appendTranscript(event.data); |
|
}); |
|
} catch (err) { |
|
clearTimeout(timeoutId); |
|
console.error('Error setting up WebRTC:', err); |
|
showError('Failed to establish connection. Please try again.'); |
|
stop(); |
|
startButton.textContent = 'Start Recording'; |
|
} |
|
} |
|
|
|
function appendTranscript(text) { |
|
const p = document.createElement('p'); |
|
p.textContent = text; |
|
transcriptDiv.appendChild(p); |
|
transcriptDiv.scrollTop = transcriptDiv.scrollHeight; |
|
} |
|
|
|
function stop() { |
|
if (animationFrame) { |
|
cancelAnimationFrame(animationFrame); |
|
animationFrame = null; |
|
} |
|
if (audioContext) { |
|
audioContext.close().catch(e => console.error("Error closing AudioContext:", e)); |
|
audioContext = null; |
|
analyser = null; |
|
audioSource = null; |
|
} |
|
if (peerConnection) { |
|
if (peerConnection.getSenders) { |
|
peerConnection.getSenders().forEach(sender => { |
|
if (sender.track) { |
|
sender.track.stop(); |
|
console.log(`Track ${sender.track.id} stopped.`); |
|
} |
|
}); |
|
} |
|
peerConnection.close(); |
|
peerConnection = null; |
|
console.log("Peer connection closed."); |
|
} |
|
audioLevel = 0; |
|
isMuted = false; |
|
updateButtonState(); |
|
} |
|
|
|
startButton.addEventListener('click', (event) => { |
|
if (event.target.closest('.mute-toggle')) { |
|
return; |
|
} |
|
|
|
if (peerConnection && peerConnection.connectionState === 'connected') { |
|
console.log("Stop button clicked"); |
|
stop(); |
|
} else if (!peerConnection || ['new', 'closed', 'failed', 'disconnected'].includes(peerConnection.connectionState)) { |
|
console.log("Start button clicked"); |
|
transcriptDiv.innerHTML = ''; |
|
setupWebRTC(); |
|
updateButtonState(); |
|
} |
|
}); |
|
</script> |
|
</body> |
|
|
|
</html> |