meeting-noter 0.7.0__py3-none-any.whl → 1.0.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of meeting-noter might be problematic. Click here for more details.

@@ -105,12 +105,17 @@ class MP3Encoder:
105
105
 
106
106
  def finalize(self) -> bytes:
107
107
  """Finalize encoding."""
108
- if self._process is not None and self._process.stdin is not None:
108
+ if self._process is not None:
109
109
  try:
110
- self._process.stdin.close()
110
+ if self._process.stdin is not None:
111
+ self._process.stdin.close()
111
112
  except Exception:
112
113
  pass
113
- self._process.wait()
114
+ try:
115
+ self._process.wait(timeout=5.0)
116
+ except subprocess.TimeoutExpired:
117
+ self._process.kill()
118
+ self._process.wait()
114
119
  self._process = None
115
120
  return b""
116
121
 
meeting_noter/cli.py CHANGED
@@ -131,8 +131,9 @@ def cli(ctx):
131
131
 
132
132
  @cli.command()
133
133
  @click.argument("name", required=False)
134
+ @click.option("--live", "-l", is_flag=True, help="Show live transcription in terminal")
134
135
  @require_setup
135
- def start(name: Optional[str]):
136
+ def start(name: Optional[str], live: bool):
136
137
  """Start an interactive foreground recording session.
137
138
 
138
139
  NAME is the meeting name (optional). If not provided, uses a timestamp
@@ -141,11 +142,14 @@ def start(name: Optional[str]):
141
142
  Examples:
142
143
  meeting-noter start # Uses timestamp name
143
144
  meeting-noter start "Weekly Standup" # Uses custom name
145
+ meeting-noter start "Meeting" --live # With live transcription
144
146
 
145
147
  Press Ctrl+C to stop recording. The recording will be automatically
146
148
  transcribed if auto_transcribe is enabled in settings.
147
149
  """
148
150
  from meeting_noter.daemon import run_foreground_capture
151
+ import threading
152
+ import time
149
153
 
150
154
  config = get_config()
151
155
  output_dir = config.recordings_dir
@@ -154,14 +158,59 @@ def start(name: Optional[str]):
154
158
  # Use default timestamp name if not provided
155
159
  meeting_name = name if name else generate_meeting_name()
156
160
 
157
- run_foreground_capture(
158
- output_dir=output_dir,
159
- meeting_name=meeting_name,
160
- auto_transcribe=config.auto_transcribe,
161
- whisper_model=config.whisper_model,
162
- transcripts_dir=config.transcripts_dir,
163
- silence_timeout_minutes=config.silence_timeout,
164
- )
161
+ # Live transcription display thread
162
+ stop_live_display = threading.Event()
163
+
164
+ def display_live_transcript():
165
+ """Background thread to display live transcription."""
166
+ live_dir = output_dir / "live"
167
+ last_content = ""
168
+
169
+ # Wait for live file to appear
170
+ while not stop_live_display.is_set():
171
+ live_files = list(live_dir.glob("*.live.txt")) if live_dir.exists() else []
172
+ if live_files:
173
+ live_file = max(live_files, key=lambda p: p.stat().st_mtime)
174
+ break
175
+ time.sleep(0.5)
176
+ else:
177
+ return
178
+
179
+ # Tail the file
180
+ while not stop_live_display.is_set():
181
+ try:
182
+ content = live_file.read_text()
183
+ if len(content) > len(last_content):
184
+ new_content = content[len(last_content):]
185
+ for line in new_content.splitlines():
186
+ # Only show timestamp lines (transcriptions)
187
+ if line.strip() and line.startswith("["):
188
+ click.echo(click.style(line, fg="cyan"))
189
+ last_content = content
190
+ except Exception:
191
+ pass
192
+ time.sleep(0.5)
193
+
194
+ # Start live display thread if requested
195
+ live_thread = None
196
+ if live:
197
+ click.echo(click.style("Live transcription enabled", fg="cyan"))
198
+ live_thread = threading.Thread(target=display_live_transcript, daemon=True)
199
+ live_thread.start()
200
+
201
+ try:
202
+ run_foreground_capture(
203
+ output_dir=output_dir,
204
+ meeting_name=meeting_name,
205
+ auto_transcribe=config.auto_transcribe,
206
+ whisper_model=config.whisper_model,
207
+ transcripts_dir=config.transcripts_dir,
208
+ silence_timeout_minutes=config.silence_timeout,
209
+ )
210
+ finally:
211
+ stop_live_display.set()
212
+ if live_thread:
213
+ live_thread.join(timeout=1.0)
165
214
 
166
215
 
167
216
  @cli.command(hidden=True) # Internal command used by watcher
@@ -355,7 +404,13 @@ def logs(follow: bool, lines: int):
355
404
  )
356
405
  @require_setup
357
406
  def list_recordings(output_dir: Optional[str], limit: int):
358
- """List recent meeting recordings."""
407
+ """List recent meeting recordings.
408
+
409
+ \b
410
+ Examples:
411
+ meeting-noter list # Show last 10 recordings
412
+ meeting-noter list -n 20 # Show last 20 recordings
413
+ """
359
414
  from meeting_noter.output.writer import list_recordings as _list_recordings
360
415
 
361
416
  config = get_config()
@@ -377,28 +432,116 @@ def list_recordings(output_dir: Optional[str], limit: int):
377
432
  default=None,
378
433
  help="Whisper model size (overrides config)",
379
434
  )
380
- @click.option(
381
- "--live", "-l",
382
- is_flag=True,
383
- help="Real-time transcription of current recording",
384
- )
385
435
  @require_setup
386
- def transcribe(file: Optional[str], output_dir: Optional[str], model: Optional[str], live: bool):
436
+ def transcribe(file: Optional[str], output_dir: Optional[str], model: Optional[str]):
387
437
  """Transcribe a meeting recording.
388
438
 
389
- If no FILE is specified, transcribes the most recent recording.
390
- Use --live for real-time transcription of an ongoing meeting.
439
+ \b
440
+ Examples:
441
+ meeting-noter transcribe # Transcribe latest recording
442
+ meeting-noter transcribe recording.mp3 # Transcribe specific file
443
+ meeting-noter transcribe -m base.en # Use larger model for accuracy
391
444
  """
392
- from meeting_noter.transcription.engine import transcribe_file, transcribe_live
445
+ from meeting_noter.transcription.engine import transcribe_file
393
446
 
394
447
  config = get_config()
395
448
  output_path = Path(output_dir) if output_dir else config.recordings_dir
396
449
  whisper_model = model or config.whisper_model
397
450
 
398
- if live:
399
- transcribe_live(output_path, whisper_model)
400
- else:
401
- transcribe_file(file, output_path, whisper_model, config.transcripts_dir)
451
+ transcribe_file(file, output_path, whisper_model, config.transcripts_dir)
452
+
453
+
454
+ @cli.command()
455
+ @require_setup
456
+ def live():
457
+ """Show live transcription of an active recording.
458
+
459
+ Displays the real-time transcript as it's being generated.
460
+ Use in a separate terminal while recording with 'meeting-noter start'.
461
+
462
+ \b
463
+ Examples:
464
+ # Terminal 1: Start recording
465
+ meeting-noter start "Team Meeting"
466
+
467
+ # Terminal 2: Watch live transcript
468
+ meeting-noter live
469
+
470
+ Or use 'meeting-noter start "name" --live' to see both in one terminal.
471
+ """
472
+ import time
473
+
474
+ config = get_config()
475
+ live_dir = config.recordings_dir / "live"
476
+
477
+ # Find the most recent .live.txt file in the live/ subfolder
478
+ if not live_dir.exists():
479
+ click.echo(click.style("No live transcript found.", fg="yellow"))
480
+ click.echo("Start a recording with: meeting-noter start")
481
+ return
482
+
483
+ live_files = sorted(
484
+ live_dir.glob("*.live.txt"),
485
+ key=lambda p: p.stat().st_mtime,
486
+ reverse=True,
487
+ )
488
+
489
+ if not live_files:
490
+ click.echo(click.style("No live transcript found.", fg="yellow"))
491
+ click.echo("Start a recording with: meeting-noter start")
492
+ return
493
+
494
+ live_file = live_files[0]
495
+
496
+ # Check if file is actively being written (modified in last 30 seconds)
497
+ file_age = time.time() - live_file.stat().st_mtime
498
+ if file_age > 30:
499
+ click.echo(click.style("No active recording found.", fg="yellow"))
500
+ click.echo(f"Most recent transcript ({live_file.name}) is {int(file_age)}s old.")
501
+ click.echo("Start a recording with: meeting-noter start")
502
+ return
503
+
504
+ click.echo(click.style("Live Transcription", fg="cyan", bold=True))
505
+ click.echo(f"Source: {live_file.name.replace('.live.txt', '.mp3')}")
506
+ click.echo("Press Ctrl+C to stop watching.\n")
507
+ click.echo("-" * 40)
508
+
509
+ # Tail the file
510
+ try:
511
+ last_content = ""
512
+ no_update_count = 0
513
+
514
+ while True:
515
+ try:
516
+ with open(live_file, "r") as f:
517
+ content = f.read()
518
+
519
+ # Print only new content
520
+ if len(content) > len(last_content):
521
+ new_content = content[len(last_content):]
522
+ # Print line by line for better formatting
523
+ for line in new_content.splitlines():
524
+ if line.strip():
525
+ click.echo(line)
526
+ last_content = content
527
+ no_update_count = 0
528
+ else:
529
+ no_update_count += 1
530
+
531
+ # Check if file hasn't been updated for 30+ seconds (recording likely ended)
532
+ file_age = time.time() - live_file.stat().st_mtime
533
+ if file_age > 30 and no_update_count > 5:
534
+ click.echo("\n" + click.style("Recording ended.", fg="yellow"))
535
+ break
536
+
537
+ except FileNotFoundError:
538
+ click.echo("\n" + click.style("Live transcript file removed.", fg="yellow"))
539
+ break
540
+
541
+ time.sleep(1)
542
+
543
+ except KeyboardInterrupt:
544
+ click.echo("\n" + click.style("Stopped watching.", fg="cyan"))
402
545
 
403
546
 
404
547
  @cli.command()
@@ -553,7 +696,7 @@ def config(key: Optional[str], value: Optional[str]):
553
696
  WATCHER_PID_FILE = Path.home() / ".meeting-noter-watcher.pid"
554
697
 
555
698
 
556
- @cli.command()
699
+ @cli.command(hidden=True)
557
700
  @click.option(
558
701
  "--foreground", "-f",
559
702
  is_flag=True,
@@ -665,7 +808,7 @@ def _run_watcher_loop():
665
808
  stop_daemon(DEFAULT_PID_FILE)
666
809
 
667
810
 
668
- @cli.command()
811
+ @cli.command(hidden=True)
669
812
  @require_setup
670
813
  def watch():
671
814
  """Watch for meetings interactively (foreground with prompts).
meeting_noter/daemon.py CHANGED
@@ -157,7 +157,11 @@ def run_daemon(
157
157
  remove_pid_file(pid_file)
158
158
 
159
159
 
160
- def _run_capture_loop(output_dir: Path, meeting_name: Optional[str] = None):
160
+ def _run_capture_loop(
161
+ output_dir: Path,
162
+ meeting_name: Optional[str] = None,
163
+ enable_live_transcription: bool = True,
164
+ ):
161
165
  """Main capture loop.
162
166
 
163
167
  Audio imports happen HERE, safely AFTER the fork.
@@ -171,6 +175,16 @@ def _run_capture_loop(output_dir: Path, meeting_name: Optional[str] = None):
171
175
 
172
176
  config = get_config()
173
177
 
178
+ # Live transcription (imported here to avoid loading Whisper before fork)
179
+ live_transcriber = None
180
+ if enable_live_transcription:
181
+ try:
182
+ from meeting_noter.transcription.live_transcription import LiveTranscriber
183
+ LiveTranscriber # Just verify import works, create later
184
+ except ImportError as e:
185
+ print(f"Live transcription not available: {e}")
186
+ enable_live_transcription = False
187
+
174
188
  print(f"Meeting Noter daemon started. Saving to {output_dir}")
175
189
  sys.stdout.flush()
176
190
  if meeting_name:
@@ -253,15 +267,42 @@ def _run_capture_loop(output_dir: Path, meeting_name: Optional[str] = None):
253
267
  print(f"Recording started: {filepath.name}")
254
268
  recording_started = True
255
269
  audio_detected = True
270
+
271
+ # Start live transcription
272
+ if enable_live_transcription:
273
+ try:
274
+ from meeting_noter.transcription.live_transcription import LiveTranscriber
275
+ live_transcriber = LiveTranscriber(
276
+ output_path=filepath,
277
+ sample_rate=sample_rate,
278
+ channels=channels,
279
+ window_seconds=5.0,
280
+ slide_seconds=2.0,
281
+ model_size=config.whisper_model,
282
+ )
283
+ live_transcriber.start()
284
+ print(f"Live transcription: {live_transcriber.live_file_path.name}")
285
+ except Exception as e:
286
+ print(f"Failed to start live transcription: {e}")
287
+ live_transcriber = None
256
288
  else:
257
289
  # Currently recording
258
290
  session.write(audio)
259
291
 
292
+ # Feed audio to live transcriber
293
+ if live_transcriber is not None:
294
+ live_transcriber.write(audio)
295
+
260
296
  if has_audio:
261
297
  audio_detected = True
262
298
 
263
299
  # Check for extended silence (meeting ended)
264
300
  if is_silence and audio_detected:
301
+ # Stop live transcription first
302
+ if live_transcriber is not None:
303
+ live_transcriber.stop()
304
+ live_transcriber = None
305
+
265
306
  filepath, duration = session.stop()
266
307
  if filepath:
267
308
  print(f"Recording saved: {filepath.name} ({duration:.1f}s)")
@@ -275,6 +316,10 @@ def _run_capture_loop(output_dir: Path, meeting_name: Optional[str] = None):
275
316
  finally:
276
317
  capture.stop()
277
318
 
319
+ # Stop live transcription
320
+ if live_transcriber is not None:
321
+ live_transcriber.stop()
322
+
278
323
  # Save any ongoing recording
279
324
  if 'session' in locals() and session.is_active:
280
325
  filepath, duration = session.stop()
@@ -377,6 +422,7 @@ def run_foreground_capture(
377
422
  whisper_model: str = "tiny.en",
378
423
  transcripts_dir: Optional[Path] = None,
379
424
  silence_timeout_minutes: int = 5,
425
+ enable_live_transcription: bool = True,
380
426
  ) -> Optional[Path]:
381
427
  """Run audio capture in foreground with a named meeting.
382
428
 
@@ -390,6 +436,7 @@ def run_foreground_capture(
390
436
  whisper_model: Whisper model to use for transcription
391
437
  transcripts_dir: Directory for transcripts
392
438
  silence_timeout_minutes: Stop after this many minutes of silence
439
+ enable_live_transcription: Whether to enable real-time transcription
393
440
 
394
441
  Returns:
395
442
  Path to the saved recording, or None if recording was too short
@@ -401,6 +448,9 @@ def run_foreground_capture(
401
448
 
402
449
  config = get_config()
403
450
 
451
+ # Initialize live transcriber
452
+ live_transcriber = None
453
+
404
454
  # Check audio device
405
455
  if not check_audio_available():
406
456
  click.echo(click.style("Error: ", fg="red") + "No audio input device found.")
@@ -461,6 +511,27 @@ def run_foreground_capture(
461
511
  filepath = session.start()
462
512
  click.echo(click.style("Recording: ", fg="green") + filepath.name)
463
513
 
514
+ # Start live transcription
515
+ if enable_live_transcription:
516
+ try:
517
+ from meeting_noter.transcription.live_transcription import LiveTranscriber
518
+ live_transcriber = LiveTranscriber(
519
+ output_path=filepath,
520
+ sample_rate=sample_rate,
521
+ channels=channels,
522
+ window_seconds=5.0,
523
+ slide_seconds=2.0,
524
+ model_size=whisper_model,
525
+ )
526
+ live_transcriber.start()
527
+ click.echo(
528
+ click.style("Live transcript: ", fg="cyan") +
529
+ str(live_transcriber.live_file_path)
530
+ )
531
+ except Exception as e:
532
+ click.echo(click.style(f"Live transcription not available: {e}", fg="yellow"))
533
+ live_transcriber = None
534
+
464
535
  while not _stop_event.is_set():
465
536
  audio = capture.get_audio(timeout=0.5)
466
537
  if audio is None:
@@ -472,6 +543,10 @@ def run_foreground_capture(
472
543
 
473
544
  session.write(audio)
474
545
 
546
+ # Feed audio to live transcriber
547
+ if live_transcriber is not None:
548
+ live_transcriber.write(audio)
549
+
475
550
  # Check for extended silence
476
551
  if silence_detector.update(audio):
477
552
  click.echo("\n" + click.style("Stopped: ", fg="yellow") + "silence timeout reached")
@@ -487,6 +562,10 @@ def run_foreground_capture(
487
562
  except Exception as e:
488
563
  click.echo(click.style(f"\nError: {e}", fg="red"))
489
564
  finally:
565
+ # Stop live transcription
566
+ if live_transcriber is not None:
567
+ live_transcriber.stop()
568
+
490
569
  capture.stop()
491
570
 
492
571
  # Save recording
@@ -169,6 +169,9 @@ def is_meeting_app_active() -> Optional[str]:
169
169
 
170
170
  # Browser-based meetings (Google Meet, etc.)
171
171
  if any(browser in owner_lower for browser in ["chrome", "safari", "firefox", "edge", "brave", "arc"]):
172
+ # Skip video streaming sites (YouTube, Vimeo, etc.)
173
+ if any(x in title_lower for x in ["youtube", "vimeo", "twitch", "netflix"]):
174
+ continue
172
175
  if title_lower.startswith("meet -") or "meet.google.com" in title_lower:
173
176
  return "Google Meet"
174
177
  if " meeting" in title_lower and any(x in title_lower for x in ["zoom", "teams", "webex"]):
@@ -0,0 +1,250 @@
1
+ """Live transcription during recording.
2
+
3
+ Buffers audio chunks and transcribes them in a background thread,
4
+ writing segments to a .live.txt file that can be tailed by the CLI.
5
+
6
+ Uses overlapping windows for lower latency: keeps a 5-second context window
7
+ but transcribes every 2 seconds, only outputting new content.
8
+ """
9
+
10
+ from __future__ import annotations
11
+
12
+ import sys
13
+ import numpy as np
14
+ from collections import deque
15
+ from pathlib import Path
16
+ from queue import Queue, Empty
17
+ from threading import Thread, Event
18
+ from typing import Optional
19
+ from datetime import datetime
20
+
21
+
22
+ class LiveTranscriber:
23
+ """Transcribes audio in real-time during recording.
24
+
25
+ Uses overlapping windows approach:
26
+ - Maintains a rolling window of audio (default 5 seconds)
27
+ - Transcribes every `slide_seconds` (default 2 seconds)
28
+ - Only outputs new segments to avoid duplicates
29
+ """
30
+
31
+ def __init__(
32
+ self,
33
+ output_path: Path,
34
+ sample_rate: int = 48000,
35
+ channels: int = 2,
36
+ window_seconds: float = 5.0,
37
+ slide_seconds: float = 2.0,
38
+ model_size: str = "tiny.en",
39
+ ):
40
+ """Initialize the live transcriber.
41
+
42
+ Args:
43
+ output_path: Path to write live transcript (will use .live.txt suffix in live/ subfolder)
44
+ sample_rate: Audio sample rate
45
+ channels: Number of audio channels
46
+ window_seconds: Size of the context window for transcription
47
+ slide_seconds: How often to transcribe (lower = more responsive, higher CPU)
48
+ model_size: Whisper model to use (tiny.en recommended for speed)
49
+ """
50
+ # Put live transcripts in a 'live/' subfolder to keep recordings folder clean
51
+ live_dir = output_path.parent / "live"
52
+ live_dir.mkdir(exist_ok=True)
53
+ self.output_path = live_dir / (output_path.stem + ".live.txt")
54
+ self.sample_rate = sample_rate
55
+ self.channels = channels
56
+ self.window_seconds = window_seconds
57
+ self.slide_seconds = slide_seconds
58
+ self.model_size = model_size
59
+
60
+ self._audio_queue: Queue[np.ndarray] = Queue()
61
+ self._stop_event = Event()
62
+ self._thread: Optional[Thread] = None
63
+ self._model = None
64
+ self._start_time: Optional[datetime] = None
65
+ self._recording_offset = 0.0 # Current position in recording (seconds)
66
+ self._last_output_end = 0.0 # End time of last outputted segment
67
+
68
+ def start(self):
69
+ """Start the live transcription thread."""
70
+ self._stop_event.clear()
71
+ self._start_time = datetime.now()
72
+ self._recording_offset = 0.0
73
+ self._last_output_end = 0.0
74
+
75
+ # Create/clear the output file
76
+ with open(self.output_path, "w") as f:
77
+ f.write(f"Live Transcription - {self._start_time.strftime('%Y-%m-%d %H:%M:%S')}\n")
78
+ f.write("-" * 40 + "\n\n")
79
+
80
+ self._thread = Thread(target=self._transcribe_loop, daemon=True)
81
+ self._thread.start()
82
+
83
+ def write(self, audio: np.ndarray):
84
+ """Add audio chunk to the transcription queue.
85
+
86
+ Args:
87
+ audio: Audio data (float32, -1 to 1)
88
+ """
89
+ if not self._stop_event.is_set():
90
+ self._audio_queue.put(audio.copy())
91
+
92
+ def stop(self):
93
+ """Stop the live transcription thread."""
94
+ self._stop_event.set()
95
+ if self._thread is not None:
96
+ self._thread.join(timeout=10.0)
97
+ self._thread = None
98
+ self._model = None
99
+
100
+ def _load_model(self):
101
+ """Load the Whisper model (lazy loading)."""
102
+ if self._model is not None:
103
+ return
104
+
105
+ try:
106
+ from faster_whisper import WhisperModel
107
+
108
+ # Check for bundled model
109
+ bundled_path = None
110
+ try:
111
+ from meeting_noter_models import get_model_path
112
+ bundled_path = get_model_path()
113
+ if not (bundled_path.exists() and (bundled_path / "model.bin").exists()):
114
+ bundled_path = None
115
+ except ImportError:
116
+ pass
117
+
118
+ if bundled_path and self.model_size == "tiny.en":
119
+ self._model = WhisperModel(
120
+ str(bundled_path),
121
+ device="cpu",
122
+ compute_type="int8",
123
+ )
124
+ else:
125
+ self._model = WhisperModel(
126
+ self.model_size,
127
+ device="cpu",
128
+ compute_type="int8",
129
+ )
130
+ except Exception as e:
131
+ print(f"Failed to load Whisper model: {e}", file=sys.stderr)
132
+ self._model = None
133
+
134
+ def _transcribe_loop(self):
135
+ """Main transcription loop with overlapping windows."""
136
+ # Rolling buffer using deque for efficient sliding
137
+ window_samples = int(self.window_seconds * self.sample_rate)
138
+ slide_samples = int(self.slide_seconds * self.sample_rate)
139
+
140
+ # Buffer holds raw audio samples
141
+ rolling_buffer: deque[float] = deque(maxlen=window_samples)
142
+ samples_since_last_transcribe = 0
143
+
144
+ # Load model on first use
145
+ self._load_model()
146
+ if self._model is None:
147
+ return
148
+
149
+ while not self._stop_event.is_set():
150
+ try:
151
+ # Collect audio chunks
152
+ try:
153
+ chunk = self._audio_queue.get(timeout=0.5)
154
+
155
+ # Add samples to rolling buffer
156
+ for sample in chunk:
157
+ rolling_buffer.append(sample)
158
+
159
+ samples_since_last_transcribe += len(chunk)
160
+ self._recording_offset += len(chunk) / self.sample_rate
161
+
162
+ except Empty:
163
+ if self._stop_event.is_set():
164
+ break
165
+ continue
166
+
167
+ # Transcribe every slide_seconds
168
+ if samples_since_last_transcribe >= slide_samples and len(rolling_buffer) >= slide_samples:
169
+ self._transcribe_window(rolling_buffer)
170
+ samples_since_last_transcribe = 0
171
+
172
+ except Exception as e:
173
+ print(f"Live transcription error: {e}", file=sys.stderr)
174
+
175
+ # Final transcription on stop
176
+ if len(rolling_buffer) > 0:
177
+ self._transcribe_window(rolling_buffer)
178
+
179
+ def _transcribe_window(self, rolling_buffer: deque):
180
+ """Transcribe the current window and output new segments."""
181
+ if not rolling_buffer or self._model is None:
182
+ return
183
+
184
+ try:
185
+ # Convert deque to numpy array (ensure float32 for Whisper)
186
+ audio = np.array(list(rolling_buffer), dtype=np.float32)
187
+
188
+ # Convert stereo to mono if needed
189
+ if self.channels == 2 and len(audio) % 2 == 0:
190
+ audio = audio.reshape(-1, 2).mean(axis=1).astype(np.float32)
191
+
192
+ # Resample to 16kHz if needed (Whisper expects 16kHz)
193
+ if self.sample_rate != 16000:
194
+ audio = self._resample(audio, self.sample_rate, 16000).astype(np.float32)
195
+
196
+ # Calculate window timing
197
+ window_duration = len(rolling_buffer) / self.sample_rate
198
+ window_start = self._recording_offset - window_duration
199
+
200
+ # Transcribe using faster-whisper
201
+ segments, _ = self._model.transcribe(
202
+ audio,
203
+ beam_size=1, # Fastest
204
+ vad_filter=True,
205
+ vad_parameters=dict(min_silence_duration_ms=200),
206
+ )
207
+
208
+ # Write only NEW segments to file
209
+ with open(self.output_path, "a") as f:
210
+ for segment in segments:
211
+ # Calculate absolute timestamp
212
+ abs_start = window_start + segment.start
213
+ abs_end = window_start + segment.end
214
+
215
+ # Only output if this segment is new (starts after last output)
216
+ if abs_start >= self._last_output_end - 0.5: # 0.5s tolerance for overlap
217
+ text = segment.text.strip()
218
+ if text:
219
+ timestamp = self._format_timestamp(abs_start)
220
+ f.write(f"{timestamp} {text}\n")
221
+ f.flush()
222
+ self._last_output_end = abs_end
223
+
224
+ except Exception as e:
225
+ print(f"Transcription error: {e}", file=sys.stderr)
226
+
227
+ @staticmethod
228
+ def _resample(audio: np.ndarray, orig_sr: int, target_sr: int) -> np.ndarray:
229
+ """Simple resampling using linear interpolation."""
230
+ if orig_sr == target_sr:
231
+ return audio
232
+
233
+ duration = len(audio) / orig_sr
234
+ target_length = int(duration * target_sr)
235
+
236
+ # Use numpy interpolation (returns float64, so cast back)
237
+ indices = np.linspace(0, len(audio) - 1, target_length)
238
+ return np.interp(indices, np.arange(len(audio)), audio).astype(np.float32)
239
+
240
+ @staticmethod
241
+ def _format_timestamp(seconds: float) -> str:
242
+ """Format seconds as [MM:SS]."""
243
+ minutes = int(seconds // 60)
244
+ secs = int(seconds % 60)
245
+ return f"[{minutes:02d}:{secs:02d}]"
246
+
247
+ @property
248
+ def live_file_path(self) -> Path:
249
+ """Get the path to the live transcript file."""
250
+ return self.output_path
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: meeting-noter
3
- Version: 0.7.0
3
+ Version: 1.0.0
4
4
  Summary: Offline meeting transcription for macOS with automatic meeting detection
5
5
  Author: Victor
6
6
  License: MIT
@@ -40,6 +40,7 @@ Requires-Dist: meeting-noter-models>=0.1.0; extra == "offline"
40
40
  Provides-Extra: dev
41
41
  Requires-Dist: pytest>=7.0; extra == "dev"
42
42
  Requires-Dist: pytest-cov; extra == "dev"
43
+ Requires-Dist: pytest-mock; extra == "dev"
43
44
  Requires-Dist: black; extra == "dev"
44
45
  Requires-Dist: ruff; extra == "dev"
45
46
  Requires-Dist: mypy; extra == "dev"
@@ -1,14 +1,14 @@
1
1
  meeting_noter/__init__.py,sha256=bLOErRC3sfnQ4a4RyZUzUljZEikXy7zOiYYUz5GytPg,103
2
2
  meeting_noter/__main__.py,sha256=6sSOqH1o3jvgvkVzsVKmF6-xVGcUAbNVQkRl2CrygdE,120
3
- meeting_noter/cli.py,sha256=OAlLc7CEgov3akDo89I0JUMl9wCXs81x3QO2mBcqLvE,27769
3
+ meeting_noter/cli.py,sha256=brLJ_2kuqEi5wq868hFaMGtKPSkyBpD-OS1u27rJb1k,32859
4
4
  meeting_noter/config.py,sha256=41LFBNp5o0IojYS5Hf0FJVIr7GNn7B5O1TJDE8SQkkk,5977
5
- meeting_noter/daemon.py,sha256=BYMc-sAz8zo9ZFMRr_Th8iw_l7rc71ldEf1mwsJJgaw,16502
5
+ meeting_noter/daemon.py,sha256=u9VrYe94o3lxabuIS9MDVPHSH7MqKqzTqGTuA7TNAIc,19767
6
6
  meeting_noter/meeting_detector.py,sha256=St0qoMkvUERP4BaxnXO1M6fZDJpWqBf9In7z2SgWcWg,10564
7
7
  meeting_noter/menubar.py,sha256=Gn6p8y5jA_HCWf1T3ademxH-vndpONHkf9vUlKs6XEo,14379
8
- meeting_noter/mic_monitor.py,sha256=zgcNypn29gWTyslalu5-JFseG47DVHEtcHCGiEUugAs,16021
8
+ meeting_noter/mic_monitor.py,sha256=P8vF4qaZcGrEzzJyVos78Vuf38NXHGNRREDsD-HyBHc,16211
9
9
  meeting_noter/audio/__init__.py,sha256=O7PU8CxHSHxMeHbc9Jdwt9kePLQzsPh81GQU7VHCtBY,44
10
10
  meeting_noter/audio/capture.py,sha256=fDrT5oXfva8vdFlht9cv60NviKbksw2QeJ8eOtI19uE,6469
11
- meeting_noter/audio/encoder.py,sha256=6UgEYLFACSQEIx2nhH1Qq-cBh3qPJziMGkrm39k6Nz8,6401
11
+ meeting_noter/audio/encoder.py,sha256=OBsgUmlZPz-YZQZ7Rp8MAlMRaQxTsccjuTgCtvRebmc,6573
12
12
  meeting_noter/audio/system_audio.py,sha256=jbHGjNCerI19weXap0a90Ik17lVTCT1hCEgRKYke-p8,13016
13
13
  meeting_noter/gui/__init__.py,sha256=z5GxxaeXyjqyEa9ox0dQxuL5u_BART0bi7cI6rfntEI,103
14
14
  meeting_noter/gui/__main__.py,sha256=A2HWdYod0bTgjQQIi21O7XpmgxLH36e_X0aygEUZLls,146
@@ -32,8 +32,9 @@ meeting_noter/resources/icon_512.png,sha256=o7X3ngYcppcIAAk9AcfPx94MUmrsPRp0qBTp
32
32
  meeting_noter/resources/icon_64.png,sha256=TqG7Awx3kK8YdiX1e_z1odZonosZyQI2trlkNZCzUoI,607
33
33
  meeting_noter/transcription/__init__.py,sha256=7GY9diP06DzFyoli41wddbrPv5bVDzH35bmnWlIJev4,29
34
34
  meeting_noter/transcription/engine.py,sha256=G9NcSS6Q-UhW7PlQ0E85hQXn6BWao64nIvyw4NR2yxI,7208
35
- meeting_noter-0.7.0.dist-info/METADATA,sha256=8V-ZO5eL8tFm3PauQU9dCZVw0ZbkyCWl6xklfcH3a-c,6952
36
- meeting_noter-0.7.0.dist-info/WHEEL,sha256=wUyA8OaulRlbfwMtmQsvNngGrxQHAvkKcvRmdizlJi0,92
37
- meeting_noter-0.7.0.dist-info/entry_points.txt,sha256=rKNhzjSF5-e3bLRr8LVe22FeiwcacXabCvNpoEXfu4I,56
38
- meeting_noter-0.7.0.dist-info/top_level.txt,sha256=9Tuq04_0SXM0OXOHVbOHkHkB5tG3fqkrMrfzCMpbLpY,14
39
- meeting_noter-0.7.0.dist-info/RECORD,,
35
+ meeting_noter/transcription/live_transcription.py,sha256=AslB1T1_gxu7eSp7xc79_2SdfGrNJq7L_8bA1t6YoU4,9277
36
+ meeting_noter-1.0.0.dist-info/METADATA,sha256=m7Pi8_-haGOHX0DbA7YXTc1KMpMWAxLHHgbE3kB3-FM,6995
37
+ meeting_noter-1.0.0.dist-info/WHEEL,sha256=wUyA8OaulRlbfwMtmQsvNngGrxQHAvkKcvRmdizlJi0,92
38
+ meeting_noter-1.0.0.dist-info/entry_points.txt,sha256=osZoOmm-UBPCJ4b6DGH6JOAm7mofM2fK06eK6blplmg,83
39
+ meeting_noter-1.0.0.dist-info/top_level.txt,sha256=9Tuq04_0SXM0OXOHVbOHkHkB5tG3fqkrMrfzCMpbLpY,14
40
+ meeting_noter-1.0.0.dist-info/RECORD,,
@@ -1,2 +1,3 @@
1
1
  [console_scripts]
2
2
  meeting-noter = meeting_noter.cli:cli
3
+ mn = meeting_noter.cli:cli