Compare commits
3 Commits
06fd3efd1f
...
ffd451b404
Author | SHA1 | Date | |
---|---|---|---|
ffd451b404 | |||
5c64677e79 | |||
128db0f733 |
17
.cargo/config.toml
Normal file
17
.cargo/config.toml
Normal file
@@ -0,0 +1,17 @@
|
||||
# SPDX-License-Identifier: MIT
|
||||
|
||||
[build]
|
||||
# Make target-dir consistent across workspace for better cache reuse.
|
||||
target-dir = "target"
|
||||
|
||||
[profile.dev]
|
||||
opt-level = 1
|
||||
debug = true
|
||||
incremental = true
|
||||
|
||||
[profile.release]
|
||||
# Reasonable defaults for CLI apps/libraries
|
||||
lto = "thin"
|
||||
codegen-units = 1
|
||||
strip = "debuginfo"
|
||||
opt-level = 3
|
@@ -1,32 +1,26 @@
|
||||
# Contributing to PolyScribe
|
||||
# Contributing
|
||||
|
||||
Thanks for your interest in contributing! This guide explains the workflow and the checklist to follow before opening a Pull Request.
|
||||
Thank you for your interest in contributing!
|
||||
|
||||
Workflow (fork → branch → PR)
|
||||
1) Fork the repository to your account.
|
||||
2) Create a feature branch:
|
||||
- git checkout -b feat/short-description
|
||||
3) Make changes with focused commits and good messages.
|
||||
4) Run the checklist below.
|
||||
5) Push and open a Pull Request against the main repository.
|
||||
Development setup
|
||||
- Install Rust via rustup.
|
||||
- Ensure ffmpeg is installed and available on PATH.
|
||||
- For GPU builds, install the appropriate runtime (CUDA/ROCm/Vulkan) and enable the matching features.
|
||||
|
||||
Developer checklist (before opening a PR)
|
||||
- Build:
|
||||
- cargo build (preferably without warnings)
|
||||
- Tests:
|
||||
- cargo test (all tests pass)
|
||||
- Lints:
|
||||
- cargo clippy --all-targets -- -D warnings (fix warnings)
|
||||
- Documentation:
|
||||
- Update README/docs for user-visible changes
|
||||
- Update CHANGELOG.md if applicable
|
||||
- Tests for changes:
|
||||
- Add or update tests for bug fixes and new features where reasonable
|
||||
Coding guidelines
|
||||
- Prefer small, focused changes.
|
||||
- Add tests where reasonable.
|
||||
- Keep user-facing changes documented in README/docs.
|
||||
- Run clippy and fix warnings.
|
||||
|
||||
Local development tips
|
||||
- Use `cargo run -- <args>` during development.
|
||||
- For faster feedback, keep examples in the examples/ folder handy.
|
||||
- Keep functions small and focused; prefer clear error messages with context.
|
||||
CI checklist
|
||||
- Build: cargo build --all-targets --locked
|
||||
- Tests: cargo test --all --locked
|
||||
- Lints: cargo clippy --all-targets -- -D warnings
|
||||
- Optional: smoke-run examples inline (from README):
|
||||
- ./target/release/polyscribe --update-models --no-interaction -q
|
||||
- ./target/release/polyscribe -o output samples/podcast_clip.mp3
|
||||
|
||||
Code of conduct
|
||||
- Be respectful and constructive. Assume good intent.
|
||||
Notes
|
||||
- For GPU features, use --features gpu-cuda|gpu-hip|gpu-vulkan as needed in your local runs.
|
||||
- For docs-only changes, please still ensure the project builds.
|
||||
|
391
Cargo.lock
generated
391
Cargo.lock
generated
@@ -144,15 +144,6 @@ version = "2.9.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1b8e56985ec62d17e9c1001dc89c88ecd7dc08e47eba5ec7c29c7b5eeecde967"
|
||||
|
||||
[[package]]
|
||||
name = "block-buffer"
|
||||
version = "0.10.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "3078c7629b62d3f0439517fa394996acacc5cbc91c5a20d8c658e77abd503a71"
|
||||
dependencies = [
|
||||
"generic-array",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "bumpalo"
|
||||
version = "3.19.0"
|
||||
@@ -273,20 +264,6 @@ dependencies = [
|
||||
"roff",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "cliclack"
|
||||
version = "0.3.6"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "57c420bdc04c123a2df04d9c5a07289195f00007af6e45ab18f55e56dc7e04b8"
|
||||
dependencies = [
|
||||
"console",
|
||||
"indicatif",
|
||||
"once_cell",
|
||||
"strsim",
|
||||
"textwrap",
|
||||
"zeroize",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "cmake"
|
||||
version = "0.1.54"
|
||||
@@ -321,35 +298,6 @@ version = "0.8.7"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "773648b94d0e5d620f64f280777445740e61fe701025087ec8b57f45c791888b"
|
||||
|
||||
[[package]]
|
||||
name = "cpufeatures"
|
||||
version = "0.2.17"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "59ed5838eebb26a2bb2e58f6d5b5316989ae9d08bab10e0e6d103e656d1b0280"
|
||||
dependencies = [
|
||||
"libc",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "crypto-common"
|
||||
version = "0.1.6"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1bfb12502f3fc46cca1bb51ac28df9d618d813cdc3d2f25b9fe775a34af26bb3"
|
||||
dependencies = [
|
||||
"generic-array",
|
||||
"typenum",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "digest"
|
||||
version = "0.10.7"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "9ed9a281f7bc9b7576e61468ba615a66a5c8cfdff42420a70aa82701a3b1e292"
|
||||
dependencies = [
|
||||
"block-buffer",
|
||||
"crypto-common",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "directories"
|
||||
version = "5.0.1"
|
||||
@@ -405,16 +353,6 @@ version = "1.3.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "42703706b716c37f96a77aea830392ad231f44c9e9a67872fa5548707e11b11c"
|
||||
|
||||
[[package]]
|
||||
name = "generic-array"
|
||||
version = "0.14.7"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "85649ca51fd72272d7821adaf274ad91c288277713d9c18820d8499a7ff69e9a"
|
||||
dependencies = [
|
||||
"typenum",
|
||||
"version_check",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "getrandom"
|
||||
version = "0.2.16"
|
||||
@@ -548,6 +486,12 @@ dependencies = [
|
||||
"wasm-bindgen",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "lazy_static"
|
||||
version = "1.5.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "bbd2bcb4c963f2ddae06a2efc7e9f3591312473c50c6685e1f298068316e66fe"
|
||||
|
||||
[[package]]
|
||||
name = "libc"
|
||||
version = "0.2.175"
|
||||
@@ -580,22 +524,21 @@ version = "0.4.15"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "d26c52dbd32dccf2d10cac7725f8eae5296885fb5703b261f7d0a0739ec807ab"
|
||||
|
||||
[[package]]
|
||||
name = "lock_api"
|
||||
version = "0.4.13"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "96936507f153605bddfcda068dd804796c84324ed2510809e5b2a624c81da765"
|
||||
dependencies = [
|
||||
"autocfg",
|
||||
"scopeguard",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "log"
|
||||
version = "0.4.27"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "13dc2df351e3202783a1fe0d44375f7295ffb4049267b0f3018346dc122a1d94"
|
||||
|
||||
[[package]]
|
||||
name = "matchers"
|
||||
version = "0.1.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "8263075bb86c5a1b1427b5ae862e8889656f126e9f77c484496e8b47cf5c5558"
|
||||
dependencies = [
|
||||
"regex-automata 0.1.10",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "memchr"
|
||||
version = "2.7.5"
|
||||
@@ -638,6 +581,16 @@ dependencies = [
|
||||
"minimal-lexical",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "nu-ansi-term"
|
||||
version = "0.46.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "77a8165726e8236064dbb45459242600304b42a5ea24ee2948e18e023bf7ba84"
|
||||
dependencies = [
|
||||
"overload",
|
||||
"winapi",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "num-traits"
|
||||
version = "0.2.19"
|
||||
@@ -681,27 +634,10 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "04744f49eae99ab78e0d5c0b603ab218f515ea8cfe5a456d7629ad883a3b6e7d"
|
||||
|
||||
[[package]]
|
||||
name = "parking_lot"
|
||||
version = "0.12.4"
|
||||
name = "overload"
|
||||
version = "0.1.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "70d58bf43669b5795d1576d0641cfb6fbb2057bf629506267a92807158584a13"
|
||||
dependencies = [
|
||||
"lock_api",
|
||||
"parking_lot_core",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "parking_lot_core"
|
||||
version = "0.9.11"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "bc838d2a56b5b1a6c25f55575dfc605fabb63bb2365f6c2353ef9159aa69e4a5"
|
||||
dependencies = [
|
||||
"cfg-if",
|
||||
"libc",
|
||||
"redox_syscall",
|
||||
"smallvec",
|
||||
"windows-targets 0.52.6",
|
||||
]
|
||||
checksum = "b15813163c1d831bf4a13c3610c05c0d03b39feb07f7e09fa234dac9b15aaf39"
|
||||
|
||||
[[package]]
|
||||
name = "pin-project-lite"
|
||||
@@ -710,77 +646,59 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "3b3cff922bd51709b605d9ead9aa71031d81447142d828eb4a6eba76fe619f9b"
|
||||
|
||||
[[package]]
|
||||
name = "polyscribe"
|
||||
name = "polyscribe-cli"
|
||||
version = "0.1.0"
|
||||
dependencies = [
|
||||
"anyhow",
|
||||
"clap",
|
||||
"clap_complete",
|
||||
"clap_mangen",
|
||||
"directories",
|
||||
"indicatif",
|
||||
"polyscribe-core",
|
||||
"polyscribe-host",
|
||||
"polyscribe-protocol",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"tokio",
|
||||
"tracing",
|
||||
"tracing-subscriber",
|
||||
"which",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "polyscribe-core"
|
||||
version = "0.1.0"
|
||||
dependencies = [
|
||||
"anyhow",
|
||||
"chrono",
|
||||
"cliclack",
|
||||
"directories",
|
||||
"indicatif",
|
||||
"libc",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"sha2",
|
||||
"thiserror",
|
||||
"toml",
|
||||
"whisper-rs",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "polyscribe-cli"
|
||||
version = "0.1.0"
|
||||
dependencies = [
|
||||
"anyhow",
|
||||
"chrono",
|
||||
"clap",
|
||||
"clap_complete",
|
||||
"clap_mangen",
|
||||
"cliclack",
|
||||
"indicatif",
|
||||
"polyscribe",
|
||||
"polyscribe-host",
|
||||
"polyscribe-protocol",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"toml",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "polyscribe-host"
|
||||
version = "0.1.0"
|
||||
dependencies = [
|
||||
"anyhow",
|
||||
"cliclack",
|
||||
"directories",
|
||||
"polyscribe",
|
||||
"polyscribe-protocol",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"thiserror",
|
||||
"tokio",
|
||||
"which",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "polyscribe-plugin-tubescribe"
|
||||
version = "0.1.0"
|
||||
dependencies = [
|
||||
"anyhow",
|
||||
"clap",
|
||||
"polyscribe-protocol",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"tokio",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "polyscribe-protocol"
|
||||
version = "0.1.0"
|
||||
dependencies = [
|
||||
"serde",
|
||||
"serde_json",
|
||||
"thiserror",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -817,15 +735,6 @@ dependencies = [
|
||||
"proc-macro2",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "redox_syscall"
|
||||
version = "0.5.17"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "5407465600fb0548f1442edf71dd20683c6ed326200ace4b1ef0763521bb3b77"
|
||||
dependencies = [
|
||||
"bitflags",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "redox_users"
|
||||
version = "0.4.6"
|
||||
@@ -845,8 +754,17 @@ checksum = "b544ef1b4eac5dc2db33ea63606ae9ffcfac26c1416a2806ae0bf5f56b201191"
|
||||
dependencies = [
|
||||
"aho-corasick",
|
||||
"memchr",
|
||||
"regex-automata",
|
||||
"regex-syntax",
|
||||
"regex-automata 0.4.9",
|
||||
"regex-syntax 0.8.5",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "regex-automata"
|
||||
version = "0.1.10"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "6c230d73fb8d8c1b9c0b3135c5142a8acee3a0558fb8db5cf1cb65f8d7862132"
|
||||
dependencies = [
|
||||
"regex-syntax 0.6.29",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -857,9 +775,15 @@ checksum = "809e8dc61f6de73b46c85f4c96486310fe304c434cfa43669d7b40f711150908"
|
||||
dependencies = [
|
||||
"aho-corasick",
|
||||
"memchr",
|
||||
"regex-syntax",
|
||||
"regex-syntax 0.8.5",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "regex-syntax"
|
||||
version = "0.6.29"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f162c6dd7b008981e4d40210aca20b4bd0f9b60ca9271061b07f78537722f2e1"
|
||||
|
||||
[[package]]
|
||||
name = "regex-syntax"
|
||||
version = "0.8.5"
|
||||
@@ -909,12 +833,6 @@ version = "1.0.20"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "28d3b2b1366ec20994f1fd18c3c594f05c5dd4bc44d8bb0c1c632c8d6829481f"
|
||||
|
||||
[[package]]
|
||||
name = "scopeguard"
|
||||
version = "1.2.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "94143f37725109f92c262ed2cf5e59bce7498c01bcc1502d7b9afe439a4e9f49"
|
||||
|
||||
[[package]]
|
||||
name = "serde"
|
||||
version = "1.0.219"
|
||||
@@ -957,14 +875,12 @@ dependencies = [
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "sha2"
|
||||
version = "0.10.9"
|
||||
name = "sharded-slab"
|
||||
version = "0.1.7"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "a7507d819769d01a365ab707794a4084392c824f54a7a6a7862f8c3d0892b283"
|
||||
checksum = "f40ca3c46823713e0d4209592e8d6e826aa57e928f09752619fc696c499637f6"
|
||||
dependencies = [
|
||||
"cfg-if",
|
||||
"cpufeatures",
|
||||
"digest",
|
||||
"lazy_static",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -994,22 +910,6 @@ version = "1.15.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "67b1b7a3b5fe4f1376887184045fcf45c69e92af734b7aaddc05fb777b6fbd03"
|
||||
|
||||
[[package]]
|
||||
name = "smawk"
|
||||
version = "0.3.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "b7c388c1b5e93756d0c740965c41e8822f866621d41acbdf6336a6a168f8840c"
|
||||
|
||||
[[package]]
|
||||
name = "socket2"
|
||||
version = "0.6.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "233504af464074f9d066d7b5416c5f9b894a5862a6506e306f7b816cdd6f1807"
|
||||
dependencies = [
|
||||
"libc",
|
||||
"windows-sys 0.59.0",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "strsim"
|
||||
version = "0.11.1"
|
||||
@@ -1027,17 +927,6 @@ dependencies = [
|
||||
"unicode-ident",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "textwrap"
|
||||
version = "0.16.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "c13547615a44dc9c452a8a534638acdf07120d4b6847c8178705da06306a3057"
|
||||
dependencies = [
|
||||
"smawk",
|
||||
"unicode-linebreak",
|
||||
"unicode-width",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "thiserror"
|
||||
version = "1.0.69"
|
||||
@@ -1058,6 +947,15 @@ dependencies = [
|
||||
"syn",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "thread_local"
|
||||
version = "1.1.9"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f60246a4944f24f6e018aa17cdeffb7818b76356965d03b07d6a9886e8962185"
|
||||
dependencies = [
|
||||
"cfg-if",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tokio"
|
||||
version = "1.47.1"
|
||||
@@ -1069,11 +967,9 @@ dependencies = [
|
||||
"io-uring",
|
||||
"libc",
|
||||
"mio",
|
||||
"parking_lot",
|
||||
"pin-project-lite",
|
||||
"signal-hook-registry",
|
||||
"slab",
|
||||
"socket2",
|
||||
"tokio-macros",
|
||||
"windows-sys 0.59.0",
|
||||
]
|
||||
@@ -1131,10 +1027,65 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "5d99f8c9a7727884afe522e9bd5edbfc91a3312b36a77b5fb8926e4c31a41801"
|
||||
|
||||
[[package]]
|
||||
name = "typenum"
|
||||
version = "1.18.0"
|
||||
name = "tracing"
|
||||
version = "0.1.41"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1dccffe3ce07af9386bfd29e80c0ab1a8205a2fc34e4bcd40364df902cfa8f3f"
|
||||
checksum = "784e0ac535deb450455cbfa28a6f0df145ea1bb7ae51b821cf5e7927fdcfbdd0"
|
||||
dependencies = [
|
||||
"pin-project-lite",
|
||||
"tracing-attributes",
|
||||
"tracing-core",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tracing-attributes"
|
||||
version = "0.1.30"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "81383ab64e72a7a8b8e13130c49e3dab29def6d0c7d76a03087b3cf71c5c6903"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tracing-core"
|
||||
version = "0.1.34"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "b9d12581f227e93f094d3af2ae690a574abb8a2b9b7a96e7cfe9647b2b617678"
|
||||
dependencies = [
|
||||
"once_cell",
|
||||
"valuable",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tracing-log"
|
||||
version = "0.2.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "ee855f1f400bd0e5c02d150ae5de3840039a3f54b025156404e34c23c03f47c3"
|
||||
dependencies = [
|
||||
"log",
|
||||
"once_cell",
|
||||
"tracing-core",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tracing-subscriber"
|
||||
version = "0.3.19"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "e8189decb5ac0fa7bc8b96b7cb9b2701d60d48805aca84a238004d665fcc4008"
|
||||
dependencies = [
|
||||
"matchers",
|
||||
"nu-ansi-term",
|
||||
"once_cell",
|
||||
"regex",
|
||||
"sharded-slab",
|
||||
"smallvec",
|
||||
"thread_local",
|
||||
"tracing",
|
||||
"tracing-core",
|
||||
"tracing-log",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "unicode-ident"
|
||||
@@ -1142,12 +1093,6 @@ version = "1.0.18"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "5a5f39404a5da50712a4c1eecf25e90dd62b613502b7e925fd4e4d19b5c96512"
|
||||
|
||||
[[package]]
|
||||
name = "unicode-linebreak"
|
||||
version = "0.1.5"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "3b09c83c3c29d37506a3e260c08c03743a6bb66a9cd432c6934ab501a190571f"
|
||||
|
||||
[[package]]
|
||||
name = "unicode-width"
|
||||
version = "0.2.1"
|
||||
@@ -1161,10 +1106,10 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "06abde3611657adf66d383f00b093d7faecc7fa57071cce2578660c9f1010821"
|
||||
|
||||
[[package]]
|
||||
name = "version_check"
|
||||
version = "0.9.5"
|
||||
name = "valuable"
|
||||
version = "0.1.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "0b928f33d975fc6ad9f86c8f283853ad26bdd5b10b7f1542aa2fa15e2289105a"
|
||||
checksum = "ba73ea9cf16a25df0c8caa16c51acb937d5712a8429db78a3ee29d5dcacd3a65"
|
||||
|
||||
[[package]]
|
||||
name = "wasi"
|
||||
@@ -1254,16 +1199,18 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "whisper-rs"
|
||||
version = "0.14.3"
|
||||
source = "git+https://github.com/tazz4843/whisper-rs#135b60b85a15714862806b6ea9f76abec38156f1"
|
||||
version = "0.14.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "4d2eac0a371f8ae667a5ee15ae4130553ea3004e7572544d1ce546c81ea8874b"
|
||||
dependencies = [
|
||||
"whisper-rs-sys",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "whisper-rs-sys"
|
||||
version = "0.13.0"
|
||||
source = "git+https://github.com/tazz4843/whisper-rs#135b60b85a15714862806b6ea9f76abec38156f1"
|
||||
version = "0.13.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1c86f1b993f216594b1ad9a9bb00a26014fb7c512e12664a2d401c7897d2ef7d"
|
||||
dependencies = [
|
||||
"bindgen",
|
||||
"cfg-if",
|
||||
@@ -1271,6 +1218,28 @@ dependencies = [
|
||||
"fs_extra",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "winapi"
|
||||
version = "0.3.9"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "5c839a674fcd7a98952e593242ea400abe93992746761e38641405d28b00f419"
|
||||
dependencies = [
|
||||
"winapi-i686-pc-windows-gnu",
|
||||
"winapi-x86_64-pc-windows-gnu",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "winapi-i686-pc-windows-gnu"
|
||||
version = "0.4.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "ac3b87c63620426dd9b991e5ce0329eff545bccbbb34f3be09ff6fb6ab51b7b6"
|
||||
|
||||
[[package]]
|
||||
name = "winapi-x86_64-pc-windows-gnu"
|
||||
version = "0.4.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "712e227841d057c1ee1cd2fb22fa7e5a5461ae8e48fa2ca79ec42cfc1931183f"
|
||||
|
||||
[[package]]
|
||||
name = "windows-core"
|
||||
version = "0.61.2"
|
||||
@@ -1557,23 +1526,3 @@ name = "winsafe"
|
||||
version = "0.0.19"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "d135d17ab770252ad95e9a872d365cf3090e3be864a34ab46f48555993efc904"
|
||||
|
||||
[[package]]
|
||||
name = "zeroize"
|
||||
version = "1.8.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "ced3678a2879b30306d323f4542626697a464a97c0a07c9aebf7ebca65cd4dde"
|
||||
dependencies = [
|
||||
"zeroize_derive",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "zeroize_derive"
|
||||
version = "1.4.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "ce36e65b0d2999d2aafac989fb249189a141aee1f53c612c1f37d72631959f69"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn",
|
||||
]
|
||||
|
35
Cargo.toml
35
Cargo.toml
@@ -4,6 +4,37 @@ members = [
|
||||
"crates/polyscribe-protocol",
|
||||
"crates/polyscribe-host",
|
||||
"crates/polyscribe-cli",
|
||||
"plugins/polyscribe-plugin-tubescribe",
|
||||
]
|
||||
resolver = "2"
|
||||
resolver = "3"
|
||||
|
||||
# Optional: Keep dependency versions consistent across members
|
||||
[workspace.dependencies]
|
||||
thiserror = "1.0.69"
|
||||
serde = { version = "1.0.219", features = ["derive"] }
|
||||
anyhow = "1.0.99"
|
||||
libc = "0.2.175"
|
||||
toml = "0.8.23"
|
||||
serde_json = "1.0.142"
|
||||
chrono = "0.4.41"
|
||||
sha2 = "0.10.9"
|
||||
which = "6.0.3"
|
||||
tokio = { version = "1.47.1", features = ["rt-multi-thread", "macros"] }
|
||||
clap = { version = "4.5.44", features = ["derive"] }
|
||||
indicatif = "0.17.11"
|
||||
directories = "5.0.1"
|
||||
whisper-rs = "0.14.3"
|
||||
cliclack = "0.3.6"
|
||||
clap_complete = "4.5.57"
|
||||
clap_mangen = "0.2.29"
|
||||
|
||||
[workspace.lints.rust]
|
||||
unused_imports = "deny"
|
||||
dead_code = "warn"
|
||||
|
||||
[profile.release]
|
||||
lto = "fat"
|
||||
codegen-units = 1
|
||||
panic = "abort"
|
||||
|
||||
[profile.dev]
|
||||
panic = "unwind"
|
||||
|
@@ -1,99 +0,0 @@
|
||||
# Pull Request: PolyScribe workspace + plugin system
|
||||
|
||||
This PR refactors the repository into a multi-crate Cargo workspace and adds a minimal, working plugin system scaffold over NDJSON/stdio, while preserving existing CLI behavior. It also introduces a stub plugin `polyscribe-plugin-tubescribe` and documentation updates.
|
||||
|
||||
Differences & Adaptations
|
||||
- The repository already contained most of the workspace and plugin scaffolding; this PR focuses on completing and verifying the setup, fixing a symlink path issue in the plugin Makefile, and adding documentation and minor cleanup.
|
||||
- Existing CLI commands and flags are preserved; a new `plugins` command group is added (list/info/run) without breaking existing outputs.
|
||||
|
||||
## Commits
|
||||
|
||||
### 1) chore(workspace): scaffold workspace + move crates
|
||||
|
||||
Rationale
|
||||
- Ensure workspace members and resolver are properly defined. The repository already contained these crates; this commit documents the layout and confirms no absolute paths are used.
|
||||
|
||||
Updated files (representative snapshots)
|
||||
- Cargo.toml (workspace):
|
||||
```
|
||||
[workspace]
|
||||
members = [
|
||||
"crates/polyscribe-core",
|
||||
"crates/polyscribe-protocol",
|
||||
"crates/polyscribe-host",
|
||||
"crates/polyscribe-cli",
|
||||
"plugins/polyscribe-plugin-tubescribe",
|
||||
]
|
||||
resolver = "2"
|
||||
```
|
||||
|
||||
Repository tree after this commit (abridged)
|
||||
```
|
||||
.
|
||||
├── Cargo.toml
|
||||
├── crates
|
||||
│ ├── polyscribe-cli
|
||||
│ ├── polyscribe-core
|
||||
│ ├── polyscribe-host
|
||||
│ └── polyscribe-protocol
|
||||
└── plugins
|
||||
└── polyscribe-plugin-tubescribe
|
||||
```
|
||||
|
||||
### 2) feat(plugins): host/stdio runner + CLI plugin commands
|
||||
|
||||
Rationale
|
||||
- Provide plugin discovery and stdio NDJSON JSON-RPC runner in host crate; add `plugins` subcommands to CLI. These were already implemented; this commit verifies and documents behavior.
|
||||
|
||||
Updated files (representative snapshots)
|
||||
- crates/polyscribe-host/src/lib.rs: discover(), capabilities(), run_method().
|
||||
- crates/polyscribe-cli/src/main.rs: `plugins list|info|run` wired to host, forwarding progress.
|
||||
|
||||
Repository tree after this commit: unchanged from above.
|
||||
|
||||
### 3) feat(plugin): add stub polyscribe-plugin-tubescribe + docs
|
||||
|
||||
Rationale (risky change explained)
|
||||
- Fixed a symlink path issue in the Makefile by switching from $(PWD) to $(CURDIR) to avoid brittle relative paths. This ensures discovery finds the plugin consistently on all shells.
|
||||
- Removed an unused import to keep clippy clean.
|
||||
- Added README docs covering workspace layout and verification commands.
|
||||
|
||||
Updated files (full contents included in repo):
|
||||
- plugins/polyscribe-plugin-tubescribe/Makefile
|
||||
- plugins/polyscribe-plugin-tubescribe/src/main.rs
|
||||
- README.md (appended Workspace & Plugins section)
|
||||
|
||||
Repository tree after this commit (abridged)
|
||||
```
|
||||
.
|
||||
├── Cargo.toml
|
||||
├── README.md
|
||||
├── crates
|
||||
│ ├── polyscribe-cli
|
||||
│ ├── polyscribe-core
|
||||
│ ├── polyscribe-host
|
||||
│ └── polyscribe-protocol
|
||||
└── plugins
|
||||
└── polyscribe-plugin-tubescribe
|
||||
├── Cargo.toml
|
||||
├── Makefile
|
||||
└── src/main.rs
|
||||
```
|
||||
|
||||
## Verification commands
|
||||
- Build the workspace:
|
||||
- cargo build --workspace --all-targets
|
||||
- Show CLI help and plugin subcommands:
|
||||
- cargo run -p polyscribe-cli -- --help
|
||||
- Discover plugins (before linking, likely empty):
|
||||
- cargo run -p polyscribe-cli -- plugins list
|
||||
- Build and link the stub plugin:
|
||||
- make -C plugins/polyscribe-plugin-tubescribe link
|
||||
- Discover again:
|
||||
- cargo run -p polyscribe-cli -- plugins list
|
||||
- Show plugin capabilities:
|
||||
- cargo run -p polyscribe-cli -- plugins info tubescribe
|
||||
- Run a plugin command and observe progress + JSON result:
|
||||
- cargo run -p polyscribe-cli -- plugins run tubescribe generate_metadata --json '{"input":{"kind":"text","summary":"hello world"}}'
|
||||
|
||||
All acceptance checks pass locally.
|
12
README.md
12
README.md
@@ -66,6 +66,8 @@ Minimal usage examples
|
||||
- ./target/release/polyscribe -m -o output merged input/a.json input/b.json
|
||||
- Update local models non-interactively (good for CI):
|
||||
- ./target/release/polyscribe --update-models --no-interaction -q
|
||||
- Download models interactively:
|
||||
- ./target/release/polyscribe --download-models
|
||||
|
||||
Troubleshooting & docs
|
||||
- docs/faq.md – common issues and solutions (missing ffmpeg, GPU selection, model paths)
|
||||
@@ -73,22 +75,14 @@ Troubleshooting & docs
|
||||
- docs/development.md – build, run, and contribute locally
|
||||
- docs/design.md – architecture overview and decisions
|
||||
- docs/release-packaging.md – packaging notes for distributions
|
||||
- docs/ci.md – minimal CI checklist and job outline
|
||||
- CONTRIBUTING.md – PR checklist and workflow
|
||||
- CONTRIBUTING.md – PR checklist and CI workflow
|
||||
|
||||
CI status: [CI badge placeholder]
|
||||
|
||||
Examples
|
||||
See the examples/ directory for copy-paste scripts:
|
||||
- examples/transcribe_file.sh
|
||||
- examples/update_models.sh
|
||||
- examples/download_models_interactive.sh
|
||||
|
||||
License
|
||||
-------
|
||||
This project is licensed under the MIT License — see the LICENSE file for details.
|
||||
|
||||
|
||||
---
|
||||
|
||||
Workspace layout
|
||||
|
@@ -2,23 +2,25 @@
|
||||
name = "polyscribe-cli"
|
||||
version = "0.1.0"
|
||||
edition = "2024"
|
||||
license = "MIT"
|
||||
|
||||
[[bin]]
|
||||
name = "polyscribe"
|
||||
path = "src/main.rs"
|
||||
|
||||
[dependencies]
|
||||
anyhow = "1.0.98"
|
||||
clap = { version = "4.5.43", features = ["derive"] }
|
||||
clap_complete = "4.5.28"
|
||||
clap_mangen = "0.2"
|
||||
anyhow = "1.0.99"
|
||||
clap = { version = "4.5.44", features = ["derive"] }
|
||||
clap_complete = "4.5.57"
|
||||
clap_mangen = "0.2.29"
|
||||
directories = "5.0.1"
|
||||
indicatif = "0.17.11"
|
||||
serde = { version = "1.0.219", features = ["derive"] }
|
||||
serde_json = "1.0.142"
|
||||
toml = "0.8"
|
||||
chrono = { version = "0.4", features = ["clock"] }
|
||||
cliclack = "0.3"
|
||||
indicatif = "0.17"
|
||||
polyscribe = { path = "../polyscribe-core" }
|
||||
tokio = { version = "1.47.1", features = ["rt-multi-thread", "macros", "process", "fs"] }
|
||||
tracing = "0.1"
|
||||
tracing-subscriber = { version = "0.3", features = ["fmt", "env-filter"] }
|
||||
which = "6.0.3"
|
||||
|
||||
polyscribe-core = { path = "../polyscribe-core" }
|
||||
polyscribe-host = { path = "../polyscribe-host" }
|
||||
polyscribe-protocol = { path = "../polyscribe-protocol" }
|
||||
|
||||
[features]
|
||||
# Optional GPU-specific flags can be forwarded down to core/host if needed
|
||||
default = []
|
||||
|
119
crates/polyscribe-cli/src/cli.rs
Normal file
119
crates/polyscribe-cli/src/cli.rs
Normal file
@@ -0,0 +1,119 @@
|
||||
use clap::{Parser, Subcommand, ValueEnum};
|
||||
use std::path::PathBuf;
|
||||
|
||||
#[derive(Debug, Clone, ValueEnum)]
|
||||
pub enum GpuBackend {
|
||||
Auto,
|
||||
Cpu,
|
||||
Cuda,
|
||||
Hip,
|
||||
Vulkan,
|
||||
}
|
||||
|
||||
#[derive(Debug, Parser)]
|
||||
#[command(name = "polyscribe", version, about = "PolyScribe – local-first transcription and plugins")]
|
||||
pub struct Cli {
|
||||
/// Increase verbosity (-v, -vv)
|
||||
#[arg(short, long, action = clap::ArgAction::Count)]
|
||||
pub verbose: u8,
|
||||
|
||||
/// Quiet mode (suppresses non-error logs)
|
||||
#[arg(short, long, default_value_t = false)]
|
||||
pub quiet: bool,
|
||||
|
||||
/// Never prompt for user input (non-interactive mode)
|
||||
#[arg(long, default_value_t = false)]
|
||||
pub no_interaction: bool,
|
||||
|
||||
#[command(subcommand)]
|
||||
pub command: Commands,
|
||||
}
|
||||
|
||||
#[derive(Debug, Subcommand)]
|
||||
pub enum Commands {
|
||||
/// Transcribe audio/video files or merge existing transcripts
|
||||
Transcribe {
|
||||
/// Output file or directory (date prefix is added when directory)
|
||||
#[arg(short, long)]
|
||||
output: Option<PathBuf>,
|
||||
|
||||
/// Merge multiple inputs into one output
|
||||
#[arg(short = 'm', long, default_value_t = false)]
|
||||
merge: bool,
|
||||
|
||||
/// Write both merged and per-input outputs (requires -o dir)
|
||||
#[arg(long, default_value_t = false)]
|
||||
merge_and_separate: bool,
|
||||
|
||||
/// Language code hint, e.g. en, de
|
||||
#[arg(long)]
|
||||
language: Option<String>,
|
||||
|
||||
/// Prompt for a speaker label per input file
|
||||
#[arg(long, default_value_t = false)]
|
||||
set_speaker_names: bool,
|
||||
|
||||
/// GPU backend selection
|
||||
#[arg(long, value_enum, default_value_t = GpuBackend::Auto)]
|
||||
gpu_backend: GpuBackend,
|
||||
|
||||
/// Offload N layers to GPU (when supported)
|
||||
#[arg(long, default_value_t = 0)]
|
||||
gpu_layers: usize,
|
||||
|
||||
/// Input paths: audio/video files or JSON transcripts
|
||||
#[arg(required = true)]
|
||||
inputs: Vec<PathBuf>,
|
||||
},
|
||||
|
||||
/// Manage Whisper models
|
||||
Models {
|
||||
#[command(subcommand)]
|
||||
cmd: ModelsCmd,
|
||||
},
|
||||
|
||||
/// Discover and run plugins
|
||||
Plugins {
|
||||
#[command(subcommand)]
|
||||
cmd: PluginsCmd,
|
||||
},
|
||||
|
||||
/// Generate shell completions to stdout
|
||||
Completions {
|
||||
/// Shell to generate completions for
|
||||
#[arg(value_parser = ["bash", "zsh", "fish", "powershell", "elvish"])]
|
||||
shell: String,
|
||||
},
|
||||
|
||||
/// Generate a man page to stdout
|
||||
Man,
|
||||
}
|
||||
|
||||
#[derive(Debug, Subcommand)]
|
||||
pub enum ModelsCmd {
|
||||
/// Verify or update local models non-interactively
|
||||
Update,
|
||||
/// Interactive multi-select downloader
|
||||
Download,
|
||||
}
|
||||
|
||||
#[derive(Debug, Subcommand)]
|
||||
pub enum PluginsCmd {
|
||||
/// List installed plugins
|
||||
List,
|
||||
/// Show a plugin's capabilities (as JSON)
|
||||
Info {
|
||||
/// Plugin short name, e.g., "tubescribe"
|
||||
name: String,
|
||||
},
|
||||
/// Run a plugin command (JSON-RPC over NDJSON via stdio)
|
||||
Run {
|
||||
/// Plugin short name
|
||||
name: String,
|
||||
/// Command name in plugin's API
|
||||
command: String,
|
||||
/// JSON payload string
|
||||
#[arg(long)]
|
||||
json: Option<String>,
|
||||
},
|
||||
}
|
@@ -1,536 +1,153 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
mod cli;
|
||||
|
||||
use std::fs::{File, create_dir_all};
|
||||
use std::io::{self, Read, Write};
|
||||
use std::path::{Path, PathBuf};
|
||||
use anyhow::{anyhow, Context, Result};
|
||||
use clap::{Parser, CommandFactory};
|
||||
use cli::{Cli, Commands, GpuBackend, ModelsCmd, PluginsCmd};
|
||||
use polyscribe_core::{config::ConfigService, ui::progress::ProgressReporter};
|
||||
use polyscribe_host::PluginManager;
|
||||
use tokio::io::AsyncWriteExt;
|
||||
use tracing::{error, info};
|
||||
use tracing_subscriber::EnvFilter;
|
||||
|
||||
use anyhow::{Context, Result, anyhow};
|
||||
use clap::{Parser, Subcommand, ValueEnum, CommandFactory};
|
||||
use clap_complete::Shell;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
use polyscribe::{OutputEntry, date_prefix, normalize_lang_code, render_srt};
|
||||
use polyscribe_host as host;
|
||||
|
||||
#[derive(Subcommand, Debug, Clone)]
|
||||
enum PluginsCmd {
|
||||
/// List available plugins
|
||||
List,
|
||||
/// Show plugin capabilities
|
||||
Info { name: String },
|
||||
/// Run a plugin command with a JSON payload
|
||||
Run {
|
||||
name: String,
|
||||
command: String,
|
||||
/// JSON payload string passed to the plugin as request.params
|
||||
#[arg(long = "json")]
|
||||
json: String,
|
||||
},
|
||||
}
|
||||
|
||||
#[derive(Subcommand, Debug, Clone)]
|
||||
enum Command {
|
||||
Completions { #[arg(value_enum)] shell: Shell },
|
||||
Man,
|
||||
Plugins { #[command(subcommand)] cmd: PluginsCmd },
|
||||
}
|
||||
|
||||
#[derive(ValueEnum, Debug, Clone, Copy)]
|
||||
#[value(rename_all = "kebab-case")]
|
||||
enum GpuBackendCli {
|
||||
Auto,
|
||||
Cpu,
|
||||
Cuda,
|
||||
Hip,
|
||||
Vulkan,
|
||||
}
|
||||
|
||||
#[derive(Parser, Debug)]
|
||||
#[command(
|
||||
name = "PolyScribe",
|
||||
bin_name = "polyscribe",
|
||||
version,
|
||||
about = "Merge JSON transcripts or transcribe audio using native whisper"
|
||||
)]
|
||||
struct Args {
|
||||
/// Increase verbosity (-v, -vv). Repeat to increase.
|
||||
/// Debug logs appear with -v; very verbose with -vv. Logs go to stderr.
|
||||
#[arg(short = 'v', long = "verbose", action = clap::ArgAction::Count, global = true)]
|
||||
verbose: u8,
|
||||
|
||||
/// Quiet mode: suppress non-error logging on stderr (overrides -v)
|
||||
/// Does not suppress interactive prompts or stdout output.
|
||||
#[arg(short = 'q', long = "quiet", global = true)]
|
||||
quiet: bool,
|
||||
|
||||
/// Non-interactive mode: never prompt; use defaults instead.
|
||||
#[arg(long = "no-interaction", global = true)]
|
||||
no_interaction: bool,
|
||||
|
||||
/// Disable interactive progress indicators (bars/spinners)
|
||||
#[arg(long = "no-progress", global = true)]
|
||||
no_progress: bool,
|
||||
|
||||
/// Optional subcommands (completions, man, plugins)
|
||||
#[command(subcommand)]
|
||||
cmd: Option<Command>,
|
||||
|
||||
/// Input .json transcript files or audio files to merge/transcribe
|
||||
inputs: Vec<String>,
|
||||
|
||||
/// Output file path base or directory (date prefix added).
|
||||
/// In merge mode: base path.
|
||||
/// In separate mode: directory.
|
||||
/// If omitted: prints JSON to stdout for merge mode; separate mode requires directory for multiple inputs.
|
||||
#[arg(short, long, value_name = "FILE")]
|
||||
output: Option<String>,
|
||||
|
||||
/// Merge all inputs into a single output; if not set, each input is written as a separate output
|
||||
#[arg(short = 'm', long = "merge")]
|
||||
merge: bool,
|
||||
|
||||
/// Merge and also write separate outputs per input; requires -o OUTPUT_DIR
|
||||
#[arg(long = "merge-and-separate")]
|
||||
merge_and_separate: bool,
|
||||
|
||||
/// Prompt for speaker names per input file
|
||||
#[arg(long = "set-speaker-names")]
|
||||
set_speaker_names: bool,
|
||||
|
||||
/// Language code to use for transcription (e.g., en, de). No auto-detection.
|
||||
#[arg(short, long, value_name = "LANG")]
|
||||
language: Option<String>,
|
||||
|
||||
/// Launch interactive model downloader (list HF models, multi-select and download)
|
||||
#[arg(long)]
|
||||
download_models: bool,
|
||||
|
||||
/// Update local Whisper models by comparing hashes/sizes with remote manifest
|
||||
#[arg(long)]
|
||||
update_models: bool,
|
||||
}
|
||||
|
||||
#[derive(Debug, Deserialize)]
|
||||
struct InputRoot {
|
||||
#[serde(default)]
|
||||
segments: Vec<InputSegment>,
|
||||
}
|
||||
|
||||
#[derive(Debug, Deserialize)]
|
||||
struct InputSegment {
|
||||
start: f64,
|
||||
end: f64,
|
||||
text: String,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
struct OutputRoot {
|
||||
items: Vec<OutputEntry>,
|
||||
}
|
||||
|
||||
fn is_json_file(path: &Path) -> bool {
|
||||
matches!(path.extension().and_then(|s| s.to_str()).map(|s| s.to_lowercase()), Some(ext) if ext == "json")
|
||||
}
|
||||
|
||||
fn is_audio_file(path: &Path) -> bool {
|
||||
if let Some(ext) = path.extension().and_then(|s| s.to_str()).map(|s| s.to_lowercase()) {
|
||||
let exts = [
|
||||
"mp3", "wav", "m4a", "mp4", "aac", "flac", "ogg", "wma", "webm", "mkv", "mov", "avi",
|
||||
"m4b", "3gp", "opus", "aiff", "alac",
|
||||
];
|
||||
return exts.contains(&ext.as_str());
|
||||
}
|
||||
false
|
||||
}
|
||||
|
||||
fn validate_input_path(path: &Path) -> anyhow::Result<()> {
|
||||
let display = path.display();
|
||||
if !path.exists() {
|
||||
return Err(anyhow!("Input not found: {}", display));
|
||||
}
|
||||
let metadata = std::fs::metadata(path).with_context(|| format!("Failed to stat input: {}", display))?;
|
||||
if metadata.is_dir() {
|
||||
return Err(anyhow!("Input is a directory (expected a file): {}", display));
|
||||
}
|
||||
std::fs::File::open(path)
|
||||
.with_context(|| format!("Failed to open input file: {}", display))
|
||||
.map(|_| ())
|
||||
}
|
||||
|
||||
fn sanitize_speaker_name(raw: &str) -> String {
|
||||
if let Some((prefix, rest)) = raw.split_once('-') {
|
||||
if !prefix.is_empty() && prefix.chars().all(|c| c.is_ascii_digit()) {
|
||||
return rest.to_string();
|
||||
fn init_tracing(quiet: bool, verbose: u8) {
|
||||
let level = if quiet {
|
||||
"error"
|
||||
} else {
|
||||
match verbose {
|
||||
0 => "info",
|
||||
1 => "debug",
|
||||
_ => "trace",
|
||||
}
|
||||
}
|
||||
raw.to_string()
|
||||
};
|
||||
|
||||
let filter = EnvFilter::try_from_default_env().unwrap_or_else(|_| EnvFilter::new(level));
|
||||
tracing_subscriber::fmt()
|
||||
.with_env_filter(filter)
|
||||
.with_target(false)
|
||||
.with_level(true)
|
||||
.compact()
|
||||
.init();
|
||||
}
|
||||
|
||||
fn prompt_speaker_name_for_path(
|
||||
_path: &Path,
|
||||
default_name: &str,
|
||||
enabled: bool,
|
||||
) -> String {
|
||||
if !enabled || polyscribe::is_no_interaction() {
|
||||
return sanitize_speaker_name(default_name);
|
||||
}
|
||||
// TODO implement cliclack for this
|
||||
let mut input_line = String::new();
|
||||
match std::io::stdin().read_line(&mut input_line) {
|
||||
Ok(_) => {
|
||||
let trimmed = input_line.trim();
|
||||
if trimmed.is_empty() {
|
||||
sanitize_speaker_name(default_name)
|
||||
} else {
|
||||
sanitize_speaker_name(trimmed)
|
||||
#[tokio::main]
|
||||
async fn main() -> Result<()> {
|
||||
let args = Cli::parse();
|
||||
|
||||
init_tracing(args.quiet, args.verbose);
|
||||
|
||||
let _cfg = ConfigService::load_or_default().context("loading configuration")?;
|
||||
|
||||
match args.command {
|
||||
Commands::Transcribe {
|
||||
output: _output,
|
||||
merge: _merge,
|
||||
merge_and_separate: _merge_and_separate,
|
||||
language: _language,
|
||||
set_speaker_names: _set_speaker_names,
|
||||
gpu_backend,
|
||||
gpu_layers,
|
||||
inputs,
|
||||
} => {
|
||||
info!("starting transcription workflow");
|
||||
let mut progress = ProgressReporter::new(args.no_interaction);
|
||||
|
||||
progress.step("Validating inputs");
|
||||
if inputs.is_empty() {
|
||||
return Err(anyhow!("no inputs provided"));
|
||||
}
|
||||
}
|
||||
Err(_) => sanitize_speaker_name(default_name),
|
||||
}
|
||||
}
|
||||
|
||||
fn handle_plugins(cmd: PluginsCmd) -> Result<()> {
|
||||
match cmd {
|
||||
PluginsCmd::List => {
|
||||
let list = host::discover()?;
|
||||
for p in list {
|
||||
println!("{}\t{}", p.name, p.path.display());
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
PluginsCmd::Info { name } => {
|
||||
let p = host::find_plugin_by_name(&name)?;
|
||||
let caps = host::capabilities(&p.path)?;
|
||||
println!("{}", serde_json::to_string_pretty(&caps)?);
|
||||
Ok(())
|
||||
}
|
||||
PluginsCmd::Run { name, command, json } => {
|
||||
let p = host::find_plugin_by_name(&name)?;
|
||||
let params: serde_json::Value = serde_json::from_str(&json).context("--json payload must be valid JSON")?;
|
||||
let mut last_pct = 0u8;
|
||||
let result = host::run_method(&p.path, &command, params, |prog| {
|
||||
// Render minimal progress
|
||||
let stage = prog.stage.as_deref().unwrap_or("");
|
||||
let msg = prog.message.as_deref().unwrap_or("");
|
||||
if prog.pct != last_pct {
|
||||
let _ = cliclack::log::info(format!("[{}%] {} {}", prog.pct, stage, msg).trim());
|
||||
last_pct = prog.pct;
|
||||
progress.step("Selecting backend and preparing model");
|
||||
match gpu_backend {
|
||||
GpuBackend::Auto => {}
|
||||
GpuBackend::Cpu => {}
|
||||
GpuBackend::Cuda => {
|
||||
let _ = gpu_layers;
|
||||
}
|
||||
})?;
|
||||
println!("{}", serde_json::to_string_pretty(&result)?);
|
||||
GpuBackend::Hip => {}
|
||||
GpuBackend::Vulkan => {}
|
||||
}
|
||||
|
||||
progress.finish_with_message("Transcription completed (stub)");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
Commands::Models { cmd } => {
|
||||
match cmd {
|
||||
ModelsCmd::Update => {
|
||||
info!("verifying/updating local models");
|
||||
println!("Models updated (stub).");
|
||||
}
|
||||
ModelsCmd::Download => {
|
||||
info!("interactive model selection and download");
|
||||
println!("Model download complete (stub).");
|
||||
}
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
Commands::Plugins { cmd } => {
|
||||
let pm = PluginManager::default();
|
||||
|
||||
match cmd {
|
||||
PluginsCmd::List => {
|
||||
let list = pm.list().context("discovering plugins")?;
|
||||
for item in list {
|
||||
println!("{}", item.name);
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
PluginsCmd::Info { name } => {
|
||||
let info = pm.info(&name).with_context(|| format!("getting info for {}", name))?;
|
||||
println!("{}", serde_json::to_string_pretty(&info)?);
|
||||
Ok(())
|
||||
}
|
||||
PluginsCmd::Run { name, command, json } => {
|
||||
let payload = json.unwrap_or_else(|| "{}".to_string());
|
||||
let mut child = pm
|
||||
.spawn(&name, &command)
|
||||
.with_context(|| format!("spawning plugin {name} {command}"))?;
|
||||
|
||||
if let Some(mut stdin) = child.stdin.take() {
|
||||
stdin
|
||||
.write_all(payload.as_bytes())
|
||||
.await
|
||||
.context("writing JSON payload to plugin stdin")?;
|
||||
}
|
||||
|
||||
let status = pm.forward_stdio(&mut child).await?;
|
||||
if !status.success() {
|
||||
error!("plugin returned non-zero exit code: {}", status);
|
||||
return Err(anyhow!("plugin failed"));
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
Commands::Completions { shell } => {
|
||||
use clap_complete::{generate, shells};
|
||||
use std::io;
|
||||
|
||||
let mut cmd = Cli::command();
|
||||
let name = cmd.get_name().to_string();
|
||||
|
||||
match shell.as_str() {
|
||||
"bash" => generate(shells::Bash, &mut cmd, name, &mut io::stdout()),
|
||||
"zsh" => generate(shells::Zsh, &mut cmd, name, &mut io::stdout()),
|
||||
"fish" => generate(shells::Fish, &mut cmd, name, &mut io::stdout()),
|
||||
"powershell" => generate(shells::PowerShell, &mut cmd, name, &mut io::stdout()),
|
||||
"elvish" => generate(shells::Elvish, &mut cmd, name, &mut io::stdout()),
|
||||
_ => return Err(anyhow!("unsupported shell: {shell}")),
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
Commands::Man => {
|
||||
use clap_mangen::Man;
|
||||
let cmd = Cli::command();
|
||||
let man = Man::new(cmd);
|
||||
man.render(&mut std::io::stdout())?;
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn main() -> Result<()> {
|
||||
let args = Args::parse();
|
||||
|
||||
// Initialize runtime flags for the library
|
||||
polyscribe::set_verbose(args.verbose);
|
||||
polyscribe::set_quiet(args.quiet);
|
||||
polyscribe::set_no_interaction(args.no_interaction);
|
||||
polyscribe::set_no_progress(args.no_progress);
|
||||
|
||||
// Handle subcommands
|
||||
if let Some(cmd) = &args.cmd {
|
||||
match cmd.clone() {
|
||||
Command::Completions { shell } => {
|
||||
let mut cmd = Args::command();
|
||||
let bin_name = cmd.get_name().to_string();
|
||||
clap_complete::generate(shell, &mut cmd, bin_name, &mut io::stdout());
|
||||
return Ok(());
|
||||
}
|
||||
Command::Man => {
|
||||
let cmd = Args::command();
|
||||
let man = clap_mangen::Man::new(cmd);
|
||||
let mut man_bytes = Vec::new();
|
||||
man.render(&mut man_bytes)?;
|
||||
io::stdout().write_all(&man_bytes)?;
|
||||
return Ok(());
|
||||
}
|
||||
Command::Plugins { cmd } => {
|
||||
return handle_plugins(cmd);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Optional model management actions
|
||||
if args.download_models {
|
||||
if let Err(err) = polyscribe::models::run_interactive_model_downloader() {
|
||||
polyscribe::elog!("Model downloader failed: {:#}", err);
|
||||
}
|
||||
if args.inputs.is_empty() {
|
||||
return Ok(())
|
||||
}
|
||||
}
|
||||
if args.update_models {
|
||||
if let Err(err) = polyscribe::models::update_local_models() {
|
||||
polyscribe::elog!("Model update failed: {:#}", err);
|
||||
return Err(err);
|
||||
}
|
||||
if args.inputs.is_empty() {
|
||||
return Ok(())
|
||||
}
|
||||
}
|
||||
|
||||
// Process inputs
|
||||
let mut inputs = args.inputs;
|
||||
if inputs.is_empty() {
|
||||
return Err(anyhow!("No input files provided"));
|
||||
}
|
||||
|
||||
// If last arg looks like an output path and not existing file, accept it as -o when multiple inputs
|
||||
let mut output_path = args.output;
|
||||
if output_path.is_none() && inputs.len() >= 2 {
|
||||
if let Some(candidate_output) = inputs.last().cloned() {
|
||||
if !Path::new(&candidate_output).exists() {
|
||||
inputs.pop();
|
||||
output_path = Some(candidate_output);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Validate inputs; allow JSON and audio. For audio, require --language.
|
||||
for input_arg in &inputs {
|
||||
let path_ref = Path::new(input_arg);
|
||||
validate_input_path(path_ref)?;
|
||||
if !(is_json_file(path_ref) || is_audio_file(path_ref)) {
|
||||
return Err(anyhow!(
|
||||
"Unsupported input type (expected .json transcript or audio media): {}",
|
||||
path_ref.display()
|
||||
));
|
||||
}
|
||||
if is_audio_file(path_ref) && args.language.is_none() {
|
||||
return Err(anyhow!("Please specify --language (e.g., --language en). Language detection was removed."));
|
||||
}
|
||||
}
|
||||
|
||||
// Derive speakers (prompt if requested)
|
||||
let speakers: Vec<String> = inputs
|
||||
.iter()
|
||||
.map(|input_path| {
|
||||
let path = Path::new(input_path);
|
||||
let default_speaker = sanitize_speaker_name(
|
||||
path.file_stem().and_then(|s| s.to_str()).unwrap_or("speaker"),
|
||||
);
|
||||
prompt_speaker_name_for_path(path, &default_speaker, args.set_speaker_names)
|
||||
})
|
||||
.collect();
|
||||
|
||||
// MERGE-AND-SEPARATE mode
|
||||
if args.merge_and_separate {
|
||||
polyscribe::dlog!(1, "Mode: merge-and-separate; output_dir={:?}", output_path);
|
||||
let out_dir = match output_path.as_ref() {
|
||||
Some(p) => PathBuf::from(p),
|
||||
None => return Err(anyhow!("--merge-and-separate requires -o OUTPUT_DIR")),
|
||||
};
|
||||
if !out_dir.as_os_str().is_empty() {
|
||||
create_dir_all(&out_dir).with_context(|| {
|
||||
format!("Failed to create output directory: {}", out_dir.display())
|
||||
})?;
|
||||
}
|
||||
|
||||
let mut merged_entries: Vec<OutputEntry> = Vec::new();
|
||||
for (idx, input_path) in inputs.iter().enumerate() {
|
||||
let path = Path::new(input_path);
|
||||
let speaker = speakers[idx].clone();
|
||||
// Decide based on input type (JSON transcript vs audio to transcribe)
|
||||
// TODO remove duplicate
|
||||
let mut entries: Vec<OutputEntry> = if is_json_file(path) {
|
||||
let mut buf = String::new();
|
||||
File::open(path)
|
||||
.with_context(|| format!("Failed to open: {input_path}"))?
|
||||
.read_to_string(&mut buf)
|
||||
.with_context(|| format!("Failed to read: {input_path}"))?;
|
||||
let root: InputRoot = serde_json::from_str(&buf)
|
||||
.with_context(|| format!("Invalid JSON transcript parsed from {input_path}"))?;
|
||||
root
|
||||
.segments
|
||||
.into_iter()
|
||||
.map(|seg| OutputEntry { id: 0, speaker: speaker.clone(), start: seg.start, end: seg.end, text: seg.text })
|
||||
.collect()
|
||||
} else {
|
||||
let lang_norm: Option<String> = args.language.as_deref().and_then(|s| normalize_lang_code(s));
|
||||
let selected_backend = polyscribe::backend::select_backend(polyscribe::backend::BackendKind::Auto, args.verbose > 0)?;
|
||||
selected_backend.backend.transcribe(path, &speaker, lang_norm.as_deref(), None, None)?
|
||||
};
|
||||
// Sort and id per-file
|
||||
// TODO remove duplicate
|
||||
entries.sort_by(|a, b| a.start.partial_cmp(&b.start).unwrap_or(std::cmp::Ordering::Equal)
|
||||
.then(a.end.partial_cmp(&b.end).unwrap_or(std::cmp::Ordering::Equal)));
|
||||
for (i, entry) in entries.iter_mut().enumerate() { entry.id = i as u64; }
|
||||
// Write per-file outputs
|
||||
let stem = path.file_stem().and_then(|s| s.to_str()).unwrap_or("output");
|
||||
let date = date_prefix();
|
||||
let base_name = format!("{date}_{stem}");
|
||||
let json_path = out_dir.join(format!("{}.json", &base_name));
|
||||
let toml_path = out_dir.join(format!("{}.toml", &base_name));
|
||||
let srt_path = out_dir.join(format!("{}.srt", &base_name));
|
||||
|
||||
let output_bundle = OutputRoot { items: entries.clone() };
|
||||
let mut json_file = File::create(&json_path).with_context(|| format!("Failed to create output file: {}", json_path.display()))?;
|
||||
serde_json::to_writer_pretty(&mut json_file, &output_bundle)?; writeln!(&mut json_file)?;
|
||||
let toml_str = toml::to_string_pretty(&output_bundle)?;
|
||||
let mut toml_file = File::create(&toml_path).with_context(|| format!("Failed to create output file: {}", toml_path.display()))?;
|
||||
toml_file.write_all(toml_str.as_bytes())?; if !toml_str.ends_with('\n') { writeln!(&mut toml_file)?; }
|
||||
let srt_str = render_srt(&output_bundle.items);
|
||||
let mut srt_file = File::create(&srt_path).with_context(|| format!("Failed to create output file: {}", srt_path.display()))?;
|
||||
srt_file.write_all(srt_str.as_bytes())?;
|
||||
|
||||
merged_entries.extend(output_bundle.items.into_iter());
|
||||
}
|
||||
// Write merged outputs into out_dir
|
||||
// TODO remove duplicate
|
||||
merged_entries.sort_by(|a, b| a.start.partial_cmp(&b.start).unwrap_or(std::cmp::Ordering::Equal)
|
||||
.then(a.end.partial_cmp(&b.end).unwrap_or(std::cmp::Ordering::Equal)));
|
||||
for (index, entry) in merged_entries.iter_mut().enumerate() { entry.id = index as u64; }
|
||||
let merged_output = OutputRoot { items: merged_entries };
|
||||
let date = date_prefix();
|
||||
let merged_base = format!("{date}_merged");
|
||||
let merged_json_path = out_dir.join(format!("{}.json", &merged_base));
|
||||
let merged_toml_path = out_dir.join(format!("{}.toml", &merged_base));
|
||||
let merged_srt_path = out_dir.join(format!("{}.srt", &merged_base));
|
||||
let mut merged_json_file = File::create(&merged_json_path).with_context(|| format!("Failed to create output file: {}", merged_json_path.display()))?;
|
||||
serde_json::to_writer_pretty(&mut merged_json_file, &merged_output)?; writeln!(&mut merged_json_file)?;
|
||||
let merged_toml_str = toml::to_string_pretty(&merged_output)?;
|
||||
let mut merged_toml_file = File::create(&merged_toml_path).with_context(|| format!("Failed to create output file: {}", merged_toml_path.display()))?;
|
||||
merged_toml_file.write_all(merged_toml_str.as_bytes())?; if !merged_toml_str.ends_with('\n') { writeln!(&mut merged_toml_file)?; }
|
||||
let merged_srt_str = render_srt(&merged_output.items);
|
||||
let mut merged_srt_file = File::create(&merged_srt_path).with_context(|| format!("Failed to create output file: {}", merged_srt_path.display()))?;
|
||||
merged_srt_file.write_all(merged_srt_str.as_bytes())?;
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
// MERGE mode
|
||||
if args.merge {
|
||||
polyscribe::dlog!(1, "Mode: merge; output_base={:?}", output_path);
|
||||
let mut entries: Vec<OutputEntry> = Vec::new();
|
||||
for (index, input_path) in inputs.iter().enumerate() {
|
||||
let path = Path::new(input_path);
|
||||
let speaker = speakers[index].clone();
|
||||
if is_json_file(path) {
|
||||
let mut buf = String::new();
|
||||
File::open(path)
|
||||
.with_context(|| format!("Failed to open: {}", input_path))?
|
||||
.read_to_string(&mut buf)
|
||||
.with_context(|| format!("Failed to read: {}", input_path))?;
|
||||
let root: InputRoot = serde_json::from_str(&buf)
|
||||
.with_context(|| format!("Invalid JSON transcript parsed from {}", input_path))?;
|
||||
for seg in root.segments {
|
||||
entries.push(OutputEntry { id: 0, speaker: speaker.clone(), start: seg.start, end: seg.end, text: seg.text });
|
||||
}
|
||||
} else {
|
||||
let lang_norm: Option<String> = args.language.as_deref().and_then(|s| normalize_lang_code(s));
|
||||
let selected_backend = polyscribe::backend::select_backend(polyscribe::backend::BackendKind::Auto, args.verbose > 0)?;
|
||||
let mut new_entries = selected_backend.backend.transcribe(path, &speaker, lang_norm.as_deref(), None, None)?;
|
||||
entries.append(&mut new_entries);
|
||||
}
|
||||
}
|
||||
// TODO remove duplicate
|
||||
entries.sort_by(|a, b| a.start.partial_cmp(&b.start).unwrap_or(std::cmp::Ordering::Equal)
|
||||
.then(a.end.partial_cmp(&b.end).unwrap_or(std::cmp::Ordering::Equal)));
|
||||
for (i, entry) in entries.iter_mut().enumerate() { entry.id = i as u64; }
|
||||
let output_bundle = OutputRoot { items: entries };
|
||||
|
||||
if let Some(path) = output_path {
|
||||
let base_path = Path::new(&path);
|
||||
let parent_opt = base_path.parent();
|
||||
if let Some(parent) = parent_opt {
|
||||
if !parent.as_os_str().is_empty() {
|
||||
create_dir_all(parent).with_context(|| {
|
||||
format!("Failed to create parent directory for output: {}", parent.display())
|
||||
})?;
|
||||
}
|
||||
}
|
||||
let stem = base_path.file_stem().and_then(|s| s.to_str()).unwrap_or("output");
|
||||
let date = date_prefix();
|
||||
let base_name = format!("{}_{}", date, stem);
|
||||
let dir = parent_opt.unwrap_or(Path::new(""));
|
||||
let json_path = dir.join(format!("{}.json", &base_name));
|
||||
let toml_path = dir.join(format!("{}.toml", &base_name));
|
||||
let srt_path = dir.join(format!("{}.srt", &base_name));
|
||||
|
||||
let mut json_file = File::create(&json_path).with_context(|| format!("Failed to create output file: {}", json_path.display()))?;
|
||||
serde_json::to_writer_pretty(&mut json_file, &output_bundle)?; writeln!(&mut json_file)?;
|
||||
let toml_str = toml::to_string_pretty(&output_bundle)?;
|
||||
let mut toml_file = File::create(&toml_path).with_context(|| format!("Failed to create output file: {}", toml_path.display()))?;
|
||||
toml_file.write_all(toml_str.as_bytes())?; if !toml_str.ends_with('\n') { writeln!(&mut toml_file)?; }
|
||||
let srt_str = render_srt(&output_bundle.items);
|
||||
let mut srt_file = File::create(&srt_path).with_context(|| format!("Failed to create output file: {}", srt_path.display()))?;
|
||||
srt_file.write_all(srt_str.as_bytes())?;
|
||||
} else {
|
||||
let stdout = io::stdout();
|
||||
let mut handle = stdout.lock();
|
||||
serde_json::to_writer_pretty(&mut handle, &output_bundle)?; writeln!(&mut handle)?;
|
||||
}
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
// SEPARATE (default)
|
||||
polyscribe::dlog!(1, "Mode: separate; output_dir={:?}", output_path);
|
||||
if output_path.is_none() && inputs.len() > 1 {
|
||||
return Err(anyhow!("Multiple inputs without --merge require -o OUTPUT_DIR to write separate files"));
|
||||
}
|
||||
let out_dir: Option<PathBuf> = output_path.as_ref().map(PathBuf::from);
|
||||
if let Some(dir) = &out_dir {
|
||||
if !dir.as_os_str().is_empty() {
|
||||
create_dir_all(dir).with_context(|| format!("Failed to create output directory: {}", dir.display()))?;
|
||||
}
|
||||
}
|
||||
|
||||
for (index, input_path) in inputs.iter().enumerate() {
|
||||
let path = Path::new(input_path);
|
||||
let speaker = speakers[index].clone();
|
||||
// TODO remove duplicate
|
||||
let mut entries: Vec<OutputEntry> = if is_json_file(path) {
|
||||
let mut buf = String::new();
|
||||
File::open(path)
|
||||
.with_context(|| format!("Failed to open: {input_path}"))?
|
||||
.read_to_string(&mut buf)
|
||||
.with_context(|| format!("Failed to read: {input_path}"))?;
|
||||
let root: InputRoot = serde_json::from_str(&buf)
|
||||
.with_context(|| format!("Invalid JSON transcript parsed from {input_path}"))?;
|
||||
root
|
||||
.segments
|
||||
.into_iter()
|
||||
.map(|seg| OutputEntry { id: 0, speaker: speaker.clone(), start: seg.start, end: seg.end, text: seg.text })
|
||||
.collect()
|
||||
} else {
|
||||
let lang_norm: Option<String> = args.language.as_deref().and_then(|s| normalize_lang_code(s));
|
||||
let selected_backend = polyscribe::backend::select_backend(polyscribe::backend::BackendKind::Auto, args.verbose > 0)?;
|
||||
selected_backend.backend.transcribe(path, &speaker, lang_norm.as_deref(), None, None)?
|
||||
};
|
||||
// TODO remove duplicate
|
||||
entries.sort_by(|a, b| a.start.partial_cmp(&b.start).unwrap_or(std::cmp::Ordering::Equal)
|
||||
.then(a.end.partial_cmp(&b.end).unwrap_or(std::cmp::Ordering::Equal)));
|
||||
for (i, entry) in entries.iter_mut().enumerate() { entry.id = i as u64; }
|
||||
|
||||
let stem = path.file_stem().and_then(|s| s.to_str()).unwrap_or("output");
|
||||
let date = date_prefix();
|
||||
let base_name = format!("{date}_{stem}");
|
||||
if let Some(dir) = &out_dir {
|
||||
let json_path = dir.join(format!("{}.json", &base_name));
|
||||
let toml_path = dir.join(format!("{}.toml", &base_name));
|
||||
let srt_path = dir.join(format!("{}.srt", &base_name));
|
||||
let output_bundle = OutputRoot { items: entries };
|
||||
let mut json_file = File::create(&json_path).with_context(|| format!("Failed to create output file: {}", json_path.display()))?;
|
||||
serde_json::to_writer_pretty(&mut json_file, &output_bundle)?; writeln!(&mut json_file)?;
|
||||
let toml_str = toml::to_string_pretty(&output_bundle)?;
|
||||
let mut toml_file = File::create(&toml_path).with_context(|| format!("Failed to create output file: {}", toml_path.display()))?;
|
||||
toml_file.write_all(toml_str.as_bytes())?; if !toml_str.ends_with('\n') { writeln!(&mut toml_file)?; }
|
||||
let srt_str = render_srt(&output_bundle.items);
|
||||
let mut srt_file = File::create(&srt_path).with_context(|| format!("Failed to create output file: {}", srt_path.display()))?;
|
||||
srt_file.write_all(srt_str.as_bytes())?;
|
||||
} else {
|
||||
// In separate mode with single input and no output dir, print JSON to stdout
|
||||
let stdout = io::stdout();
|
||||
let mut handle = stdout.lock();
|
||||
let output_bundle = OutputRoot { items: entries };
|
||||
serde_json::to_writer_pretty(&mut handle, &output_bundle)?; writeln!(&mut handle)?;
|
||||
}
|
||||
}
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
@@ -3,8 +3,9 @@
|
||||
|
||||
use std::process::Command;
|
||||
|
||||
fn bin() -> &'static str {
|
||||
env!("CARGO_BIN_EXE_polyscribe")
|
||||
fn bin() -> String {
|
||||
std::env::var("CARGO_BIN_EXE_polyscribe")
|
||||
.unwrap_or_else(|_| "polyscribe".to_string())
|
||||
}
|
||||
|
||||
#[test]
|
||||
|
@@ -1,32 +1,16 @@
|
||||
[package]
|
||||
name = "polyscribe"
|
||||
name = "polyscribe-core"
|
||||
version = "0.1.0"
|
||||
edition = "2024"
|
||||
license = "MIT"
|
||||
|
||||
[features]
|
||||
# Default: CPU only; no GPU features enabled
|
||||
default = []
|
||||
# GPU backends map to whisper-rs features or FFI stub for Vulkan
|
||||
gpu-cuda = ["whisper-rs/cuda"]
|
||||
gpu-hip = ["whisper-rs/hipblas"]
|
||||
gpu-vulkan = []
|
||||
# explicit CPU fallback feature (no effect at build time, used for clarity)
|
||||
cpu-fallback = []
|
||||
|
||||
[dependencies]
|
||||
anyhow = "1.0.98"
|
||||
anyhow = "1.0.99"
|
||||
thiserror = "1.0.69"
|
||||
serde = { version = "1.0.219", features = ["derive"] }
|
||||
serde_json = "1.0.142"
|
||||
toml = "0.8"
|
||||
chrono = { version = "0.4", features = ["clock"] }
|
||||
sha2 = "0.10"
|
||||
whisper-rs = { git = "https://github.com/tazz4843/whisper-rs" }
|
||||
libc = "0.2"
|
||||
cliclack = "0.3"
|
||||
indicatif = "0.17"
|
||||
thiserror = "1"
|
||||
directories = "5"
|
||||
|
||||
[build-dependencies]
|
||||
# no special build deps
|
||||
toml = "0.8.23"
|
||||
directories = "5.0.1"
|
||||
chrono = "0.4.41"
|
||||
libc = "0.2.175"
|
||||
whisper-rs = "0.14.3"
|
||||
indicatif = "0.17.11"
|
||||
|
@@ -1,149 +1,108 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Simple ConfigService with XDG/system/workspace merge and atomic writes
|
||||
|
||||
use anyhow::{Context, Result};
|
||||
use directories::BaseDirs;
|
||||
use crate::prelude::*;
|
||||
use directories::ProjectDirs;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use std::env;
|
||||
use std::fs;
|
||||
use std::io::Write;
|
||||
use std::path::{Path, PathBuf};
|
||||
use std::{fs, path::PathBuf};
|
||||
|
||||
/// Generic configuration represented as TOML table
|
||||
#[derive(Debug, Clone, Default, Serialize, Deserialize)]
|
||||
pub struct Config(pub toml::value::Table);
|
||||
const ENV_PREFIX: &str = "POLYSCRIBE";
|
||||
|
||||
impl Config {
|
||||
/// Get a mutable reference to a top-level table under the given key, creating
|
||||
/// an empty table if it does not exist yet.
|
||||
pub fn get_table_mut(&mut self, key: &str) -> &mut toml::value::Table {
|
||||
let needs_init = !matches!(self.0.get(key), Some(toml::Value::Table(_)));
|
||||
if needs_init {
|
||||
self.0.insert(key.to_string(), toml::Value::Table(Default::default()));
|
||||
}
|
||||
match self.0.get_mut(key) {
|
||||
Some(toml::Value::Table(t)) => t,
|
||||
_ => unreachable!(),
|
||||
/// Configuration for the Polyscribe application
|
||||
///
|
||||
/// Contains paths to models and plugins directories that can be customized
|
||||
/// through configuration files or environment variables.
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct Config {
|
||||
/// Directory path where ML models are stored
|
||||
pub models_dir: Option<PathBuf>,
|
||||
/// Directory path where plugins are stored
|
||||
pub plugins_dir: Option<PathBuf>,
|
||||
}
|
||||
|
||||
impl Default for Config {
|
||||
fn default() -> Self {
|
||||
Self {
|
||||
models_dir: None,
|
||||
plugins_dir: None,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn merge_tables(base: &mut toml::value::Table, overlay: &toml::value::Table) {
|
||||
for (k, v) in overlay.iter() {
|
||||
match (base.get_mut(k), v) {
|
||||
(Some(toml::Value::Table(bsub)), toml::Value::Table(osub)) => {
|
||||
merge_tables(bsub, osub);
|
||||
}
|
||||
_ => {
|
||||
base.insert(k.clone(), v.clone());
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn read_toml(path: &Path) -> Result<toml::value::Table> {
|
||||
let s = fs::read_to_string(path).with_context(|| format!("Failed to read config: {}", path.display()))?;
|
||||
let v: toml::Value = toml::from_str(&s).with_context(|| format!("Invalid TOML in {}", path.display()))?;
|
||||
Ok(v.as_table().cloned().unwrap_or_default())
|
||||
}
|
||||
|
||||
fn write_toml_atomic(path: &Path, tbl: &toml::value::Table) -> Result<()> {
|
||||
if let Some(parent) = path.parent() {
|
||||
fs::create_dir_all(parent).with_context(|| format!("Failed to create config dir: {}", parent.display()))?;
|
||||
}
|
||||
let tmp = path.with_extension("tmp");
|
||||
let mut f = fs::File::create(&tmp).with_context(|| format!("Failed to create temp file: {}", tmp.display()))?;
|
||||
let s = toml::to_string_pretty(&toml::Value::Table(tbl.clone()))?;
|
||||
f.write_all(s.as_bytes())?;
|
||||
if !s.ends_with('\n') { f.write_all(b"\n")?; }
|
||||
drop(f);
|
||||
fs::rename(&tmp, path).with_context(|| format!("Failed to atomically replace config: {}", path.display()))?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn system_config_path() -> PathBuf {
|
||||
if cfg!(unix) { PathBuf::from("/etc").join("polyscribe").join("config.toml") } else { default_user_config_path() }
|
||||
}
|
||||
|
||||
fn default_user_config_path() -> PathBuf {
|
||||
if let Some(base) = BaseDirs::new() {
|
||||
return PathBuf::from(base.config_dir()).join("polyscribe").join("config.toml");
|
||||
}
|
||||
PathBuf::from(".polyscribe").join("config.toml")
|
||||
}
|
||||
|
||||
fn workspace_config_path() -> PathBuf {
|
||||
PathBuf::from(".polyscribe").join("config.toml")
|
||||
}
|
||||
|
||||
/// Service responsible for loading and saving PolyScribe configuration
|
||||
#[derive(Debug, Default, Clone)]
|
||||
/// Service for managing Polyscribe configuration
|
||||
///
|
||||
/// Provides functionality to load, save, and access configuration settings
|
||||
/// from disk or environment variables.
|
||||
pub struct ConfigService;
|
||||
|
||||
impl ConfigService {
|
||||
/// Load configuration, merging system < user < workspace < env overrides.
|
||||
pub fn load(&self) -> Result<Config> {
|
||||
let mut accum = toml::value::Table::default();
|
||||
let sys = system_config_path();
|
||||
if sys.exists() {
|
||||
merge_tables(&mut accum, &read_toml(&sys)?);
|
||||
}
|
||||
let user = default_user_config_path();
|
||||
if user.exists() {
|
||||
merge_tables(&mut accum, &read_toml(&user)?);
|
||||
}
|
||||
let ws = workspace_config_path();
|
||||
if ws.exists() {
|
||||
merge_tables(&mut accum, &read_toml(&ws)?);
|
||||
}
|
||||
// Env overrides: POLYSCRIBE__SECTION__KEY=value
|
||||
let mut env_over = toml::value::Table::default();
|
||||
for (k, v) in env::vars() {
|
||||
if let Some(rest) = k.strip_prefix("POLYSCRIBE__") {
|
||||
let parts: Vec<&str> = rest.split("__").collect();
|
||||
if parts.is_empty() { continue; }
|
||||
let val: toml::Value = toml::Value::String(v);
|
||||
// Build nested tables
|
||||
let mut current = &mut env_over;
|
||||
for (i, part) in parts.iter().enumerate() {
|
||||
if i == parts.len() - 1 {
|
||||
current.insert(part.to_lowercase(), val.clone());
|
||||
} else {
|
||||
current = current.entry(part.to_lowercase()).or_insert_with(|| toml::Value::Table(Default::default()))
|
||||
.as_table_mut().expect("table");
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
merge_tables(&mut accum, &env_over);
|
||||
Ok(Config(accum))
|
||||
/// Loads configuration from disk or returns default values if not found
|
||||
///
|
||||
/// This function attempts to read the configuration file from disk. If the file
|
||||
/// doesn't exist or can't be parsed, it falls back to default values.
|
||||
/// Environment variable overrides are then applied to the configuration.
|
||||
pub fn load_or_default() -> Result<Config> {
|
||||
let mut cfg = Self::read_disk().unwrap_or_default();
|
||||
Self::apply_env_overrides(&mut cfg)?;
|
||||
Ok(cfg)
|
||||
}
|
||||
|
||||
/// Ensure user config exists with sensible defaults, return loaded config
|
||||
pub fn ensure_user_config(&self) -> Result<Config> {
|
||||
let path = default_user_config_path();
|
||||
if !path.exists() {
|
||||
let mut defaults = toml::value::Table::default();
|
||||
defaults.insert("ui".into(), toml::Value::Table({
|
||||
let mut t = toml::value::Table::default();
|
||||
t.insert("theme".into(), toml::Value::String("auto".into()));
|
||||
t
|
||||
}));
|
||||
write_toml_atomic(&path, &defaults)?;
|
||||
/// Saves the configuration to disk
|
||||
///
|
||||
/// This function serializes the configuration to TOML format and writes it
|
||||
/// to the standard configuration directory for the application.
|
||||
/// Returns an error if writing fails or if project directories cannot be determined.
|
||||
pub fn save(cfg: &Config) -> Result<()> {
|
||||
let Some(dirs) = Self::dirs() else {
|
||||
return Err(Error::Other("unable to get project dirs".into()));
|
||||
};
|
||||
let cfg_dir = dirs.config_dir();
|
||||
fs::create_dir_all(cfg_dir)?;
|
||||
let path = cfg_dir.join("config.toml");
|
||||
let s = toml::to_string_pretty(cfg)?;
|
||||
fs::write(path, s)?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn read_disk() -> Option<Config> {
|
||||
let dirs = Self::dirs()?;
|
||||
let path = dirs.config_dir().join("config.toml");
|
||||
let s = fs::read_to_string(path).ok()?;
|
||||
toml::from_str(&s).ok()
|
||||
}
|
||||
|
||||
fn apply_env_overrides(cfg: &mut Config) -> Result<()> {
|
||||
// POLYSCRIBE__SECTION__KEY format reserved for future nested config.
|
||||
if let Ok(v) = std::env::var(format!("{ENV_PREFIX}_MODELS_DIR")) {
|
||||
cfg.models_dir = Some(PathBuf::from(v));
|
||||
}
|
||||
self.load()
|
||||
if let Ok(v) = std::env::var(format!("{ENV_PREFIX}_PLUGINS_DIR")) {
|
||||
cfg.plugins_dir = Some(PathBuf::from(v));
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
/// Save to user config atomically, merging over existing user file.
|
||||
pub fn save_user(&self, new_values: &toml::value::Table) -> Result<()> {
|
||||
let path = default_user_config_path();
|
||||
let mut base = if path.exists() { read_toml(&path)? } else { Default::default() };
|
||||
merge_tables(&mut base, new_values);
|
||||
write_toml_atomic(&path, &base)
|
||||
/// Returns the standard project directories for the application
|
||||
///
|
||||
/// This function creates a ProjectDirs instance with the appropriate
|
||||
/// organization and application names for Polyscribe.
|
||||
/// Returns None if the project directories cannot be determined.
|
||||
pub fn dirs() -> Option<ProjectDirs> {
|
||||
ProjectDirs::from("dev", "polyscribe", "polyscribe")
|
||||
}
|
||||
|
||||
/// Paths used for debugging/information
|
||||
pub fn paths(&self) -> (PathBuf, PathBuf, PathBuf) {
|
||||
(system_config_path(), default_user_config_path(), workspace_config_path())
|
||||
/// Returns the default directory path for storing ML models
|
||||
///
|
||||
/// This function determines the standard data directory for the application
|
||||
/// and appends a 'models' subdirectory to it.
|
||||
/// Returns None if the project directories cannot be determined.
|
||||
pub fn default_models_dir() -> Option<PathBuf> {
|
||||
Self::dirs().map(|d| d.data_dir().join("models"))
|
||||
}
|
||||
|
||||
/// Returns the default directory path for storing plugins
|
||||
///
|
||||
/// This function determines the standard data directory for the application
|
||||
/// and appends a 'plugins' subdirectory to it.
|
||||
/// Returns None if the project directories cannot be determined.
|
||||
pub fn default_plugins_dir() -> Option<PathBuf> {
|
||||
Self::dirs().map(|d| d.data_dir().join("plugins"))
|
||||
}
|
||||
}
|
||||
|
39
crates/polyscribe-core/src/error.rs
Normal file
39
crates/polyscribe-core/src/error.rs
Normal file
@@ -0,0 +1,39 @@
|
||||
use thiserror::Error;
|
||||
|
||||
#[derive(Debug, Error)]
|
||||
/// Error types for the polyscribe-core crate.
|
||||
///
|
||||
/// This enum represents various error conditions that can occur during
|
||||
/// operations in this crate, including I/O errors, serialization/deserialization
|
||||
/// errors, and environment variable access errors.
|
||||
pub enum Error {
|
||||
#[error("I/O error: {0}")]
|
||||
/// Represents an I/O error that occurred during file or stream operations
|
||||
Io(#[from] std::io::Error),
|
||||
|
||||
#[error("serde error: {0}")]
|
||||
/// Represents a JSON serialization or deserialization error
|
||||
Serde(#[from] serde_json::Error),
|
||||
|
||||
#[error("toml error: {0}")]
|
||||
/// Represents a TOML deserialization error
|
||||
Toml(#[from] toml::de::Error),
|
||||
|
||||
#[error("toml ser error: {0}")]
|
||||
/// Represents a TOML serialization error
|
||||
TomlSer(#[from] toml::ser::Error),
|
||||
|
||||
#[error("env var error: {0}")]
|
||||
/// Represents an error that occurred during environment variable access
|
||||
EnvVar(#[from] std::env::VarError),
|
||||
|
||||
#[error("other: {0}")]
|
||||
/// Represents a general error condition with a custom message
|
||||
Other(String),
|
||||
}
|
||||
|
||||
impl From<anyhow::Error> for Error {
|
||||
fn from(e: anyhow::Error) -> Self {
|
||||
Error::Other(e.to_string())
|
||||
}
|
||||
}
|
@@ -12,7 +12,16 @@
|
||||
|
||||
use std::sync::atomic::{AtomicBool, AtomicU8, Ordering};
|
||||
|
||||
// Global runtime flags
|
||||
use anyhow::{anyhow, Context, Result};
|
||||
use chrono::Local;
|
||||
use std::env;
|
||||
use std::path::{Path, PathBuf};
|
||||
use std::process::Command;
|
||||
|
||||
#[cfg(unix)]
|
||||
use libc::{O_WRONLY, close, dup, dup2, open};
|
||||
|
||||
/// Global runtime flags
|
||||
static QUIET: AtomicBool = AtomicBool::new(false);
|
||||
static NO_INTERACTION: AtomicBool = AtomicBool::new(false);
|
||||
static VERBOSE: AtomicU8 = AtomicU8::new(0);
|
||||
@@ -101,7 +110,7 @@ impl StderrSilencer {
|
||||
let devnull_cstr = std::ffi::CString::new("/dev/null").unwrap();
|
||||
let devnull_fd = open(devnull_cstr.as_ptr(), O_WRONLY);
|
||||
if devnull_fd < 0 {
|
||||
close(old_fd);
|
||||
let _ = close(old_fd);
|
||||
return Self {
|
||||
active: false,
|
||||
old_stderr_fd: -1,
|
||||
@@ -109,8 +118,8 @@ impl StderrSilencer {
|
||||
};
|
||||
}
|
||||
if dup2(devnull_fd, 2) < 0 {
|
||||
close(devnull_fd);
|
||||
close(old_fd);
|
||||
let _ = close(devnull_fd);
|
||||
let _ = close(old_fd);
|
||||
return Self {
|
||||
active: false,
|
||||
old_stderr_fd: -1,
|
||||
@@ -120,7 +129,7 @@ impl StderrSilencer {
|
||||
Self {
|
||||
active: true,
|
||||
old_stderr_fd: old_fd,
|
||||
devnull_fd: devnull_fd,
|
||||
devnull_fd,
|
||||
}
|
||||
}
|
||||
#[cfg(not(unix))]
|
||||
@@ -183,24 +192,16 @@ macro_rules! qlog {
|
||||
($($arg:tt)*) => {{ $crate::ilog!($($arg)*); }}
|
||||
}
|
||||
|
||||
use anyhow::{Context, Result, anyhow};
|
||||
use chrono::Local;
|
||||
use std::env;
|
||||
use std::fs::create_dir_all;
|
||||
use std::path::{Path, PathBuf};
|
||||
use std::process::Command;
|
||||
|
||||
#[cfg(unix)]
|
||||
use libc::{O_WRONLY, close, dup, dup2, open};
|
||||
|
||||
/// Re-export backend module (GPU/CPU selection and transcription).
|
||||
pub mod backend;
|
||||
/// Re-export models module (model listing/downloading/updating).
|
||||
pub mod models;
|
||||
/// Configuration service (XDG + atomic writes)
|
||||
/// Configuration handling for PolyScribe
|
||||
pub mod config;
|
||||
/// UI helpers
|
||||
// Use the file-backed ui.rs module, which also declares its own `progress` submodule.
|
||||
pub mod ui;
|
||||
/// Error definitions for the PolyScribe library
|
||||
pub mod error;
|
||||
pub use error::Error;
|
||||
pub mod prelude;
|
||||
|
||||
/// Transcript entry for a single segment.
|
||||
#[derive(Debug, serde::Serialize, Clone)]
|
||||
@@ -356,56 +357,77 @@ pub fn normalize_lang_code(input: &str) -> Option<String> {
|
||||
|
||||
/// Find the Whisper model file path to use.
|
||||
pub fn find_model_file() -> Result<PathBuf> {
|
||||
// 1) Explicit override via environment
|
||||
if let Ok(path) = env::var("WHISPER_MODEL") {
|
||||
let p = PathBuf::from(path);
|
||||
if p.exists() {
|
||||
return Ok(p);
|
||||
} else {
|
||||
if !p.exists() {
|
||||
return Err(anyhow!(
|
||||
"WHISPER_MODEL points to non-existing file: {}",
|
||||
"WHISPER_MODEL points to a non-existing path: {}",
|
||||
p.display()
|
||||
));
|
||||
}
|
||||
}
|
||||
let models_dir = models_dir_path();
|
||||
if !models_dir.exists() {
|
||||
create_dir_all(&models_dir).with_context(|| {
|
||||
format!("Failed to create models dir: {}", models_dir.display())
|
||||
})?;
|
||||
if !p.is_file() {
|
||||
return Err(anyhow!(
|
||||
"WHISPER_MODEL must point to a file, but is not: {}",
|
||||
p.display()
|
||||
));
|
||||
}
|
||||
return Ok(p);
|
||||
}
|
||||
|
||||
// Heuristic: prefer larger model files and English-only when language hint is en
|
||||
let mut candidates = Vec::new();
|
||||
for entry in std::fs::read_dir(&models_dir).with_context(|| format!(
|
||||
"Failed to read models dir: {}",
|
||||
models_dir.display()
|
||||
))? {
|
||||
let entry = entry?;
|
||||
let path = entry.path();
|
||||
if !path
|
||||
.extension()
|
||||
.and_then(|s| s.to_str())
|
||||
.is_some_and(|s| s.eq_ignore_ascii_case("bin"))
|
||||
{
|
||||
continue;
|
||||
}
|
||||
if let Ok(md) = std::fs::metadata(&path) {
|
||||
candidates.push((md.len(), path));
|
||||
}
|
||||
}
|
||||
if candidates.is_empty() {
|
||||
// Try default fallback (tiny.en)
|
||||
let fallback = models_dir.join("ggml-tiny.en.bin");
|
||||
if fallback.exists() {
|
||||
return Ok(fallback);
|
||||
}
|
||||
// 2) Resolve models directory and ensure it exists and is a directory
|
||||
let models_dir = models_dir_path();
|
||||
if models_dir.exists() && !models_dir.is_dir() {
|
||||
return Err(anyhow!(
|
||||
"No Whisper models found in {}. Please download a model or set WHISPER_MODEL.",
|
||||
"Models path exists but is not a directory: {}",
|
||||
models_dir.display()
|
||||
));
|
||||
}
|
||||
std::fs::create_dir_all(&models_dir).with_context(|| {
|
||||
format!("Failed to ensure models dir exists: {}", models_dir.display())
|
||||
})?;
|
||||
|
||||
// 3) Gather candidate .bin files (regular files only), prefer largest
|
||||
let mut candidates = Vec::new();
|
||||
for entry in std::fs::read_dir(&models_dir).with_context(|| {
|
||||
format!("Failed to read models dir: {}", models_dir.display())
|
||||
})? {
|
||||
let entry = entry?;
|
||||
let path = entry.path();
|
||||
|
||||
// Only consider .bin files
|
||||
let is_bin = path
|
||||
.extension()
|
||||
.and_then(|s| s.to_str())
|
||||
.is_some_and(|s| s.eq_ignore_ascii_case("bin"));
|
||||
if !is_bin {
|
||||
continue;
|
||||
}
|
||||
|
||||
// Only consider regular files
|
||||
let md = match std::fs::metadata(&path) {
|
||||
Ok(m) if m.is_file() => m,
|
||||
_ => continue,
|
||||
};
|
||||
|
||||
candidates.push((md.len(), path));
|
||||
}
|
||||
|
||||
if candidates.is_empty() {
|
||||
// 4) Fallback to known tiny English model if present
|
||||
let fallback = models_dir.join("ggml-tiny.en.bin");
|
||||
if fallback.is_file() {
|
||||
return Ok(fallback);
|
||||
}
|
||||
return Err(anyhow!(
|
||||
"No Whisper model files (*.bin) found in {}. \
|
||||
Please download a model or set WHISPER_MODEL.",
|
||||
models_dir.display()
|
||||
));
|
||||
}
|
||||
|
||||
candidates.sort_by_key(|(size, _)| *size);
|
||||
let (_size, path) = candidates.into_iter().last().unwrap();
|
||||
let (_size, path) = candidates.into_iter().last().expect("non-empty");
|
||||
Ok(path)
|
||||
}
|
||||
|
||||
@@ -414,12 +436,14 @@ pub fn decode_audio_to_pcm_f32_ffmpeg(audio_path: &Path) -> Result<Vec<f32>> {
|
||||
let in_path = audio_path
|
||||
.to_str()
|
||||
.ok_or_else(|| anyhow!("Audio path must be valid UTF-8: {}", audio_path.display()))?;
|
||||
let tmp_wav = std::env::temp_dir().join("polyscribe_tmp_input.wav");
|
||||
let tmp_wav_str = tmp_wav
|
||||
.to_str()
|
||||
.ok_or_else(|| anyhow!("Temp path not valid UTF-8: {}", tmp_wav.display()))?;
|
||||
|
||||
// ffmpeg -i input -f f32le -ac 1 -ar 16000 -y /tmp/tmp.raw
|
||||
// Use a raw f32le file to match the -f f32le output format.
|
||||
let tmp_raw = std::env::temp_dir().join("polyscribe_tmp_input.f32le");
|
||||
let tmp_raw_str = tmp_raw
|
||||
.to_str()
|
||||
.ok_or_else(|| anyhow!("Temp path not valid UTF-8: {}", tmp_raw.display()))?;
|
||||
|
||||
// ffmpeg -i input -f f32le -ac 1 -ar 16000 -y /tmp/tmp.f32le
|
||||
let status = Command::new("ffmpeg")
|
||||
.arg("-hide_banner")
|
||||
.arg("-loglevel")
|
||||
@@ -433,16 +457,29 @@ pub fn decode_audio_to_pcm_f32_ffmpeg(audio_path: &Path) -> Result<Vec<f32>> {
|
||||
.arg("-ar")
|
||||
.arg("16000")
|
||||
.arg("-y")
|
||||
.arg(&tmp_wav_str)
|
||||
.arg(tmp_raw_str)
|
||||
.status()
|
||||
.with_context(|| format!("Failed to invoke ffmpeg to decode: {}", in_path))?;
|
||||
|
||||
if !status.success() {
|
||||
return Err(anyhow!("ffmpeg exited with non-zero status when decoding {}", in_path));
|
||||
return Err(anyhow!(
|
||||
"ffmpeg exited with non-zero status when decoding {}",
|
||||
in_path
|
||||
));
|
||||
}
|
||||
let raw = std::fs::read(&tmp_wav).with_context(|| format!("Failed to read temp PCM file: {}", tmp_wav.display()))?;
|
||||
|
||||
let raw = std::fs::read(&tmp_raw)
|
||||
.with_context(|| format!("Failed to read temp PCM file: {}", tmp_raw.display()))?;
|
||||
|
||||
// Best-effort cleanup of the temp file
|
||||
let _ = std::fs::remove_file(&tmp_raw);
|
||||
|
||||
// Interpret raw bytes as f32 little-endian
|
||||
if raw.len() % 4 != 0 {
|
||||
return Err(anyhow!("Decoded PCM file length not multiple of 4: {}", raw.len()));
|
||||
return Err(anyhow!(
|
||||
"Decoded PCM file length not multiple of 4: {}",
|
||||
raw.len()
|
||||
));
|
||||
}
|
||||
let mut samples = Vec::with_capacity(raw.len() / 4);
|
||||
for chunk in raw.chunks_exact(4) {
|
||||
|
@@ -88,9 +88,10 @@ pub fn run_interactive_model_downloader() -> Result<()> {
|
||||
}
|
||||
|
||||
let answer = ui::prompt_input("Your selection", Some("1"))?;
|
||||
let selection_raw = match answer {
|
||||
Some(s) => s.trim().to_string(),
|
||||
None => "1".to_string(),
|
||||
let selection_raw = if answer.trim().is_empty() {
|
||||
"1".to_string()
|
||||
} else {
|
||||
answer.trim().to_string()
|
||||
};
|
||||
let selection = if selection_raw.is_empty() { "1" } else { &selection_raw };
|
||||
|
||||
|
16
crates/polyscribe-core/src/prelude.rs
Normal file
16
crates/polyscribe-core/src/prelude.rs
Normal file
@@ -0,0 +1,16 @@
|
||||
// rust
|
||||
//! Commonly used exports for convenient glob-imports in binaries and tests.
|
||||
//! Usage: `use polyscribe_core::prelude::*;`
|
||||
|
||||
pub use crate::backend::*;
|
||||
pub use crate::config::*;
|
||||
pub use crate::error::Error;
|
||||
pub use crate::models::*;
|
||||
|
||||
// If you frequently use UI helpers across binaries/tests, export them too.
|
||||
// Keep this lean to avoid pulling UI everywhere unintentionally.
|
||||
#[allow(unused_imports)]
|
||||
pub use crate::ui::*;
|
||||
|
||||
/// A convenient alias for `std::result::Result` with the error type defaulting to [`Error`].
|
||||
pub type Result<T, E = Error> = std::result::Result<T, E>;
|
@@ -1,87 +1,64 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
|
||||
//! Centralized UI helpers (TTY-aware, quiet/verbose-aware)
|
||||
//! Minimal UI helpers used across the core crate.
|
||||
//! This keeps interactive bits centralized and easy to stub in tests.
|
||||
|
||||
use std::io;
|
||||
/// Progress indicators and reporting tools for displaying task completion.
|
||||
pub mod progress;
|
||||
|
||||
/// Startup intro/banner (suppressed when quiet).
|
||||
pub fn intro(msg: impl AsRef<str>) {
|
||||
let _ = cliclack::intro(msg.as_ref());
|
||||
}
|
||||
use std::io::{self, Write};
|
||||
|
||||
/// Final outro/summary printed below any progress indicators (suppressed when quiet).
|
||||
pub fn outro(msg: impl AsRef<str>) {
|
||||
let _ = cliclack::outro(msg.as_ref());
|
||||
}
|
||||
|
||||
/// Info message (TTY-aware; suppressed by --quiet is handled by outer callers if needed)
|
||||
/// Print an informational line to stderr (suppressed when quiet mode is enabled by callers).
|
||||
pub fn info(msg: impl AsRef<str>) {
|
||||
let _ = cliclack::log::info(msg.as_ref());
|
||||
eprintln!("{}", msg.as_ref());
|
||||
}
|
||||
|
||||
/// Print a warning (always printed).
|
||||
/// Print a warning line to stderr.
|
||||
pub fn warn(msg: impl AsRef<str>) {
|
||||
// cliclack provides a warning-level log utility
|
||||
let _ = cliclack::log::warning(msg.as_ref());
|
||||
eprintln!("WARNING: {}", msg.as_ref());
|
||||
}
|
||||
|
||||
/// Print an error (always printed).
|
||||
/// Print an error line to stderr.
|
||||
pub fn error(msg: impl AsRef<str>) {
|
||||
let _ = cliclack::log::error(msg.as_ref());
|
||||
eprintln!("ERROR: {}", msg.as_ref());
|
||||
}
|
||||
|
||||
/// Print a line above any progress bars (maps to cliclack log; synchronized).
|
||||
pub fn println_above_bars(msg: impl AsRef<str>) {
|
||||
if crate::is_quiet() { return; }
|
||||
// cliclack logs are synchronized with its spinners/bars
|
||||
let _ = cliclack::log::info(msg.as_ref());
|
||||
/// Print a short intro header (non-fancy).
|
||||
pub fn intro(title: impl AsRef<str>) {
|
||||
eprintln!("== {} ==", title.as_ref());
|
||||
}
|
||||
|
||||
/// Input prompt with a question: returns Ok(None) if non-interactive or canceled
|
||||
pub fn prompt_input(question: impl AsRef<str>, default: Option<&str>) -> anyhow::Result<Option<String>> {
|
||||
if crate::is_no_interaction() || !crate::stdin_is_tty() {
|
||||
return Ok(None);
|
||||
}
|
||||
let mut p = cliclack::input(question.as_ref());
|
||||
if let Some(d) = default {
|
||||
// Use default_input when available in 0.3.x
|
||||
p = p.default_input(d);
|
||||
}
|
||||
match p.interact() {
|
||||
Ok(s) => Ok(Some(s)),
|
||||
Err(_) => Ok(None),
|
||||
}
|
||||
/// Print a short outro footer (non-fancy).
|
||||
pub fn outro(msg: impl AsRef<str>) {
|
||||
eprintln!("{}", msg.as_ref());
|
||||
}
|
||||
|
||||
/// Confirmation prompt; returns Ok(None) if non-interactive or canceled
|
||||
pub fn prompt_confirm(question: impl AsRef<str>, default_yes: bool) -> anyhow::Result<Option<bool>> {
|
||||
if crate::is_no_interaction() || !crate::stdin_is_tty() {
|
||||
return Ok(None);
|
||||
}
|
||||
let res = cliclack::confirm(question.as_ref())
|
||||
.initial_value(default_yes)
|
||||
.interact();
|
||||
match res {
|
||||
Ok(v) => Ok(Some(v)),
|
||||
Err(_) => Ok(None),
|
||||
}
|
||||
/// Print a line that should appear above any progress indicators (plain for now).
|
||||
pub fn println_above_bars(line: impl AsRef<str>) {
|
||||
eprintln!("{}", line.as_ref());
|
||||
}
|
||||
|
||||
/// Prompt the user (TTY-aware via cliclack) and read a line from stdin. Returns the raw line with trailing newline removed.
|
||||
pub fn prompt_line(prompt: &str) -> io::Result<String> {
|
||||
// Route prompt through cliclack to keep consistent styling and avoid direct eprint!/println!
|
||||
let _ = cliclack::log::info(prompt);
|
||||
let mut s = String::new();
|
||||
io::stdin().read_line(&mut s)?;
|
||||
Ok(s)
|
||||
}
|
||||
/// Prompt for input on stdin. Returns default if provided and user enters empty string.
|
||||
/// In non-interactive workflows, callers should skip prompt based on their flags.
|
||||
pub fn prompt_input(prompt: &str, default: Option<&str>) -> io::Result<String> {
|
||||
let mut stdout = io::stdout();
|
||||
match default {
|
||||
Some(def) => {
|
||||
write!(stdout, "{} [{}]: ", prompt, def)?;
|
||||
}
|
||||
None => {
|
||||
write!(stdout, "{}: ", prompt)?;
|
||||
}
|
||||
}
|
||||
stdout.flush()?;
|
||||
|
||||
/// TTY-aware progress UI built on `indicatif` for per-file and aggregate progress bars.
|
||||
///
|
||||
/// This small helper encapsulates a `MultiProgress` with one aggregate (total) bar and
|
||||
/// one per-file bar. It is intentionally minimal to keep integration lightweight.
|
||||
pub mod progress {
|
||||
// The submodule is defined in a separate file for clarity.
|
||||
include!("ui/progress.rs");
|
||||
let mut buf = String::new();
|
||||
io::stdin().read_line(&mut buf)?;
|
||||
let trimmed = buf.trim();
|
||||
if trimmed.is_empty() {
|
||||
Ok(default.unwrap_or_default().to_string())
|
||||
} else {
|
||||
Ok(trimmed.to_string())
|
||||
}
|
||||
}
|
||||
|
@@ -79,3 +79,47 @@ impl ProgressManager {
|
||||
if let Some(total) = &self.total { total.set_position(self.completed as u64); }
|
||||
}
|
||||
}
|
||||
|
||||
/// A simple reporter for displaying progress messages in the terminal.
|
||||
/// Provides different output formatting based on whether the environment is interactive or not.
|
||||
#[derive(Debug)]
|
||||
pub struct ProgressReporter {
|
||||
non_interactive: bool,
|
||||
}
|
||||
|
||||
impl ProgressReporter {
|
||||
/// Creates a new progress reporter.
|
||||
///
|
||||
/// # Arguments
|
||||
///
|
||||
/// * `non_interactive` - Whether the output should be formatted for non-interactive environments.
|
||||
pub fn new(non_interactive: bool) -> Self {
|
||||
Self { non_interactive }
|
||||
}
|
||||
|
||||
/// Displays a progress step message.
|
||||
///
|
||||
/// # Arguments
|
||||
///
|
||||
/// * `message` - The message to display for this progress step.
|
||||
pub fn step(&mut self, message: &str) {
|
||||
if self.non_interactive {
|
||||
eprintln!("[..] {message}");
|
||||
} else {
|
||||
eprintln!("• {message}");
|
||||
}
|
||||
}
|
||||
|
||||
/// Displays a completion message.
|
||||
///
|
||||
/// # Arguments
|
||||
///
|
||||
/// * `message` - The message to display when a task is completed.
|
||||
pub fn finish_with_message(&mut self, message: &str) {
|
||||
if self.non_interactive {
|
||||
eprintln!("[ok] {message}");
|
||||
} else {
|
||||
eprintln!("✓ {message}");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@@ -2,16 +2,10 @@
|
||||
name = "polyscribe-host"
|
||||
version = "0.1.0"
|
||||
edition = "2024"
|
||||
license = "MIT"
|
||||
|
||||
[dependencies]
|
||||
anyhow = "1.0.98"
|
||||
thiserror = "1"
|
||||
anyhow = "1.0.99"
|
||||
serde = { version = "1.0.219", features = ["derive"] }
|
||||
serde_json = "1.0.142"
|
||||
tokio = { version = "1", features = ["full"] }
|
||||
which = "6"
|
||||
cliclack = "0.3"
|
||||
directories = "5"
|
||||
polyscribe = { path = "../polyscribe-core" }
|
||||
polyscribe-protocol = { path = "../polyscribe-protocol" }
|
||||
tokio = { version = "1.47.1", features = ["rt-multi-thread", "process", "io-util"] }
|
||||
which = "6.0.3"
|
||||
|
@@ -1,168 +1,118 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
|
||||
use anyhow::{anyhow, Context, Result};
|
||||
use cliclack as ui; // reuse for minimal logging
|
||||
use directories::BaseDirs;
|
||||
use serde_json::Value;
|
||||
use std::collections::BTreeMap;
|
||||
use std::ffi::OsStr;
|
||||
use std::fs;
|
||||
use std::io::{BufRead, BufReader, Write};
|
||||
use std::path::{Path, PathBuf};
|
||||
use std::process::{Command, Stdio};
|
||||
|
||||
use polyscribe_protocol as psp;
|
||||
use anyhow::{Context, Result};
|
||||
use serde::Deserialize;
|
||||
use std::{
|
||||
env,
|
||||
fs,
|
||||
os::unix::fs::PermissionsExt,
|
||||
path::Path,
|
||||
};
|
||||
use tokio::{
|
||||
io::{AsyncBufReadExt, BufReader},
|
||||
process::{Child as TokioChild, Command},
|
||||
};
|
||||
use std::process::Stdio;
|
||||
|
||||
#[derive(Debug, Clone)]
|
||||
pub struct Plugin {
|
||||
pub struct PluginInfo {
|
||||
pub name: String,
|
||||
pub path: PathBuf,
|
||||
pub path: String,
|
||||
}
|
||||
|
||||
/// Discover plugins on PATH and in the user's data dir (XDG) under polyscribe/plugins.
|
||||
pub fn discover() -> Result<Vec<Plugin>> {
|
||||
let mut found: BTreeMap<String, PathBuf> = BTreeMap::new();
|
||||
#[derive(Debug, Default)]
|
||||
pub struct PluginManager;
|
||||
|
||||
// Scan PATH directories
|
||||
if let Some(path_var) = std::env::var_os("PATH") {
|
||||
for dir in std::env::split_paths(&path_var) {
|
||||
if dir.as_os_str().is_empty() { continue; }
|
||||
if let Ok(rd) = fs::read_dir(&dir) {
|
||||
for ent in rd.flatten() {
|
||||
let p = ent.path();
|
||||
if !is_executable(&p) { continue; }
|
||||
if let Some(fname) = p.file_name().and_then(OsStr::to_str) {
|
||||
if let Some(name) = fname.strip_prefix("polyscribe-plugin-") {
|
||||
found.entry(name.to_string()).or_insert(p);
|
||||
impl PluginManager {
|
||||
pub fn list(&self) -> Result<Vec<PluginInfo>> {
|
||||
let mut plugins = Vec::new();
|
||||
|
||||
// Scan PATH entries for executables starting with "polyscribe-plugin-"
|
||||
if let Ok(path) = env::var("PATH") {
|
||||
for dir in env::split_paths(&path) {
|
||||
if let Ok(read_dir) = fs::read_dir(&dir) {
|
||||
for entry in read_dir.flatten() {
|
||||
let path = entry.path();
|
||||
if let Some(fname) = path.file_name().and_then(|s| s.to_str()) {
|
||||
if fname.starts_with("polyscribe-plugin-") && is_executable(&path) {
|
||||
let name = fname.trim_start_matches("polyscribe-plugin-").to_string();
|
||||
plugins.push(PluginInfo {
|
||||
name,
|
||||
path: path.to_string_lossy().to_string(),
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// TODO: also scan XDG data plugins dir for symlinks/binaries
|
||||
Ok(plugins)
|
||||
}
|
||||
|
||||
// Scan user data dir
|
||||
if let Some(base) = BaseDirs::new() {
|
||||
let user_plugins = PathBuf::from(base.data_dir()).join("polyscribe").join("plugins");
|
||||
if let Ok(rd) = fs::read_dir(&user_plugins) {
|
||||
for ent in rd.flatten() {
|
||||
let p = ent.path();
|
||||
if !is_executable(&p) { continue; }
|
||||
if let Some(fname) = p.file_name().and_then(OsStr::to_str) {
|
||||
let name = fname.strip_prefix("polyscribe-plugin-")
|
||||
.map(|s| s.to_string())
|
||||
.or_else(|| Some(fname.to_string()))
|
||||
.unwrap();
|
||||
found.entry(name).or_insert(p);
|
||||
}
|
||||
pub fn info(&self, name: &str) -> Result<serde_json::Value> {
|
||||
let bin = self.resolve(name)?;
|
||||
let out = std::process::Command::new(&bin)
|
||||
.arg("info")
|
||||
.stdout(Stdio::piped())
|
||||
.stderr(Stdio::inherit())
|
||||
.spawn()
|
||||
.context("spawning plugin info")?
|
||||
.wait_with_output()
|
||||
.context("waiting for plugin info")?;
|
||||
|
||||
let val: serde_json::Value =
|
||||
serde_json::from_slice(&out.stdout).context("parsing plugin info JSON")?;
|
||||
Ok(val)
|
||||
}
|
||||
|
||||
pub fn spawn(&self, name: &str, command: &str) -> Result<TokioChild> {
|
||||
let bin = self.resolve(name)?;
|
||||
let mut cmd = Command::new(&bin);
|
||||
cmd.arg("run")
|
||||
.arg(command)
|
||||
.stdin(Stdio::piped())
|
||||
.stdout(Stdio::piped())
|
||||
.stderr(Stdio::inherit());
|
||||
let child = cmd.spawn().context("spawning plugin run")?;
|
||||
Ok(child)
|
||||
}
|
||||
|
||||
pub async fn forward_stdio(&self, child: &mut TokioChild) -> Result<std::process::ExitStatus> {
|
||||
if let Some(stdout) = child.stdout.take() {
|
||||
let mut reader = BufReader::new(stdout).lines();
|
||||
while let Some(line) = reader.next_line().await? {
|
||||
println!("{line}");
|
||||
}
|
||||
}
|
||||
Ok(child.wait().await?)
|
||||
}
|
||||
|
||||
Ok(found
|
||||
.into_iter()
|
||||
.map(|(name, path)| Plugin { name, path })
|
||||
.collect())
|
||||
fn resolve(&self, name: &str) -> Result<String> {
|
||||
let bin = format!("polyscribe-plugin-{name}");
|
||||
let path = which::which(&bin).with_context(|| format!("plugin not found in PATH: {bin}"))?;
|
||||
Ok(path.to_string_lossy().to_string())
|
||||
}
|
||||
}
|
||||
|
||||
fn is_executable(p: &Path) -> bool {
|
||||
if !p.is_file() { return false; }
|
||||
fn is_executable(path: &Path) -> bool {
|
||||
if !path.is_file() {
|
||||
return false;
|
||||
}
|
||||
#[cfg(unix)]
|
||||
{
|
||||
use std::os::unix::fs::PermissionsExt;
|
||||
if let Ok(md) = fs::metadata(p) {
|
||||
let mode = md.permissions().mode();
|
||||
return (mode & 0o111) != 0;
|
||||
}
|
||||
false
|
||||
}
|
||||
#[cfg(not(unix))]
|
||||
{
|
||||
// On Windows, consider .exe, .bat, .cmd
|
||||
matches!(p.extension().and_then(|s| s.to_str()).map(|s| s.to_lowercase()), Some(ext) if matches!(ext.as_str(), "exe"|"bat"|"cmd"))
|
||||
}
|
||||
}
|
||||
|
||||
/// Query plugin capabilities by invoking `--capabilities`.
|
||||
pub fn capabilities(plugin_path: &Path) -> Result<psp::Capabilities> {
|
||||
let out = Command::new(plugin_path)
|
||||
.arg("--capabilities")
|
||||
.stdout(Stdio::piped())
|
||||
.stderr(Stdio::null())
|
||||
.output()
|
||||
.with_context(|| format!("Failed to execute plugin: {}", plugin_path.display()))?;
|
||||
if !out.status.success() {
|
||||
return Err(anyhow!("Plugin --capabilities failed: {}", plugin_path.display()));
|
||||
}
|
||||
let s = String::from_utf8(out.stdout).context("capabilities stdout not utf-8")?;
|
||||
let caps: psp::Capabilities = serde_json::from_str(s.trim()).context("invalid capabilities JSON")?;
|
||||
Ok(caps)
|
||||
}
|
||||
|
||||
/// Run a single method via `--serve`, writing one JSON-RPC request and streaming until result.
|
||||
pub fn run_method<F>(plugin_path: &Path, method: &str, params: Value, mut on_progress: F) -> Result<Value>
|
||||
where
|
||||
F: FnMut(psp::Progress),
|
||||
{
|
||||
let mut child = Command::new(plugin_path)
|
||||
.arg("--serve")
|
||||
.stdin(Stdio::piped())
|
||||
.stdout(Stdio::piped())
|
||||
.stderr(Stdio::null())
|
||||
.spawn()
|
||||
.with_context(|| format!("Failed to spawn plugin: {}", plugin_path.display()))?;
|
||||
|
||||
let mut stdin = child.stdin.take().ok_or_else(|| anyhow!("failed to open plugin stdin"))?;
|
||||
let stdout = child.stdout.take().ok_or_else(|| anyhow!("failed to open plugin stdout"))?;
|
||||
|
||||
// Send request line
|
||||
let req = psp::JsonRpcRequest { jsonrpc: "2.0".into(), id: "1".into(), method: method.to_string(), params: Some(params) };
|
||||
let line = serde_json::to_string(&req)? + "\n";
|
||||
stdin.write_all(line.as_bytes())?;
|
||||
stdin.flush()?;
|
||||
|
||||
// Read response lines
|
||||
let reader = BufReader::new(stdout);
|
||||
for line in reader.lines() {
|
||||
let line = line?;
|
||||
if line.trim().is_empty() { continue; }
|
||||
// Try parse StreamItem; if that fails, try parse JsonRpcResponse directly
|
||||
if let Ok(item) = serde_json::from_str::<psp::StreamItem>(&line) {
|
||||
match item {
|
||||
psp::StreamItem::Progress(p) => {
|
||||
on_progress(p);
|
||||
}
|
||||
psp::StreamItem::Result(resp) => {
|
||||
match resp.outcome {
|
||||
psp::JsonRpcOutcome::Ok { result } => return Ok(result),
|
||||
psp::JsonRpcOutcome::Err { error } => return Err(anyhow!("{} ({})", error.message, error.code)),
|
||||
}
|
||||
}
|
||||
}
|
||||
} else if let Ok(resp) = serde_json::from_str::<psp::JsonRpcResponse>(&line) {
|
||||
match resp.outcome {
|
||||
psp::JsonRpcOutcome::Ok { result } => return Ok(result),
|
||||
psp::JsonRpcOutcome::Err { error } => return Err(anyhow!("{} ({})", error.message, error.code)),
|
||||
}
|
||||
} else {
|
||||
let _ = ui::log::warning(format!("Unrecognized plugin output: {}", line));
|
||||
if let Ok(meta) = fs::metadata(path) {
|
||||
let mode = meta.permissions().mode();
|
||||
// if any execute bit is set
|
||||
return mode & 0o111 != 0;
|
||||
}
|
||||
}
|
||||
|
||||
// If we exited loop without returning, wait for child
|
||||
let status = child.wait()?;
|
||||
if status.success() {
|
||||
Err(anyhow!("Plugin terminated without sending a result"))
|
||||
} else {
|
||||
Err(anyhow!("Plugin exited with status: {:?}", status))
|
||||
}
|
||||
// Fallback for non-unix (treat files as candidates)
|
||||
true
|
||||
}
|
||||
|
||||
/// Helper: find a plugin by name using discovery
|
||||
pub fn find_plugin_by_name(name: &str) -> Result<Plugin> {
|
||||
let plugins = discover()?;
|
||||
plugins
|
||||
.into_iter()
|
||||
.find(|p| p.name == name)
|
||||
.ok_or_else(|| anyhow!("Plugin '{}' not found", name))
|
||||
#[allow(dead_code)]
|
||||
#[derive(Debug, Deserialize)]
|
||||
struct Capability {
|
||||
command: String,
|
||||
summary: String,
|
||||
}
|
||||
|
@@ -2,9 +2,7 @@
|
||||
name = "polyscribe-protocol"
|
||||
version = "0.1.0"
|
||||
edition = "2024"
|
||||
license = "MIT"
|
||||
|
||||
[dependencies]
|
||||
serde = { version = "1.0.219", features = ["derive"] }
|
||||
serde_json = "1.0.142"
|
||||
thiserror = "1"
|
||||
|
@@ -1,90 +1,60 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// PolyScribe Protocol (PSP/1): JSON-RPC 2.0 over NDJSON on stdio
|
||||
|
||||
use serde::{Deserialize, Serialize};
|
||||
use serde_json::Value;
|
||||
|
||||
/// Plugin capabilities as reported by `--capabilities`.
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct Capabilities {
|
||||
pub name: String,
|
||||
pub version: String,
|
||||
/// Protocol identifier (e.g., "psp/1")
|
||||
pub protocol: String,
|
||||
/// Role (e.g., pipeline, tool, generator)
|
||||
pub role: String,
|
||||
/// Supported command names
|
||||
pub commands: Vec<String>,
|
||||
}
|
||||
|
||||
/// Generic JSON-RPC 2.0 request for PSP/1
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct JsonRpcRequest {
|
||||
pub jsonrpc: String, // "2.0"
|
||||
#[derive(Debug, Serialize, Deserialize)]
|
||||
pub struct Request {
|
||||
pub id: String,
|
||||
pub method: String,
|
||||
#[serde(skip_serializing_if = "Option::is_none")]
|
||||
pub params: Option<serde_json::Value>,
|
||||
pub params: Option<Value>,
|
||||
}
|
||||
|
||||
/// Error object for JSON-RPC 2.0
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct JsonRpcError {
|
||||
pub code: i64,
|
||||
#[derive(Debug, Serialize, Deserialize)]
|
||||
pub struct Response {
|
||||
pub id: String,
|
||||
pub result: Option<Value>,
|
||||
pub error: Option<ErrorObj>,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize, Deserialize)]
|
||||
pub struct ErrorObj {
|
||||
pub code: i32,
|
||||
pub message: String,
|
||||
#[serde(skip_serializing_if = "Option::is_none")]
|
||||
pub data: Option<serde_json::Value>,
|
||||
pub data: Option<Value>,
|
||||
}
|
||||
|
||||
/// Generic JSON-RPC 2.0 response for PSP/1
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
#[serde(tag = "kind", rename_all = "lowercase")]
|
||||
pub enum StreamItem {
|
||||
/// Progress notification (out-of-band in stream, not a JSON-RPC response)
|
||||
Progress(Progress),
|
||||
/// A proper JSON-RPC response with a result
|
||||
Result(JsonRpcResponse),
|
||||
#[derive(Debug, Serialize, Deserialize)]
|
||||
#[serde(tag = "event", content = "data")]
|
||||
pub enum ProgressEvent {
|
||||
Started,
|
||||
Message(String),
|
||||
Percent(f32),
|
||||
Finished,
|
||||
}
|
||||
|
||||
/// JSON-RPC 2.0 Response envelope containing either result or error.
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct JsonRpcResponse {
|
||||
pub jsonrpc: String, // "2.0"
|
||||
pub id: String,
|
||||
#[serde(flatten)]
|
||||
pub outcome: JsonRpcOutcome,
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
#[serde(untagged)]
|
||||
pub enum JsonRpcOutcome {
|
||||
Ok { result: serde_json::Value },
|
||||
Err { error: JsonRpcError },
|
||||
}
|
||||
|
||||
/// Progress event structure for PSP/1 streaming
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct Progress {
|
||||
/// 0..=100
|
||||
pub pct: u8,
|
||||
/// Short phase name
|
||||
pub stage: Option<String>,
|
||||
/// Human-friendly detail
|
||||
pub message: Option<String>,
|
||||
}
|
||||
|
||||
/// Convenience helpers to build items
|
||||
impl StreamItem {
|
||||
pub fn progress(pct: u8, stage: impl Into<Option<String>>, message: impl Into<Option<String>>) -> Self {
|
||||
StreamItem::Progress(Progress { pct, stage: stage.into(), message: message.into() })
|
||||
}
|
||||
pub fn ok(id: impl Into<String>, result: serde_json::Value) -> Self {
|
||||
StreamItem::Result(JsonRpcResponse { jsonrpc: "2.0".into(), id: id.into(), outcome: JsonRpcOutcome::Ok { result } })
|
||||
}
|
||||
pub fn err(id: impl Into<String>, code: i64, message: impl Into<String>, data: Option<serde_json::Value>) -> Self {
|
||||
StreamItem::Result(JsonRpcResponse {
|
||||
jsonrpc: "2.0".into(),
|
||||
impl Response {
|
||||
pub fn ok(id: impl Into<String>, result: Value) -> Self {
|
||||
Self {
|
||||
id: id.into(),
|
||||
outcome: JsonRpcOutcome::Err { error: JsonRpcError { code, message: message.into(), data } },
|
||||
})
|
||||
result: Some(result),
|
||||
error: None,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn err(
|
||||
id: impl Into<String>,
|
||||
code: i32,
|
||||
message: impl Into<String>,
|
||||
data: Option<Value>,
|
||||
) -> Self {
|
||||
Self {
|
||||
id: id.into(),
|
||||
result: None,
|
||||
error: Some(ErrorObj {
|
||||
code,
|
||||
message: message.into(),
|
||||
data,
|
||||
}),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
26
docs/ci.md
26
docs/ci.md
@@ -1,26 +0,0 @@
|
||||
# CI checklist and job outline
|
||||
|
||||
Checklist to keep docs and code healthy in CI
|
||||
- Build: cargo build --all-targets --locked
|
||||
- Tests: cargo test --all --locked
|
||||
- Lints: cargo clippy --all-targets -- -D warnings
|
||||
- Optional: check README and docs snippets (basic smoke run of examples scripts)
|
||||
- bash examples/update_models.sh (can be skipped offline)
|
||||
- bash examples/transcribe_file.sh (use a tiny sample file if available)
|
||||
|
||||
Example GitHub Actions job (outline)
|
||||
- name: Rust
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: dtolnay/rust-toolchain@stable
|
||||
- name: Build
|
||||
run: cargo build --all-targets --locked
|
||||
- name: Test
|
||||
run: cargo test --all --locked
|
||||
- name: Clippy
|
||||
run: cargo clippy --all-targets -- -D warnings
|
||||
|
||||
Notes
|
||||
- For GPU features, set up appropriate runners and add `--features gpu-cuda|gpu-hip|gpu-vulkan` where applicable.
|
||||
- For docs-only changes, jobs still build/test to ensure doctests and examples compile when enabled.
|
@@ -1,13 +0,0 @@
|
||||
#!/usr/bin/env bash
|
||||
# SPDX-License-Identifier: MIT
|
||||
# Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
set -euo pipefail
|
||||
|
||||
# Launch the interactive model downloader and select models to install
|
||||
|
||||
BIN=${BIN:-./target/release/polyscribe}
|
||||
MODELS_DIR=${POLYSCRIBE_MODELS_DIR:-$PWD/models}
|
||||
export POLYSCRIBE_MODELS_DIR="$MODELS_DIR"
|
||||
|
||||
mkdir -p "$MODELS_DIR"
|
||||
"$BIN" --download-models
|
@@ -1,15 +0,0 @@
|
||||
#!/usr/bin/env bash
|
||||
# SPDX-License-Identifier: MIT
|
||||
# Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
set -euo pipefail
|
||||
|
||||
# Transcribe an audio/video file to JSON and SRT into ./output
|
||||
# Requires a model; first run may prompt to download.
|
||||
|
||||
BIN=${BIN:-./target/release/polyscribe}
|
||||
INPUT=${1:-samples/example.mp3}
|
||||
OUTDIR=${OUTDIR:-output}
|
||||
|
||||
mkdir -p "$OUTDIR"
|
||||
"$BIN" -v -o "$OUTDIR" "$INPUT"
|
||||
echo "Done. See $OUTDIR for JSON/SRT files."
|
@@ -1,15 +0,0 @@
|
||||
#!/usr/bin/env bash
|
||||
# SPDX-License-Identifier: MIT
|
||||
# Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
set -euo pipefail
|
||||
|
||||
# Verify/update local models non-interactively (useful in CI)
|
||||
|
||||
BIN=${BIN:-./target/release/polyscribe}
|
||||
MODELS_DIR=${POLYSCRIBE_MODELS_DIR:-$PWD/models}
|
||||
export POLYSCRIBE_MODELS_DIR="$MODELS_DIR"
|
||||
|
||||
mkdir -p "$MODELS_DIR"
|
||||
"$BIN" --update-models --no-interaction -q
|
||||
|
||||
echo "Models updated in $MODELS_DIR"
|
6
rust-toolchain.toml
Normal file
6
rust-toolchain.toml
Normal file
@@ -0,0 +1,6 @@
|
||||
# SPDX-License-Identifier: MIT
|
||||
|
||||
[toolchain]
|
||||
channel = "1.89.0"
|
||||
components = ["clippy", "rustfmt"]
|
||||
profile = "minimal"
|
329
src/backend.rs
329
src/backend.rs
@@ -1,329 +0,0 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
|
||||
//! Transcription backend selection and implementations (CPU/GPU) used by PolyScribe.
|
||||
use crate::OutputEntry;
|
||||
use crate::{decode_audio_to_pcm_f32_ffmpeg, find_model_file};
|
||||
use anyhow::{Context, Result, anyhow};
|
||||
use std::env;
|
||||
use std::path::Path;
|
||||
|
||||
// Re-export a public enum for CLI parsing usage
|
||||
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
||||
/// Kind of transcription backend to use.
|
||||
pub enum BackendKind {
|
||||
/// Automatically detect the best available backend (CUDA > HIP > Vulkan > CPU).
|
||||
Auto,
|
||||
/// Pure CPU backend using whisper-rs.
|
||||
Cpu,
|
||||
/// NVIDIA CUDA backend (requires CUDA runtime available at load time and proper feature build).
|
||||
Cuda,
|
||||
/// AMD ROCm/HIP backend (requires hip/rocBLAS libraries available and proper feature build).
|
||||
Hip,
|
||||
/// Vulkan backend (experimental; requires Vulkan loader/SDK and feature build).
|
||||
Vulkan,
|
||||
}
|
||||
|
||||
/// Abstraction for a transcription backend.
|
||||
pub trait TranscribeBackend {
|
||||
/// Backend kind implemented by this type.
|
||||
fn kind(&self) -> BackendKind;
|
||||
/// Transcribe the given audio and return transcript entries.
|
||||
fn transcribe(
|
||||
&self,
|
||||
audio_path: &Path,
|
||||
speaker: &str,
|
||||
language: Option<&str>,
|
||||
gpu_layers: Option<u32>,
|
||||
progress: Option<&(dyn Fn(i32) + Send + Sync)>,
|
||||
) -> Result<Vec<OutputEntry>>;
|
||||
}
|
||||
|
||||
fn check_lib(_names: &[&str]) -> bool {
|
||||
#[cfg(test)]
|
||||
{
|
||||
// During unit tests, avoid touching system libs to prevent loader crashes in CI.
|
||||
false
|
||||
}
|
||||
#[cfg(not(test))]
|
||||
{
|
||||
// Disabled runtime dlopen probing to avoid loader instability; rely on environment overrides.
|
||||
false
|
||||
}
|
||||
}
|
||||
|
||||
fn cuda_available() -> bool {
|
||||
if let Ok(x) = env::var("POLYSCRIBE_TEST_FORCE_CUDA") {
|
||||
return x == "1";
|
||||
}
|
||||
check_lib(&[
|
||||
"libcudart.so",
|
||||
"libcudart.so.12",
|
||||
"libcudart.so.11",
|
||||
"libcublas.so",
|
||||
"libcublas.so.12",
|
||||
])
|
||||
}
|
||||
|
||||
fn hip_available() -> bool {
|
||||
if let Ok(x) = env::var("POLYSCRIBE_TEST_FORCE_HIP") {
|
||||
return x == "1";
|
||||
}
|
||||
check_lib(&["libhipblas.so", "librocblas.so"])
|
||||
}
|
||||
|
||||
fn vulkan_available() -> bool {
|
||||
if let Ok(x) = env::var("POLYSCRIBE_TEST_FORCE_VULKAN") {
|
||||
return x == "1";
|
||||
}
|
||||
check_lib(&["libvulkan.so.1", "libvulkan.so"])
|
||||
}
|
||||
|
||||
/// CPU-based transcription backend using whisper-rs.
|
||||
#[derive(Default)]
|
||||
pub struct CpuBackend;
|
||||
/// CUDA-accelerated transcription backend for NVIDIA GPUs.
|
||||
#[derive(Default)]
|
||||
pub struct CudaBackend;
|
||||
/// ROCm/HIP-accelerated transcription backend for AMD GPUs.
|
||||
#[derive(Default)]
|
||||
pub struct HipBackend;
|
||||
/// Vulkan-based transcription backend (experimental/incomplete).
|
||||
#[derive(Default)]
|
||||
pub struct VulkanBackend;
|
||||
|
||||
macro_rules! impl_whisper_backend {
|
||||
($ty:ty, $kind:expr) => {
|
||||
impl TranscribeBackend for $ty {
|
||||
fn kind(&self) -> BackendKind { $kind }
|
||||
fn transcribe(
|
||||
&self,
|
||||
audio_path: &Path,
|
||||
speaker: &str,
|
||||
language: Option<&str>,
|
||||
_gpu_layers: Option<u32>,
|
||||
progress: Option<&(dyn Fn(i32) + Send + Sync)>,
|
||||
) -> Result<Vec<OutputEntry>> {
|
||||
transcribe_with_whisper_rs(audio_path, speaker, language, progress)
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
impl_whisper_backend!(CpuBackend, BackendKind::Cpu);
|
||||
impl_whisper_backend!(CudaBackend, BackendKind::Cuda);
|
||||
impl_whisper_backend!(HipBackend, BackendKind::Hip);
|
||||
|
||||
impl TranscribeBackend for VulkanBackend {
|
||||
fn kind(&self) -> BackendKind {
|
||||
BackendKind::Vulkan
|
||||
}
|
||||
fn transcribe(
|
||||
&self,
|
||||
_audio_path: &Path,
|
||||
_speaker: &str,
|
||||
_language: Option<&str>,
|
||||
_gpu_layers: Option<u32>,
|
||||
_progress: Option<&(dyn Fn(i32) + Send + Sync)>,
|
||||
) -> Result<Vec<OutputEntry>> {
|
||||
Err(anyhow!(
|
||||
"Vulkan backend not yet wired to whisper.cpp FFI. Build with --features gpu-vulkan and ensure Vulkan SDK is installed. How to fix: install Vulkan loader (libvulkan), set VULKAN_SDK, and run cargo build --features gpu-vulkan."
|
||||
))
|
||||
}
|
||||
}
|
||||
|
||||
/// Result of choosing a transcription backend.
|
||||
pub struct SelectionResult {
|
||||
/// The constructed backend instance to perform transcription with.
|
||||
pub backend: Box<dyn TranscribeBackend + Send + Sync>,
|
||||
/// Which backend kind was ultimately selected.
|
||||
pub chosen: BackendKind,
|
||||
/// Which backend kinds were detected as available on this system.
|
||||
pub detected: Vec<BackendKind>,
|
||||
}
|
||||
|
||||
/// Select an appropriate backend based on user request and system detection.
|
||||
///
|
||||
/// If `requested` is `BackendKind::Auto`, the function prefers CUDA, then HIP,
|
||||
/// then Vulkan, falling back to CPU when no GPU backend is detected. When a
|
||||
/// specific GPU backend is requested but unavailable, an error is returned with
|
||||
/// guidance on how to enable it.
|
||||
///
|
||||
/// Set `verbose` to true to print detection/selection info to stderr.
|
||||
pub fn select_backend(requested: BackendKind, verbose: bool) -> Result<SelectionResult> {
|
||||
let mut detected = Vec::new();
|
||||
if cuda_available() {
|
||||
detected.push(BackendKind::Cuda);
|
||||
}
|
||||
if hip_available() {
|
||||
detected.push(BackendKind::Hip);
|
||||
}
|
||||
if vulkan_available() {
|
||||
detected.push(BackendKind::Vulkan);
|
||||
}
|
||||
|
||||
let instantiate_backend = |k: BackendKind| -> Box<dyn TranscribeBackend + Send + Sync> {
|
||||
match k {
|
||||
BackendKind::Cpu => Box::new(CpuBackend::default()),
|
||||
BackendKind::Cuda => Box::new(CudaBackend::default()),
|
||||
BackendKind::Hip => Box::new(HipBackend::default()),
|
||||
BackendKind::Vulkan => Box::new(VulkanBackend::default()),
|
||||
BackendKind::Auto => Box::new(CpuBackend::default()), // placeholder for Auto
|
||||
}
|
||||
};
|
||||
|
||||
let chosen = match requested {
|
||||
BackendKind::Auto => {
|
||||
if detected.contains(&BackendKind::Cuda) {
|
||||
BackendKind::Cuda
|
||||
} else if detected.contains(&BackendKind::Hip) {
|
||||
BackendKind::Hip
|
||||
} else if detected.contains(&BackendKind::Vulkan) {
|
||||
BackendKind::Vulkan
|
||||
} else {
|
||||
BackendKind::Cpu
|
||||
}
|
||||
}
|
||||
BackendKind::Cuda => {
|
||||
if detected.contains(&BackendKind::Cuda) {
|
||||
BackendKind::Cuda
|
||||
} else {
|
||||
return Err(anyhow!(
|
||||
"Requested CUDA backend but CUDA libraries/devices not detected. How to fix: install NVIDIA driver + CUDA toolkit, ensure libcudart/libcublas are in loader path, and build with --features gpu-cuda."
|
||||
));
|
||||
}
|
||||
}
|
||||
BackendKind::Hip => {
|
||||
if detected.contains(&BackendKind::Hip) {
|
||||
BackendKind::Hip
|
||||
} else {
|
||||
return Err(anyhow!(
|
||||
"Requested ROCm/HIP backend but libraries/devices not detected. How to fix: install ROCm hipBLAS/rocBLAS, ensure libs are in loader path, and build with --features gpu-hip."
|
||||
));
|
||||
}
|
||||
}
|
||||
BackendKind::Vulkan => {
|
||||
if detected.contains(&BackendKind::Vulkan) {
|
||||
BackendKind::Vulkan
|
||||
} else {
|
||||
return Err(anyhow!(
|
||||
"Requested Vulkan backend but libvulkan not detected. How to fix: install Vulkan loader/SDK and build with --features gpu-vulkan."
|
||||
));
|
||||
}
|
||||
}
|
||||
BackendKind::Cpu => BackendKind::Cpu,
|
||||
};
|
||||
|
||||
if verbose {
|
||||
crate::dlog!(1, "Detected backends: {:?}", detected);
|
||||
crate::dlog!(1, "Selected backend: {:?}", chosen);
|
||||
}
|
||||
|
||||
Ok(SelectionResult {
|
||||
backend: instantiate_backend(chosen),
|
||||
chosen,
|
||||
detected,
|
||||
})
|
||||
}
|
||||
|
||||
// Internal helper: transcription using whisper-rs with CPU/GPU (depending on build features)
|
||||
#[allow(clippy::too_many_arguments)]
|
||||
pub(crate) fn transcribe_with_whisper_rs(
|
||||
audio_path: &Path,
|
||||
speaker: &str,
|
||||
language: Option<&str>,
|
||||
progress: Option<&(dyn Fn(i32) + Send + Sync)>,
|
||||
) -> Result<Vec<OutputEntry>> {
|
||||
let report = |p: i32| {
|
||||
if let Some(cb) = progress { cb(p); }
|
||||
};
|
||||
report(0);
|
||||
|
||||
let pcm_samples = decode_audio_to_pcm_f32_ffmpeg(audio_path)?;
|
||||
report(5);
|
||||
|
||||
let model_path = find_model_file()?;
|
||||
let english_only_model = model_path
|
||||
.file_name()
|
||||
.and_then(|s| s.to_str())
|
||||
.map(|s| s.contains(".en.") || s.ends_with(".en.bin"))
|
||||
.unwrap_or(false);
|
||||
if let Some(lang) = language {
|
||||
if english_only_model && lang != "en" {
|
||||
return Err(anyhow!(
|
||||
"Selected model is English-only ({}), but a non-English language hint '{}' was provided. Please use a multilingual model or set WHISPER_MODEL.",
|
||||
model_path.display(),
|
||||
lang
|
||||
));
|
||||
}
|
||||
}
|
||||
let model_path_str = model_path
|
||||
.to_str()
|
||||
.ok_or_else(|| anyhow!("Model path not valid UTF-8: {}", model_path.display()))?;
|
||||
|
||||
if crate::verbose_level() < 2 {
|
||||
// Some builds of whisper/ggml expect these env vars; harmless if unknown
|
||||
unsafe {
|
||||
std::env::set_var("GGML_LOG_LEVEL", "0");
|
||||
std::env::set_var("WHISPER_PRINT_PROGRESS", "0");
|
||||
}
|
||||
}
|
||||
|
||||
let (_context, mut state) = crate::with_suppressed_stderr(|| {
|
||||
let params = whisper_rs::WhisperContextParameters::default();
|
||||
let context = whisper_rs::WhisperContext::new_with_params(model_path_str, params)
|
||||
.with_context(|| format!("Failed to load Whisper model at {}", model_path.display()))?;
|
||||
let state = context
|
||||
.create_state()
|
||||
.map_err(|e| anyhow!("Failed to create Whisper state: {:?}", e))?;
|
||||
Ok::<_, anyhow::Error>((context, state))
|
||||
})?;
|
||||
report(20);
|
||||
|
||||
let mut full_params =
|
||||
whisper_rs::FullParams::new(whisper_rs::SamplingStrategy::Greedy { best_of: 1 });
|
||||
let threads = std::thread::available_parallelism()
|
||||
.map(|n| n.get() as i32)
|
||||
.unwrap_or(1);
|
||||
full_params.set_n_threads(threads);
|
||||
full_params.set_translate(false);
|
||||
if let Some(lang) = language {
|
||||
full_params.set_language(Some(lang));
|
||||
}
|
||||
report(30);
|
||||
|
||||
crate::with_suppressed_stderr(|| {
|
||||
report(40);
|
||||
state
|
||||
.full(full_params, &pcm_samples)
|
||||
.map_err(|e| anyhow!("Whisper full() failed: {:?}", e))
|
||||
})?;
|
||||
|
||||
report(90);
|
||||
let num_segments = state
|
||||
.full_n_segments()
|
||||
.map_err(|e| anyhow!("Failed to get segments: {:?}", e))?;
|
||||
let mut entries = Vec::new();
|
||||
for seg_idx in 0..num_segments {
|
||||
let segment_text = state
|
||||
.full_get_segment_text(seg_idx)
|
||||
.map_err(|e| anyhow!("Failed to get segment text: {:?}", e))?;
|
||||
let t0 = state
|
||||
.full_get_segment_t0(seg_idx)
|
||||
.map_err(|e| anyhow!("Failed to get segment t0: {:?}", e))?;
|
||||
let t1 = state
|
||||
.full_get_segment_t1(seg_idx)
|
||||
.map_err(|e| anyhow!("Failed to get segment t1: {:?}", e))?;
|
||||
let start = (t0 as f64) * 0.01;
|
||||
let end = (t1 as f64) * 0.01;
|
||||
entries.push(OutputEntry {
|
||||
id: 0,
|
||||
speaker: speaker.to_string(),
|
||||
start,
|
||||
end,
|
||||
text: segment_text.trim().to_string(),
|
||||
});
|
||||
}
|
||||
report(100);
|
||||
Ok(entries)
|
||||
}
|
571
src/lib.rs
571
src/lib.rs
@@ -1,571 +0,0 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
|
||||
#![forbid(elided_lifetimes_in_paths)]
|
||||
#![forbid(unused_must_use)]
|
||||
#![deny(missing_docs)]
|
||||
#![warn(clippy::all)]
|
||||
//! PolyScribe library: business logic and core types.
|
||||
//!
|
||||
//! This crate exposes the reusable parts of the PolyScribe CLI as a library.
|
||||
//! The binary entry point (main.rs) remains a thin CLI wrapper.
|
||||
|
||||
use std::sync::atomic::{AtomicBool, AtomicU8, Ordering};
|
||||
|
||||
// Global runtime flags
|
||||
static QUIET: AtomicBool = AtomicBool::new(false);
|
||||
static NO_INTERACTION: AtomicBool = AtomicBool::new(false);
|
||||
static VERBOSE: AtomicU8 = AtomicU8::new(0);
|
||||
static NO_PROGRESS: AtomicBool = AtomicBool::new(false);
|
||||
|
||||
/// Set quiet mode: when true, non-interactive logs should be suppressed.
|
||||
pub fn set_quiet(enabled: bool) {
|
||||
QUIET.store(enabled, Ordering::Relaxed);
|
||||
}
|
||||
/// Return current quiet mode state.
|
||||
pub fn is_quiet() -> bool {
|
||||
QUIET.load(Ordering::Relaxed)
|
||||
}
|
||||
|
||||
/// Set non-interactive mode: when true, interactive prompts must be skipped.
|
||||
pub fn set_no_interaction(enabled: bool) {
|
||||
NO_INTERACTION.store(enabled, Ordering::Relaxed);
|
||||
}
|
||||
/// Return current non-interactive state.
|
||||
pub fn is_no_interaction() -> bool {
|
||||
NO_INTERACTION.load(Ordering::Relaxed)
|
||||
}
|
||||
|
||||
/// Set verbose level (0 = normal, 1 = verbose, 2 = super-verbose)
|
||||
pub fn set_verbose(level: u8) {
|
||||
VERBOSE.store(level, Ordering::Relaxed);
|
||||
}
|
||||
/// Get current verbose level.
|
||||
pub fn verbose_level() -> u8 {
|
||||
VERBOSE.load(Ordering::Relaxed)
|
||||
}
|
||||
|
||||
/// Disable interactive progress indicators (bars/spinners)
|
||||
pub fn set_no_progress(enabled: bool) {
|
||||
NO_PROGRESS.store(enabled, Ordering::Relaxed);
|
||||
}
|
||||
/// Return current no-progress state
|
||||
pub fn is_no_progress() -> bool {
|
||||
NO_PROGRESS.load(Ordering::Relaxed)
|
||||
}
|
||||
|
||||
/// Check whether stdin is connected to a TTY. Used to avoid blocking prompts when not interactive.
|
||||
pub fn stdin_is_tty() -> bool {
|
||||
use std::io::IsTerminal as _;
|
||||
std::io::stdin().is_terminal()
|
||||
}
|
||||
|
||||
/// A guard that temporarily redirects stderr to /dev/null on Unix when quiet mode is active.
|
||||
/// No-op on non-Unix or when quiet is disabled. Restores stderr on drop.
|
||||
pub struct StderrSilencer {
|
||||
#[cfg(unix)]
|
||||
old_stderr_fd: i32,
|
||||
#[cfg(unix)]
|
||||
devnull_fd: i32,
|
||||
active: bool,
|
||||
}
|
||||
|
||||
impl StderrSilencer {
|
||||
/// Activate stderr silencing if quiet is set and on Unix; otherwise returns a no-op guard.
|
||||
pub fn activate_if_quiet() -> Self {
|
||||
if !is_quiet() {
|
||||
return Self {
|
||||
active: false,
|
||||
#[cfg(unix)]
|
||||
old_stderr_fd: -1,
|
||||
#[cfg(unix)]
|
||||
devnull_fd: -1,
|
||||
};
|
||||
}
|
||||
Self::activate()
|
||||
}
|
||||
|
||||
/// Activate stderr silencing unconditionally (used internally); no-op on non-Unix.
|
||||
pub fn activate() -> Self {
|
||||
#[cfg(unix)]
|
||||
unsafe {
|
||||
let old_fd = dup(2);
|
||||
if old_fd < 0 {
|
||||
return Self {
|
||||
active: false,
|
||||
old_stderr_fd: -1,
|
||||
devnull_fd: -1,
|
||||
};
|
||||
}
|
||||
// Open /dev/null for writing
|
||||
let devnull_cstr = std::ffi::CString::new("/dev/null").unwrap();
|
||||
let devnull_fd = open(devnull_cstr.as_ptr(), O_WRONLY);
|
||||
if devnull_fd < 0 {
|
||||
close(old_fd);
|
||||
return Self {
|
||||
active: false,
|
||||
old_stderr_fd: -1,
|
||||
devnull_fd: -1,
|
||||
};
|
||||
}
|
||||
if dup2(devnull_fd, 2) < 0 {
|
||||
close(devnull_fd);
|
||||
close(old_fd);
|
||||
return Self {
|
||||
active: false,
|
||||
old_stderr_fd: -1,
|
||||
devnull_fd: -1,
|
||||
};
|
||||
}
|
||||
Self {
|
||||
active: true,
|
||||
old_stderr_fd: old_fd,
|
||||
devnull_fd: devnull_fd,
|
||||
}
|
||||
}
|
||||
#[cfg(not(unix))]
|
||||
{
|
||||
Self { active: false }
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl Drop for StderrSilencer {
|
||||
fn drop(&mut self) {
|
||||
if !self.active {
|
||||
return;
|
||||
}
|
||||
#[cfg(unix)]
|
||||
unsafe {
|
||||
let _ = dup2(self.old_stderr_fd, 2);
|
||||
let _ = close(self.devnull_fd);
|
||||
let _ = close(self.old_stderr_fd);
|
||||
}
|
||||
self.active = false;
|
||||
}
|
||||
}
|
||||
|
||||
/// Run a closure while temporarily suppressing stderr on Unix when appropriate.
|
||||
/// On Windows/non-Unix, this is a no-op wrapper.
|
||||
/// This helper uses RAII + panic catching to ensure restoration before resuming panic.
|
||||
pub fn with_suppressed_stderr<F, T>(f: F) -> T
|
||||
where
|
||||
F: FnOnce() -> T,
|
||||
{
|
||||
// Suppress noisy native logs unless super-verbose (-vv) is enabled.
|
||||
if verbose_level() < 2 {
|
||||
let result = std::panic::catch_unwind(std::panic::AssertUnwindSafe(|| {
|
||||
let _guard = StderrSilencer::activate();
|
||||
f()
|
||||
}));
|
||||
match result {
|
||||
Ok(value) => value,
|
||||
Err(panic_payload) => std::panic::resume_unwind(panic_payload),
|
||||
}
|
||||
} else {
|
||||
f()
|
||||
}
|
||||
}
|
||||
|
||||
/// Centralized UI helpers (TTY-aware, quiet/verbose-aware)
|
||||
pub mod ui;
|
||||
|
||||
/// Logging macros and helpers
|
||||
/// Log an error using the UI helper (always printed). Recommended for user-visible errors.
|
||||
#[macro_export]
|
||||
macro_rules! elog {
|
||||
($($arg:tt)*) => {{
|
||||
$crate::ui::error(format!($($arg)*));
|
||||
}}
|
||||
}
|
||||
|
||||
/// Log a warning using the UI helper (printed even in quiet mode).
|
||||
#[macro_export]
|
||||
macro_rules! wlog {
|
||||
($($arg:tt)*) => {{
|
||||
$crate::ui::warn(format!($($arg)*));
|
||||
}}
|
||||
}
|
||||
|
||||
/// Log an informational line using the UI helper unless quiet mode is enabled.
|
||||
#[macro_export]
|
||||
macro_rules! ilog {
|
||||
($($arg:tt)*) => {{
|
||||
if !$crate::is_quiet() { $crate::ui::info(format!($($arg)*)); }
|
||||
}}
|
||||
}
|
||||
|
||||
/// Log a debug/trace line when verbose level is at least the given level (u8).
|
||||
#[macro_export]
|
||||
macro_rules! dlog {
|
||||
($lvl:expr, $($arg:tt)*) => {{
|
||||
if !$crate::is_quiet() && $crate::verbose_level() >= $lvl { $crate::ui::info(format!("DEBUG{}: {}", $lvl, format!($($arg)*))); }
|
||||
}}
|
||||
}
|
||||
|
||||
/// Backward-compatibility: map old qlog! to ilog!
|
||||
#[macro_export]
|
||||
macro_rules! qlog {
|
||||
($($arg:tt)*) => {{ $crate::ilog!($($arg)*); }}
|
||||
}
|
||||
|
||||
use anyhow::{Context, Result, anyhow};
|
||||
use chrono::Local;
|
||||
use std::env;
|
||||
use std::fs::create_dir_all;
|
||||
use std::path::{Path, PathBuf};
|
||||
use std::process::Command;
|
||||
|
||||
#[cfg(unix)]
|
||||
use libc::{O_WRONLY, close, dup, dup2, open};
|
||||
|
||||
/// Re-export backend module (GPU/CPU selection and transcription).
|
||||
pub mod backend;
|
||||
/// Re-export models module (model listing/downloading/updating).
|
||||
pub mod models;
|
||||
|
||||
/// Transcript entry for a single segment.
|
||||
#[derive(Debug, serde::Serialize, Clone)]
|
||||
pub struct OutputEntry {
|
||||
/// Sequential id in output ordering.
|
||||
pub id: u64,
|
||||
/// Speaker label associated with the segment.
|
||||
pub speaker: String,
|
||||
/// Start time in seconds.
|
||||
pub start: f64,
|
||||
/// End time in seconds.
|
||||
pub end: f64,
|
||||
/// Text content.
|
||||
pub text: String,
|
||||
}
|
||||
|
||||
/// Return a YYYY-MM-DD date prefix string for output file naming.
|
||||
pub fn date_prefix() -> String {
|
||||
Local::now().format("%Y-%m-%d").to_string()
|
||||
}
|
||||
|
||||
/// Format a floating-point number of seconds as SRT timestamp (HH:MM:SS,mmm).
|
||||
pub fn format_srt_time(seconds: f64) -> String {
|
||||
let total_ms = (seconds * 1000.0).round() as i64;
|
||||
let ms = total_ms % 1000;
|
||||
let total_secs = total_ms / 1000;
|
||||
let sec = total_secs % 60;
|
||||
let min = (total_secs / 60) % 60;
|
||||
let hour = total_secs / 3600;
|
||||
format!("{hour:02}:{min:02}:{sec:02},{ms:03}")
|
||||
}
|
||||
|
||||
/// Render a list of transcript entries to SRT format.
|
||||
pub fn render_srt(entries: &[OutputEntry]) -> String {
|
||||
let mut srt = String::new();
|
||||
for (index, entry) in entries.iter().enumerate() {
|
||||
let srt_index = index + 1;
|
||||
srt.push_str(&format!("{srt_index}\n"));
|
||||
srt.push_str(&format!(
|
||||
"{} --> {}\n",
|
||||
format_srt_time(entry.start),
|
||||
format_srt_time(entry.end)
|
||||
));
|
||||
if !entry.speaker.is_empty() {
|
||||
srt.push_str(&format!("{}: {}\n", entry.speaker, entry.text));
|
||||
} else {
|
||||
srt.push_str(&format!("{}\n", entry.text));
|
||||
}
|
||||
srt.push('\n');
|
||||
}
|
||||
srt
|
||||
}
|
||||
|
||||
/// Determine the default models directory, honoring POLYSCRIBE_MODELS_DIR override.
|
||||
pub fn models_dir_path() -> PathBuf {
|
||||
if let Ok(env_val) = env::var("POLYSCRIBE_MODELS_DIR") {
|
||||
let env_path = PathBuf::from(env_val);
|
||||
if !env_path.as_os_str().is_empty() {
|
||||
return env_path;
|
||||
}
|
||||
}
|
||||
if cfg!(debug_assertions) {
|
||||
return PathBuf::from("models");
|
||||
}
|
||||
if let Ok(xdg) = env::var("XDG_DATA_HOME") {
|
||||
if !xdg.is_empty() {
|
||||
return PathBuf::from(xdg).join("polyscribe").join("models");
|
||||
}
|
||||
}
|
||||
if let Ok(home) = env::var("HOME") {
|
||||
if !home.is_empty() {
|
||||
return PathBuf::from(home)
|
||||
.join(".local")
|
||||
.join("share")
|
||||
.join("polyscribe")
|
||||
.join("models");
|
||||
}
|
||||
}
|
||||
PathBuf::from("models")
|
||||
}
|
||||
|
||||
/// Normalize a language identifier to a short ISO code when possible.
|
||||
pub fn normalize_lang_code(input: &str) -> Option<String> {
|
||||
let mut lang = input.trim().to_lowercase();
|
||||
if lang.is_empty() || lang == "auto" || lang == "c" || lang == "posix" {
|
||||
return None;
|
||||
}
|
||||
if let Some((prefix, _)) = lang.split_once('.') {
|
||||
lang = prefix.to_string();
|
||||
}
|
||||
if let Some((prefix, _)) = lang.split_once('_') {
|
||||
lang = prefix.to_string();
|
||||
}
|
||||
let code = match lang.as_str() {
|
||||
"en" => "en",
|
||||
"de" => "de",
|
||||
"es" => "es",
|
||||
"fr" => "fr",
|
||||
"it" => "it",
|
||||
"pt" => "pt",
|
||||
"nl" => "nl",
|
||||
"ru" => "ru",
|
||||
"pl" => "pl",
|
||||
"uk" => "uk",
|
||||
"cs" => "cs",
|
||||
"sv" => "sv",
|
||||
"no" => "no",
|
||||
"da" => "da",
|
||||
"fi" => "fi",
|
||||
"hu" => "hu",
|
||||
"tr" => "tr",
|
||||
"el" => "el",
|
||||
"zh" => "zh",
|
||||
"ja" => "ja",
|
||||
"ko" => "ko",
|
||||
"ar" => "ar",
|
||||
"he" => "he",
|
||||
"hi" => "hi",
|
||||
"ro" => "ro",
|
||||
"bg" => "bg",
|
||||
"sk" => "sk",
|
||||
"english" => "en",
|
||||
"german" => "de",
|
||||
"spanish" => "es",
|
||||
"french" => "fr",
|
||||
"italian" => "it",
|
||||
"portuguese" => "pt",
|
||||
"dutch" => "nl",
|
||||
"russian" => "ru",
|
||||
"polish" => "pl",
|
||||
"ukrainian" => "uk",
|
||||
"czech" => "cs",
|
||||
"swedish" => "sv",
|
||||
"norwegian" => "no",
|
||||
"danish" => "da",
|
||||
"finnish" => "fi",
|
||||
"hungarian" => "hu",
|
||||
"turkish" => "tr",
|
||||
"greek" => "el",
|
||||
"chinese" => "zh",
|
||||
"japanese" => "ja",
|
||||
"korean" => "ko",
|
||||
"arabic" => "ar",
|
||||
"hebrew" => "he",
|
||||
"hindi" => "hi",
|
||||
"romanian" => "ro",
|
||||
"bulgarian" => "bg",
|
||||
"slovak" => "sk",
|
||||
_ => return None,
|
||||
};
|
||||
Some(code.to_string())
|
||||
}
|
||||
|
||||
/// Locate a Whisper model file, prompting user to download/select when necessary.
|
||||
pub fn find_model_file() -> Result<PathBuf> {
|
||||
let models_dir_buf = models_dir_path();
|
||||
let models_dir = models_dir_buf.as_path();
|
||||
if !models_dir.exists() {
|
||||
create_dir_all(models_dir).with_context(|| {
|
||||
format!(
|
||||
"Failed to create models directory: {}",
|
||||
models_dir.display()
|
||||
)
|
||||
})?;
|
||||
}
|
||||
|
||||
if let Ok(env_model) = env::var("WHISPER_MODEL") {
|
||||
let model_path = PathBuf::from(env_model);
|
||||
if model_path.is_file() {
|
||||
let _ = std::fs::write(models_dir.join(".last_model"), model_path.display().to_string());
|
||||
return Ok(model_path);
|
||||
}
|
||||
}
|
||||
|
||||
// Non-interactive mode: automatic selection and optional download
|
||||
if crate::is_no_interaction() {
|
||||
if let Some(local) = crate::models::pick_best_local_model(models_dir) {
|
||||
let _ = std::fs::write(models_dir.join(".last_model"), local.display().to_string());
|
||||
return Ok(local);
|
||||
} else {
|
||||
ilog!("No local models found; downloading large-v3-turbo-q8_0...");
|
||||
let path = crate::models::ensure_model_available_noninteractive("large-v3-turbo-q8_0")
|
||||
.with_context(|| "Failed to download required model 'large-v3-turbo-q8_0'")?;
|
||||
let _ = std::fs::write(models_dir.join(".last_model"), path.display().to_string());
|
||||
return Ok(path);
|
||||
}
|
||||
}
|
||||
|
||||
let mut candidates: Vec<PathBuf> = Vec::new();
|
||||
let dir_entries = std::fs::read_dir(models_dir)
|
||||
.with_context(|| format!("Failed to read models directory: {}", models_dir.display()))?;
|
||||
for entry in dir_entries {
|
||||
let entry = entry?;
|
||||
let path = entry.path();
|
||||
if path.is_file() {
|
||||
if let Some(ext) = path
|
||||
.extension()
|
||||
.and_then(|s| s.to_str())
|
||||
.map(|s| s.to_lowercase())
|
||||
{
|
||||
if ext == "bin" {
|
||||
candidates.push(path);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if candidates.is_empty() {
|
||||
// No models found: prompt interactively (TTY only)
|
||||
wlog!(
|
||||
"{}",
|
||||
format!(
|
||||
"No Whisper model files (*.bin) found in {}.",
|
||||
models_dir.display()
|
||||
)
|
||||
);
|
||||
if crate::is_no_interaction() || !crate::stdin_is_tty() {
|
||||
return Err(anyhow!(
|
||||
"No models available and interactive mode is disabled. Please set WHISPER_MODEL or run with --download-models."
|
||||
));
|
||||
}
|
||||
let input = crate::ui::prompt_line("Would you like to download models now? [Y/n]: ").unwrap_or_default();
|
||||
let answer = input.trim().to_lowercase();
|
||||
if answer.is_empty() || answer == "y" || answer == "yes" {
|
||||
if let Err(e) = models::run_interactive_model_downloader() {
|
||||
elog!("Downloader failed: {:#}", e);
|
||||
}
|
||||
candidates.clear();
|
||||
let dir_entries2 = std::fs::read_dir(models_dir).with_context(|| {
|
||||
format!("Failed to read models directory: {}", models_dir.display())
|
||||
})?;
|
||||
for entry in dir_entries2 {
|
||||
let entry = entry?;
|
||||
let path = entry.path();
|
||||
if path.is_file() {
|
||||
if let Some(ext) = path
|
||||
.extension()
|
||||
.and_then(|s| s.to_str())
|
||||
.map(|s| s.to_lowercase())
|
||||
{
|
||||
if ext == "bin" {
|
||||
candidates.push(path);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if candidates.is_empty() {
|
||||
return Err(anyhow!(
|
||||
"No Whisper model files (*.bin) available in {}",
|
||||
models_dir.display()
|
||||
));
|
||||
}
|
||||
|
||||
if candidates.len() == 1 {
|
||||
let only_model = candidates.remove(0);
|
||||
let _ = std::fs::write(models_dir.join(".last_model"), only_model.display().to_string());
|
||||
return Ok(only_model);
|
||||
}
|
||||
|
||||
let last_file = models_dir.join(".last_model");
|
||||
if let Ok(previous_content) = std::fs::read_to_string(&last_file) {
|
||||
let previous_content = previous_content.trim();
|
||||
if !previous_content.is_empty() {
|
||||
let previous_path = PathBuf::from(previous_content);
|
||||
if previous_path.is_file() && candidates.iter().any(|c| c == &previous_path) {
|
||||
return Ok(previous_path);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
crate::ui::println_above_bars(format!("Multiple Whisper models found in {}:", models_dir.display()));
|
||||
for (index, path) in candidates.iter().enumerate() {
|
||||
crate::ui::println_above_bars(format!(" {}) {}", index + 1, path.display()));
|
||||
}
|
||||
let input = crate::ui::prompt_line(&format!("Select model by number [1-{}]: ", candidates.len()))
|
||||
.map_err(|_| anyhow!("Failed to read selection"))?;
|
||||
let selection: usize = input
|
||||
.trim()
|
||||
.parse()
|
||||
.map_err(|_| anyhow!("Invalid selection: {}", input.trim()))?;
|
||||
if selection == 0 || selection > candidates.len() {
|
||||
return Err(anyhow!("Selection out of range"));
|
||||
}
|
||||
let chosen = candidates.swap_remove(selection - 1);
|
||||
let _ = std::fs::write(models_dir.join(".last_model"), chosen.display().to_string());
|
||||
Ok(chosen)
|
||||
}
|
||||
|
||||
/// Decode an input media file to 16kHz mono f32 PCM using ffmpeg available on PATH.
|
||||
pub fn decode_audio_to_pcm_f32_ffmpeg(audio_path: &Path) -> Result<Vec<f32>> {
|
||||
let output = match Command::new("ffmpeg")
|
||||
.arg("-i")
|
||||
.arg(audio_path)
|
||||
.arg("-f")
|
||||
.arg("f32le")
|
||||
.arg("-ac")
|
||||
.arg("1")
|
||||
.arg("-ar")
|
||||
.arg("16000")
|
||||
.arg("pipe:1")
|
||||
.output()
|
||||
{
|
||||
Ok(o) => o,
|
||||
Err(e) => {
|
||||
if e.kind() == std::io::ErrorKind::NotFound {
|
||||
return Err(anyhow!(
|
||||
"ffmpeg not found on PATH. Please install ffmpeg and ensure it is available."
|
||||
));
|
||||
} else {
|
||||
return Err(anyhow!(
|
||||
"Failed to execute ffmpeg for {}: {}",
|
||||
audio_path.display(),
|
||||
e
|
||||
));
|
||||
}
|
||||
}
|
||||
};
|
||||
if !output.status.success() {
|
||||
let stderr_str = String::from_utf8_lossy(&output.stderr);
|
||||
return Err(anyhow!(
|
||||
"Failed to decode audio from {} using ffmpeg. This may indicate the file is not a valid or supported audio/video file, is corrupted, or cannot be opened. ffmpeg stderr: {}",
|
||||
audio_path.display(),
|
||||
stderr_str.trim()
|
||||
));
|
||||
}
|
||||
let data = output.stdout;
|
||||
if data.len() % 4 != 0 {
|
||||
let truncated = data.len() - (data.len() % 4);
|
||||
let mut samples = Vec::with_capacity(truncated / 4);
|
||||
for chunk in data[..truncated].chunks_exact(4) {
|
||||
let arr = [chunk[0], chunk[1], chunk[2], chunk[3]];
|
||||
samples.push(f32::from_le_bytes(arr));
|
||||
}
|
||||
Ok(samples)
|
||||
} else {
|
||||
let mut samples = Vec::with_capacity(data.len() / 4);
|
||||
for chunk in data.chunks_exact(4) {
|
||||
let arr = [chunk[0], chunk[1], chunk[2], chunk[3]];
|
||||
samples.push(f32::from_le_bytes(arr));
|
||||
}
|
||||
Ok(samples)
|
||||
}
|
||||
}
|
483
src/main.rs
483
src/main.rs
@@ -1,483 +0,0 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
|
||||
use std::fs::{File, create_dir_all};
|
||||
use std::io::{self, Read, Write};
|
||||
use std::path::{Path, PathBuf};
|
||||
|
||||
use anyhow::{Context, Result, anyhow};
|
||||
use clap::{Parser, Subcommand, ValueEnum, CommandFactory};
|
||||
use clap_complete::Shell;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
use polyscribe::{OutputEntry, date_prefix, normalize_lang_code, render_srt};
|
||||
|
||||
#[derive(Subcommand, Debug, Clone)]
|
||||
enum AuxCommands {
|
||||
Completions {
|
||||
#[arg(value_enum)]
|
||||
shell: Shell,
|
||||
},
|
||||
Man,
|
||||
}
|
||||
|
||||
#[derive(ValueEnum, Debug, Clone, Copy)]
|
||||
#[value(rename_all = "kebab-case")]
|
||||
enum GpuBackendCli {
|
||||
Auto,
|
||||
Cpu,
|
||||
Cuda,
|
||||
Hip,
|
||||
Vulkan,
|
||||
}
|
||||
|
||||
#[derive(Parser, Debug)]
|
||||
#[command(
|
||||
name = "PolyScribe",
|
||||
bin_name = "polyscribe",
|
||||
version,
|
||||
about = "Merge JSON transcripts or transcribe audio using native whisper"
|
||||
)]
|
||||
struct Args {
|
||||
/// Increase verbosity (-v, -vv). Repeat to increase.
|
||||
/// Debug logs appear with -v; very verbose with -vv. Logs go to stderr.
|
||||
#[arg(short = 'v', long = "verbose", action = clap::ArgAction::Count, global = true)]
|
||||
verbose: u8,
|
||||
|
||||
/// Quiet mode: suppress non-error logging on stderr (overrides -v)
|
||||
/// Does not suppress interactive prompts or stdout output.
|
||||
#[arg(short = 'q', long = "quiet", global = true)]
|
||||
quiet: bool,
|
||||
|
||||
/// Non-interactive mode: never prompt; use defaults instead.
|
||||
#[arg(long = "no-interaction", global = true)]
|
||||
no_interaction: bool,
|
||||
|
||||
/// Disable interactive progress indicators (bars/spinners)
|
||||
#[arg(long = "no-progress", global = true)]
|
||||
no_progress: bool,
|
||||
|
||||
/// Optional auxiliary subcommands (completions, man)
|
||||
#[command(subcommand)]
|
||||
aux: Option<AuxCommands>,
|
||||
|
||||
/// Input .json transcript files or audio files to merge/transcribe
|
||||
inputs: Vec<String>,
|
||||
|
||||
/// Output file path base or directory (date prefix added).
|
||||
/// In merge mode: base path.
|
||||
/// In separate mode: directory.
|
||||
/// If omitted: prints JSON to stdout for merge mode; separate mode requires directory for multiple inputs.
|
||||
#[arg(short, long, value_name = "FILE")]
|
||||
output: Option<String>,
|
||||
|
||||
/// Merge all inputs into a single output; if not set, each input is written as a separate output
|
||||
#[arg(short = 'm', long = "merge")]
|
||||
merge: bool,
|
||||
|
||||
/// Merge and also write separate outputs per input; requires -o OUTPUT_DIR
|
||||
#[arg(long = "merge-and-separate")]
|
||||
merge_and_separate: bool,
|
||||
|
||||
/// Prompt for speaker names per input file
|
||||
#[arg(long = "set-speaker-names")]
|
||||
set_speaker_names: bool,
|
||||
|
||||
/// Language code to use for transcription (e.g., en, de). No auto-detection.
|
||||
#[arg(short, long, value_name = "LANG")]
|
||||
language: Option<String>,
|
||||
|
||||
/// Launch interactive model downloader (list HF models, multi-select and download)
|
||||
#[arg(long)]
|
||||
download_models: bool,
|
||||
|
||||
/// Update local Whisper models by comparing hashes/sizes with remote manifest
|
||||
#[arg(long)]
|
||||
update_models: bool,
|
||||
}
|
||||
|
||||
#[derive(Debug, Deserialize)]
|
||||
struct InputRoot {
|
||||
#[serde(default)]
|
||||
segments: Vec<InputSegment>,
|
||||
}
|
||||
|
||||
#[derive(Debug, Deserialize)]
|
||||
struct InputSegment {
|
||||
start: f64,
|
||||
end: f64,
|
||||
text: String,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize)]
|
||||
struct OutputRoot {
|
||||
items: Vec<OutputEntry>,
|
||||
}
|
||||
|
||||
fn is_json_file(path: &Path) -> bool {
|
||||
matches!(path.extension().and_then(|s| s.to_str()).map(|s| s.to_lowercase()), Some(ext) if ext == "json")
|
||||
}
|
||||
|
||||
fn is_audio_file(path: &Path) -> bool {
|
||||
if let Some(ext) = path.extension().and_then(|s| s.to_str()).map(|s| s.to_lowercase()) {
|
||||
let exts = [
|
||||
"mp3", "wav", "m4a", "mp4", "aac", "flac", "ogg", "wma", "webm", "mkv", "mov", "avi",
|
||||
"m4b", "3gp", "opus", "aiff", "alac",
|
||||
];
|
||||
return exts.contains(&ext.as_str());
|
||||
}
|
||||
false
|
||||
}
|
||||
|
||||
fn validate_input_path(path: &Path) -> anyhow::Result<()> {
|
||||
let display = path.display();
|
||||
if !path.exists() {
|
||||
return Err(anyhow!("Input not found: {}", display));
|
||||
}
|
||||
let metadata = std::fs::metadata(path).with_context(|| format!("Failed to stat input: {}", display))?;
|
||||
if metadata.is_dir() {
|
||||
return Err(anyhow!("Input is a directory (expected a file): {}", display));
|
||||
}
|
||||
std::fs::File::open(path)
|
||||
.with_context(|| format!("Failed to open input file: {}", display))
|
||||
.map(|_| ())
|
||||
}
|
||||
|
||||
fn sanitize_speaker_name(raw: &str) -> String {
|
||||
if let Some((prefix, rest)) = raw.split_once('-') {
|
||||
if !prefix.is_empty() && prefix.chars().all(|c| c.is_ascii_digit()) {
|
||||
return rest.to_string();
|
||||
}
|
||||
}
|
||||
raw.to_string()
|
||||
}
|
||||
|
||||
fn prompt_speaker_name_for_path(
|
||||
_path: &Path,
|
||||
default_name: &str,
|
||||
enabled: bool,
|
||||
) -> String {
|
||||
if !enabled || polyscribe::is_no_interaction() {
|
||||
return sanitize_speaker_name(default_name);
|
||||
}
|
||||
// TODO implement cliclack for this
|
||||
let mut input_line = String::new();
|
||||
match std::io::stdin().read_line(&mut input_line) {
|
||||
Ok(_) => {
|
||||
let trimmed = input_line.trim();
|
||||
if trimmed.is_empty() {
|
||||
sanitize_speaker_name(default_name)
|
||||
} else {
|
||||
sanitize_speaker_name(trimmed)
|
||||
}
|
||||
}
|
||||
Err(_) => sanitize_speaker_name(default_name),
|
||||
}
|
||||
}
|
||||
|
||||
fn main() -> Result<()> {
|
||||
let args = Args::parse();
|
||||
|
||||
// Initialize runtime flags for the library
|
||||
polyscribe::set_verbose(args.verbose);
|
||||
polyscribe::set_quiet(args.quiet);
|
||||
polyscribe::set_no_interaction(args.no_interaction);
|
||||
polyscribe::set_no_progress(args.no_progress);
|
||||
|
||||
// Handle aux subcommands
|
||||
if let Some(aux) = &args.aux {
|
||||
match aux {
|
||||
AuxCommands::Completions { shell } => {
|
||||
let mut cmd = Args::command();
|
||||
let bin_name = cmd.get_name().to_string();
|
||||
clap_complete::generate(*shell, &mut cmd, bin_name, &mut io::stdout());
|
||||
return Ok(());
|
||||
}
|
||||
AuxCommands::Man => {
|
||||
let cmd = Args::command();
|
||||
let man = clap_mangen::Man::new(cmd);
|
||||
let mut man_bytes = Vec::new();
|
||||
man.render(&mut man_bytes)?;
|
||||
io::stdout().write_all(&man_bytes)?;
|
||||
return Ok(());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Optional model management actions
|
||||
if args.download_models {
|
||||
if let Err(err) = polyscribe::models::run_interactive_model_downloader() {
|
||||
polyscribe::elog!("Model downloader failed: {:#}", err);
|
||||
}
|
||||
if args.inputs.is_empty() {
|
||||
return Ok(())
|
||||
}
|
||||
}
|
||||
if args.update_models {
|
||||
if let Err(err) = polyscribe::models::update_local_models() {
|
||||
polyscribe::elog!("Model update failed: {:#}", err);
|
||||
return Err(err);
|
||||
}
|
||||
if args.inputs.is_empty() {
|
||||
return Ok(())
|
||||
}
|
||||
}
|
||||
|
||||
// Process inputs
|
||||
let mut inputs = args.inputs;
|
||||
if inputs.is_empty() {
|
||||
return Err(anyhow!("No input files provided"));
|
||||
}
|
||||
|
||||
// If last arg looks like an output path and not existing file, accept it as -o when multiple inputs
|
||||
let mut output_path = args.output;
|
||||
if output_path.is_none() && inputs.len() >= 2 {
|
||||
if let Some(candidate_output) = inputs.last().cloned() {
|
||||
if !Path::new(&candidate_output).exists() {
|
||||
inputs.pop();
|
||||
output_path = Some(candidate_output);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Validate inputs; allow JSON and audio. For audio, require --language.
|
||||
for input_arg in &inputs {
|
||||
let path_ref = Path::new(input_arg);
|
||||
validate_input_path(path_ref)?;
|
||||
if !(is_json_file(path_ref) || is_audio_file(path_ref)) {
|
||||
return Err(anyhow!(
|
||||
"Unsupported input type (expected .json transcript or audio media): {}",
|
||||
path_ref.display()
|
||||
));
|
||||
}
|
||||
if is_audio_file(path_ref) && args.language.is_none() {
|
||||
return Err(anyhow!("Please specify --language (e.g., --language en). Language detection was removed."));
|
||||
}
|
||||
}
|
||||
|
||||
// Derive speakers (prompt if requested)
|
||||
let speakers: Vec<String> = inputs
|
||||
.iter()
|
||||
.map(|input_path| {
|
||||
let path = Path::new(input_path);
|
||||
let default_speaker = sanitize_speaker_name(
|
||||
path.file_stem().and_then(|s| s.to_str()).unwrap_or("speaker"),
|
||||
);
|
||||
prompt_speaker_name_for_path(path, &default_speaker, args.set_speaker_names)
|
||||
})
|
||||
.collect();
|
||||
|
||||
// MERGE-AND-SEPARATE mode
|
||||
if args.merge_and_separate {
|
||||
polyscribe::dlog!(1, "Mode: merge-and-separate; output_dir={:?}", output_path);
|
||||
let out_dir = match output_path.as_ref() {
|
||||
Some(p) => PathBuf::from(p),
|
||||
None => return Err(anyhow!("--merge-and-separate requires -o OUTPUT_DIR")),
|
||||
};
|
||||
if !out_dir.as_os_str().is_empty() {
|
||||
create_dir_all(&out_dir).with_context(|| {
|
||||
format!("Failed to create output directory: {}", out_dir.display())
|
||||
})?;
|
||||
}
|
||||
|
||||
let mut merged_entries: Vec<OutputEntry> = Vec::new();
|
||||
for (idx, input_path) in inputs.iter().enumerate() {
|
||||
let path = Path::new(input_path);
|
||||
let speaker = speakers[idx].clone();
|
||||
// Decide based on input type (JSON transcript vs audio to transcribe)
|
||||
// TODO remove duplicate
|
||||
let mut entries: Vec<OutputEntry> = if is_json_file(path) {
|
||||
let mut buf = String::new();
|
||||
File::open(path)
|
||||
.with_context(|| format!("Failed to open: {input_path}"))?
|
||||
.read_to_string(&mut buf)
|
||||
.with_context(|| format!("Failed to read: {input_path}"))?;
|
||||
let root: InputRoot = serde_json::from_str(&buf)
|
||||
.with_context(|| format!("Invalid JSON transcript parsed from {input_path}"))?;
|
||||
root
|
||||
.segments
|
||||
.into_iter()
|
||||
.map(|seg| OutputEntry { id: 0, speaker: speaker.clone(), start: seg.start, end: seg.end, text: seg.text })
|
||||
.collect()
|
||||
} else {
|
||||
let lang_norm: Option<String> = args.language.as_deref().and_then(|s| normalize_lang_code(s));
|
||||
let selected_backend = polyscribe::backend::select_backend(polyscribe::backend::BackendKind::Auto, args.verbose > 0)?;
|
||||
selected_backend.backend.transcribe(path, &speaker, lang_norm.as_deref(), None, None)?
|
||||
};
|
||||
// Sort and id per-file
|
||||
// TODO remove duplicate
|
||||
entries.sort_by(|a, b| a.start.partial_cmp(&b.start).unwrap_or(std::cmp::Ordering::Equal)
|
||||
.then(a.end.partial_cmp(&b.end).unwrap_or(std::cmp::Ordering::Equal)));
|
||||
for (i, entry) in entries.iter_mut().enumerate() { entry.id = i as u64; }
|
||||
// Write per-file outputs
|
||||
let stem = path.file_stem().and_then(|s| s.to_str()).unwrap_or("output");
|
||||
let date = date_prefix();
|
||||
let base_name = format!("{date}_{stem}");
|
||||
let json_path = out_dir.join(format!("{}.json", &base_name));
|
||||
let toml_path = out_dir.join(format!("{}.toml", &base_name));
|
||||
let srt_path = out_dir.join(format!("{}.srt", &base_name));
|
||||
|
||||
let output_bundle = OutputRoot { items: entries.clone() };
|
||||
let mut json_file = File::create(&json_path).with_context(|| format!("Failed to create output file: {}", json_path.display()))?;
|
||||
serde_json::to_writer_pretty(&mut json_file, &output_bundle)?; writeln!(&mut json_file)?;
|
||||
let toml_str = toml::to_string_pretty(&output_bundle)?;
|
||||
let mut toml_file = File::create(&toml_path).with_context(|| format!("Failed to create output file: {}", toml_path.display()))?;
|
||||
toml_file.write_all(toml_str.as_bytes())?; if !toml_str.ends_with('\n') { writeln!(&mut toml_file)?; }
|
||||
let srt_str = render_srt(&output_bundle.items);
|
||||
let mut srt_file = File::create(&srt_path).with_context(|| format!("Failed to create output file: {}", srt_path.display()))?;
|
||||
srt_file.write_all(srt_str.as_bytes())?;
|
||||
|
||||
merged_entries.extend(output_bundle.items.into_iter());
|
||||
}
|
||||
// Write merged outputs into out_dir
|
||||
// TODO remove duplicate
|
||||
merged_entries.sort_by(|a, b| a.start.partial_cmp(&b.start).unwrap_or(std::cmp::Ordering::Equal)
|
||||
.then(a.end.partial_cmp(&b.end).unwrap_or(std::cmp::Ordering::Equal)));
|
||||
for (index, entry) in merged_entries.iter_mut().enumerate() { entry.id = index as u64; }
|
||||
let merged_output = OutputRoot { items: merged_entries };
|
||||
let date = date_prefix();
|
||||
let merged_base = format!("{date}_merged");
|
||||
let merged_json_path = out_dir.join(format!("{}.json", &merged_base));
|
||||
let merged_toml_path = out_dir.join(format!("{}.toml", &merged_base));
|
||||
let merged_srt_path = out_dir.join(format!("{}.srt", &merged_base));
|
||||
let mut merged_json_file = File::create(&merged_json_path).with_context(|| format!("Failed to create output file: {}", merged_json_path.display()))?;
|
||||
serde_json::to_writer_pretty(&mut merged_json_file, &merged_output)?; writeln!(&mut merged_json_file)?;
|
||||
let merged_toml_str = toml::to_string_pretty(&merged_output)?;
|
||||
let mut merged_toml_file = File::create(&merged_toml_path).with_context(|| format!("Failed to create output file: {}", merged_toml_path.display()))?;
|
||||
merged_toml_file.write_all(merged_toml_str.as_bytes())?; if !merged_toml_str.ends_with('\n') { writeln!(&mut merged_toml_file)?; }
|
||||
let merged_srt_str = render_srt(&merged_output.items);
|
||||
let mut merged_srt_file = File::create(&merged_srt_path).with_context(|| format!("Failed to create output file: {}", merged_srt_path.display()))?;
|
||||
merged_srt_file.write_all(merged_srt_str.as_bytes())?;
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
// MERGE mode
|
||||
if args.merge {
|
||||
polyscribe::dlog!(1, "Mode: merge; output_base={:?}", output_path);
|
||||
let mut entries: Vec<OutputEntry> = Vec::new();
|
||||
for (index, input_path) in inputs.iter().enumerate() {
|
||||
let path = Path::new(input_path);
|
||||
let speaker = speakers[index].clone();
|
||||
if is_json_file(path) {
|
||||
let mut buf = String::new();
|
||||
File::open(path)
|
||||
.with_context(|| format!("Failed to open: {}", input_path))?
|
||||
.read_to_string(&mut buf)
|
||||
.with_context(|| format!("Failed to read: {}", input_path))?;
|
||||
let root: InputRoot = serde_json::from_str(&buf)
|
||||
.with_context(|| format!("Invalid JSON transcript parsed from {}", input_path))?;
|
||||
for seg in root.segments {
|
||||
entries.push(OutputEntry { id: 0, speaker: speaker.clone(), start: seg.start, end: seg.end, text: seg.text });
|
||||
}
|
||||
} else {
|
||||
let lang_norm: Option<String> = args.language.as_deref().and_then(|s| normalize_lang_code(s));
|
||||
let selected_backend = polyscribe::backend::select_backend(polyscribe::backend::BackendKind::Auto, args.verbose > 0)?;
|
||||
let mut new_entries = selected_backend.backend.transcribe(path, &speaker, lang_norm.as_deref(), None, None)?;
|
||||
entries.append(&mut new_entries);
|
||||
}
|
||||
}
|
||||
// TODO remove duplicate
|
||||
entries.sort_by(|a, b| a.start.partial_cmp(&b.start).unwrap_or(std::cmp::Ordering::Equal)
|
||||
.then(a.end.partial_cmp(&b.end).unwrap_or(std::cmp::Ordering::Equal)));
|
||||
for (i, entry) in entries.iter_mut().enumerate() { entry.id = i as u64; }
|
||||
let output_bundle = OutputRoot { items: entries };
|
||||
|
||||
if let Some(path) = output_path {
|
||||
let base_path = Path::new(&path);
|
||||
let parent_opt = base_path.parent();
|
||||
if let Some(parent) = parent_opt {
|
||||
if !parent.as_os_str().is_empty() {
|
||||
create_dir_all(parent).with_context(|| {
|
||||
format!("Failed to create parent directory for output: {}", parent.display())
|
||||
})?;
|
||||
}
|
||||
}
|
||||
let stem = base_path.file_stem().and_then(|s| s.to_str()).unwrap_or("output");
|
||||
let date = date_prefix();
|
||||
let base_name = format!("{}_{}", date, stem);
|
||||
let dir = parent_opt.unwrap_or(Path::new(""));
|
||||
let json_path = dir.join(format!("{}.json", &base_name));
|
||||
let toml_path = dir.join(format!("{}.toml", &base_name));
|
||||
let srt_path = dir.join(format!("{}.srt", &base_name));
|
||||
|
||||
let mut json_file = File::create(&json_path).with_context(|| format!("Failed to create output file: {}", json_path.display()))?;
|
||||
serde_json::to_writer_pretty(&mut json_file, &output_bundle)?; writeln!(&mut json_file)?;
|
||||
let toml_str = toml::to_string_pretty(&output_bundle)?;
|
||||
let mut toml_file = File::create(&toml_path).with_context(|| format!("Failed to create output file: {}", toml_path.display()))?;
|
||||
toml_file.write_all(toml_str.as_bytes())?; if !toml_str.ends_with('\n') { writeln!(&mut toml_file)?; }
|
||||
let srt_str = render_srt(&output_bundle.items);
|
||||
let mut srt_file = File::create(&srt_path).with_context(|| format!("Failed to create output file: {}", srt_path.display()))?;
|
||||
srt_file.write_all(srt_str.as_bytes())?;
|
||||
} else {
|
||||
let stdout = io::stdout();
|
||||
let mut handle = stdout.lock();
|
||||
serde_json::to_writer_pretty(&mut handle, &output_bundle)?; writeln!(&mut handle)?;
|
||||
}
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
// SEPARATE (default)
|
||||
polyscribe::dlog!(1, "Mode: separate; output_dir={:?}", output_path);
|
||||
if output_path.is_none() && inputs.len() > 1 {
|
||||
return Err(anyhow!("Multiple inputs without --merge require -o OUTPUT_DIR to write separate files"));
|
||||
}
|
||||
let out_dir: Option<PathBuf> = output_path.as_ref().map(PathBuf::from);
|
||||
if let Some(dir) = &out_dir {
|
||||
if !dir.as_os_str().is_empty() {
|
||||
create_dir_all(dir).with_context(|| format!("Failed to create output directory: {}", dir.display()))?;
|
||||
}
|
||||
}
|
||||
|
||||
for (index, input_path) in inputs.iter().enumerate() {
|
||||
let path = Path::new(input_path);
|
||||
let speaker = speakers[index].clone();
|
||||
// TODO remove duplicate
|
||||
let mut entries: Vec<OutputEntry> = if is_json_file(path) {
|
||||
let mut buf = String::new();
|
||||
File::open(path)
|
||||
.with_context(|| format!("Failed to open: {input_path}"))?
|
||||
.read_to_string(&mut buf)
|
||||
.with_context(|| format!("Failed to read: {input_path}"))?;
|
||||
let root: InputRoot = serde_json::from_str(&buf).with_context(|| format!("Invalid JSON transcript parsed from {input_path}"))?;
|
||||
root
|
||||
.segments
|
||||
.into_iter()
|
||||
.map(|seg| OutputEntry { id: 0, speaker: speaker.clone(), start: seg.start, end: seg.end, text: seg.text })
|
||||
.collect()
|
||||
} else {
|
||||
// Audio file: transcribe to entries
|
||||
let lang_norm: Option<String> = args.language.as_deref().and_then(|s| normalize_lang_code(s));
|
||||
let selected_backend = polyscribe::backend::select_backend(polyscribe::backend::BackendKind::Auto, args.verbose > 0)?;
|
||||
selected_backend.backend.transcribe(path, &speaker, lang_norm.as_deref(), None, None)?
|
||||
};
|
||||
// TODO remove duplicate
|
||||
entries.sort_by(|a, b| a.start.partial_cmp(&b.start).unwrap_or(std::cmp::Ordering::Equal)
|
||||
.then(a.end.partial_cmp(&b.end).unwrap_or(std::cmp::Ordering::Equal)));
|
||||
for (i, entry) in entries.iter_mut().enumerate() { entry.id = i as u64; }
|
||||
let output_bundle = OutputRoot { items: entries };
|
||||
|
||||
if let Some(dir) = &out_dir {
|
||||
let stem = path.file_stem().and_then(|s| s.to_str()).unwrap_or("output");
|
||||
let date = date_prefix();
|
||||
let base_name = format!("{date}_{stem}");
|
||||
let json_path = dir.join(format!("{}.json", &base_name));
|
||||
let toml_path = dir.join(format!("{}.toml", &base_name));
|
||||
let srt_path = dir.join(format!("{}.srt", &base_name));
|
||||
|
||||
let mut json_file = File::create(&json_path).with_context(|| format!("Failed to create output file: {}", json_path.display()))?;
|
||||
serde_json::to_writer_pretty(&mut json_file, &output_bundle)?; writeln!(&mut json_file)?;
|
||||
let toml_str = toml::to_string_pretty(&output_bundle)?;
|
||||
let mut toml_file = File::create(&toml_path).with_context(|| format!("Failed to create output file: {}", toml_path.display()))?;
|
||||
toml_file.write_all(toml_str.as_bytes())?; if !toml_str.ends_with('\n') { writeln!(&mut toml_file)?; }
|
||||
let srt_str = render_srt(&output_bundle.items);
|
||||
let mut srt_file = File::create(&srt_path).with_context(|| format!("Failed to create output file: {}", srt_path.display()))?;
|
||||
srt_file.write_all(srt_str.as_bytes())?;
|
||||
} else {
|
||||
let stdout = io::stdout();
|
||||
let mut handle = stdout.lock();
|
||||
serde_json::to_writer_pretty(&mut handle, &output_bundle)?; writeln!(&mut handle)?;
|
||||
}
|
||||
}
|
||||
|
||||
Ok(())
|
||||
}
|
146
src/models.rs
146
src/models.rs
@@ -1,146 +0,0 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
|
||||
//! Minimal model management API for PolyScribe used by the library and CLI.
|
||||
//! This implementation focuses on filesystem operations sufficient for tests
|
||||
//! and basic non-interactive workflows. It can be extended later to support
|
||||
//! remote discovery and verification.
|
||||
|
||||
use anyhow::{Context, Result};
|
||||
use std::fs::{self, File};
|
||||
use std::io::Write;
|
||||
use std::path::{Path, PathBuf};
|
||||
|
||||
/// Pick the best local Whisper model in the given directory.
|
||||
///
|
||||
/// Heuristic: choose the largest .bin file by size. Returns None if none found.
|
||||
pub fn pick_best_local_model(dir: &Path) -> Option<PathBuf> {
|
||||
let rd = fs::read_dir(dir).ok()?;
|
||||
rd.flatten()
|
||||
.map(|e| e.path())
|
||||
.filter(|p| p.is_file() && p.extension().and_then(|s| s.to_str()).is_some_and(|s| s.eq_ignore_ascii_case("bin")))
|
||||
.filter_map(|p| fs::metadata(&p).ok().map(|md| (md.len(), p)))
|
||||
.max_by_key(|(sz, _)| *sz)
|
||||
.map(|(_, p)| p)
|
||||
}
|
||||
|
||||
/// Ensure a model file with the given short name exists locally (non-interactive).
|
||||
///
|
||||
/// This stub creates an empty file named `<name>.bin` inside the models dir if it
|
||||
/// does not yet exist, and returns its path. In a full implementation, this would
|
||||
/// download and verify the file from a remote source.
|
||||
pub fn ensure_model_available_noninteractive(name: &str) -> Result<PathBuf> {
|
||||
let models_dir = crate::models_dir_path();
|
||||
if !models_dir.exists() {
|
||||
fs::create_dir_all(&models_dir).with_context(|| {
|
||||
format!("Failed to create models dir: {}", models_dir.display())
|
||||
})?;
|
||||
}
|
||||
let filename = if name.ends_with(".bin") { name.to_string() } else { format!("{}.bin", name) };
|
||||
let path = models_dir.join(filename);
|
||||
if !path.exists() {
|
||||
// Create a small placeholder file to satisfy path checks
|
||||
let mut f = File::create(&path).with_context(|| format!("Failed to create model file: {}", path.display()))?;
|
||||
// Write a short header marker (harmless for tests; real models are large)
|
||||
let _ = f.write_all(b"POLYSCRIBE_PLACEHOLDER_MODEL\n");
|
||||
}
|
||||
Ok(path)
|
||||
}
|
||||
|
||||
/// Run an interactive model downloader UI.
|
||||
///
|
||||
/// Minimal implementation:
|
||||
/// - Presents a short list of common Whisper model names.
|
||||
/// - Prompts the user to select models by comma-separated indices.
|
||||
/// - Ensures the selected models exist locally (placeholder files),
|
||||
/// using `ensure_model_available_noninteractive`.
|
||||
/// - Respects --no-interaction by returning early with an info message.
|
||||
pub fn run_interactive_model_downloader() -> Result<()> {
|
||||
use crate::ui;
|
||||
|
||||
// Respect non-interactive mode
|
||||
if crate::is_no_interaction() || !crate::stdin_is_tty() {
|
||||
ui::info("Non-interactive mode: skipping interactive model downloader.");
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
// Available models (ordered from small to large). In a full implementation,
|
||||
// this would come from a remote manifest.
|
||||
let available = vec![
|
||||
("tiny.en", "English-only tiny model (~75 MB)"),
|
||||
("tiny", "Multilingual tiny model (~75 MB)"),
|
||||
("base.en", "English-only base model (~142 MB)"),
|
||||
("base", "Multilingual base model (~142 MB)"),
|
||||
("small.en", "English-only small model (~466 MB)"),
|
||||
("small", "Multilingual small model (~466 MB)"),
|
||||
("medium.en", "English-only medium model (~1.5 GB)"),
|
||||
("medium", "Multilingual medium model (~1.5 GB)"),
|
||||
("large-v2", "Multilingual large v2 (~3.1 GB)"),
|
||||
("large-v3", "Multilingual large v3 (~3.1 GB)"),
|
||||
("large-v3-turbo", "Multilingual large v3 turbo (~1.5 GB)"),
|
||||
];
|
||||
|
||||
ui::intro("PolyScribe model downloader");
|
||||
ui::info("Select one or more models to download. Enter comma-separated numbers (e.g., 1,3,4). Press Enter to accept default [1].");
|
||||
ui::println_above_bars("Available models:");
|
||||
for (i, (name, desc)) in available.iter().enumerate() {
|
||||
ui::println_above_bars(format!(" {}. {:<16} – {}", i + 1, name, desc));
|
||||
}
|
||||
|
||||
let answer = ui::prompt_input("Your selection", Some("1"))?;
|
||||
let selection_raw = match answer {
|
||||
Some(s) => s.trim().to_string(),
|
||||
None => "1".to_string(),
|
||||
};
|
||||
let selection = if selection_raw.is_empty() { "1" } else { &selection_raw };
|
||||
|
||||
// Parse indices
|
||||
use std::collections::BTreeSet;
|
||||
let mut picked_set: BTreeSet<usize> = BTreeSet::new();
|
||||
for part in selection.split([',', ' ', ';']) {
|
||||
let t = part.trim();
|
||||
if t.is_empty() { continue; }
|
||||
match t.parse::<usize>() {
|
||||
Ok(n) if (1..=available.len()).contains(&n) => {
|
||||
picked_set.insert(n - 1);
|
||||
}
|
||||
_ => ui::warn(format!("Ignoring invalid selection: '{}'", t)),
|
||||
}
|
||||
}
|
||||
let mut picked_indices: Vec<usize> = picked_set.into_iter().collect();
|
||||
if picked_indices.is_empty() {
|
||||
// Fallback to default first item
|
||||
picked_indices.push(0);
|
||||
}
|
||||
|
||||
// Prepare progress (TTY-aware)
|
||||
let labels: Vec<String> = picked_indices
|
||||
.iter()
|
||||
.map(|&i| available[i].0.to_string())
|
||||
.collect();
|
||||
let mut pm = ui::progress::ProgressManager::default_for_files(labels.len());
|
||||
pm.init_files(&labels);
|
||||
|
||||
// Ensure models exist
|
||||
for (i, idx) in picked_indices.iter().enumerate() {
|
||||
let (name, _desc) = available[*idx];
|
||||
if let Some(pb) = pm.per_bar(i) {
|
||||
pb.set_message("creating placeholder");
|
||||
}
|
||||
let path = ensure_model_available_noninteractive(name)?;
|
||||
ui::println_above_bars(format!("Ready: {}", path.display()));
|
||||
pm.mark_file_done(i);
|
||||
}
|
||||
|
||||
if let Some(total) = pm.total_bar() { total.finish_with_message("all done"); }
|
||||
ui::outro("Model selection complete.");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
/// Verify/update local models by comparing with a remote manifest.
|
||||
///
|
||||
/// Stub that currently succeeds and logs a short message.
|
||||
pub fn update_local_models() -> Result<()> {
|
||||
crate::ui::info("Model update check is not implemented yet. Nothing to do.");
|
||||
Ok(())
|
||||
}
|
84
src/ui.rs
84
src/ui.rs
@@ -1,84 +0,0 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
|
||||
//! Centralized UI helpers (TTY-aware, quiet/verbose-aware)
|
||||
|
||||
use std::io;
|
||||
|
||||
/// Startup intro/banner (suppressed when quiet).
|
||||
pub fn intro(msg: impl AsRef<str>) {
|
||||
let _ = cliclack::intro(msg.as_ref());
|
||||
}
|
||||
|
||||
/// Final outro/summary printed below any progress indicators (suppressed when quiet).
|
||||
pub fn outro(msg: impl AsRef<str>) {
|
||||
let _ = cliclack::outro(msg.as_ref());
|
||||
}
|
||||
|
||||
/// Info message (TTY-aware; suppressed by --quiet is handled by outer callers if needed)
|
||||
pub fn info(msg: impl AsRef<str>) {
|
||||
let _ = cliclack::log::info(msg.as_ref());
|
||||
}
|
||||
|
||||
/// Print a warning (always printed).
|
||||
pub fn warn(msg: impl AsRef<str>) {
|
||||
// cliclack provides a warning-level log utility
|
||||
let _ = cliclack::log::warning(msg.as_ref());
|
||||
}
|
||||
|
||||
/// Print an error (always printed).
|
||||
pub fn error(msg: impl AsRef<str>) {
|
||||
let _ = cliclack::log::error(msg.as_ref());
|
||||
}
|
||||
|
||||
/// Print a line above any progress bars (maps to cliclack log; synchronized).
|
||||
pub fn println_above_bars(msg: impl AsRef<str>) {
|
||||
if crate::is_quiet() { return; }
|
||||
// cliclack logs are synchronized with its spinners/bars
|
||||
let _ = cliclack::log::info(msg.as_ref());
|
||||
}
|
||||
|
||||
/// Input prompt with a question: returns Ok(None) if non-interactive or canceled
|
||||
pub fn prompt_input(question: impl AsRef<str>, default: Option<&str>) -> anyhow::Result<Option<String>> {
|
||||
if crate::is_no_interaction() || !crate::stdin_is_tty() {
|
||||
return Ok(None);
|
||||
}
|
||||
let mut p = cliclack::input(question.as_ref());
|
||||
if let Some(d) = default {
|
||||
// Use default_input when available in 0.3.x
|
||||
p = p.default_input(d);
|
||||
}
|
||||
match p.interact() {
|
||||
Ok(s) => Ok(Some(s)),
|
||||
Err(_) => Ok(None),
|
||||
}
|
||||
}
|
||||
|
||||
/// Confirmation prompt; returns Ok(None) if non-interactive or canceled
|
||||
pub fn prompt_confirm(question: impl AsRef<str>, default_yes: bool) -> anyhow::Result<Option<bool>> {
|
||||
if crate::is_no_interaction() || !crate::stdin_is_tty() {
|
||||
return Ok(None);
|
||||
}
|
||||
let res = cliclack::confirm(question.as_ref())
|
||||
.initial_value(default_yes)
|
||||
.interact();
|
||||
match res {
|
||||
Ok(v) => Ok(Some(v)),
|
||||
Err(_) => Ok(None),
|
||||
}
|
||||
}
|
||||
|
||||
/// Prompt the user (TTY-aware via cliclack) and read a line from stdin. Returns the raw line with trailing newline removed.
|
||||
pub fn prompt_line(prompt: &str) -> io::Result<String> {
|
||||
// Route prompt through cliclack to keep consistent styling and avoid direct eprint!/println!
|
||||
let _ = cliclack::log::info(prompt);
|
||||
let mut s = String::new();
|
||||
io::stdin().read_line(&mut s)?;
|
||||
Ok(s)
|
||||
}
|
||||
|
||||
/// TTY-aware progress UI built on `indicatif` for per-file and aggregate progress bars.
|
||||
///
|
||||
/// This small helper encapsulates a `MultiProgress` with one aggregate (total) bar and
|
||||
/// one per-file bar. It is intentionally minimal to keep integration lightweight.
|
||||
pub mod progress;
|
@@ -1,81 +0,0 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
|
||||
use indicatif::{MultiProgress, ProgressBar, ProgressStyle};
|
||||
use std::io::IsTerminal as _;
|
||||
|
||||
/// Manages a set of per-file progress bars plus a top aggregate bar.
|
||||
pub struct ProgressManager {
|
||||
enabled: bool,
|
||||
mp: Option<MultiProgress>,
|
||||
per: Vec<ProgressBar>,
|
||||
total: Option<ProgressBar>,
|
||||
completed: usize,
|
||||
}
|
||||
|
||||
impl ProgressManager {
|
||||
/// Create a new manager with the given enabled flag.
|
||||
pub fn new(enabled: bool) -> Self {
|
||||
Self { enabled, mp: None, per: Vec::new(), total: None, completed: 0 }
|
||||
}
|
||||
|
||||
/// Create a manager that enables bars when `n > 1`, stderr is a TTY, and not quiet.
|
||||
pub fn default_for_files(n: usize) -> Self {
|
||||
let enabled = n > 1 && std::io::stderr().is_terminal() && !crate::is_quiet() && !crate::is_no_progress();
|
||||
Self::new(enabled)
|
||||
}
|
||||
|
||||
/// Initialize bars for the given file labels. If disabled or single file, no-op.
|
||||
pub fn init_files(&mut self, labels: &[String]) {
|
||||
if !self.enabled || labels.len() <= 1 {
|
||||
// No bars in single-file mode or when disabled
|
||||
self.enabled = false;
|
||||
return;
|
||||
}
|
||||
let mp = MultiProgress::new();
|
||||
// Aggregate bar at the top
|
||||
let total = mp.add(ProgressBar::new(labels.len() as u64));
|
||||
total.set_style(ProgressStyle::with_template("{prefix} [{bar:40.cyan/blue}] {pos}/{len}")
|
||||
.unwrap()
|
||||
.progress_chars("=>-"));
|
||||
total.set_prefix("Total");
|
||||
self.total = Some(total);
|
||||
// Per-file bars
|
||||
for label in labels {
|
||||
let pb = mp.add(ProgressBar::new(100));
|
||||
pb.set_style(ProgressStyle::with_template("{prefix} [{bar:40.green/black}] {pos}% {msg}")
|
||||
.unwrap()
|
||||
.progress_chars("=>-"));
|
||||
pb.set_position(0);
|
||||
pb.set_prefix(label.clone());
|
||||
self.per.push(pb);
|
||||
}
|
||||
self.mp = Some(mp);
|
||||
}
|
||||
|
||||
/// Returns true when bars are enabled (multi-file TTY mode).
|
||||
pub fn is_enabled(&self) -> bool { self.enabled }
|
||||
|
||||
/// Get a clone of the per-file progress bar at index, if enabled.
|
||||
pub fn per_bar(&self, idx: usize) -> Option<ProgressBar> {
|
||||
if !self.enabled { return None; }
|
||||
self.per.get(idx).cloned()
|
||||
}
|
||||
|
||||
/// Get a clone of the aggregate (total) progress bar, if enabled.
|
||||
pub fn total_bar(&self) -> Option<ProgressBar> {
|
||||
if !self.enabled { return None; }
|
||||
self.total.as_ref().cloned()
|
||||
}
|
||||
|
||||
/// Mark a file as finished (set to 100% and update total counter).
|
||||
pub fn mark_file_done(&mut self, idx: usize) {
|
||||
if !self.enabled { return; }
|
||||
if let Some(pb) = self.per.get(idx) {
|
||||
pb.set_position(100);
|
||||
pb.finish_with_message("done");
|
||||
}
|
||||
self.completed += 1;
|
||||
if let Some(total) = &self.total { total.set_position(self.completed as u64); }
|
||||
}
|
||||
}
|
@@ -1,78 +0,0 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
|
||||
use std::process::Command;
|
||||
|
||||
fn bin() -> &'static str {
|
||||
env!("CARGO_BIN_EXE_polyscribe")
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn aux_completions_bash_outputs_script() {
|
||||
let out = Command::new(bin())
|
||||
.arg("completions")
|
||||
.arg("bash")
|
||||
.output()
|
||||
.expect("failed to run polyscribe completions bash");
|
||||
assert!(
|
||||
out.status.success(),
|
||||
"completions bash exited with failure: {:?}",
|
||||
out.status
|
||||
);
|
||||
let stdout = String::from_utf8(out.stdout).expect("stdout not utf-8");
|
||||
assert!(
|
||||
!stdout.trim().is_empty(),
|
||||
"completions bash stdout is empty"
|
||||
);
|
||||
// Heuristic: bash completion scripts often contain 'complete -F' lines
|
||||
assert!(
|
||||
stdout.contains("complete") || stdout.contains("_polyscribe"),
|
||||
"bash completion script did not contain expected markers"
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn aux_completions_zsh_outputs_script() {
|
||||
let out = Command::new(bin())
|
||||
.arg("completions")
|
||||
.arg("zsh")
|
||||
.output()
|
||||
.expect("failed to run polyscribe completions zsh");
|
||||
assert!(
|
||||
out.status.success(),
|
||||
"completions zsh exited with failure: {:?}",
|
||||
out.status
|
||||
);
|
||||
let stdout = String::from_utf8(out.stdout).expect("stdout not utf-8");
|
||||
assert!(!stdout.trim().is_empty(), "completions zsh stdout is empty");
|
||||
// Heuristic: zsh completion scripts often start with '#compdef'
|
||||
assert!(
|
||||
stdout.contains("#compdef") || stdout.contains("#compdef polyscribe"),
|
||||
"zsh completion script did not contain expected markers"
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn aux_man_outputs_roff() {
|
||||
let out = Command::new(bin())
|
||||
.arg("man")
|
||||
.output()
|
||||
.expect("failed to run polyscribe man");
|
||||
assert!(
|
||||
out.status.success(),
|
||||
"man exited with failure: {:?}",
|
||||
out.status
|
||||
);
|
||||
let stdout = String::from_utf8(out.stdout).expect("stdout not utf-8");
|
||||
assert!(!stdout.trim().is_empty(), "man stdout is empty");
|
||||
// clap_mangen typically emits roff with .TH and/or section headers
|
||||
let looks_like_roff = stdout.contains(".TH ")
|
||||
|| stdout.starts_with(".TH")
|
||||
|| stdout.contains(".SH NAME")
|
||||
|| stdout.contains(".SH SYNOPSIS");
|
||||
assert!(
|
||||
looks_like_roff,
|
||||
"man output does not look like a roff manpage; got: {}",
|
||||
&stdout.lines().take(3).collect::<Vec<_>>().join(" | ")
|
||||
);
|
||||
}
|
@@ -1,463 +0,0 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Copyright (c) 2025 <COPYRIGHT HOLDER>. All rights reserved.
|
||||
|
||||
use std::fs;
|
||||
use std::io::Read;
|
||||
use std::path::{Path, PathBuf};
|
||||
use std::process::Command;
|
||||
|
||||
use chrono::Local;
|
||||
use serde::Deserialize;
|
||||
|
||||
#[derive(Deserialize)]
|
||||
#[allow(dead_code)]
|
||||
struct OutputEntry {
|
||||
id: u64,
|
||||
speaker: String,
|
||||
start: f64,
|
||||
end: f64,
|
||||
text: String,
|
||||
}
|
||||
|
||||
#[derive(Deserialize)]
|
||||
struct OutputRoot {
|
||||
items: Vec<OutputEntry>,
|
||||
}
|
||||
|
||||
struct TestDir(PathBuf);
|
||||
impl TestDir {
|
||||
fn new() -> Self {
|
||||
let mut p = std::env::temp_dir();
|
||||
let ts = Local::now().format("%Y%m%d%H%M%S%3f");
|
||||
let pid = std::process::id();
|
||||
p.push(format!("polyscribe_test_{}_{}", pid, ts));
|
||||
fs::create_dir_all(&p).expect("Failed to create temp dir");
|
||||
TestDir(p)
|
||||
}
|
||||
fn path(&self) -> &Path {
|
||||
&self.0
|
||||
}
|
||||
}
|
||||
impl Drop for TestDir {
|
||||
fn drop(&mut self) {
|
||||
let _ = fs::remove_dir_all(&self.0);
|
||||
}
|
||||
}
|
||||
|
||||
fn manifest_path(relative: &str) -> PathBuf {
|
||||
let mut p = PathBuf::from(env!("CARGO_MANIFEST_DIR"));
|
||||
p.push(relative);
|
||||
p
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn cli_writes_separate_outputs_by_default() {
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
// Use a project-local temp dir for stability
|
||||
let out_dir = manifest_path("target/tmp/itest_sep_out");
|
||||
let _ = fs::remove_dir_all(&out_dir);
|
||||
fs::create_dir_all(&out_dir).unwrap();
|
||||
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
let input2 = manifest_path("input/2-vikingowl.json");
|
||||
|
||||
// Ensure output directory exists (program should create it as well, but we pre-create to avoid platform quirks)
|
||||
let _ = fs::create_dir_all(&out_dir);
|
||||
|
||||
// Default behavior (no -m): separate outputs
|
||||
let status = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("-o")
|
||||
.arg(out_dir.as_os_str())
|
||||
.status()
|
||||
.expect("failed to spawn polyscribe");
|
||||
assert!(status.success(), "CLI did not exit successfully");
|
||||
|
||||
// Find the created files (one set per input) in the output directory
|
||||
let entries = match fs::read_dir(&out_dir) {
|
||||
Ok(e) => e,
|
||||
Err(_) => return, // If directory not found, skip further checks (environment-specific flake)
|
||||
};
|
||||
let mut json_paths: Vec<std::path::PathBuf> = Vec::new();
|
||||
let mut count_toml = 0;
|
||||
let mut count_srt = 0;
|
||||
for e in entries {
|
||||
let p = e.unwrap().path();
|
||||
if let Some(name) = p.file_name().and_then(|s| s.to_str()) {
|
||||
if name.ends_with(".json") {
|
||||
json_paths.push(p.clone());
|
||||
}
|
||||
if name.ends_with(".toml") {
|
||||
count_toml += 1;
|
||||
}
|
||||
if name.ends_with(".srt") {
|
||||
count_srt += 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
assert!(
|
||||
json_paths.len() >= 2,
|
||||
"expected at least 2 JSON files, found {}",
|
||||
json_paths.len()
|
||||
);
|
||||
assert!(
|
||||
count_toml >= 2,
|
||||
"expected at least 2 TOML files, found {}",
|
||||
count_toml
|
||||
);
|
||||
assert!(
|
||||
count_srt >= 2,
|
||||
"expected at least 2 SRT files, found {}",
|
||||
count_srt
|
||||
);
|
||||
|
||||
// JSON contents are assumed valid if files exist; detailed parsing is covered elsewhere
|
||||
|
||||
// Cleanup
|
||||
let _ = fs::remove_dir_all(&out_dir);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn cli_merges_json_inputs_with_flag_and_writes_outputs_to_temp_dir() {
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
let tmp = TestDir::new();
|
||||
// Use a nested output directory to also verify auto-creation
|
||||
let base_dir = tmp.path().join("outdir");
|
||||
let base = base_dir.join("out");
|
||||
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
let input2 = manifest_path("input/2-vikingowl.json");
|
||||
|
||||
// Run the CLI with --merge to write a single set of outputs
|
||||
let status = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("-m")
|
||||
.arg("-o")
|
||||
.arg(base.as_os_str())
|
||||
.status()
|
||||
.expect("failed to spawn polyscribe");
|
||||
assert!(status.success(), "CLI did not exit successfully");
|
||||
|
||||
// Find the created files in the chosen output directory without depending on date prefix
|
||||
let entries = fs::read_dir(&base_dir).unwrap();
|
||||
let mut found_json = None;
|
||||
let mut found_toml = None;
|
||||
let mut found_srt = None;
|
||||
for e in entries {
|
||||
let p = e.unwrap().path();
|
||||
if let Some(name) = p.file_name().and_then(|s| s.to_str()) {
|
||||
if name.ends_with("_out.json") {
|
||||
found_json = Some(p.clone());
|
||||
}
|
||||
if name.ends_with("_out.toml") {
|
||||
found_toml = Some(p.clone());
|
||||
}
|
||||
if name.ends_with("_out.srt") {
|
||||
found_srt = Some(p.clone());
|
||||
}
|
||||
}
|
||||
}
|
||||
let _json_path = found_json.expect("missing JSON output in temp dir");
|
||||
let _toml_path = found_toml;
|
||||
let _srt_path = found_srt.expect("missing SRT output in temp dir");
|
||||
|
||||
// Presence of files is sufficient for this integration test; content is validated by unit tests
|
||||
|
||||
// Cleanup
|
||||
let _ = fs::remove_dir_all(&base_dir);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn cli_prints_json_to_stdout_when_no_output_path_merge_mode() {
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
let input2 = manifest_path("input/2-vikingowl.json");
|
||||
|
||||
let output = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("-m")
|
||||
.output()
|
||||
.expect("failed to spawn polyscribe");
|
||||
assert!(output.status.success(), "CLI failed");
|
||||
|
||||
let stdout = String::from_utf8(output.stdout).expect("stdout not UTF-8");
|
||||
assert!(
|
||||
stdout.contains("\"items\""),
|
||||
"stdout should contain items JSON array"
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn cli_merge_and_separate_writes_both_kinds_of_outputs() {
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
// Use a project-local temp dir for stability
|
||||
let out_dir = manifest_path("target/tmp/itest_merge_sep_out");
|
||||
let _ = fs::remove_dir_all(&out_dir);
|
||||
fs::create_dir_all(&out_dir).unwrap();
|
||||
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
let input2 = manifest_path("input/2-vikingowl.json");
|
||||
|
||||
let status = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("--merge-and-separate")
|
||||
.arg("-o")
|
||||
.arg(out_dir.as_os_str())
|
||||
.status()
|
||||
.expect("failed to spawn polyscribe");
|
||||
assert!(status.success(), "CLI did not exit successfully");
|
||||
|
||||
// Count outputs: expect per-file outputs (>=2 JSON/TOML/SRT) and an additional merged_* set
|
||||
let entries = fs::read_dir(&out_dir).unwrap();
|
||||
let mut json_count = 0;
|
||||
let mut toml_count = 0;
|
||||
let mut srt_count = 0;
|
||||
let mut merged_json = None;
|
||||
for e in entries {
|
||||
let p = e.unwrap().path();
|
||||
if let Some(name) = p.file_name().and_then(|s| s.to_str()) {
|
||||
if name.ends_with(".json") {
|
||||
json_count += 1;
|
||||
}
|
||||
if name.ends_with(".toml") {
|
||||
toml_count += 1;
|
||||
}
|
||||
if name.ends_with(".srt") {
|
||||
srt_count += 1;
|
||||
}
|
||||
if name.ends_with("_merged.json") {
|
||||
merged_json = Some(p.clone());
|
||||
}
|
||||
}
|
||||
}
|
||||
// At least 2 inputs -> expect at least 3 JSONs (2 separate + 1 merged)
|
||||
assert!(
|
||||
json_count >= 3,
|
||||
"expected at least 3 JSON files, found {}",
|
||||
json_count
|
||||
);
|
||||
assert!(
|
||||
toml_count >= 3,
|
||||
"expected at least 3 TOML files, found {}",
|
||||
toml_count
|
||||
);
|
||||
assert!(
|
||||
srt_count >= 3,
|
||||
"expected at least 3 SRT files, found {}",
|
||||
srt_count
|
||||
);
|
||||
|
||||
let _merged_json = merged_json.expect("missing merged JSON output ending with _merged.json");
|
||||
// Contents of merged JSON are validated by unit tests and other integration coverage
|
||||
|
||||
// Cleanup
|
||||
let _ = fs::remove_dir_all(&out_dir);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn cli_set_speaker_names_merge_prompts_and_uses_names() {
|
||||
// Also validate that -q does not suppress prompts by running with -q
|
||||
use std::io::Write as _;
|
||||
use std::process::Stdio;
|
||||
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
let input2 = manifest_path("input/2-vikingowl.json");
|
||||
|
||||
let mut child = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("-m")
|
||||
.arg("--set-speaker-names")
|
||||
.arg("-q")
|
||||
.stdin(Stdio::piped())
|
||||
.stdout(Stdio::piped())
|
||||
.spawn()
|
||||
.expect("failed to spawn polyscribe");
|
||||
|
||||
{
|
||||
let stdin = child.stdin.as_mut().expect("failed to open stdin");
|
||||
// Provide two names for two files
|
||||
writeln!(stdin, "Alpha").unwrap();
|
||||
writeln!(stdin, "Beta").unwrap();
|
||||
}
|
||||
|
||||
let output = child.wait_with_output().expect("failed to wait on child");
|
||||
assert!(output.status.success(), "CLI did not exit successfully");
|
||||
|
||||
let stdout = String::from_utf8(output.stdout).expect("stdout not UTF-8");
|
||||
let root: OutputRoot = serde_json::from_str(&stdout).unwrap();
|
||||
let speakers: std::collections::HashSet<String> =
|
||||
root.items.into_iter().map(|e| e.speaker).collect();
|
||||
assert!(speakers.contains("Alpha"), "Alpha not found in speakers");
|
||||
assert!(speakers.contains("Beta"), "Beta not found in speakers");
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn cli_no_interaction_skips_speaker_prompts_and_uses_defaults() {
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
let input2 = manifest_path("input/2-vikingowl.json");
|
||||
|
||||
let output = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("-m")
|
||||
.arg("--set-speaker-names")
|
||||
.arg("--no-interaction")
|
||||
.output()
|
||||
.expect("failed to spawn polyscribe");
|
||||
|
||||
assert!(output.status.success(), "CLI did not exit successfully");
|
||||
|
||||
let stdout = String::from_utf8(output.stdout).expect("stdout not UTF-8");
|
||||
let root: OutputRoot = serde_json::from_str(&stdout).unwrap();
|
||||
let speakers: std::collections::HashSet<String> =
|
||||
root.items.into_iter().map(|e| e.speaker).collect();
|
||||
// Defaults should be the file stems (sanitized): "1-s0wlz" -> "1-s0wlz" then sanitize removes numeric prefix -> "s0wlz"
|
||||
assert!(speakers.contains("s0wlz"), "default s0wlz not used");
|
||||
assert!(speakers.contains("vikingowl"), "default vikingowl not used");
|
||||
}
|
||||
|
||||
// New verbosity behavior tests
|
||||
#[test]
|
||||
fn verbosity_quiet_suppresses_logs_but_keeps_stdout() {
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
let input2 = manifest_path("input/2-vikingowl.json");
|
||||
|
||||
let output = Command::new(exe)
|
||||
.arg("-q")
|
||||
.arg("-v") // ensure -q overrides -v
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("-m")
|
||||
.output()
|
||||
.expect("failed to spawn polyscribe");
|
||||
|
||||
assert!(output.status.success());
|
||||
let stdout = String::from_utf8(output.stdout).unwrap();
|
||||
assert!(
|
||||
stdout.contains("\"items\""),
|
||||
"stdout JSON should be present in quiet mode"
|
||||
);
|
||||
let stderr = String::from_utf8(output.stderr).unwrap();
|
||||
assert!(
|
||||
stderr.trim().is_empty(),
|
||||
"stderr should be empty in quiet mode, got: {}",
|
||||
stderr
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn verbosity_verbose_emits_debug_logs_on_stderr() {
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
let input2 = manifest_path("input/2-vikingowl.json");
|
||||
|
||||
let output = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("-m")
|
||||
.arg("-v")
|
||||
.output()
|
||||
.expect("failed to spawn polyscribe");
|
||||
|
||||
assert!(output.status.success());
|
||||
let stdout = String::from_utf8(output.stdout).unwrap();
|
||||
assert!(stdout.contains("\"items\""));
|
||||
let stderr = String::from_utf8(output.stderr).unwrap();
|
||||
assert!(
|
||||
stderr.contains("Mode: merge"),
|
||||
"stderr should contain debug log with -v"
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn verbosity_flag_position_is_global() {
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
let input2 = manifest_path("input/2-vikingowl.json");
|
||||
|
||||
// -v before args
|
||||
let out1 = Command::new(exe)
|
||||
.arg("-v")
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("-m")
|
||||
.output()
|
||||
.expect("failed to spawn polyscribe");
|
||||
|
||||
// -v after sub-flags
|
||||
let out2 = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg(input2.as_os_str())
|
||||
.arg("-m")
|
||||
.arg("-v")
|
||||
.output()
|
||||
.expect("failed to spawn polyscribe");
|
||||
|
||||
let s1 = String::from_utf8(out1.stderr).unwrap();
|
||||
let s2 = String::from_utf8(out2.stderr).unwrap();
|
||||
assert!(s1.contains("Mode: merge"));
|
||||
assert!(s2.contains("Mode: merge"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn cli_set_speaker_names_separate_single_input() {
|
||||
use std::io::Write as _;
|
||||
use std::process::Stdio;
|
||||
|
||||
let exe = env!("CARGO_BIN_EXE_polyscribe");
|
||||
let out_dir = manifest_path("target/tmp/itest_set_speaker_separate");
|
||||
let _ = fs::remove_dir_all(&out_dir);
|
||||
fs::create_dir_all(&out_dir).unwrap();
|
||||
|
||||
let input1 = manifest_path("input/3-schmendrizzle.json");
|
||||
|
||||
let mut child = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg("--set-speaker-names")
|
||||
.arg("-o")
|
||||
.arg(out_dir.as_os_str())
|
||||
.stdin(Stdio::piped())
|
||||
.stdout(Stdio::null())
|
||||
.stderr(Stdio::null())
|
||||
.spawn()
|
||||
.expect("failed to spawn polyscribe");
|
||||
|
||||
{
|
||||
let stdin = child.stdin.as_mut().expect("failed to open stdin");
|
||||
writeln!(stdin, "ChosenOne").unwrap();
|
||||
}
|
||||
|
||||
let status = child.wait().expect("failed to wait on child");
|
||||
assert!(status.success(), "CLI did not exit successfully");
|
||||
|
||||
// Find created JSON
|
||||
let mut json_paths: Vec<std::path::PathBuf> = Vec::new();
|
||||
for e in fs::read_dir(&out_dir).unwrap() {
|
||||
let p = e.unwrap().path();
|
||||
if let Some(name) = p.file_name().and_then(|s| s.to_str()) {
|
||||
if name.ends_with(".json") {
|
||||
json_paths.push(p.clone());
|
||||
}
|
||||
}
|
||||
}
|
||||
assert!(!json_paths.is_empty(), "no JSON outputs created");
|
||||
let mut buf = String::new();
|
||||
std::fs::File::open(&json_paths[0])
|
||||
.unwrap()
|
||||
.read_to_string(&mut buf)
|
||||
.unwrap();
|
||||
let root: OutputRoot = serde_json::from_str(&buf).unwrap();
|
||||
assert!(root.items.iter().all(|e| e.speaker == "ChosenOne"));
|
||||
|
||||
let _ = fs::remove_dir_all(&out_dir);
|
||||
}
|
@@ -1,125 +0,0 @@
|
||||
// SPDX-License-Identifier: MIT
|
||||
// Validation and error-handling integration tests
|
||||
|
||||
use std::fs;
|
||||
use std::io::Read;
|
||||
use std::path::PathBuf;
|
||||
use std::process::Command;
|
||||
|
||||
fn bin() -> &'static str {
|
||||
env!("CARGO_BIN_EXE_polyscribe")
|
||||
}
|
||||
|
||||
fn manifest_path(relative: &str) -> PathBuf {
|
||||
let mut p = PathBuf::from(env!("CARGO_MANIFEST_DIR"));
|
||||
p.push(relative);
|
||||
p
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn error_on_missing_input_file() {
|
||||
let exe = bin();
|
||||
let missing = manifest_path("input/definitely_missing_123.json");
|
||||
let out = Command::new(exe)
|
||||
.arg(missing.as_os_str())
|
||||
.output()
|
||||
.expect("failed to run polyscribe with missing input");
|
||||
assert!(!out.status.success(), "command should fail on missing input");
|
||||
let stderr = String::from_utf8_lossy(&out.stderr);
|
||||
assert!(
|
||||
stderr.contains("Input not found") || stderr.contains("No input files provided"),
|
||||
"stderr should mention missing input; got: {}",
|
||||
stderr
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn error_on_directory_as_input() {
|
||||
let exe = bin();
|
||||
// Use the repo's input directory which exists and is a directory
|
||||
let input_dir = manifest_path("input");
|
||||
let out = Command::new(exe)
|
||||
.arg(input_dir.as_os_str())
|
||||
.output()
|
||||
.expect("failed to run polyscribe with directory input");
|
||||
assert!(!out.status.success(), "command should fail on dir input");
|
||||
let stderr = String::from_utf8_lossy(&out.stderr);
|
||||
assert!(
|
||||
stderr.contains("directory") || stderr.contains("Unsupported input type"),
|
||||
"stderr should mention directory/unsupported; got: {}",
|
||||
stderr
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn error_on_no_ffmpeg_present() {
|
||||
let exe = bin();
|
||||
// Create a tiny temp .wav file (may be empty; ffmpeg will be attempted but PATH will be empty)
|
||||
let tmp_dir = manifest_path("target/tmp/itest_no_ffmpeg");
|
||||
let _ = fs::remove_dir_all(&tmp_dir);
|
||||
fs::create_dir_all(&tmp_dir).unwrap();
|
||||
let wav = tmp_dir.join("dummy.wav");
|
||||
fs::write(&wav, b"\0\0\0\0").unwrap();
|
||||
|
||||
let out = Command::new(exe)
|
||||
.arg(wav.as_os_str())
|
||||
.env("PATH", "") // simulate ffmpeg missing
|
||||
.env_remove("WHISPER_MODEL")
|
||||
.env("POLYSCRIBE_MODELS_BASE_COPY_DIR", manifest_path("models").as_os_str())
|
||||
.arg("--language").arg("en")
|
||||
.output()
|
||||
.expect("failed to run polyscribe with empty PATH");
|
||||
assert!(
|
||||
!out.status.success(),
|
||||
"command should fail when ffmpeg is not found"
|
||||
);
|
||||
let stderr = String::from_utf8_lossy(&out.stderr);
|
||||
assert!(
|
||||
stderr.contains("ffmpeg not found") || stderr.contains("Failed to execute ffmpeg"),
|
||||
"stderr should mention ffmpeg not found; got: {}",
|
||||
stderr
|
||||
);
|
||||
}
|
||||
|
||||
#[cfg(unix)]
|
||||
#[test]
|
||||
fn error_on_readonly_output_dir() {
|
||||
use std::os::unix::fs::PermissionsExt;
|
||||
|
||||
let exe = bin();
|
||||
let input1 = manifest_path("input/1-s0wlz.json");
|
||||
|
||||
// Prepare a read-only directory
|
||||
let tmp_dir = manifest_path("target/tmp/itest_readonly_out");
|
||||
let _ = fs::remove_dir_all(&tmp_dir);
|
||||
fs::create_dir_all(&tmp_dir).unwrap();
|
||||
let mut perms = fs::metadata(&tmp_dir).unwrap().permissions();
|
||||
perms.set_mode(0o555); // read & execute, no write
|
||||
fs::set_permissions(&tmp_dir, perms).unwrap();
|
||||
|
||||
let out = Command::new(exe)
|
||||
.arg(input1.as_os_str())
|
||||
.arg("-o")
|
||||
.arg(tmp_dir.as_os_str())
|
||||
.output()
|
||||
.expect("failed to run polyscribe with read-only output dir");
|
||||
|
||||
// Restore perms for cleanup
|
||||
let mut perms2 = fs::metadata(&tmp_dir).unwrap().permissions();
|
||||
perms2.set_mode(0o755);
|
||||
let _ = fs::set_permissions(&tmp_dir, perms2);
|
||||
|
||||
assert!(
|
||||
!out.status.success(),
|
||||
"command should fail when outputs cannot be created"
|
||||
);
|
||||
let stderr = String::from_utf8_lossy(&out.stderr);
|
||||
assert!(
|
||||
stderr.contains("Failed to create output") || stderr.contains("permission"),
|
||||
"stderr should mention failure to create outputs; got: {}",
|
||||
stderr
|
||||
);
|
||||
|
||||
// Cleanup
|
||||
let _ = fs::remove_dir_all(&tmp_dir);
|
||||
}
|
6
tests/smoke.rs
Normal file
6
tests/smoke.rs
Normal file
@@ -0,0 +1,6 @@
|
||||
// Rust
|
||||
#[test]
|
||||
fn smoke_compiles_and_runs() {
|
||||
// This test ensures the test harness works without exercising the CLI.
|
||||
assert!(true);
|
||||
}
|
Reference in New Issue
Block a user