graph TD;
A[Input file] --> B["Convert to text and chapterize (Optionally extract out the cover image)"];
B --> C[Transform to audio files];
- Python 3.10+ (This program was tested on 3.12)
- (Optional) Install espeak-ng (On Debian/Ubuntu, run
apt install -y espeak-ng
) - (Development only) uv
- You don't need to clone this repository and you can install either way:
- Using
pip
:python -m pip install audiobook-generator
(virtual environment highly recommended) - Using
pipx
:pipx install audiobook-generator
- NOTE For Windows users, there is one extra step needed to make cuda(Nvidia) GPU is used when available:
- If using
pip
and virtual environment, run this after the abovepip install
command (with the virtual environment activated first)pip install torch --index-url https://download.pytorch.org/whl/cu124 --force
- If using
pipx
, run this command instead:pipx runpip audiobook-generator install torch --index-url https://download.pytorch.org/whl/cu124 --force
- Technical details on why this is needed is described at the "Why you need that extra pip install step for Windows?" section.
- If using
- Using
- Convert your epub file to audiobooks via the command
abg <epub path> <audio output directory>
- If you want to see all the command line switches, just run
abg -h
- This program uses
uv
for dependency management and execution in development, install it first if you haven't done so. - To run the program from its source:
- Clone this repository and
cd
inside. - (Do it ONCE only at the first time) Run
uv sync
to create the virtual environment in the.env
directory and download all the dependencies. - Then run the following command
uv run -m audiobook_generator.main ...
- Clone this repository and
- Click this button
to open the Colab notebook.
- Upload your epub file to the root directory of the Colab runtime.
- Run the code cells in sequence, and after you click the run button of the last cell, you will be prompted to give the notebook access to your Google Drive, which the fully converted audiobook will be uploaded to.
- Create 2 environments
testpypi
andpypi
at https://github.com/houtianze/audiobook-generator/settings/environments - Configure the Publisher settings at testpypi and pypi accordingly:
- testpypi:
Repository: houtianze/audiobook-generator Workflow: python-publish-testpypi.yml Environment name: testpypi
- pypi:
Repository: houtianze/audiobook-generator Workflow: python-publish.yml Environment name: pypi
- Tag a new version
git tag v1.x.y
- Push to GitHub
git push --tags
- Create a release on GitHub using the tag either using its website or run
gh release create v1.x.y --generate-notes
(You need to install the GitHub CLI from https://cli.github.com/ and auth yourself first) - Relax and let
The selection to run the model on CPU or GPU is automatic, meaning:
- On Windows/WSL/Linux, If you have Nvidia graphic card with the driver properly installed, the model will be loaded to GPU (cuda) and executed, otherwise, the CPU is used (which is slower compared to GPU)
- On Mac, you need to set the environment variable
PYTORCH_ENABLE_MPS_FALLBACK=1
for it to run on GPU (because at the time of writing, the MPS support in PyTorch is not complete and it won't work without the CPU fallback), otherwise it will run on CPU.
(Thanks to @notimp for spotting this issue.)
If you go to pytorch, you will see that to install pytorch (only) on Windows, you need to specify the --index-url
parameter (e.g. pip3 install torch --index-url https://download.pytorch.org/whl/cu124
). When using uv
for development, this is handled by this section of the pyproject.toml
file:
[tool.uv.sources]
torch = [
{ index = "pytorch-cu124", marker = "sys_platform == 'win32'" },
]
[[tool.uv.index]]
name = "pytorch-cu124"
url = "https://download.pytorch.org/whl/cu124"
explicit = true
So running under uv
in development, the torch dependency are installed correctly. However when it's packaged and published to PyPI, it seems that this special specification of the torch index part is not respected by pip
and when you run pip
(or pipx
) install, it just runs pip install torch
on Windows without that --index-url
parameter, which installs a version that doesn't support cuda/GPU. Currently I don't see how I can resolve this in packaging as I guess python package specification may not support different dependency installation parameters on different platforms, or maybe I haven't digged deep enough. So for now, this extra step is required to install the correct version of torch on Windows.
- You can use this epub for testing: