automated scanning tool that integrates widely-used media management
services with various media servers for seamless media organization
Report Bug
Β·
Request Feature
.
Documentation
.
autopulseUI
autopulse is a simple project, designed after the deprecation of autoscan and a lot of inspiration comes from there. The goal is to provide a simple, efficient, and reliable way to update your media library reducing full library scans. A key feature is the ability to provide a hash of the file to the API, which will then wait for the file to match that hash before updating targets.
Following autoscan, we use the following terminology:
- Trigger: A trigger is a specification for an endpoint that will be called when a file is ready to be processed
- Target: A target is a specification for a library that will be updated when a file is ready to be processed
- Plex
- Jellyfin
- Emby
- Command
- Tdarr
- FileFlows
- Another autopulse instance
- Rewrites: rewrites the path provided by the trigger to the path expected by the target
- Integration: integrates with Sonarr, Radarr, Plex, Jellyfin, and more in the future
- Checks: checks the file exists before updating the target and optionally waits for the file to match a provided hash
- Reliability: uses a database to store the state of the scan requests
- Webhooks: allow for notifications to be sent when a file is ready to be processed with webhooks such as Discord
- User-Interface: provides a simple web interface to view/add scan requests
The easiest way to get started with autopulse is to use the provided docker image
latest
- full image with support for postgres/sqlitelatest-postgres
- smaller image that only supports Postgreslatest-sqlite
- smaller image that only supports SQLiteui
- self-hostable UI for autopulse
Append -arm64
to the tag to get the arm64 image
latest-arm64
- full image with support for postgres/sqlitelatest-postgres-arm64
- smaller image that only supports Postgreslatest-sqlite-arm64
- smaller image that only supports SQLite
One exception is the ui
tag which is a joint image for both architectures
ui
Here is a sample docker-compose.yml
# create a network
$ docker network create autopulse
# postgres database
$ docker run -d --net autopulse --name postgres -e POSTGRES_PASSWORD=autopulse -e POSTGRES_DB=autopulse postgres
$ docker run -d --net autopulse -e AUTOPULSE__APP__DATABASE_URL=postgres://postgres:autopulse@postgresql/autopulse --name autopulse danonline/autopulse
# sqlite database
$ docker run -d --net autopulse -e AUTOPULSE__APP__DATABASE_URL=sqlite://database.db --name autopulse danonline/autopulse
# or in-memory
$ docker run -d --net autopulse -e AUTOPULSE__APP__DATABASE_URL=sqlite://:memory: --name autopulse danonline/autopulse
All the documentation can be found on the website
Here's some quick links:
autopulse requires a configuration file to run. By default, it looks for config.toml
in the current working directory. You can override the default values using a config file or by setting environment variables in the format of: AUTOPULSE__{SECTION}__{KEY}
.
For example: AUTOPULSE__APP__DATABASE_URL
An example has been provided in the example directory
Note: You can provide the config with
json
,toml
,yaml
,json5
,ron
, orini
format
autopulse uses basic authorization for the API. You can set the username and password in the config file or by setting the AUTOPULSE__AUTH__USERNAME
and AUTOPULSE__AUTH__PASSWORD
environment variables.
auth:
username: terry
password: yoghurt
Note: By default the username and password are
admin
andpassword
respectively, however it is HIGHLY recommended to change these values if you are exposing the API to the internet.
triggers:
my_sonarr:
type: "sonarr"
rewrite:
from: "/downloads"
to: "/tvshows"
my_radarr:
type: "radarr"
rewrite:
from: "/downloads"
to: "/movies"
my_lidarr:
type: "lidarr"
rewrite:
from: "/downloads"
to: "/music"
my_readarr:
type: "readarr"
rewrite:
from: "/downloads"
to: "/books"
my_notify:
type: "notify"
paths:
- "/watch"
rewrite:
from: "/watch"
to: "/media"
webhooks:
my_discord:
type: "discord"
url: "https://discord.com/api/webhooks/1234567890/abcdefg"
targets:
my_plex:
type: "plex"
url: "http://plex:32400"
token: "<your_token>"
my_jellyfin:
type: "jellyfin"
url: "http://jellyfin:8096"
token: "<your_token>"
my_command:
type: "command"
raw: "echo $FILE_PATH >> list.txt"
By default a manual
endpoint is provided which can be used to manually trigger a scan. This can be useful for testing or for when you want to trigger a scan without waiting for a file to be ready.
$ curl -u 'admin:password' 'http://localhost:8080/manual?path=/path/to/file&hash=1234567890'
# or
$ curl -H 'Authorization: Basic <base_64_encoded_login>' 'http://localhost:8080/manual?path=/path/to/file&hash=1234567890'
The autopulse ui is a simple web interface that allows you to view and add scan requests. It is available hosted on Cloudflare Pages at autopulseui.pages.dev or you can host it yourself using the provided docker image. Note that requests are made server-side so you do not need to expose your autopulse instance to the internet, only the UI.
Variable | Description | Example |
---|---|---|
FORCE_DEFAULT_SERVER_URL |
Forces the default server URL to be used | true |
DEFAULT_SERVER_URL |
The default server URL to use | http://localhost:2875 |
FORCE_AUTH |
Forces the UI to use auth from env | true |
FORCE_SERVER_URL |
Forces the server url | true |
FORCE_USERNAME |
Forces the username | true |
FORCE_PASSWORD |
Forces the password | true |
Force a default server URL
FORCE_DEFAULT_SERVER_URL=true
DEFAULT_SERVER_URL=http://localhost:2875
Force the UI to use the provided auth
FORCE_AUTH=true
FORCE_SERVER_URL=https://localhost:2875
FORCE_USERNAME=admin
FORCE_PASSWORD=password
- Add more triggers
- Lidarr
- Readarr
- inotify
- Move triggers to structs
- Hooks
- Add/Found/Processed hooks
- Move Webhooks to hook
- Command hook
- Add more targets
- Emby
- Add more webhooks
- Generic JSON
- Add more options
- Cleanup duration (currently 10 days)
- Jellyfin
metadataRefreshMode
currently set toFullRefresh
- Plex refresh
- Databases
- SQLite
- [-] MySQL - linking mysql for alpine docker image is quite complex, so for now not supported unless someone can figure it out
- UI
- Add/View scan requests
- Add/View triggers
- Add/View targets
- Add/View webhooks
Contributions are what make the open-source community such an amazing place to learn, inspire, and create. Any contributions you make are greatly appreciated.
This project follows the Conventional Commits specification
- Fork the Project
- Create your Feature Branch (
git checkout -b feat/AmazingFeature
) - Commit your Changes (
git commit -m 'feat: add some AmazingFeature'
) - Push to the Branch (
git push origin feat/AmazingFeature
) - Open a Pull Request
# clone the repo
$ git clone https://github.com/dan-online/autopulse.git
$ cd autopulse
# basic easy config
$ cat <<EOF > config.toml
[app]
database_url = "sqlite://data/test.sqlite"
log_level = "trace"
EOF
# easy start using vendored/bundled dependencies
$ cargo run --features vendored
# or if you have the dependencies installed (libql-dev, libsqlite3-dev)
$ cargo run
# or if you only have one of the dependencies installed
$ cargo run --no-default-features --features sqlite # for sqlite
$ cargo run --no-default-features --features postgres # for postgres
Distributed under the MIT License. See LICENSE
for more information.
DanCodes - [email protected]
Project Link: https://github.com/dan-online/autopulse