Need help with homescripts?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

animosity22
786 Stars 129 Forks 239 Commits 1 Opened issues

Description

My Scripts for Plex / Emby with Dropbox and rclone

Services available

!
?

Need anything else?

Contributors list

# 29,022
openvpn...
google-...
plex-me...
golang
155 commits
# 102,192
openvpn...
Go
vuejs
Vue.js
8 commits
# 301,139
Shell
openvpn...
plex-me...
Python
3 commits
# 38,312
Laravel
curl
bitmex
bitfine...
1 commit
# 80,438
C#
bittorr...
usenet
unix
1 commit
# 245,231
reddit
JavaFX
JavaScr...
plex
1 commit
# 291,407
php-pro...
Twitter
faceboo...
HTML
1 commit
# 365,065
JavaScr...
plex
mariadb
nginx
1 commit

Homescripts README

This is my configuration that works for me the best for my use case of an all in one media server. This is not meant to be a tutorial by any means as it does require some knowledge to get setup. I'm happy to help out as best as I can and welcome any updates/fixes/pulls to make it better and more helpful to others.

I use the latest rclone stable version downloaded direclty via the script install as package managers are frequently out of date and not maintained.

Change Log

Home Configuration

  • Verizon Gigabit Fios
  • Dropbox with encrypted media folder
  • Ubuntu 20.04
  • Intel(R) Core(TM) i7-7700 CPU @ 3.60GHz
  • 32 GB of Memory
  • 250 GB SSD Storage for my root
  • 1TB SSD for rclone Disk Caching
  • 6TB mirrored for staging

Dropbox

I migrated away from Google Drive to Dropbox as there still is an Enterprise Standard plan that seems to be unlimited space but I disliked the upload and download limits so I made the change. Dropbox is similiar to API usage compared to Google but there is not a pacer by default I work around that by setting a limit for transactions per second on the API via my mount command. API usage in Dropbox is tied to each application registration so I seperate out my apps and use one for uploading, one for movies and one for television shows as to never overload a particular one and allow easy reporting in the console.

My Workflow

I use Sonarr and Radarr in conjuction with NZBGet and qBittorrent to get my media.

My normal flow is that they grab a file, download it and place it in

/media
under the correct spot of
/media/TV
/media/Movies
, that is locally stored underneath the covers.

Every night, a rclone upload scripts moves files from local to my Google Drive. To achieve this, I use rclone to mount my Google Drive and mergerfs to combine a local disk and Google Drive together to provide a single access point for all services.

rclone and MergerFS

Installation

My Linux setup:

No LSB modules are available.
Distributor ID: Ubuntu
Description:    Ubuntu 20.04.3 LTS
Release:    20.04
Codename:   focal

Fuse needs to be installed for a rclone mount to function.

allow-other
is for my use and not recommended for shared servers as it allows any user to see a rclone mount. I am on a dedicated server that only I use so that is why I uncomment it.
$ sudo apt install fuse

You need to make the change to /etc/fuse.conf to

allow_other
by uncommenting the last line or removing the # from the last line.
sudo vi /etc/fuse.conf
[email protected]:~# cat /etc/fuse.conf
# /etc/fuse.conf - Configuration file for Filesystem in Userspace (FUSE)

Set the maximum number of FUSE mounts allowed to non-root users.

The default is 1000.

#mount_max = 1000

Allow non-root users to specify the allow_other or allow_root mount options.

user_allow_other

After fuse is installed, I install mergerfs. I would grab the proper package from the GitHub repository as packages from default are out-of-date.

My use case for mergerfs is that I always want to write to the local disk first and all my applications (Sonarr/Radarr/Plex/Emby/any application) all point directly to

/media
. For them it's not relevant if the file is local or remote as they should act the same.
/media
    /local (local disk)
    /DB_Movies (rclone mount)
    /DB_TV (rclone mount)

My

rclone.conf
has an entry for the Google Drive connection and and encrypted folder in my Google Drive called
media
. I mount media with a rclone script to display the decrypted contents on my server.

My rclone looks like: rclone.conf

This is my current rclone.service file which has the mount settings documented in it rclone.service

They are all mounted via systemd scripts. rclone is mounted first followed by the mergerfs mount.

My media starts up items in order: 1) rclone-movies service This is a standard rclone mount, the post execution command allows for the caching of the file structure in a single systemd file that simplies the process.

2) rclone-tv service This is a standard rclone mount, the post execution command allows for the caching of the file structure in a single systemd file that simplies the process.

3) mergerfs service This needs to be named the same as the mount point for the mount to work properly. I use

/media
so the file is named accordingly.

mergerfs configuration

This is located over here if you want to request help or compile from source [email protected].

I found unionfs to not do what I wanted and I can't stand the hidden files so for my use, it's much easier to configure and use mergerfs.

The following options always write to the first disk in the mount as with post 2.25 there are some changes with the settings so I had to add a few things that were default before.

/usr/bin/mergerfs /local:/DB_Movies:/DB_TV /media -o rw,use_ino,allow_other,func.getattr=newest,category.action=all,category.create=ff,cache.files=auto-full,dropcacheonclose=true

Important items:

  • use_ino
    is for hard linking with Sonarr/Radarr.
  • cache.files=auto-full
    uses memory for caching and helps out a bit if you have extra memory to spare.
  • category.action=all
    ,
    category.create=ff
    says to always create directories / files on the first listed mount point and for my configuration that is
    /cache
  • if you are reading directly from your rclone mount, you don't need to worry about any of mergerfs' settings.

Scheduled Nightly Uploads

I move my files to my GD every night via a cron job and an

upload cloud
script.

This is my cron entry:

# Cloud Upload
12 3 * * * /opt/homescripts/scripts/upload_cloud

This is my only upload to Dropbox as it is done on a nightly basis.

Plex Tweaks

If you have a lot of directories and files, it might be helpful to increase the file watchers by adding this to your

/etc/sysctl.conf
and rebooting. I chose the number below as I wanted to have plenty of headroom over the default option.
# Plex optimizations
fs.inotify.max_user_watches=262144

These tips and more for Linux can be found at the Plex Forum Linux Tips.

Plex

  • Enable Thumbnail previews
    - off: This creates a full read of the file to generate the preview and is set per library that is setup
  • Perform extensive media analysis during maintenance
    - off: This is listed under Scheduled Tasks and does a full download of files and is ony used for bandwidth analysis when streaming.

Sonarr/Radarr

  • Analyze video files
    - off: This also fully downloads files to perform analysis and should be turned off as this happens frequently on library refreshes if left on.

Caddy Proxy Server

I use Caddy to server majority of my things as I plug directly into GitHub oAuth2 for authentication. I can toggle CDN on and off via the proxy in the DNS.

My configuration is here.

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.