First things first !
sudo apt update
sudo apt upgrade
sudo apt-get install build-essential git
#!/bin/sh | |
# | |
# INFO | |
# | |
# This works if sonarr and radarr are set up to have a Category of sonarr and radarr respectively | |
# If you are using other Categories to save your automated downloads, update the script where you see: | |
# "radarr"|"sonarr") | |
# This script will not touch anything outside those Categories |
# Update & Upgrade the System | |
sudo apt-get update | |
sudo apt-get upgrade | |
# Install dependencies there might be more based on your system | |
# However below instructions are for the fresh Ubuntu install/server | |
# Please carefully watch the logs because if something could not be install | |
# You have to make sure it is installed properly by trying the command or that particular | |
# dependency again |
This document details how I setup LE on my server. Firstly, install the client as described on http://letsencrypt.readthedocs.org/en/latest/using.html and make sure you can execute it. I put it in /root/letsencrypt
.
As it is not possible to change the ports used for the standalone
authenticator and I already have a nginx running on port 80/443, I opted to use the webroot
method for each of my domains (note that LE does not issue wildcard certificates by design, so you probably want to get a cert for www.example.com
and example.com
).
For this, I placed config files into etc/letsencrypt/configs
, named after <domain>.conf
. The files are simple:
#! /bin/sh | |
# ZFS health version for Ubuntu | |
# You must install the package "dateuils" from the univers | |
# | |
# Based on Calomel.org | |
# https://calomel.org/zfs_health_check_script.html | |
# FreeBSD ZFS Health Check script | |
# zfs_health.sh @ Version 0.16 | |
# Check health of ZFS volumes and drives. On any faults send email. |
For this configuration you can use web server you like, i decided, because i work mostly with it to use nginx.
Generally, properly configured nginx can handle up to 400K to 500K requests per second (clustered), most what i saw is 50K to 80K (non-clustered) requests per second and 30% CPU load, course, this was 2 x Intel Xeon
with HyperThreading enabled, but it can work without problem on slower machines.
You must understand that this config is used in testing environment and not in production so you will need to find a way to implement most of those features best possible for your servers.
# This hosts file is brought to you by Dan Pollock and can be found at | |
# http://someonewhocares.org/hosts/ | |
# You are free to copy and distribute this file for non-commercial uses, | |
# as long the original URL and attribution is included. | |
#<localhost> | |
127.0.0.1 localhost | |
127.0.0.1 localhost.localdomain | |
255.255.255.255 broadcasthost | |
::1 localhost |
In one of my pet projects, I redirect all requests to index.php, which then decides what to do with it:
This snippet in your .htaccess will ensure that all requests for files and folders that does not exists will be redirected to index.php:
RewriteEngine on
RewriteCond %{REQUEST_FILENAME} !-d