I set this up so that lists can be added very easily, and the end result has (mildy) nice filenames.
Unparsed Lists are mirrored in the mirroredlists directory.
Lists that have to be extracted have a methodology as well.
This runs daily!
I am also fairly certain that I have more blocked domains than I have seen throughout the web.
My wife hasn't complained about websites being blocked, and I haven't had any major issues.
EDIT:
To clarify, adding 1111ALLPARSEDLISTS1111 does NOT give you the 3 million domains.
PSA
I missed a line in my installer, and the crojob will still work,,, however if you want it to update correctly, either reinstall or update the cronjob to
There are 2 txt files you have parsed from Easylist-GER, but I don't think that the first bigger one is parsed from the original Easylist-GER file, or am I wrong? The second file has only 59 entries?
maybe you have some time to take a look.
If in an enterprise environment, why not have the logs and lists, the ones that get abused the most, stored on a share or iSCSI target from a corporate NAS ?
eg.
mount -t nfs /NAS/pihole/var/log /var/log mount -t nfs /NAS/pihole/etc/pihole /etc/pihole
In fstab of course.
Or use formatted iSCSI targets and mount similar.
Or setup free Xen virtualisation, a baremetal type 1 hypervisor, on two hosts so you can do failover/balancing and dont depend on SD cards anymore.
The Xenserver distro (Citrix owned) saves you setting up the entire Xen-project hypervisor environment manually, with I believe a nice GUI and so on.
I think that you also have to add a -L option in parser.sh in the curl activity for a file that is moved on the website and generate a 3XX response code. I have had that with one of the files and with the -L option I could download / process the file.
For example sudo curl --silent -L ad (Daniel Apatin) ยท GitHub.........
This script is a work in progress, and the mirroredlists is there to have a reference of what a list was before parsings,, it also gives credit where it is deserved.