mirror of
https://github.com/uklans/cache-domains
synced 2025-08-03 14:51:23 +02:00
Compare commits
4 Commits
eb56d83265
...
081b485677
Author | SHA1 | Date | |
---|---|---|---|
|
081b485677 | ||
|
098c1e9c54 | ||
|
27f066bb8b | ||
|
b6ddac9531 |
@ -2,7 +2,13 @@ root = true
|
||||
|
||||
[*]
|
||||
indent_style = space
|
||||
indent_size = 4
|
||||
indent_size = 2
|
||||
trim_trailing_whitespace = true
|
||||
end_of_line = lf
|
||||
insert_final_newline = true
|
||||
|
||||
[*.md]
|
||||
indent_size = 0
|
||||
|
||||
[*.sh]
|
||||
indent_style = tab
|
55
.github/workflows/mega-linter.yml
vendored
Normal file
55
.github/workflows/mega-linter.yml
vendored
Normal file
@ -0,0 +1,55 @@
|
||||
# MegaLinter GitHub Action configuration file
|
||||
# More info at https://megalinter.io
|
||||
---
|
||||
name: MegaLinter
|
||||
|
||||
on:
|
||||
pull_request:
|
||||
branches:
|
||||
- master
|
||||
|
||||
env:
|
||||
GITHUB_STATUS_REPORTER: true
|
||||
|
||||
concurrency:
|
||||
group: ${{ github.ref }}-${{ github.workflow }}
|
||||
cancel-in-progress: true
|
||||
|
||||
jobs:
|
||||
megalinter:
|
||||
name: MegaLinter
|
||||
runs-on: ubuntu-latest
|
||||
|
||||
permissions:
|
||||
contents: write
|
||||
issues: write
|
||||
pull-requests: write
|
||||
|
||||
steps:
|
||||
- name: Checkout Code
|
||||
uses: actions/checkout@v4
|
||||
with:
|
||||
token: ${{ secrets.PAT || secrets.GITHUB_TOKEN }}
|
||||
fetch-depth: 0
|
||||
|
||||
- name: MegaLinter
|
||||
uses: oxsecurity/megalinter@v8
|
||||
id: ml
|
||||
env:
|
||||
VALIDATE_ALL_CODEBASE: >-
|
||||
${{
|
||||
github.event_name == 'push' &&
|
||||
github.ref == 'refs/heads/master'
|
||||
}}
|
||||
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
||||
- name: Archive production artifacts
|
||||
uses: actions/upload-artifact@v4
|
||||
if: success() || failure()
|
||||
with:
|
||||
name: MegaLinter reports
|
||||
include-hidden-files: "true"
|
||||
path: |
|
||||
megalinter-reports
|
||||
mega-linter.log
|
1
.gitignore
vendored
1
.gitignore
vendored
@ -1,2 +1,3 @@
|
||||
scripts/output
|
||||
scripts/config.json
|
||||
megalinter-reports/
|
||||
|
19
.lefthook.yml
Normal file
19
.lefthook.yml
Normal file
@ -0,0 +1,19 @@
|
||||
pre-commit:
|
||||
parallel: true
|
||||
jobs:
|
||||
- name: sort cache domains
|
||||
run: jq ".cache_domains |= sort_by(.name)" {all_files} > {all_files}.tmp && mv {all_files}.tmp {all_files}
|
||||
glob: "cache_domains.json"
|
||||
|
||||
- name: sort cdns
|
||||
run: |
|
||||
for f in {all_files}; do
|
||||
sort -uo ${f} ${f}
|
||||
done
|
||||
env:
|
||||
LC_ALL: C
|
||||
glob: "*.txt"
|
||||
|
||||
- name: megalinter
|
||||
run: docker run --rm -e USER=$(id -u) -e GROUP=$(id -g) -v "$PWD:/tmp/lint" oxsecurity/megalinter:v8
|
||||
stage_fixed: true
|
39
.mega-linter.yml
Normal file
39
.mega-linter.yml
Normal file
@ -0,0 +1,39 @@
|
||||
# Configuration file for MegaLinter
|
||||
#
|
||||
# See all available variables at https://megalinter.io/latest/config-file/ and in
|
||||
# linters documentation
|
||||
|
||||
APPLY_FIXES: all
|
||||
FLAVOR_SUGGESTIONS: false
|
||||
PRINT_ALPACA: false
|
||||
SHOW_ELAPSED_TIME: true
|
||||
|
||||
ENABLE:
|
||||
- ACTION
|
||||
- BASH
|
||||
- EDITORCONFIG
|
||||
- JSON
|
||||
- MARKDOWN
|
||||
- SPELL
|
||||
- YAML
|
||||
|
||||
DISABLE_LINTERS:
|
||||
- JSON_V8R
|
||||
- MARKDOWN_MARKDOWNLINT
|
||||
- SPELL_CSPELL
|
||||
- YAML_V8R
|
||||
|
||||
PRE_COMMANDS:
|
||||
- command: apk add --no-cache yq
|
||||
cwd: root
|
||||
- command: yq -i ".cache_domains |= sort_by(.name)" cache_domains.json
|
||||
cwd: workspace
|
||||
- command: |
|
||||
for f in *.txt; do
|
||||
sort -uo ${f} ${f}
|
||||
done
|
||||
cwd: workspace
|
||||
|
||||
POST_COMMANDS:
|
||||
- command: find . -user root -group root -exec chown ${USER}:${GROUP} {} \;
|
||||
cwd: workspace
|
10
blizzard.txt
10
blizzard.txt
@ -1,11 +1,11 @@
|
||||
dist.blizzard.com
|
||||
dist.blizzard.com.edgesuite.net
|
||||
*.cdn.blizzard.com
|
||||
blizzard.vo.llnwd.net
|
||||
blzddist1-a.akamaihd.net
|
||||
blzddist2-a.akamaihd.net
|
||||
blzddist3-a.akamaihd.net
|
||||
cdn.blizzard.com
|
||||
dist.blizzard.com
|
||||
dist.blizzard.com.edgesuite.net
|
||||
edge.blizzard.top.comcast.net
|
||||
level3.blizzard.com
|
||||
nydus.battle.net
|
||||
edge.blizzard.top.comcast.net
|
||||
cdn.blizzard.com
|
||||
*.cdn.blizzard.com
|
||||
|
@ -1,136 +1,136 @@
|
||||
{
|
||||
"cache_domains": [
|
||||
{
|
||||
"name": "arenanet",
|
||||
"description": "CDN for guild wars, HoT",
|
||||
"domain_files": ["arenanet.txt"]
|
||||
},
|
||||
{
|
||||
"name": "blizzard",
|
||||
"description": "CDN for blizzard/battle.net",
|
||||
"domain_files": ["blizzard.txt"]
|
||||
},
|
||||
{
|
||||
"name": "cod",
|
||||
"description": "CDN for Call of Duty",
|
||||
"domain_files": ["cod.txt"]
|
||||
},
|
||||
{
|
||||
"name": "bsg",
|
||||
"description": "CDN for Battle State Games, Tarkov",
|
||||
"domain_files": ["bsg.txt"]
|
||||
},
|
||||
{
|
||||
"name": "cityofheroes",
|
||||
"description": "CDN for City of Heroes (Homecoming)",
|
||||
"domain_files": ["cityofheroes.txt"]
|
||||
},
|
||||
{
|
||||
"name": "daybreak",
|
||||
"description": "Daybreak games CDN",
|
||||
"domain_files": ["daybreak.txt"]
|
||||
},
|
||||
{
|
||||
"name": "epicgames",
|
||||
"description": "CDN for Epic Games",
|
||||
"domain_files": ["epicgames.txt"]
|
||||
},
|
||||
{
|
||||
"name": "frontier",
|
||||
"description": "CDN for frontier games",
|
||||
"domain_files": ["frontier.txt"]
|
||||
},
|
||||
{
|
||||
"name": "nexusmods",
|
||||
"description": "Nexus mods / skyrim content",
|
||||
"domain_files": ["nexusmods.txt"]
|
||||
},
|
||||
{
|
||||
"name": "neverwinter",
|
||||
"description": "Cryptic CDN for Neverwinter",
|
||||
"domain_files": ["neverwinter.txt"]
|
||||
},
|
||||
{
|
||||
"name": "nintendo",
|
||||
"description": "CDN for Nintendo consoles and download servers",
|
||||
"domain_files": ["nintendo.txt"]
|
||||
},
|
||||
{
|
||||
"name": "origin",
|
||||
"description": "CDN for origin",
|
||||
"notes": "Should be enabled for HTTP traffic only or with a HTTPS proxy else origin client download fails",
|
||||
"mixed_content": true,
|
||||
"domain_files": ["origin.txt"]
|
||||
},
|
||||
{
|
||||
"name": "pathofexile",
|
||||
"description": "CDN for Path Of Exile",
|
||||
"domain_files": ["pathofexile.txt"]
|
||||
},
|
||||
{
|
||||
"name": "renegadex",
|
||||
"description": "CDN for Renegade X",
|
||||
"domain_files": ["renegadex.txt"]
|
||||
},
|
||||
{
|
||||
"name": "riot",
|
||||
"description": "CDN for riot games",
|
||||
"domain_files": ["riot.txt"]
|
||||
},
|
||||
{
|
||||
"name": "rockstar",
|
||||
"description": "CDN for rockstar games",
|
||||
"domain_files": ["rockstar.txt"]
|
||||
},
|
||||
{
|
||||
"name": "sony",
|
||||
"description": "CDN for sony / playstation",
|
||||
"domain_files": ["sony.txt"]
|
||||
},
|
||||
{
|
||||
"name": "square",
|
||||
"description": "CDN for Final Fantasy XIV",
|
||||
"domain_files": ["square.txt"]
|
||||
},
|
||||
{
|
||||
"name": "steam",
|
||||
"description": "CDN for steam platform",
|
||||
"domain_files": ["steam.txt"]
|
||||
},
|
||||
{
|
||||
"name": "uplay",
|
||||
"description": "CDN for uplay downloader",
|
||||
"domain_files": ["uplay.txt"]
|
||||
},
|
||||
{
|
||||
"name": "teso",
|
||||
"description": "CDN for The Elder Scrolls Online",
|
||||
"domain_files": ["teso.txt"]
|
||||
},
|
||||
{
|
||||
"name": "warframe",
|
||||
"description": "CDN for Warframe",
|
||||
"domain_files": ["warframe.txt"]
|
||||
},
|
||||
{
|
||||
"name": "wargaming",
|
||||
"description": "CDN for wargaming.net",
|
||||
"domain_files": ["wargaming.net.txt"]
|
||||
},
|
||||
{
|
||||
"name": "wsus",
|
||||
"description": "CDN for windows updates",
|
||||
"domain_files": ["windowsupdates.txt"]
|
||||
},
|
||||
{
|
||||
"name": "xboxlive",
|
||||
"description": "CDN for xboxlive",
|
||||
"domain_files": ["xboxlive.txt"]
|
||||
},
|
||||
{
|
||||
"name": "test",
|
||||
"description": "Test CDN, recommended to enable for additional diagnostics",
|
||||
"domain_files": ["test.txt"]
|
||||
}
|
||||
]
|
||||
"cache_domains": [
|
||||
{
|
||||
"name": "arenanet",
|
||||
"description": "CDN for Guild Wars, HoT",
|
||||
"domain_files": ["arenanet.txt"]
|
||||
},
|
||||
{
|
||||
"name": "blizzard",
|
||||
"description": "CDN for Blizzard/Battle.net",
|
||||
"domain_files": ["blizzard.txt"]
|
||||
},
|
||||
{
|
||||
"name": "bsg",
|
||||
"description": "CDN for Battle State Games, Tarkov",
|
||||
"domain_files": ["bsg.txt"]
|
||||
},
|
||||
{
|
||||
"name": "cityofheroes",
|
||||
"description": "CDN for City of Heroes (Homecoming)",
|
||||
"domain_files": ["cityofheroes.txt"]
|
||||
},
|
||||
{
|
||||
"name": "cod",
|
||||
"description": "CDN for Call of Duty",
|
||||
"domain_files": ["cod.txt"]
|
||||
},
|
||||
{
|
||||
"name": "daybreak",
|
||||
"description": "CDN for Daybreak Games",
|
||||
"domain_files": ["daybreak.txt"]
|
||||
},
|
||||
{
|
||||
"name": "epicgames",
|
||||
"description": "CDN for Epic Games",
|
||||
"domain_files": ["epicgames.txt"]
|
||||
},
|
||||
{
|
||||
"name": "frontier",
|
||||
"description": "CDN for Frontier Games",
|
||||
"domain_files": ["frontier.txt"]
|
||||
},
|
||||
{
|
||||
"name": "neverwinter",
|
||||
"description": "Cryptic CDN for Neverwinter",
|
||||
"domain_files": ["neverwinter.txt"]
|
||||
},
|
||||
{
|
||||
"name": "nexusmods",
|
||||
"description": "CDN for Nexus Mods/Skyrim content",
|
||||
"domain_files": ["nexusmods.txt"]
|
||||
},
|
||||
{
|
||||
"name": "nintendo",
|
||||
"description": "CDN for Nintendo consoles",
|
||||
"domain_files": ["nintendo.txt"]
|
||||
},
|
||||
{
|
||||
"name": "origin",
|
||||
"description": "CDN for Origin",
|
||||
"notes": "Should be enabled for HTTP traffic only or with a HTTPS proxy else origin client download fails",
|
||||
"mixed_content": true,
|
||||
"domain_files": ["origin.txt"]
|
||||
},
|
||||
{
|
||||
"name": "pathofexile",
|
||||
"description": "CDN for Path Of Exile",
|
||||
"domain_files": ["pathofexile.txt"]
|
||||
},
|
||||
{
|
||||
"name": "renegadex",
|
||||
"description": "CDN for Renegade X",
|
||||
"domain_files": ["renegadex.txt"]
|
||||
},
|
||||
{
|
||||
"name": "riot",
|
||||
"description": "CDN for Riot Games",
|
||||
"domain_files": ["riot.txt"]
|
||||
},
|
||||
{
|
||||
"name": "rockstar",
|
||||
"description": "CDN for Rockstar Games",
|
||||
"domain_files": ["rockstar.txt"]
|
||||
},
|
||||
{
|
||||
"name": "sony",
|
||||
"description": "CDN for Sony/PSN",
|
||||
"domain_files": ["sony.txt"]
|
||||
},
|
||||
{
|
||||
"name": "square",
|
||||
"description": "CDN for Final Fantasy XIV",
|
||||
"domain_files": ["square.txt"]
|
||||
},
|
||||
{
|
||||
"name": "steam",
|
||||
"description": "CDN for Steam",
|
||||
"domain_files": ["steam.txt"]
|
||||
},
|
||||
{
|
||||
"name": "teso",
|
||||
"description": "CDN for The Elder Scrolls Online",
|
||||
"domain_files": ["teso.txt"]
|
||||
},
|
||||
{
|
||||
"name": "test",
|
||||
"description": "CDN for Testing, recommended to enable for additional diagnostics",
|
||||
"domain_files": ["test.txt"]
|
||||
},
|
||||
{
|
||||
"name": "uplay",
|
||||
"description": "CDN for Ubisoft",
|
||||
"domain_files": ["uplay.txt"]
|
||||
},
|
||||
{
|
||||
"name": "warframe",
|
||||
"description": "CDN for Warframe",
|
||||
"domain_files": ["warframe.txt"]
|
||||
},
|
||||
{
|
||||
"name": "wargaming",
|
||||
"description": "CDN for Wargaming.net",
|
||||
"domain_files": ["wargaming.net.txt"]
|
||||
},
|
||||
{
|
||||
"name": "wsus",
|
||||
"description": "CDN for Windows Updates",
|
||||
"domain_files": ["windowsupdates.txt"]
|
||||
},
|
||||
{
|
||||
"name": "xboxlive",
|
||||
"description": "CDN for Xbox Live",
|
||||
"domain_files": ["xboxlive.txt"]
|
||||
}
|
||||
]
|
||||
}
|
||||
|
@ -1,4 +1,4 @@
|
||||
cdn-eu1.homecomingservers.com
|
||||
cdn-na1.homecomingservers.com
|
||||
cdn-na2.homecomingservers.com
|
||||
cdn-na3.homecomingservers.com
|
||||
cdn-eu1.homecomingservers.com
|
||||
|
@ -1,6 +1,2 @@
|
||||
# Daybreak games
|
||||
#
|
||||
# Do NOT cache manifest.patch.daybreakgames.com
|
||||
#
|
||||
#PS2
|
||||
pls.patch.daybreakgames.com
|
||||
|
@ -1,14 +1,14 @@
|
||||
cdn1.epicgames.com
|
||||
cdn2.epicgames.com
|
||||
cdn.unrealengine.com
|
||||
cdn1.epicgames.com
|
||||
cdn1.unrealengine.com
|
||||
cdn2.epicgames.com
|
||||
cdn2.unrealengine.com
|
||||
cdn3.unrealengine.com
|
||||
cloudflare.epicgamescdn.com
|
||||
download.epicgames.com
|
||||
download2.epicgames.com
|
||||
download3.epicgames.com
|
||||
download4.epicgames.com
|
||||
egdownload.fastly-edge.com
|
||||
epicgames-download1.akamaized.net
|
||||
fastly-download.epicgames.com
|
||||
cloudflare.epicgamescdn.com
|
||||
egdownload.fastly-edge.com
|
||||
|
10
nintendo.txt
10
nintendo.txt
@ -1,11 +1,11 @@
|
||||
*.hac.lp1.d4c.nintendo.net
|
||||
*.hac.lp1.eshop.nintendo.net
|
||||
*.wup.eshop.nintendo.net
|
||||
*.wup.shop.nintendo.net
|
||||
ccs.cdn.wup.shop.nintendo.net.edgesuite.net
|
||||
ecs-lp1.hac.shop.nintendo.net
|
||||
geisha-wup.cdn.nintendo.net
|
||||
geisha-wup.cdn.nintendo.net.edgekey.net
|
||||
idbe-wup.cdn.nintendo.net
|
||||
idbe-wup.cdn.nintendo.net.edgekey.net
|
||||
ecs-lp1.hac.shop.nintendo.net
|
||||
receive-lp1.dg.srv.nintendo.net
|
||||
*.wup.shop.nintendo.net
|
||||
*.wup.eshop.nintendo.net
|
||||
*.hac.lp1.d4c.nintendo.net
|
||||
*.hac.lp1.eshop.nintendo.net
|
||||
|
@ -1,3 +1,3 @@
|
||||
origin-a.akamaihd.net
|
||||
lvlt.cdn.ea.com
|
||||
cdn-patch.swtor.com
|
||||
lvlt.cdn.ea.com
|
||||
origin-a.akamaihd.net
|
||||
|
@ -1,2 +1,2 @@
|
||||
patches.totemarts.services
|
||||
patches.totemarts.games
|
||||
patches.totemarts.services
|
||||
|
4
riot.txt
4
riot.txt
@ -1,5 +1,5 @@
|
||||
*.dyn.riotcdn.net
|
||||
l3cdn.riotgames.com
|
||||
worldwide.l3cdn.riotgames.com
|
||||
riotgamespatcher-a.akamaihd.net
|
||||
riotgamespatcher-a.akamaihd.net.edgesuite.net
|
||||
*.dyn.riotcdn.net
|
||||
worldwide.l3cdn.riotgames.com
|
||||
|
@ -1,20 +1,20 @@
|
||||
{
|
||||
"combined_output": false,
|
||||
"ips": {
|
||||
"steam": ["10.10.3.10", "10.10.3.11"],
|
||||
"origin": "10.10.3.12",
|
||||
"blizzard": "10.10.3.13",
|
||||
"windows": "10.10.3.14",
|
||||
"riot": "10.10.3.15",
|
||||
"generic": "10.10.3.16"
|
||||
},
|
||||
"cache_domains": {
|
||||
"default": "generic",
|
||||
"blizzard": "blizzard",
|
||||
"origin": "origin",
|
||||
"riot": "riot",
|
||||
"steam": "steam",
|
||||
"wsus": "windows",
|
||||
"xboxlive": "windows"
|
||||
}
|
||||
"combined_output": false,
|
||||
"ips": {
|
||||
"steam": ["10.10.3.10", "10.10.3.11"],
|
||||
"origin": "10.10.3.12",
|
||||
"blizzard": "10.10.3.13",
|
||||
"windows": "10.10.3.14",
|
||||
"riot": "10.10.3.15",
|
||||
"generic": "10.10.3.16"
|
||||
},
|
||||
"cache_domains": {
|
||||
"default": "generic",
|
||||
"blizzard": "blizzard",
|
||||
"origin": "origin",
|
||||
"riot": "riot",
|
||||
"steam": "steam",
|
||||
"wsus": "windows",
|
||||
"xboxlive": "windows"
|
||||
}
|
||||
}
|
||||
|
@ -3,75 +3,76 @@ basedir=".."
|
||||
outputdir="output/adguardhome"
|
||||
path="${basedir}/cache_domains.json"
|
||||
|
||||
export IFS=' '
|
||||
export IFS=" "
|
||||
|
||||
test=$(which jq);
|
||||
if [ $? -gt 0 ] ; then
|
||||
echo "This script requires jq to be installed."
|
||||
echo "Your package manager should be able to find it"
|
||||
exit 1
|
||||
if ! command -v jq >/dev/null; then
|
||||
cat <<-EOF
|
||||
This script requires jq to be installed.
|
||||
Your package manager should be able to find it
|
||||
EOF
|
||||
exit 1
|
||||
fi
|
||||
|
||||
cachenamedefault="disabled"
|
||||
combinedoutput=$(jq -r ".combined_output" config.json)
|
||||
|
||||
while read line; do
|
||||
ip=$(jq ".ips[\"${line}\"]" config.json)
|
||||
declare "cacheip${line}"="${ip}"
|
||||
done <<< $(jq -r '.ips | to_entries[] | .key' config.json)
|
||||
while read -r line; do
|
||||
ip=$(jq ".ips[\"${line}\"]" config.json)
|
||||
declare "cacheip${line}"="${ip}"
|
||||
done <<<"$(jq -r ".ips | to_entries[] | .key" config.json)"
|
||||
|
||||
while read line; do
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename$line"="$name"
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' config.json)
|
||||
while read -r line; do
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename${line}"="${name}"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" config.json)"
|
||||
|
||||
rm -rf ${outputdir}
|
||||
mkdir -p ${outputdir}
|
||||
while read entry; do
|
||||
unset cacheip
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[$entry].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename="cachenamedefault"
|
||||
fi
|
||||
if [[ ${!cachename} == "disabled" ]]; then
|
||||
continue;
|
||||
fi
|
||||
cacheipname="cacheip${!cachename}"
|
||||
cacheip=$(jq -r 'if type == "array" then .[] else . end' <<< ${!cacheipname} | xargs)
|
||||
while read fileid; do
|
||||
while read filename; do
|
||||
destfilename=$(echo ${filename} | sed -e 's/txt/conf/')
|
||||
outputfile=${outputdir}/${destfilename}
|
||||
touch ${outputfile}
|
||||
while read fileentry; do
|
||||
# Ignore comments and newlines
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
domainprefix="|"
|
||||
if [[ $fileentry =~ ^\*\. ]]; then
|
||||
domainprefix="||"
|
||||
fi
|
||||
parsed=$(echo ${fileentry} | sed -e "s/^\*\.//")
|
||||
if grep -q "${domainprefix}${parsed}^\$dnsrewrite" ${outputfile}; then
|
||||
continue
|
||||
fi
|
||||
for i in ${cacheip}; do
|
||||
echo "${domainprefix}${parsed}^\$dnsrewrite=${i}" >> ${outputfile}
|
||||
echo "${domainprefix}${parsed}^\$dnstype=AAAA" >> ${outputfile}
|
||||
done
|
||||
done <<< $(cat ${basedir}/$filename | sort);
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files[${fileid}]" ${path})
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' ${path})
|
||||
while read -r entry; do
|
||||
unset cacheip
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[$entry].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename=${cachenamedefault}
|
||||
fi
|
||||
if [[ ${cachename} == "disabled" ]]; then
|
||||
continue
|
||||
fi
|
||||
cacheipname="cacheip${!cachename}"
|
||||
cacheip=$(jq -r "if type == \"array\" then .[] else . end" <<<"${!cacheipname}" | xargs)
|
||||
while read -r fileid; do
|
||||
while read -r filename; do
|
||||
destfilename=${filename//txt/conf}
|
||||
outputfile=${outputdir}/${destfilename}
|
||||
touch "${outputfile}"
|
||||
while read -r fileentry; do
|
||||
# Ignore comments and newlines
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
domainprefix="|"
|
||||
if [[ $fileentry =~ ^\*\. ]]; then
|
||||
domainprefix="||"
|
||||
fi
|
||||
parsed="${fileentry#\*\.}"
|
||||
if grep -q "${domainprefix}${parsed}^\$dnsrewrite" "${outputfile}"; then
|
||||
continue
|
||||
fi
|
||||
for i in ${cacheip}; do
|
||||
echo "${domainprefix}${parsed}^\$dnsrewrite=${i}" >>"${outputfile}"
|
||||
echo "${domainprefix}${parsed}^\$dnstype=AAAA" >>"${outputfile}"
|
||||
done
|
||||
done <<<"$(cat ${basedir}/"${filename}" | sort)"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files[${fileid}]" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" ${path})"
|
||||
|
||||
if [[ ${combinedoutput} == "true" ]]; then
|
||||
for file in ${outputdir}/*; do f=${file//${outputdir}\/} && f=${f//.conf} && echo "# ${f^}" >> ${outputdir}/lancache.conf && cat ${file} >> ${outputdir}/lancache.conf && rm ${file}; done
|
||||
for file in "${outputdir}"/*; do f=${file//${outputdir}\//} && f=${f//.conf/} && echo "# ${f^}" >>${outputdir}/lancache.conf && cat "${file}" >>${outputdir}/lancache.conf && rm "${file}"; done
|
||||
fi
|
||||
|
||||
cat << EOF
|
||||
cat <<EOF
|
||||
Configuration generation completed.
|
||||
|
||||
Please copy the following files:
|
||||
|
@ -3,72 +3,73 @@ basedir=".."
|
||||
outputdir="output/dnsmasq"
|
||||
path="${basedir}/cache_domains.json"
|
||||
|
||||
export IFS=' '
|
||||
export IFS=" "
|
||||
|
||||
test=$(which jq);
|
||||
if [ $? -gt 0 ] ; then
|
||||
echo "This script requires jq to be installed."
|
||||
echo "Your package manager should be able to find it"
|
||||
exit 1
|
||||
if ! command -v jq >/dev/null; then
|
||||
cat <<-EOF
|
||||
This script requires jq to be installed.
|
||||
Your package manager should be able to find it
|
||||
EOF
|
||||
exit 1
|
||||
fi
|
||||
|
||||
cachenamedefault="disabled"
|
||||
combinedoutput=$(jq -r ".combined_output" config.json)
|
||||
|
||||
while read -r line; do
|
||||
ip=$(jq ".ips[\"${line}\"]" config.json)
|
||||
declare "cacheip${line}"="${ip}"
|
||||
done <<< $(jq -r '.ips | to_entries[] | .key' config.json)
|
||||
ip=$(jq ".ips[\"${line}\"]" config.json)
|
||||
declare "cacheip${line}"="${ip}"
|
||||
done <<<"$(jq -r ".ips | to_entries[] | .key" config.json)"
|
||||
|
||||
while read -r line; do
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename${line}"="${name}"
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' config.json)
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename${line}"="${name}"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" config.json)"
|
||||
|
||||
rm -rf ${outputdir}
|
||||
mkdir -p ${outputdir}
|
||||
while read -r entry; do
|
||||
unset cacheip
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[${entry}].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename="cachenamedefault"
|
||||
fi
|
||||
if [[ ${!cachename} == "disabled" ]]; then
|
||||
continue;
|
||||
fi
|
||||
cacheipname="cacheip${!cachename}"
|
||||
cacheip=$(jq -r 'if type == "array" then .[] else . end' <<< ${!cacheipname} | xargs)
|
||||
while read -r fileid; do
|
||||
while read -r filename; do
|
||||
destfilename=$(echo ${filename} | sed -e 's/txt/conf/')
|
||||
outputfile=${outputdir}/${destfilename}
|
||||
touch ${outputfile}
|
||||
while read -r fileentry; do
|
||||
# Ignore comments, newlines and wildcards
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
parsed=$(echo ${fileentry} | sed -e "s/^\*\.//")
|
||||
for i in ${cacheip}; do
|
||||
if ! grep -qx "address=/${parsed}/${i}" "${outputfile}"; then
|
||||
echo "address=/${parsed}/${i}" >> "${outputfile}"
|
||||
fi
|
||||
if ! grep -qx "local=/${parsed}/" "${outputfile}"; then
|
||||
echo "local=/${parsed}/" >> "${outputfile}"
|
||||
fi
|
||||
done
|
||||
done <<< $(cat ${basedir}/${filename} | sort);
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files[$fileid]" ${path})
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' ${path})
|
||||
unset cacheip
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[${entry}].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename=${cachenamedefault}
|
||||
fi
|
||||
if [[ ${cachename} == "disabled" ]]; then
|
||||
continue
|
||||
fi
|
||||
cacheipname="cacheip${!cachename}"
|
||||
cacheip=$(jq -r "if type == \"array\" then .[] else . end" <<<"${!cacheipname}" | xargs)
|
||||
while read -r fileid; do
|
||||
while read -r filename; do
|
||||
destfilename=${filename//txt/conf}
|
||||
outputfile=${outputdir}/${destfilename}
|
||||
touch "${outputfile}"
|
||||
while read -r fileentry; do
|
||||
# Ignore comments, newlines and wildcards
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
parsed="${fileentry#\*\.}"
|
||||
for i in ${cacheip}; do
|
||||
if ! grep -qx "address=/${parsed}/${i}" "${outputfile}"; then
|
||||
echo "address=/${parsed}/${i}" >>"${outputfile}"
|
||||
fi
|
||||
if ! grep -qx "local=/${parsed}/" "${outputfile}"; then
|
||||
echo "local=/${parsed}/" >>"${outputfile}"
|
||||
fi
|
||||
done
|
||||
done <<<"$(cat ${basedir}/"${filename}" | sort)"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files[${fileid}]" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" ${path})"
|
||||
|
||||
if [[ ${combinedoutput} == "true" ]]; then
|
||||
for file in ${outputdir}/*; do f=${file//${outputdir}\/} && f=${f//.conf} && echo "# ${f^}" >> ${outputdir}/lancache.conf && cat ${file} >> ${outputdir}/lancache.conf && rm ${file}; done
|
||||
for file in "${outputdir}"/*; do f=${file//${outputdir}\//} && f=${f//.conf/} && echo "# ${f^}" >>${outputdir}/lancache.conf && cat "${file}" >>${outputdir}/lancache.conf && rm "${file}"; done
|
||||
fi
|
||||
|
||||
cat << EOF
|
||||
cat <<EOF
|
||||
Configuration generation completed.
|
||||
|
||||
Please copy the following files:
|
||||
|
160
scripts/create-nextdns.py
Normal file
160
scripts/create-nextdns.py
Normal file
@ -0,0 +1,160 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
import json
|
||||
import os
|
||||
import random
|
||||
import sys
|
||||
import time
|
||||
import urllib.error
|
||||
import urllib.request
|
||||
|
||||
# Configuration
|
||||
CACHE_DOMAINS_JSON = "cache_domains.json"
|
||||
|
||||
# Timing / Retry configuration
|
||||
DELAY_SECONDS = 1 # delay between requests
|
||||
MAX_RETRIES = 5 # maximum retries if rate limited
|
||||
BACKOFF_FACTOR = 2 # exponential backoff multiplier
|
||||
USER_AGENT = "python-requests/2.31.0" # for some reason, Cloudflare will block urllib's default user-agent
|
||||
|
||||
|
||||
def post_redirect(base_url, api_key, payload):
|
||||
"""Post a redirect entry using urllib.request. Implements retry logic if rate limited (429) is encountered."""
|
||||
headers = {
|
||||
"X-Api-Key": api_key,
|
||||
"Content-Type": "application/json",
|
||||
"User-Agent": USER_AGENT,
|
||||
}
|
||||
data = json.dumps(payload).encode("utf-8")
|
||||
retries = 0
|
||||
current_delay = DELAY_SECONDS
|
||||
|
||||
while retries <= MAX_RETRIES:
|
||||
req = urllib.request.Request(
|
||||
url=base_url, data=data, headers=headers, method="POST"
|
||||
)
|
||||
try:
|
||||
with urllib.request.urlopen(req) as response:
|
||||
status_code = response.getcode()
|
||||
if status_code in [200, 201]:
|
||||
return True, response.read().decode("utf-8")
|
||||
else:
|
||||
# For non-success status codes (other than 429)
|
||||
return False, f"Unexpected status code: {status_code}"
|
||||
except urllib.error.HTTPError as e:
|
||||
if e.code == 429:
|
||||
print(f"[*] Rate limited, waiting for {current_delay} seconds ...")
|
||||
time.sleep(current_delay)
|
||||
retries += 1
|
||||
current_delay *= BACKOFF_FACTOR
|
||||
else:
|
||||
try:
|
||||
error_body = e.read().decode("utf-8")
|
||||
except Exception:
|
||||
error_body = "No response body"
|
||||
return False, f"HTTPError {e.code}, response: {error_body}"
|
||||
except urllib.error.URLError as e:
|
||||
return False, f"URLError: {e.reason}"
|
||||
|
||||
return False, "Max retries exceeded"
|
||||
|
||||
|
||||
def run(profile_id, api_key, redirect_ip):
|
||||
base_url = f"https://api.nextdns.io/profiles/{profile_id}/rewrites/"
|
||||
|
||||
# Read cache_domains.json
|
||||
try:
|
||||
with open(CACHE_DOMAINS_JSON, "r") as f:
|
||||
cache_data = json.load(f)
|
||||
except Exception as e:
|
||||
print(f"[-] Failed to load {CACHE_DOMAINS_JSON}: {e}")
|
||||
return
|
||||
|
||||
# Set to deduplicate domains
|
||||
all_domains = set()
|
||||
|
||||
# Process each CDN entry in the JSON
|
||||
for entry in cache_data.get("cache_domains", []):
|
||||
domain_files = entry.get("domain_files", [])
|
||||
for file_name in domain_files:
|
||||
if os.path.exists(file_name):
|
||||
with open(file_name, "r") as file:
|
||||
# Read each line; ignore blank lines or comment lines
|
||||
for line in file:
|
||||
line = line.strip()
|
||||
if not line or line.startswith("#") or line.startswith("//"):
|
||||
continue
|
||||
all_domains.add(line.lstrip("*."))
|
||||
else:
|
||||
print(f"[-] File '{file_name}' not found, skipping ...")
|
||||
|
||||
print("[*] Collected domains:")
|
||||
for domain in sorted(all_domains):
|
||||
print(f" - {domain}")
|
||||
|
||||
# Retrieve the existing rewrite entries from NextDNS API
|
||||
headers = {
|
||||
"X-Api-Key": api_key,
|
||||
"User-Agent": USER_AGENT,
|
||||
}
|
||||
req = urllib.request.Request(url=base_url, headers=headers, method="GET")
|
||||
try:
|
||||
with urllib.request.urlopen(req) as response:
|
||||
if response.getcode() != 200:
|
||||
resp_body = response.read().decode("utf-8")
|
||||
print(
|
||||
f"[-] Failed to get existing redirects, status code: {response.getcode()}, response: {resp_body}"
|
||||
)
|
||||
return
|
||||
resp_data = json.loads(response.read().decode("utf-8"))
|
||||
except urllib.error.HTTPError as e:
|
||||
try:
|
||||
error_body = e.read().decode("utf-8")
|
||||
except Exception:
|
||||
error_body = "No response body"
|
||||
print(
|
||||
f"[-] Failed to get existing redirects, status code: {e.code}, response: {error_body}"
|
||||
)
|
||||
return
|
||||
except urllib.error.URLError as e:
|
||||
print(f"[-] Failed to get existing redirects, URLError: {e.reason}")
|
||||
return
|
||||
|
||||
data = resp_data.get("data", [])
|
||||
existing_domains = {entry.get("name") for entry in data}
|
||||
print("\n[*] Existing domains:")
|
||||
for domain in sorted(existing_domains):
|
||||
print(f" - {domain}")
|
||||
|
||||
# For each domain, if missing in NextDNS, post a new redirect
|
||||
for domain in all_domains:
|
||||
if domain in existing_domains:
|
||||
print(f"[*] Domain '{domain}' already exists, skipping...")
|
||||
continue
|
||||
|
||||
payload = {
|
||||
"name": domain,
|
||||
"content": redirect_ip,
|
||||
}
|
||||
|
||||
print(f"[+] Adding '{domain}'...")
|
||||
success, post_resp = post_redirect(base_url, api_key, payload)
|
||||
if not success:
|
||||
print(f"[-] Failed to add redirect for '{domain}', response: {post_resp}")
|
||||
|
||||
# Delay between API calls to prevent triggering rate limits
|
||||
time.sleep(DELAY_SECONDS + random.uniform(0, 1))
|
||||
|
||||
print("\n[+] Done!")
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
if len(sys.argv) != 4:
|
||||
print("Usage: create-nextdns.py <profile_id> <api_key> <redirect_ip>")
|
||||
sys.exit(1)
|
||||
|
||||
profile_id = sys.argv[1]
|
||||
api_key = sys.argv[2]
|
||||
redirect_ip = sys.argv[3]
|
||||
|
||||
run(profile_id, api_key, redirect_ip)
|
@ -4,97 +4,97 @@ outputdir="output/rpz"
|
||||
path="${basedir}/cache_domains.json"
|
||||
basedomain=${1:-lancache.net}
|
||||
|
||||
export IFS=' '
|
||||
export IFS=" "
|
||||
|
||||
test=$(which jq);
|
||||
if [ $? -gt 0 ] ; then
|
||||
echo "This script requires jq to be installed."
|
||||
echo "Your package manager should be able to find it"
|
||||
exit 1
|
||||
if ! command -v jq >/dev/null; then
|
||||
cat <<-EOF
|
||||
This script requires jq to be installed.
|
||||
Your package manager should be able to find it
|
||||
EOF
|
||||
exit 1
|
||||
fi
|
||||
|
||||
cachenamedefault="disabled"
|
||||
|
||||
while read line; do
|
||||
ip=$(jq ".ips[\"${line}\"]" config.json)
|
||||
declare "cacheip${line}"="${ip}"
|
||||
done <<< $(jq -r '.ips | to_entries[] | .key' config.json)
|
||||
while read -r line; do
|
||||
ip=$(jq ".ips[\"${line}\"]" config.json)
|
||||
declare "cacheip${line}"="${ip}"
|
||||
done <<<"$(jq -r ".ips | to_entries[] | .key" config.json)"
|
||||
|
||||
while read line; do
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename${line}"="${name}"
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' config.json)
|
||||
while read -r line; do
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename${line}"="${name}"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" config.json)"
|
||||
|
||||
rm -rf ${outputdir}
|
||||
mkdir -p ${outputdir}
|
||||
outputfile=${outputdir}/db.rpz.${basedomain}
|
||||
cat > ${outputfile} << EOF
|
||||
cat >"${outputfile}" <<EOF
|
||||
\$TTL 60 ; default TTL
|
||||
\$ORIGIN rpz.${basedomain}.
|
||||
@ SOA ns1.${basedomain}. admin.${basedomain}. (
|
||||
@ SOA ns1.${basedomain}. admin.${basedomain}. (
|
||||
$(date +%Y%m%d01) ; serial
|
||||
604800 ; refresh (1 week)
|
||||
600 ; retry (10 mins)
|
||||
600 ; expire (10 mins)
|
||||
600 ; minimum (10 mins)
|
||||
)
|
||||
NS ns1.${basedomain}.
|
||||
NS ns2.${basedomain}.
|
||||
604800 ; refresh (1 week)
|
||||
600 ; retry (10 mins)
|
||||
600 ; expire (10 mins)
|
||||
600 ; minimum (10 mins)
|
||||
)
|
||||
NS ns1.${basedomain}.
|
||||
NS ns2.${basedomain}.
|
||||
|
||||
EOF
|
||||
|
||||
while read entry; do
|
||||
unset cacheip
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[${entry}].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename="cachenamedefault"
|
||||
fi
|
||||
if [[ ${!cachename} == "disabled" ]]; then
|
||||
continue;
|
||||
fi
|
||||
cacheipname="cacheip${!cachename}"
|
||||
cacheip=$(jq -r 'if type == "array" then .[] else . end' <<< ${!cacheipname} | xargs)
|
||||
while read fileid; do
|
||||
while read filename; do
|
||||
echo "" >> ${outputfile}
|
||||
echo "; $(echo ${filename} | sed -e 's/.txt$//')" >> ${outputfile}
|
||||
destfilename=$(echo ${filename} | sed -e 's/txt/conf/')
|
||||
while read fileentry; do
|
||||
# Ignore comments and newlines
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
parsed=$(echo ${fileentry})
|
||||
if grep -qx "^\"${parsed}\". " ${outputfile}; then
|
||||
continue
|
||||
fi
|
||||
t=""
|
||||
for i in ${cacheip}; do
|
||||
# only one cname per domain is allowed
|
||||
if [[ ${t} = "CNAME" ]]; then
|
||||
continue
|
||||
fi
|
||||
# for cnames you must use a fqdn with trailing dot
|
||||
t="CNAME"
|
||||
if [[ ${i} =~ ^[0-9\.]+$ ]] ; then
|
||||
t="A"
|
||||
elif [[ ! ${i} =~ \.$ ]] ; then
|
||||
i="${i}."
|
||||
fi
|
||||
printf "%-50s IN %s %s\n" \
|
||||
"${parsed}" \
|
||||
"${t}" \
|
||||
"${i}" \
|
||||
>> ${outputfile}
|
||||
done
|
||||
done <<< $(cat ${basedir}/${filename} | sort);
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files[${fileid}]" ${path})
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' ${path})
|
||||
while read -r entry; do
|
||||
unset cacheip
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[${entry}].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename=${cachenamedefault}
|
||||
fi
|
||||
if [[ ${cachename} == "disabled" ]]; then
|
||||
continue
|
||||
fi
|
||||
cacheipname="cacheip${!cachename}"
|
||||
cacheip=$(jq -r "if type == \"array\" then .[] else . end" <<<"${!cacheipname}" | xargs)
|
||||
while read -r fileid; do
|
||||
while read -r filename; do
|
||||
echo "" >>"${outputfile}"
|
||||
echo "; ${filename//.txt//}" >>"${outputfile}"
|
||||
while read -r fileentry; do
|
||||
# Ignore comments and newlines
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
parsed=${fileentry}
|
||||
if grep -qx "^\"${parsed}\". " "${outputfile}"; then
|
||||
continue
|
||||
fi
|
||||
t=""
|
||||
for i in ${cacheip}; do
|
||||
# only one cname per domain is allowed
|
||||
if [[ ${t} = "CNAME" ]]; then
|
||||
continue
|
||||
fi
|
||||
# for cnames you must use a fqdn with trailing dot
|
||||
t="CNAME"
|
||||
if [[ ${i} =~ ^[0-9\.]+$ ]]; then
|
||||
t="A"
|
||||
elif [[ ! ${i} =~ \.$ ]]; then
|
||||
i="${i}."
|
||||
fi
|
||||
printf "%-50s IN %s %s\n" \
|
||||
"${parsed}" \
|
||||
"${t}" \
|
||||
"${i}" \
|
||||
>>"${outputfile}"
|
||||
done
|
||||
done <<<"$(cat ${basedir}/"${filename}" | sort)"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files[${fileid}]" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" ${path})"
|
||||
|
||||
cat << EOF
|
||||
cat <<EOF
|
||||
Configuration generation completed.
|
||||
|
||||
Please include the rpz zone in your bind configuration"
|
||||
@ -102,12 +102,12 @@ Please include the rpz zone in your bind configuration"
|
||||
- configure the zone and use it
|
||||
|
||||
options {
|
||||
[...]
|
||||
response-policy {zone "rpz.${basedomain}";};
|
||||
[...]
|
||||
[...]
|
||||
response-policy {zone "rpz.${basedomain}";};
|
||||
[...]
|
||||
}
|
||||
zone "rpz.$basedomain" {
|
||||
type master;
|
||||
file "/etc/bind/db.rpz.${basedomain}";
|
||||
type master;
|
||||
file "/etc/bind/db.rpz.${basedomain}";
|
||||
};
|
||||
EOF
|
||||
|
@ -4,65 +4,65 @@ outputdir="output/squid"
|
||||
path="${basedir}/cache_domains.json"
|
||||
REGEX="^\\*\\.(.*)$"
|
||||
|
||||
export IFS=' '
|
||||
export IFS=" "
|
||||
|
||||
test=$(which jq);
|
||||
if [ $? -gt 0 ] ; then
|
||||
echo "This script requires jq to be installed."
|
||||
echo "Your package manager should be able to find it"
|
||||
exit 1
|
||||
if ! command -v jq >/dev/null; then
|
||||
cat <<-EOF
|
||||
This script requires jq to be installed.
|
||||
Your package manager should be able to find it
|
||||
EOF
|
||||
exit 1
|
||||
fi
|
||||
|
||||
cachenamedefault="disabled"
|
||||
|
||||
while read -r line; do
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename${line}"="${name}"
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' config.json)
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename${line}"="${name}"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" config.json)"
|
||||
|
||||
rm -rf ${outputdir}
|
||||
mkdir -p ${outputdir}
|
||||
while read -r entry; do
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[$entry].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename="cachenamedefault"
|
||||
fi
|
||||
if [[ ${!cachename} == "disabled" ]]; then
|
||||
continue;
|
||||
fi
|
||||
while read -r fileid; do
|
||||
while read -r filename; do
|
||||
destfilename=$(echo ${!cachename}.txt)
|
||||
outputfile=${outputdir}/${destfilename}
|
||||
touch ${outputfile}
|
||||
while read -r fileentry; do
|
||||
# Ignore comments
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
# Ha wildcards to squid wildcards
|
||||
parsed=$(echo ${fileentry} | sed -e "s/^\*\./\./")
|
||||
# If we have cdn.thing and *.cdn.thing in cache_domains
|
||||
# Squid requires ONLY cdn.thing
|
||||
#
|
||||
# If the fileentry starts with *.cdn.thing
|
||||
if [[ ${fileentry} =~ $REGEX ]]; then
|
||||
# Does the cache_domains file also contain cdn.thing
|
||||
grep "${BASH_REMATCH[1]}" ${basedir}/${filename} | grep -v "${fileentry}" > /dev/null
|
||||
if [[ $? -eq 0 ]]; then
|
||||
# Skip *.cdn.thing as cdn.thing will be collected earlier/later
|
||||
continue
|
||||
fi
|
||||
fi
|
||||
echo "${parsed}" >> "${outputfile}"
|
||||
done <<< $(cat ${basedir}/${filename} | sort);
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files[${fileid}]" ${path})
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' ${path})
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[$entry].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename=${cachenamedefault}
|
||||
fi
|
||||
if [[ ${cachename} == "disabled" ]]; then
|
||||
continue
|
||||
fi
|
||||
while read -r fileid; do
|
||||
while read -r filename; do
|
||||
destfilename=${!cachename}.txt
|
||||
outputfile=${outputdir}/${destfilename}
|
||||
touch "${outputfile}"
|
||||
while read -r fileentry; do
|
||||
# Ignore comments
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
# Ha wildcards to squid wildcards
|
||||
parsed="${fileentry#\*}"
|
||||
# If we have cdn.thing and *.cdn.thing in cache_domains
|
||||
# Squid requires ONLY cdn.thing
|
||||
#
|
||||
# If the fileentry starts with *.cdn.thing
|
||||
if [[ ${fileentry} =~ $REGEX ]]; then
|
||||
# Does the cache_domains file also contain cdn.thing
|
||||
if grep "${BASH_REMATCH[1]}" ${basedir}/"${filename}" | grep -v "${fileentry}" >/dev/null; then
|
||||
# Skip *.cdn.thing as cdn.thing will be collected earlier/later
|
||||
continue
|
||||
fi
|
||||
fi
|
||||
echo "${parsed}" >>"${outputfile}"
|
||||
done <<<"$(cat ${basedir}/"${filename}" | sort)"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files[${fileid}]" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" ${path})"
|
||||
|
||||
cat << EOF
|
||||
cat <<EOF
|
||||
Configuration generation completed.
|
||||
|
||||
Please copy the following files:
|
||||
|
@ -3,74 +3,75 @@ basedir=".."
|
||||
outputdir="output/unbound"
|
||||
path="${basedir}/cache_domains.json"
|
||||
|
||||
export IFS=' '
|
||||
export IFS=" "
|
||||
|
||||
test=$(which jq);
|
||||
if [ $? -gt 0 ] ; then
|
||||
echo "This script requires jq to be installed."
|
||||
echo "Your package manager should be able to find it"
|
||||
exit 1
|
||||
if ! command -v jq >/dev/null; then
|
||||
cat <<-EOF
|
||||
This script requires jq to be installed.
|
||||
Your package manager should be able to find it
|
||||
EOF
|
||||
exit 1
|
||||
fi
|
||||
|
||||
cachenamedefault="disabled"
|
||||
combinedoutput=$(jq -r ".combined_output" config.json)
|
||||
|
||||
while read line; do
|
||||
ip=$(jq ".ips[\"${line}\"]" config.json)
|
||||
declare "cacheip${line}"="${ip}"
|
||||
done <<< $(jq -r '.ips | to_entries[] | .key' config.json)
|
||||
while read -r line; do
|
||||
ip=$(jq ".ips[\"${line}\"]" config.json)
|
||||
declare "cacheip${line}"="${ip}"
|
||||
done <<<"$(jq -r ".ips | to_entries[] | .key" config.json)"
|
||||
|
||||
while read line; do
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename${line}"="${name}"
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' config.json)
|
||||
while read -r line; do
|
||||
name=$(jq -r ".cache_domains[\"${line}\"]" config.json)
|
||||
declare "cachename${line}"="${name}"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" config.json)"
|
||||
|
||||
rm -rf ${outputdir}
|
||||
mkdir -p ${outputdir}
|
||||
while read entry; do
|
||||
unset cacheip
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[${entry}].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename="cachenamedefault"
|
||||
fi
|
||||
if [[ ${!cachename} == "disabled" ]]; then
|
||||
continue;
|
||||
fi
|
||||
cacheipname="cacheip${!cachename}"
|
||||
cacheip=$(jq -r 'if type == "array" then .[] else . end' <<< ${!cacheipname} | xargs)
|
||||
while read fileid; do
|
||||
while read filename; do
|
||||
destfilename=$(echo ${filename} | sed -e 's/txt/conf/')
|
||||
outputfile=${outputdir}/${destfilename}
|
||||
touch ${outputfile}
|
||||
while read fileentry; do
|
||||
# Ignore comments and newlines
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
parsed=$(echo ${fileentry} | sed -e "s/^\*\.//")
|
||||
if grep -qx " local-zone: \"${parsed}\" redirect" ${outputfile}; then
|
||||
continue
|
||||
fi
|
||||
if [[ $(head -n 1 ${outputfile}) != "server:" ]]; then
|
||||
echo "server:" >> ${outputfile}
|
||||
fi
|
||||
echo " local-zone: \"${parsed}\" redirect" >> ${outputfile}
|
||||
for i in ${cacheip}; do
|
||||
echo " local-data: \"${parsed} 30 IN A ${i}\"" >> ${outputfile}
|
||||
done
|
||||
done <<< $(cat ${basedir}/${filename} | sort);
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files[${fileid}]" ${path})
|
||||
done <<< $(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})
|
||||
done <<< $(jq -r '.cache_domains | to_entries[] | .key' ${path})
|
||||
while read -r entry; do
|
||||
unset cacheip
|
||||
unset cachename
|
||||
key=$(jq -r ".cache_domains[${entry}].name" ${path})
|
||||
cachename="cachename${key}"
|
||||
if [ -z "${!cachename}" ]; then
|
||||
cachename=${cachenamedefault}
|
||||
fi
|
||||
if [[ ${cachename} == "disabled" ]]; then
|
||||
continue
|
||||
fi
|
||||
cacheipname="cacheip${!cachename}"
|
||||
cacheip=$(jq -r "if type == \"array\" then .[] else . end" <<<"${!cacheipname}" | xargs)
|
||||
while read -r fileid; do
|
||||
while read -r filename; do
|
||||
destfilename=${filename//txt/conf}
|
||||
outputfile=${outputdir}/${destfilename}
|
||||
touch "${outputfile}"
|
||||
while read -r fileentry; do
|
||||
# Ignore comments and newlines
|
||||
if [[ ${fileentry} == \#* ]] || [[ -z ${fileentry} ]]; then
|
||||
continue
|
||||
fi
|
||||
parsed="${fileentry#\*\.}"
|
||||
if grep -qx " local-zone: \"${parsed}\" redirect" "${outputfile}"; then
|
||||
continue
|
||||
fi
|
||||
if [[ $(head -n 1 "${outputfile}") != "server:" ]]; then
|
||||
echo "server:" >>"${outputfile}"
|
||||
fi
|
||||
echo " local-zone: \"${parsed}\" redirect" >>"${outputfile}"
|
||||
for i in ${cacheip}; do
|
||||
echo " local-data: \"${parsed} 30 IN A ${i}\"" >>"${outputfile}"
|
||||
done
|
||||
done <<<"$(cat ${basedir}/"${filename}" | sort)"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files[${fileid}]" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains[${entry}].domain_files | to_entries[] | .key" ${path})"
|
||||
done <<<"$(jq -r ".cache_domains | to_entries[] | .key" ${path})"
|
||||
|
||||
if [[ ${combinedoutput} == "true" ]]; then
|
||||
for file in ${outputdir}/*; do f=${file//${outputdir}\/} && f=${f//.conf} && echo "# ${f^}" >> ${outputdir}/lancache.conf && cat ${file} >> ${outputdir}/lancache.conf && rm ${file}; done
|
||||
for file in "${outputdir}"/*; do f=${file//${outputdir}\//} && f=${f//.conf/} && echo "# ${f^}" >>${outputdir}/lancache.conf && cat "${file}" >>${outputdir}/lancache.conf && rm "${file}"; done
|
||||
fi
|
||||
|
||||
cat << EOF
|
||||
cat <<EOF
|
||||
Configuration generation completed.
|
||||
|
||||
Please copy the following files:
|
||||
|
18
sony.txt
18
sony.txt
@ -1,16 +1,16 @@
|
||||
gs2.ww.prod.dl.playstation.net
|
||||
*.gs2.ww.prod.dl.playstation.net
|
||||
*.gs2.sonycoment.loris-e.llnwd.net
|
||||
gs2-ww-prod.psn.akadns.net
|
||||
*.gs2-ww-prod.psn.akadns.net
|
||||
gs2.ww.prod.dl.playstation.net.edgesuite.net
|
||||
*.gs2.sonycoment.loris-e.llnwd.net
|
||||
*.gs2.ww.prod.dl.playstation.net
|
||||
*.gs2.ww.prod.dl.playstation.net.edgesuite.net
|
||||
gs-sec.ww.np.dl.playstation.net
|
||||
gs2-ww-prod.psn.akadns.net
|
||||
gs2.ww.prod.dl.playstation.net
|
||||
gs2.ww.prod.dl.playstation.net.edgesuite.net
|
||||
gst.prod.dl.playstation.net
|
||||
playstation4.sony.akadns.net
|
||||
psnobj.prod.dl.playstation.net
|
||||
sgst.prod.dl.playstation.net
|
||||
theia.dl.playstation.net
|
||||
tmdb.np.dl.playstation.net
|
||||
gs-sec.ww.np.dl.playstation.net
|
||||
uef.np.dl.playstation.net
|
||||
gst.prod.dl.playstation.net
|
||||
vulcan.dl.playstation.net
|
||||
sgst.prod.dl.playstation.net
|
||||
psnobj.prod.dl.playstation.net
|
||||
|
6
test.txt
6
test.txt
@ -1,6 +1,6 @@
|
||||
canary.uklans.net
|
||||
litmus.uklans.net
|
||||
trigger.uklans.net
|
||||
canary.lancache.net
|
||||
canary.uklans.net
|
||||
litmus.lancache.net
|
||||
litmus.uklans.net
|
||||
trigger.lancache.net
|
||||
trigger.uklans.net
|
||||
|
@ -1,16 +1,16 @@
|
||||
dl-wot-ak.wargaming.net
|
||||
dl-wot-cdx.wargaming.net
|
||||
dl-wot-gc.wargaming.net
|
||||
dl-wot-se.wargaming.net
|
||||
dl-wowp-ak.wargaming.net
|
||||
dl-wowp-cdx.wargaming.net
|
||||
dl-wowp-gc.wargaming.net
|
||||
dl-wowp-se.wargaming.net
|
||||
dl-wows-ak.wargaming.net
|
||||
dl-wows-cdx.wargaming.net
|
||||
dl-wows-gc.wargaming.net
|
||||
dl-wows-se.wargaming.net
|
||||
dl2.wargaming.net
|
||||
wg.gcdn.co
|
||||
wgus-wotasia.wargaming.net
|
||||
dl-wot-ak.wargaming.net
|
||||
dl-wot-gc.wargaming.net
|
||||
dl-wot-se.wargaming.net
|
||||
dl-wot-cdx.wargaming.net
|
||||
dl-wows-ak.wargaming.net
|
||||
dl-wows-gc.wargaming.net
|
||||
dl-wows-se.wargaming.net
|
||||
dl-wows-cdx.wargaming.net
|
||||
dl-wowp-ak.wargaming.net
|
||||
dl-wowp-gc.wargaming.net
|
||||
dl-wowp-se.wargaming.net
|
||||
dl-wowp-cdx.wargaming.net
|
||||
wgus-woteu.wargaming.net
|
||||
|
@ -1,13 +1,13 @@
|
||||
*.windowsupdate.com
|
||||
*.dl.delivery.mp.microsoft.com
|
||||
dl.delivery.mp.microsoft.com
|
||||
*.update.microsoft.com
|
||||
*.do.dsp.mp.microsoft.com
|
||||
*.microsoft.com.edgesuite.net
|
||||
*.update.microsoft.com
|
||||
*.windowsupdate.com
|
||||
amupdatedl.microsoft.com
|
||||
amupdatedl2.microsoft.com
|
||||
amupdatedl3.microsoft.com
|
||||
amupdatedl4.microsoft.com
|
||||
amupdatedl5.microsoft.com
|
||||
dl.delivery.mp.microsoft.com
|
||||
officecdn.microsoft.com
|
||||
officecdn.microsoft.com.edgesuite.net
|
||||
|
@ -1,7 +1,7 @@
|
||||
assets1.xboxlive.com
|
||||
assets2.xboxlive.com
|
||||
xbox-mbr.xboxlive.com
|
||||
assets1.xboxlive.com.nsatc.net
|
||||
assets2.xboxlive.com
|
||||
d1.xboxlive.com
|
||||
xbox-mbr.xboxlive.com
|
||||
xvcf1.xboxlive.com
|
||||
xvcf2.xboxlive.com
|
||||
d1.xboxlive.com
|
||||
|
Loading…
Reference in New Issue
Block a user