Hi Simon,
Easy way I could think of is to use a linux script to query the proxy to get the results and parse them accordingly. Manage to test the below with my lab proxy to find it able to pull it. Steps below
Create a file called "urls.txt" will all the urls withe one per line
- Enable HTTP-Console in ProxySG
- Create a bash script "script.sh" with below content. Change the IP address to that of proxy's
#!/bin/bash
url="http://192.168.1.10:8081/ContentFilter/TestUrl"
for i in $(cat urls.txt); do
content="$(curl -u admin:admin -s "$url/$i" | grep Blue |sed -n -e 's/^.*Coat: //p')")"
echo "$i" "$content" >> Output.txt
done
- Put script.sh and urls.txt in same directory.
- Execute script by using ./script.sh (Assuming that chmod is used to set executable permission)
- Output will be dumped in a file "Output.txt"
My Lab testing gave me the output in below format with one url per line
[root@localhost ~]# cat Output.txt
http://www.google.com Search Engines/Portals
https://www.yahoo.com Search Engines/Portals
https://www.symantec.com Technology/Internet
http://www.mathrubhumi.com News/Media
http://edition.cnn.com News/Media
http://www.playboy.com Adult/Mature Content; Entertainment
Note: This script is made after a good amount of search in Google and my minimal knowledge on bash. This might or might not work as expected