[squid-users] Regex optimization

Yuri Voinov yvoinov at gmail.com
Fri Jun 17 08:00:25 UTC 2016


-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256
 
Shallalist seems most acceptable free blacklist. With fine grained
categories.

Of course, it require human power to maintain.

If this one does not will to do, just sit back and wait for someone to
do - as is usually done in open source software, of course, it will be a
joke. There is a form to send and revoke URL. They just need to use it.
They themselves are not required to maintain a list of current.

Personally, I have been actively taking part in updating the Shalla
blacklist of the last 5 years and I am happy with it. Especially
considering its price.

So.....

17.06.2016 3:11, Benjamin E. Nichols пишет:
>
>
>
> On 6/16/2016 3:28 PM, Yuri Voinov wrote:
> I propose to nominate for the second place of the contest "The most
> inefficient use of computing resources - 2016." :-!:-D
>
> Because first place already occuped. :-D 30 millions pornsites in one
> squid's ACL and 7 minutes for squid -k refresh. 8-)
> > Yeah and Ill bet about 27 Million of them are dead, expired, parked
or redirected because your list sucks.
>
> > If you really intend to use blacklists tailored for Squid proxy
Native ACL, we are the leading and only provider of such lists.
> > And we actually query each domain daily with batch updates, dead
domains are placed into a holding pool to be queried again cyclically
and re added
> > as necessary.
>
> > Shallalist is a joke, urlblacklist is garbage, if you are serious
and need a better blacklist, we would be happy to serve you.
>
>
>
> 17.06.2016 1:20, Antony Stone пишет:
> >>> On Thursday 16 June 2016 at 21:11:50, Alfredo Rezinovsky wrote:
> >>>
> >>>> Well.. I tried.
> >>>> I need to ban 8613 URLs. Because a law.
> >>> Have you considered
https://www.urlfilterdb.com/products/ufdbguard.html ?
> >>>
> >>>> If I put one per line in a file and set the filename for an
url_regex acl
> >>>> it works. But when the traffic goes up the cpu load goes 100%
(even using
> >>>> workers) and the proxy turns unusable.
> >>> Er, I'm not surprised.
> >>>
> >>>> I tested and saw my squid can't parse regexes with more than 8192
> >>>> characters.
> >>>> I managed to combine the 8000 uris in 34 regexes using a ruby gem,
> and the
> >>>> cpu load stays almost at the same level it is without any acl (same
> >>>> traffic).
> >>> That must be *way* past anything to be described as "maintainable".
> >>>
> >>>> the regex is:
> >>> Er, thanks, that confirms my suspicions above :)
> >>>
> >>>
> >>> Antony.
> >>>
>>
>>
>>
>> _______________________________________________
>> squid-users mailing list
>> squid-users at lists.squid-cache.org
>> http://lists.squid-cache.org/listinfo/squid-users
>
> --
> Signed,
>
> Benjamin E. Nichols
> http://www.squidblacklist.org
>
> 1-405-397-1360
>
>
> _______________________________________________
> squid-users mailing list
> squid-users at lists.squid-cache.org
> http://lists.squid-cache.org/listinfo/squid-users

-----BEGIN PGP SIGNATURE-----
Version: GnuPG v2
 
iQEcBAEBCAAGBQJXY64ZAAoJENNXIZxhPexGMhkIAK/hrf9wFVKwt83jRxl4T++/
2jgcuqaXb78MazN6wdNr5HZrshRGayxDZsLbF0xb7cdrA0lVYSQgnWX0SEomZZ8E
Y+gDaemoojE8d8Wqy3zLP57BzziLMs2ymVmQRnQ0265wRoCVkCLKAR4Dzod9eXS/
L115pErOcbZc0nZ3QkXfbuE1ol3gfHL1tQN5Gfm4IMxgtp1gdnxfg9BUCzMaicR+
Y7WkZqCdzA0kcLjOgvDVzudTBsKhRx+YHIjVr72seUMEu/EYXZjBWYZRKXCRDT7h
gHIzS+ESw91LCKs8VXZ9ZBEf2RclSvFWzK9qpbR2vgYFztjsFkXAmg5UNOkASX4=
=FKFP
-----END PGP SIGNATURE-----

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.squid-cache.org/pipermail/squid-users/attachments/20160617/473c7df4/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: 0x613DEC46.asc
Type: application/pgp-keys
Size: 2437 bytes
Desc: not available
URL: <http://lists.squid-cache.org/pipermail/squid-users/attachments/20160617/473c7df4/attachment-0001.key>


More information about the squid-users mailing list