[squid-users] Sudden but sustained high bandwidth usage

joe chip_pop at hotmail.com
Sun Mar 13 13:52:23 UTC 2016


regarding swapfail 
after i suffer alot even on latest squid v
what i found is  if you have lets say 32geg ram and you specify cache_mem 10
GB or whatever size you have
if it reach that  it start happening swap fail mostly on fast smole object
like   js file or jpg  not more then 100k max 
and same problem cache_dir aufs /mnt/cache-a 500000 when you reach max
specify size dose not Mather if diskd or aufs probably rock also

what im trying to say is  during that period wen max storage happen it start
fkp 
reason  unknown

1= replacing file  last freq..used it delete from storage and it dose not
delete the object info from swap.state result swap fail

2= might be write issue  =  saving the object info to the swap.state and it
its not done on storage wen  storage full it get fkp

3 = i don't know if may be if   that developer guys should check on source
code if the write to swap.state happen before the file get stored wish is
not good
it should be  wen the file get stored and signal the write to swap.state to
save the object detail in it not  befor
as i says its what my testing and experiment that i have reach to those
point. 
 i might be wrong



--
View this message in context: http://squid-web-proxy-cache.1019090.n4.nabble.com/Sudden-but-sustained-high-bandwidth-usage-tp4676366p4676635.html
Sent from the Squid - Users mailing list archive at Nabble.com.


More information about the squid-users mailing list