[19:50] <snail> mōrena
[20:01] <ojwb> morning
[20:02] <ibeardslee> morning
[20:58] <chilts> morning
[21:09] <ajmitch> morning
[21:10] <snail> got bitten badly by google last week. turns out that their RSS web crawler ignores robots.txt files. asking for an RSS feed of 5000 epubs every 3 minutes without caching makes my server sad when someone leaves their google homepage open for a couple of days
[21:10] <ojwb> eep
[21:11]  * ojwb can see an argument for ignoring robots.txt there, but hitting every 3 minutes isn't sane
[21:17] <snail> the answer, naturally, is caching
[21:17] <snail> :)
[21:39] <mwhudson> morning
[22:45] <hads> It's interesting to find weaknesses due to third parties activities. I discovered that I wasn't caching nicegear's rather heavy 404 page and just about brought the server down when some PCI-DSS company ran a stupid vulnerability scan.
[22:45] <hads> And, morning.
[22:46] <lifeless> snail: well, maybe make epub cheaper to deliver ?
[22:58] <snail> lifeless: the issue was that the RSS feed was based on a solr search. the ePubs are already stored as flat files on the disk
[23:05] <lifeless> ah
[23:13] <Atamira> morning
[23:22] <kcj> Morning.