I would recommend a small daemon who logtails the existing squid log
file and pumps the data into the database of choice.
The File::Tail perl module is a good help for this job. Kind of a tail
-f on steroids.
Regards
Henrik
On Thursday 05 June 2003 18.21, Lucas Brasilino wrote:
> Hi All:
>
> Maybe I'm crossposting but anyway....
> I'm intending do develop an application to store in
> a database all users traffic in a way to easier create reports.
> I'm considering two solutions:
>
> 1) Create a daemon which will read all squids logs by a pipe
> in /dev, configure this pipe as squid's log file and store
> users_ip and URL in the database, or
>
> 2) Create a daemon which will connect to squid's port and
> send a "GET cache_object://hostname/filedescriptors",
> and store only Description field which content is "^http://"
> (URL's).
>
> I'm wondering which will be the better way to accomplish my
> duty. Suggestion?? :)
>
> PS: as a database I'll use Berkeley DB or GDBM, which are MUCH MORE
> lighter.
>
>
> thanks in advance
-- Donations welcome if you consider my Free Squid support helpful. https://www.paypal.com/xclick/business=hno%40squid-cache.org If you need commercial Squid support or cost effective Squid or firewall appliances please refer to MARA Systems AB, Sweden http://www.marasystems.com/, info@marasystems.comReceived on Thu Jun 05 2003 - 16:01:44 MDT
This archive was generated by hypermail pre-2.1.9 : Tue Dec 09 2003 - 17:17:16 MST