usr/sbin/eximstats000075500000450406147204333040010270 0ustar00#!/usr/local/cpanel/3rdparty/perl/536/bin/perl # Copyright (c) The Exim Maintainers 2023 # Copyright (c) 2001-2017 University of Cambridge. # See the file NOTICE for conditions of use and distribution. # SPDX-License-Identifier: GPL-2.0-or-later # Perl script to generate statistics from one or more Exim log files. # Usage: eximstats [] ... # 1996-05-21: Ignore lines not starting with valid date/time, just in case # these get into a log file. # 1996-11-19: Add the -h option to control the size of the histogram, # and optionally turn it off. # Use some Perl 5 things; it should be everywhere by now. # Add the Perl -w option and rewrite so no warnings are given. # Add the -t option to control the length of the "top" listing. # Add the -ne, -nt options to turn off errors and transport # information. # Add information about length of time on queue, and -q to # control the intervals and turn it off. # Add count and percentage of delayed messages to the Received # line. # Show total number of errors. # Add count and percentage of messages with errors to Received # line. # Add information about relaying and -nr to suppress it. # 1997-02-03 Merged in some of the things Nigel Metheringham had done: # Re-worded headings # Added received histogram as well as delivered # Added local senders' league table # Added local recipients' league table # 1997-03-10 Fixed typo "destinationss" # Allow for intermediate address between final and original # when testing for relaying # Give better message when no input # 1997-04-24 Fixed bug in layout of error listing that was depending on # text length (output line got repeated). # 1997-05-06 Bug in option decoding when only one option. # Overflow bug when handling very large volumes. # 1997-10-28 Updated to handle revised log format that might show # HELO name as well as host name before IP number # 1998-01-26 Bugs in the function for calculating the number of seconds # since 1970 from a log date # 1998-02-02 Delivery to :blackhole: doesn't have a T= entry in the log # line; cope with this, thereby avoiding undefined problems # Very short log line gave substring error # 1998-02-03 A routed delivery to a local transport may not have <> in the # log line; terminate the address at white space, not < # 1998-09-07 If first line of input was a => line, $thissize was undefined; # ensure it is zero. # 1998-12-21 Adding of $thissize from => line should have been adding $size. # Oops. Should have looked more closely when fixing the previous # bug! # 1999-11-12 Increased the field widths for printed integers; numbers are # bigger than originally envisaged. # 2001-03-21 Converted seconds() routine to use Time::Local, fixing a bug # whereby seconds($timestamp) - id_seconds($id) gave an # incorrect result. # Added POD documentation. # Moved usage instructions into help() subroutine. # Added 'use strict' and declared all global variables. # Added '-html' flag and resultant code. # Added '-cache' flag and resultant code. # Added add_volume() routine and converted all volume variables # to use it, fixing the overflow problems for individual hosts # on large sites. # Converted all volume output to GB/MB/KB as appropriate. # Don't store local user stats if -nfl is specified. # Modifications done by: Steve Campbell () # 2001-04-02 Added the -t_remote_users flag. Steve Campbell. # 2001-10-15 Added the -domain flag. Steve Campbell. # 2001-10-16 Accept files on STDIN or on the command line. Steve Campbell. # 2001-10-21 Removed -domain flag and added -bydomain, -byhost, and -byemail. # We now generate our main parsing subroutine as an eval statement # which improves performance dramatically when not all the results # are required. We also cache the last timestamp to time conversion. # # NOTE: 'Top 50 destinations by (message count|volume)' lines are # now 'Top N (host|email|domain) destinations by (message count|volume)' # where N is the topcount. Steve Campbell. # # 2001-10-30 V1.16 Joachim Wieland. # Fixed minor bugs in add_volume() when taking over this version # for use in Exim 4: -w gave uninitialized value warnings in # two situations: for the first addition to a counter, and if # there were never any gigabytes, thereby leaving the $gigs # value unset. # Initialized $last_timestamp to stop a -w uninitialized warning. # Minor layout tweak for grand totals (nitpicking). # Put the IP addresses for relaying stats in [] and separated by # a space from the domain name. # Removed the IPv4-specific address test when picking out addresses # for relaying. Anything inside [] is OK. # # 2002-07-02 Philip Hazel # Fixed "uninitialized variable" message that occurred for relay # messages that arrived from H=[1.2.3.4] hosts (no name shown). # This bug didn't affect the output. # # 2002-04-15 V1.17 Joachim Wieland. # Added -charts, -chartdir. -chartrel options which use # GD::Graph modules to create graphical charts of the statistics. # # 2002-04-15 V1.18 Steve Campbell. # Added a check for $domain to to stop a -w uninitialized warning. # Added -byemaildomain option. # Only print HTML header links to included tables! # # 2002-08-02 V1.19 Steve Campbell. # Changed the debug mode to dump the parser onto STDERR rather # than STDOUT. Documented the -d flag into the help(). # Rejoined the divergent 2002-04-15 and 2002-07-02 releases. # # 2002-08-21 V1.20 Steve Campbell. # Added the '-merge' option to allow merging of previous reports. # Fixed a missing semicolon when doing -bydomain. # Make volume charts plot the data gigs and bytes rather than just bytes. # Only process log lines with $flag =~ /<=|=>|->|==|\*\*|Co/ # Converted Emaildomain to Edomain - the column header was too wide! # This changes the text output slightly. You can revert to the old # column widths by changing $COLUMN_WIDTHS to 7; # # 2002-09-04 V1.21 Andreas J Mueller # Local deliveries domain now defaults to 'localdomain'. # Don't match F= when looking for the user. # # 2002-09-05 V1.22 Steve Campbell # Fixed a perl 5.005 incompatibility problem ('our' variables). # # 2002-09-11 V1.23 Steve Campbell # Stopped -charts option from throwing errors on null data. # Don't print out 'Errors encountered' unless there are any. # 2002-10-21 V1.23a Philip Hazel - patch from Tony Finch put in until # Steve's eximstats catches up. # Handle log files that include the timezone after the timestamp. # Switch to assuming that log timestamps are in local time, with # an option for UTC timestamps, as in Exim itself. # # 2003-02-05 V1.24 Steve Campbell # Added in Sergey Sholokh's code to convert '<' and '>' characters # in HTML output. Also added code to convert them back with -merge. # Fixed timestamp offsets to convert to seconds rather than minutes. # Updated -merge to work with output files using timezones. # Added caching to speed up the calculation of timezone offsets. # # 2003-02-07 V1.25 Steve Campbell # Optimised the usage of mktime() in the seconds subroutine. # Removed the now redundant '-cache' option. # html2txt() now explicitly matches HTML tags. # Implemented a new sorting algorithm - the top_n_sort() routine. # Added Danny Carroll's '-nvr' flag and code. # # 2003-03-13 V1.26 Steve Campbell # Implemented HTML compliance changes recommended by Bernard Massot. # Bug fix to allow top_n_sort() to handle null keys. # Convert all domains and edomains to lowercase. # Remove preceding dots from domains. # # 2003-03-13 V1.27 Steve Campbell # Replaced border attributes with 'border=1', as recommended by # Bernard Massot. # # 2003-06-03 V1.28 John Newman # Added in the ability to skip over the parsing and evaluation of # specific transports as passed to eximstats via the new "-nt/.../" # command line argument. This new switch allows the viewing of # not more accurate statistics but more applicable statistics when # special transports are in use (ie; SpamAssassin). We need to be # able to ignore transports such as this otherwise the resulting # local deliveries are significantly skewed (doubled)... # # 2003-11-06 V1.29 Steve Campbell # Added the '-pattern "Description" "/pattern/"' option. # # 2004-02-17 V1.30 Steve Campbell # Added warnings if required GD::Graph modules are not available or # insufficient -chart* options are specified. # # 2004-02-20 V1.31 Andrea Balzi # Only show the Local Sender/Destination links if the tables exist. # # 2004-07-05 V1.32 Steve Campbell # Fix '-merge -h0' divide by zero error. # # 2004-07-15 V1.33 Steve Campbell # Documentation update - I've converted the subroutine # documentation from POD to comments. # # 2004-12-10 V1.34 Steve Campbell # Eximstats can now parse syslog lines as well as mainlog lines. # # 2004-12-20 V1.35 Wouter Verhelst # Pie charts by volume were actually generated by count. Fixed. # # 2005-02-07 V1.36 Gregor Herrmann / Steve Campbell # Added average sizes to HTML Top tables. # # 2005-04-26 V1.37 Frank Heydlauf # Added -xls and the ability to specify output files. # # 2005-04-29 V1.38 Steve Campbell # Use FileHandles for outputting results. # Allow any combination of xls, txt, and html output. # Fixed display of large numbers with -nvr option # Fixed merging of reports with empty tables. # # 2005-05-27 V1.39 Steve Campbell # Added the -include_original_destination flag # Removed tabs and trailing whitespace. # # 2005-06-03 V1.40 Steve Campbell # Whilst parsing the mainlog(s), store information about # the messages in a hash of arrays rather than using # individual hashes. This is a bit cleaner and results in # dramatic memory savings, albeit at a slight CPU cost. # # 2005-06-15 V1.41 Steve Campbell # Added the -show_rt flag. # Added the -show_dt flag. # # 2005-06-24 V1.42 Steve Campbell # Added Histograms for user specified patterns. # # 2005-06-30 V1.43 Steve Campbell # Bug fix for V1.42 with -h0 specified. Spotted by Chris Lear. # # 2005-07-26 V1.44 Steve Campbell # Use a glob alias rather than an array ref in the generated # parser. This improves both readability and performance. # # 2005-09-30 V1.45 Marco Gaiarin / Steve Campbell # Collect SpamAssassin and rejection statistics. # Don't display local sender or destination tables unless # there is data to show. # Added average volumes into the top table text output. # # 2006-02-07 V1.46 Steve Campbell # Collect data on the number of addresses (recipients) # as well as the number of messages. # # 2006-05-05 V1.47 Steve Campbell # Added 'Message too big' to the list of mail rejection # reasons (thanks to Marco Gaiarin). # # 2006-06-05 V1.48 Steve Campbell # Mainlog lines which have GMT offsets and are too short to # have a flag are now skipped. # # 2006-11-10 V1.49 Alain Williams # Added the -emptyok flag. # # 2006-11-16 V1.50 Steve Campbell # Fixes for obtaining the IP address from reject messages. # # 2006-11-27 V1.51 Steve Campbell # Another update for obtaining the IP address from reject messages. # # 2006-11-27 V1.52 Steve Campbell # Tally any reject message containing SpamAssassin. # # 2007-01-31 V1.53 Philip Hazel # Allow for [pid] after date in log lines # # 2007-02-14 V1.54 Daniel Tiefnig # Improved the '($parent) =' pattern match. # # 2007-03-19 V1.55 Steve Campbell # Differentiate between permanent and temporary rejects. # # 2007-03-29 V1.56 Jez Hancock # Fixed some broken HTML links and added missing column headers. # # 2007-03-30 V1.57 Steve Campbell # Fixed Grand Total Summary Domains, Edomains, and Email columns # for Rejects, Temp Rejects, Ham, and Spam rows. # # 2007-04-11 V1.58 Steve Campbell # Fix to get <> and blackhole to show in edomain tables. # # 2007-09-20 V1.59 Steve Campbell # Added the -bylocaldomain option # # 2007-09-20 V1.60 Heiko Schlittermann # Fix for misinterpreted log lines # # 2013-01-14 V1.61 Steve Campbell # Watch out for senders sending "HELO [IpAddr]" # # # For documentation on the logfile format, see # http://www.exim.org/exim-html-4.50/doc/html/spec_48.html#IX2793 =head1 NAME eximstats - generates statistics from Exim mainlog or syslog files. =head1 SYNOPSIS eximstats [Output] [Options] mainlog1 mainlog2 ... eximstats -merge [Options] report.1.txt report.2.txt ... > weekly_report.txt =head2 Output: =over 4 =item B<-txt> Output the results in plain text to STDOUT. =item B<-txt>=I Output the results in plain text. Filename '-' for STDOUT is accepted. =item B<-html> Output the results in HTML to STDOUT. =item B<-html>=I Output the results in HTML. Filename '-' for STDOUT is accepted. =item B<-xls> Output the results in Excel compatible Format to STDOUT. Requires the Spreadsheet::WriteExcel CPAN module. =item B<-xls>=I Output the results in Excel compatible format. Filename '-' for STDOUT is accepted. =back =head2 Options: =over 4 =item B<-h>I histogram divisions per hour. The default is 1, and 0 suppresses histograms. Valid values are: 0, 1, 2, 3, 5, 10, 15, 20, 30 or 60. =item B<-ne> Don't display error information. =item B<-nr> Don't display relaying information. =item B<-nr>I Don't display relaying information that matches. =item B<-nt> Don't display transport information. =item B<-nt>I Don't display transport information that matches =item B<-q>I List of times for queuing information single 0 item suppresses. =item B<-t>I Display top sources/destinations default is 50, 0 suppresses top listing. =item B<-tnl> Omit local sources/destinations in top listing. =item B<-t_remote_users> Include remote users in the top source/destination listings. =item B<-include_original_destination> Include the original destination email addresses rather than just using the final ones. Useful for finding out which of your mailing lists are receiving mail. =item B<-show_dt>I Show the delivery times (B
)for all the messages. Exim must have been configured to use the +deliver_time logging option for this option to work. I is an optional list of times. Eg -show_dt1,2,4,8 will show the number of messages with delivery times under 1 second, 2 seconds, 4 seconds, 8 seconds, and over 8 seconds. =item B<-show_rt>I Show the receipt times for all the messages. The receipt time is defined as the Completed hh:mm:ss - queue_time_overall - the Receipt hh:mm:ss. These figures will be skewed by pipelined messages so might not be that useful. Exim must have been configured to use the +queue_time_overall logging option for this option to work. I is an optional list of times. Eg -show_rt1,2,4,8 will show the number of messages with receipt times under 1 second, 2 seconds, 4 seconds, 8 seconds, and over 8 seconds. =item B<-byhost> Show results by sending host. This may be combined with B<-bydomain> and/or B<-byemail> and/or B<-byedomain>. If none of these options are specified, then B<-byhost> is assumed as a default. =item B<-bydomain> Show results by sending domain. May be combined with B<-byhost> and/or B<-byemail> and/or B<-byedomain>. =item B<-byemail> Show results by sender's email address. May be combined with B<-byhost> and/or B<-bydomain> and/or B<-byedomain>. =item B<-byemaildomain> or B<-byedomain> Show results by sender's email domain. May be combined with B<-byhost> and/or B<-bydomain> and/or B<-byemail>. =item B<-pattern> I I Look for the specified pattern and count the number of lines in which it appears. This option can be specified multiple times. Eg: -pattern 'Refused connections' '/refused connection/' =item B<-merge> This option allows eximstats to merge old eximstat reports together. Eg: eximstats mainlog.sun > report.sun.txt eximstats mainlog.mon > report.mon.txt eximstats mainlog.tue > report.tue.txt eximstats mainlog.wed > report.web.txt eximstats mainlog.thu > report.thu.txt eximstats mainlog.fri > report.fri.txt eximstats mainlog.sat > report.sat.txt eximstats -merge report.*.txt > weekly_report.txt eximstats -merge -html report.*.txt > weekly_report.html =over 4 =item * You can merge text or html reports and output the results as text or html. =item * You can use all the normal eximstat output options, but only data included in the original reports can be shown! =item * When merging reports, some loss of accuracy may occur in the top I lists. This will be towards the ends of the lists. =item * The order of items in the top I lists may vary when the data volumes round to the same value. =back =item B<-charts> Create graphical charts to be displayed in HTML output. Only valid in combination with I<-html>. This requires the following modules which can be obtained from http://www.cpan.org/modules/01modules.index.html =over 4 =item GD =item GDTextUtil =item GDGraph =back To install these, download and unpack them, then use the normal perl installation procedure: perl Makefile.PL make make test make install =item B<-chartdir>I Create the charts in the directory =item B<-chartrel>I Specify the relative directory for the "img src=" tags from where to include the charts =item B<-emptyok> Specify that it's OK to not find any valid log lines. Without this we will output an error message if we don't find any. =item B<-d> Debug flag. This outputs the eval()'d parser onto STDOUT which makes it easier to trap errors in the eval section. Remember to add 1 to the line numbers to allow for the title! =back =head1 DESCRIPTION Eximstats parses exim mainlog and syslog files to output a statistical analysis of the messages processed. By default, a text analysis is generated, but you can request other output formats using flags. See the help (B<-help>) to learn about how to create charts from the tables. =head1 AUTHOR There is a website at https://www.exim.org - this contains details of the mailing list exim-users@exim.org. =head1 TO DO This program does not perfectly handle messages whose received and delivered log lines are in different files, which can happen when you have multiple mail servers and a message cannot be immediately delivered. Fixing this could be tricky... Merging of xls files is not (yet) possible. Be free to implement :) =cut use warnings; use integer; BEGIN { pop @INC if $INC[-1] eq '.' }; use strict; use IO::File; use File::Basename; # use Time::Local; # PH/FANF use POSIX; if (@ARGV and ($ARGV[0] eq '--version' || $ARGV[0] eq '-v')) { print basename($0) . ": $0\n", "build: 4.98\n", "perl(runtime): $]\n"; exit 0; } use vars qw($HAVE_GD_Graph_pie $HAVE_GD_Graph_linespoints $HAVE_Spreadsheet_WriteExcel); eval { require GD::Graph::pie; }; $HAVE_GD_Graph_pie = $@ ? 0 : 1; eval { require GD::Graph::linespoints; }; $HAVE_GD_Graph_linespoints = $@ ? 0 : 1; eval { require Spreadsheet::WriteExcel; }; $HAVE_Spreadsheet_WriteExcel = $@ ? 0 : 1; ################################################## # Static data # ################################################## # 'use vars' instead of 'our' as perl5.005 is still in use out there! use vars qw(@tab62 @days_per_month $gig); use vars qw($VERSION); use vars qw($COLUMN_WIDTHS); use vars qw($WEEK $DAY $HOUR $MINUTE); @tab62 = (0,1,2,3,4,5,6,7,8,9,0,0,0,0,0,0, # 0-9 0,10,11,12,13,14,15,16,17,18,19,20, # A-K 21,22,23,24,25,26,27,28,29,30,31,32, # L-W 33,34,35, 0, 0, 0, 0, 0, # X-Z 0,36,37,38,39,40,41,42,43,44,45,46, # a-k 47,48,49,50,51,52,53,54,55,56,57,58, # l-w 59,60,61); # x-z @days_per_month = (0, 31, 59, 90, 120, 151, 181, 212, 243, 273, 304, 334); $gig = 1024 * 1024 * 1024; $VERSION = '1.61'; # How much space do we allow for the Hosts/Domains/Emails/Edomains column headers? $COLUMN_WIDTHS = 8; $MINUTE = 60; $HOUR = 60 * $MINUTE; $DAY = 24 * $HOUR; $WEEK = 7 * $DAY; # Declare global variables. use vars qw($total_received_data $total_received_data_gigs $total_received_count); use vars qw($total_delivered_data $total_delivered_data_gigs $total_delivered_messages $total_delivered_addresses); use vars qw(%timestamp2time); #Hash of timestamp => time. use vars qw($last_timestamp $last_time); #The last time conversion done. use vars qw($last_date $date_seconds); #The last date conversion done. use vars qw($last_offset $offset_seconds); #The last time offset conversion done. use vars qw($localtime_offset); use vars qw($i); #General loop counter. use vars qw($debug); #Debug mode? use vars qw($ntopchart); #How many entries should make it into the chart? use vars qw($gddirectory); #Where to put files from GD::Graph # SpamAssassin variables use vars qw($spam_score $spam_score_gigs); use vars qw($ham_score $ham_score_gigs); use vars qw(%ham_count_by_ip %spam_count_by_ip); use vars qw(%rejected_count_by_ip %rejected_count_by_reason); use vars qw(%temporarily_rejected_count_by_ip %temporarily_rejected_count_by_reason); #For use in Spreadsheet::WriteExcel use vars qw($workbook $ws_global $ws_relayed $ws_errors); use vars qw($row $col $row_hist $col_hist); use vars qw($run_hist); use vars qw($f_default $f_header1 $f_header2 $f_header2_m $f_headertab $f_percent); #Format Header # Output FileHandles use vars qw($txt_fh $htm_fh $xls_fh); $ntopchart = 5; # The following are parameters whose values are # set by command line switches: use vars qw($show_errors $show_relay $show_transport $transport_pattern); use vars qw($topcount $local_league_table $include_remote_users $do_local_domain); use vars qw($hist_opt $hist_interval $hist_number $volume_rounding $emptyOK); use vars qw($relay_pattern @queue_times @user_patterns @user_descriptions); use vars qw(@rcpt_times @delivery_times); use vars qw($include_original_destination); use vars qw($txt_fh $htm_fh $xls_fh); use vars qw(%do_sender); #Do sender by Host, Domain, Email, and/or Edomain tables. use vars qw($charts $chartrel $chartdir $charts_option_specified); use vars qw($merge_reports); #Merge old reports ? # The following are modified in the parse() routine, and # referred to in the print_*() routines. use vars qw($delayed_count $relayed_unshown $begin $end); use vars qw(%messages @message); use vars qw(%received_count %received_data %received_data_gigs); use vars qw(%delivered_messages %delivered_data %delivered_data_gigs %delivered_addresses); use vars qw(%received_count_user %received_data_user %received_data_gigs_user); use vars qw(%delivered_messages_user %delivered_addresses_user %delivered_data_user %delivered_data_gigs_user); use vars qw(%delivered_messages_local_domain %delivered_addresses_local_domain %delivered_data_local_domain %delivered_data_gigs_local_domain); use vars qw(%transported_count %transported_data %transported_data_gigs); use vars qw(%relayed %errors_count $message_errors); use vars qw(@qt_all_bin @qt_remote_bin); use vars qw($qt_all_overflow $qt_remote_overflow); use vars qw(@dt_all_bin @dt_remote_bin %rcpt_times_bin); use vars qw($dt_all_overflow $dt_remote_overflow %rcpt_times_overflow); use vars qw(@received_interval_count @delivered_interval_count); use vars qw(@user_pattern_totals @user_pattern_interval_count); use vars qw(%report_totals); # Enumerations use vars qw($SIZE $FROM_HOST $FROM_ADDRESS $ARRIVAL_TIME $REMOTE_DELIVERED $PROTOCOL); use vars qw($DELAYED $HAD_ERROR); $SIZE = 0; $FROM_HOST = 1; $FROM_ADDRESS = 2; $ARRIVAL_TIME = 3; $REMOTE_DELIVERED = 4; $DELAYED = 5; $HAD_ERROR = 6; $PROTOCOL = 7; ################################################## # Subroutines # ################################################## ####################################################################### # get_filehandle($file,\%output_files); # Return a filehandle writing to $file. # # If %output_files is defined, check that $output_files{$file} # doesn't exist and die if it does, or set it if it doesn't. ####################################################################### sub get_filehandle { my($file,$output_files_href) = @_; $file = '-' if ($file eq ''); if (defined $output_files_href) { die "You can only output to '$file' once! Use -h for help.\n" if exists $output_files_href->{$file}; $output_files_href->{$file} = 1; } if ($file eq '-') { return \*STDOUT; } if (-e $file) { unlink $file or die "Failed to rm $file: $!"; } my $fh = new IO::File $file, O_WRONLY|O_CREAT|O_EXCL; die "new IO::File $file failed: $!" unless (defined $fh); return $fh; } ####################################################################### # volume_rounded(); # # $rounded_volume = volume_rounded($bytes,$gigabytes); # # Given a data size in bytes, round it to KB, MB, or GB # as appropriate. # # Eg 12000 => 12KB, 15000000 => 14GB, etc. # # Note: I've experimented with Math::BigInt and it results in a 33% # performance degredation as opposed to storing numbers split into # bytes and gigabytes. ####################################################################### sub volume_rounded { my($x,$g) = @_; $x = 0 unless $x; $g = 0 unless $g; my($rounded); while ($x > $gig) { $g++; $x -= $gig; } if ($volume_rounding) { # Values < 1 GB if ($g <= 0) { if ($x < 10000) { $rounded = sprintf("%6d", $x); } elsif ($x < 10000000) { $rounded = sprintf("%4dKB", ($x + 512)/1024); } else { $rounded = sprintf("%4dMB", ($x + 512*1024)/(1024*1024)); } } # Values between 1GB and 10GB are printed in MB elsif ($g < 10) { $rounded = sprintf("%4dMB", ($g * 1024) + ($x + 512*1024)/(1024*1024)); } else { # Handle values over 10GB $rounded = sprintf("%4dGB", $g + ($x + $gig/2)/$gig); } } else { # We don't want any rounding to be done. # and we don't need broken formatted output which on one hand avoids numbers from # being interpreted as string by Spreadsheet Calculators, on the other hand # breaks if more than 4 digits! -> flexible length instead of fixed length # Format the return value at the output routine! -fh #$rounded = sprintf("%d", ($g * $gig) + $x); no integer; $rounded = sprintf("%.0f", ($g * $gig) + $x); } return $rounded; } ####################################################################### # un_round(); # # un_round($rounded_volume,\$bytes,\$gigabytes); # # Given a volume in KB, MB or GB, as generated by volume_rounded(), # do the reverse transformation and convert it back into Bytes and Gigabytes. # These are added to the $bytes and $gigabytes parameters. # # Given a data size in bytes, round it to KB, MB, or GB # as appropriate. # # EG: 500 => (500,0), 14GB => (0,14), etc. ####################################################################### sub un_round { my($rounded,$bytes_sref,$gigabytes_sref) = @_; if ($rounded =~ /(\d+)GB/) { $$gigabytes_sref += $1; } elsif ($rounded =~ /(\d+)MB/) { $$gigabytes_sref += $1 / 1024; $$bytes_sref += (($1 % 1024 ) * 1024 * 1024); } elsif ($rounded =~ /(\d+)KB/) { $$gigabytes_sref += $1 / (1024 * 1024); $$bytes_sref += ($1 % (1024 * 1024) * 1024); } elsif ($rounded =~ /(\d+)/) { # We need to turn off integer in case we are merging an -nvr report. no integer; $$gigabytes_sref += int($1 / $gig); $$bytes_sref += $1 % $gig; } #Now reduce the bytes down to less than 1GB. add_volume($bytes_sref,$gigabytes_sref,0) if ($$bytes_sref > $gig); } ####################################################################### # add_volume(); # # add_volume(\$bytes,\$gigs,$size); # # Add $size to $bytes/$gigs where this is a number split into # bytes ($bytes) and gigabytes ($gigs). This is significantly # faster than using Math::BigInt. ####################################################################### sub add_volume { my($bytes_ref,$gigs_ref,$size) = @_; $$bytes_ref = 0 if ! defined $$bytes_ref; $$gigs_ref = 0 if ! defined $$gigs_ref; $$bytes_ref += $size; while ($$bytes_ref > $gig) { $$gigs_ref++; $$bytes_ref -= $gig; } } ####################################################################### # format_time(); # # $formatted_time = format_time($seconds); # # Given a time in seconds, break it down into # weeks, days, hours, minutes, and seconds. # # Eg 12005 => 3h20m5s ####################################################################### sub format_time { my($t) = pop @_; my($s) = $t % 60; $t /= 60; my($m) = $t % 60; $t /= 60; my($h) = $t % 24; $t /= 24; my($d) = $t % 7; my($w) = $t/7; my($p) = ""; $p .= "$w"."w" if $w > 0; $p .= "$d"."d" if $d > 0; $p .= "$h"."h" if $h > 0; $p .= "$m"."m" if $m > 0; $p .= "$s"."s" if $s > 0 || $p eq ""; $p; } ####################################################################### # unformat_time(); # # $seconds = unformat_time($formatted_time); # # Given a time in weeks, days, hours, minutes, or seconds, convert it to seconds. # # Eg 3h20m5s => 12005 ####################################################################### sub unformat_time { my($formatted_time) = pop @_; my $time = 0; while ($formatted_time =~ s/^(\d+)([wdhms]?)//) { $time += $1 if ($2 eq '' || $2 eq 's'); $time += $1 * 60 if ($2 eq 'm'); $time += $1 * 60 * 60 if ($2 eq 'h'); $time += $1 * 60 * 60 * 24 if ($2 eq 'd'); $time += $1 * 60 * 60 * 24 * 7 if ($2 eq 'w'); } $time; } ####################################################################### # seconds(); # # $time = seconds($timestamp); # # Given a time-of-day timestamp, convert it into a time() value using # POSIX::mktime. We expect the timestamp to be of the form # "$year-$mon-$day $hour:$min:$sec", with month going from 1 to 12, # and the year to be absolute (we do the necessary conversions). The # seconds value can be followed by decimals, which we ignore. The # timestamp may be followed with an offset from UTC like "+$hh$mm"; if the # offset is not present, and we have not been told that the log is in UTC # (with the -utc option), then we adjust the time by the current local # time offset so that it can be compared with the time recorded in message # IDs, which is UTC. # # To improve performance, we only use mktime on the date ($year-$mon-$day), # and only calculate it if the date is different to the previous time we # came here. We then add on seconds for the '$hour:$min:$sec'. # # We also store the results of the last conversion done, and only # recalculate if the date is different. # # We used to have the '-cache' flag which would store the results of the # mktime() call. However, the current way of just using mktime() on the # date obsoletes this. ####################################################################### sub seconds { my($timestamp) = @_; # Is the timestamp the same as the last one? return $last_time if ($last_timestamp eq $timestamp); return 0 unless ($timestamp =~ /^((\d{4})\-(\d\d)-(\d\d))\s(\d\d):(\d\d):(\d\d)(?:\.\d+)?( ([+-])(\d\d)(\d\d))?/o); unless ($last_date eq $1) { $last_date = $1; my(@timestamp) = (0,0,0,$4,$3,$2); $timestamp[5] -= 1900; $timestamp[4]--; $date_seconds = mktime(@timestamp); } my $time = $date_seconds + ($5 * 3600) + ($6 * 60) + $7; # SC. Use caching. Also note we want seconds not minutes. #my($this_offset) = ($10 * 60 + $12) * ($9 . "1") if defined $8; if (defined $8 && ($8 ne $last_offset)) { $last_offset = $8; $offset_seconds = ($10 * 60 + $11) * 60; $offset_seconds = -$offset_seconds if ($9 eq '-'); } if (defined $8) { #$time -= $this_offset; $time -= $offset_seconds; } elsif (defined $localtime_offset) { $time -= $localtime_offset; } # Store the last timestamp received. $last_timestamp = $timestamp; $last_time = $time; $time; } ####################################################################### # id_seconds(); # # $time = id_seconds($message_id); # # Given a message ID, convert it into a time() value. ####################################################################### sub id_seconds { my($sub_id) = substr((pop @_), 0, 6); my($s) = 0; my(@c) = split(//, $sub_id); while($#c >= 0) { $s = $s * 62 + $tab62[ord(shift @c) - ord('0')] } $s; } ####################################################################### # wdhms_seconds(); # # $seconds = wdhms_seconds($string); # # Convert a string in a week/day/hour/minute/second format (eg 4h10s) # into seconds. ####################################################################### sub wdhms_seconds { if ($_[0] =~ /^(?:(\d+)w)?(?:(\d+)d)?(?:(\d+)h)?(?:(\d+)m)?(?:(\d+)s)?/) { return((($1||0) * $WEEK) + (($2||0) * $DAY) + (($3||0) * $HOUR) + (($4||0) * $MINUTE) + ($5||0)); } return undef; } ####################################################################### # queue_time(); # # $queued = queue_time($completed_tod, $arrival_time, $id); # # Given the completed time of day and either the arrival time # (preferred), or the message ID, calculate how long the message has # been on the queue. # ####################################################################### sub queue_time { my($completed_tod, $arrival_time, $id) = @_; # Note: id_seconds() benchmarks as 42% slower than seconds() # and computing the time accounts for a significant portion of # the run time. if (defined $arrival_time) { return(seconds($completed_tod) - seconds($arrival_time)); } else { return(seconds($completed_tod) - id_seconds($id)); } } ####################################################################### # calculate_localtime_offset(); # # $localtime_offset = calculate_localtime_offset(); # # Calculate the the localtime offset from gmtime in seconds. # # $localtime = time() + $localtime_offset. # # These are the same semantics as ISO 8601 and RFC 2822 timezone offsets. # (West is negative, East is positive.) ####################################################################### # $localtime = gmtime() + $localtime_offset. OLD COMMENT # This subroutine commented out as it's not currently in use. #sub calculate_localtime_offset { # # Pick an arbitrary date, convert it to localtime & gmtime, and return the difference. # my (@sample_date) = (0,0,0,5,5,100); # my $localtime = timelocal(@sample_date); # my $gmtime = timegm(@sample_date); # my $offset = $localtime - $gmtime; # return $offset; #} sub calculate_localtime_offset { # Assume that the offset at the moment is valid across the whole # period covered by the logs that we're analysing. This may not # be true around the time the clocks change in spring or autumn. my $utc = time; # mktime works on local time and gmtime works in UTC my $local = mktime(gmtime($utc)); return $local - $utc; } ####################################################################### # print_duration_table(); # # print_duration_table($title, $message_type, \@times, \@values, $overflow); # # Print a table showing how long a particular step took for # the messages. The parameters are: # $title Eg "Time spent on the queue" # $message_type Eg "Remote" # \@times The maximum time a message took for it to increment # the corresponding @values counter. # \@values An array of message counters. # $overflow The number of messages which exceeded the maximum # time. ####################################################################### sub print_duration_table { no integer; my($title, $message_type, $times_aref, $values_aref, $overflow) = @_; my(@chartdatanames); my(@chartdatavals); my $printed_one = 0; my $cumulative_percent = 0; my $queue_total = $overflow; map {$queue_total += $_} @$values_aref; my $temp = "$title: $message_type"; my $txt_format = "%5s %4s %6d %5.1f%% %5.1f%%\n"; my $htm_format = "%s %s%d%5.1f%%%5.1f%%\n"; # write header printf $txt_fh ("%s\n%s\n\n", $temp, "-" x length($temp)) if $txt_fh; if ($htm_fh) { print $htm_fh "

$temp

\n"; print $htm_fh "
\n"; print $htm_fh "\n"; } if ($xls_fh) { $ws_global->write($row++, $col, "$title: ".$message_type, $f_header2); my @content=("Time", "Messages", "Percentage", "Cumulative Percentage"); &set_worksheet_line($ws_global, $row++, 1, \@content, $f_headertab); } for ($i = 0; $i <= $#$times_aref; ++$i) { if ($$values_aref[$i] > 0) { my $percent = ($values_aref->[$i] * 100)/$queue_total; $cumulative_percent += $percent; my @content=($printed_one? " " : "Under", format_time($times_aref->[$i]), $values_aref->[$i], $percent, $cumulative_percent); if ($htm_fh) { printf $htm_fh ($htm_format, @content); if (!defined($values_aref->[$i])) { print $htm_fh "Not defined"; } } if ($txt_fh) { printf $txt_fh ($txt_format, @content); if (!defined($times_aref->[$i])) { print $txt_fh "Not defined"; } } if ($xls_fh) { no integer; &set_worksheet_line($ws_global, $row, 0, [@content[0,1,2]], $f_default); &set_worksheet_line($ws_global, $row++, 3, [$content[3]/100,$content[4]/100], $f_percent); if (!defined($times_aref->[$i])) { $col=0; $ws_global->write($row++, $col, "Not defined" ); } } push(@chartdatanames, ($printed_one? "" : "Under") . format_time($times_aref->[$i])); push(@chartdatavals, $$values_aref[$i]); $printed_one = 1; } } if ($overflow && $overflow > 0) { my $percent = ($overflow * 100)/$queue_total; $cumulative_percent += $percent; my @content = ("Over ", format_time($times_aref->[-1]), $overflow, $percent, $cumulative_percent); printf $txt_fh ($txt_format, @content) if $txt_fh; printf $htm_fh ($htm_format, @content) if $htm_fh; if ($xls_fh) { &set_worksheet_line($ws_global, $row, 0, [@content[0,1,2]], $f_default); &set_worksheet_line($ws_global, $row++, 3, [$content[3]/100,$content[4]/100], $f_percent); } } push(@chartdatanames, "Over " . format_time($times_aref->[-1])); push(@chartdatavals, $overflow); #printf("Unknown %6d\n", $queue_unknown) if $queue_unknown > 0; if ($htm_fh) { print $htm_fh "
TimeMessagesPercentageCumulative Percentage
"; if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals > 0)) { my @data = ( \@chartdatanames, \@chartdatavals ); my $graph = GD::Graph::pie->new(200, 200); my $pngname = "$title-$message_type.png"; $pngname =~ s/[^\w\-\.]/_/; my $graph_title = "$title ($message_type)"; $graph->set(title => $graph_title) if (length($graph_title) < 21); my $gd = $graph->plot(\@data) or warn($graph->error); if ($gd) { open(IMG, ">$chartdir/$pngname") or die "Could not write $chartdir/$pngname: $!\n"; binmode IMG; print IMG $gd->png; close IMG; print $htm_fh ""; } } print $htm_fh "
\n"; } if ($xls_fh) { $row++; } print $txt_fh "\n" if $txt_fh; print $htm_fh "\n" if $htm_fh; } ####################################################################### # print_histogram(); # # print_histogram('Deliveries|Messages received|$pattern', $unit, @interval_count); # # Print a histogram of the messages delivered/received per time slot # (hour by default). ####################################################################### sub print_histogram { my($text, $unit, @interval_count) = @_; my(@chartdatanames); my(@chartdatavals); my($maxd) = 0; # save first row of print_histogram for xls output if (!$run_hist) { $row_hist = $row; } else { $row = $row_hist; } for ($i = 0; $i < $hist_number; $i++) { $maxd = $interval_count[$i] if $interval_count[$i] > $maxd; } my $scale = int(($maxd + 25)/50); $scale = 1 if $scale == 0; if ($scale != 1) { if ($unit !~ s/y$/ies/) { $unit .= 's'; } } # make and output title my $title = sprintf("$text per %s", ($hist_interval == 60)? "hour" : ($hist_interval == 1)? "minute" : "$hist_interval minutes"); my $txt_htm_title = $title . " (each dot is $scale $unit)"; printf $txt_fh ("%s\n%s\n\n", $txt_htm_title, "-" x length($txt_htm_title)) if $txt_fh; if ($htm_fh) { print $htm_fh "

$txt_htm_title

\n"; print $htm_fh "\n"; print $htm_fh "
\n";
}

if ($xls_fh) {
  $title =~ s/Messages/Msg/ ;
  $row += 2;
  $ws_global->write($row++, $col_hist+1, $title, $f_headertab);
}


my $hour = 0;
my $minutes = 0;
for ($i = 0; $i < $hist_number; $i++) {
  my $c = $interval_count[$i];

  # If the interval is an hour (the maximum) print the starting and
  # ending hours as a label. Otherwise print the starting hour and
  # minutes, which take up the same space.

  my $temp;
  if ($hist_opt == 1) {
    $temp = sprintf("%02d-%02d", $hour, $hour + 1);

    print $txt_fh $temp if $txt_fh;
    print $htm_fh $temp if $htm_fh;

    if ($xls_fh) {
      if ($run_hist==0) {
        # only on first run
        $ws_global->write($row, 0, [$temp], $f_default);
      }
    }

    push(@chartdatanames, $temp);
    $hour++;
  }
  else {
    if ($minutes == 0)
      { $temp = sprintf("%02d:%02d", $hour, $minutes) }
    else
      { $temp = sprintf("  :%02d", $minutes) }

    print $txt_fh $temp if $txt_fh;
    print $htm_fh $temp if $htm_fh;
    if (($xls_fh) and ($run_hist==0)) {
      # only on first run
      $temp = sprintf("%02d:%02d", $hour, $minutes);
      $ws_global->write($row, 0, [$temp], $f_default);
    }

    push(@chartdatanames, $temp);
    $minutes += $hist_interval;
    if ($minutes >= 60) {
      $minutes = 0;
      $hour++;
    }
  }
  push(@chartdatavals, $c);

  printf $txt_fh (" %6d %s\n", $c, "." x ($c/$scale)) if $txt_fh;
  printf $htm_fh (" %6d %s\n", $c, "." x ($c/$scale)) if $htm_fh;
  $ws_global->write($row++, $col_hist+1, [$c], $f_default) if $xls_fh;

} #end for

printf $txt_fh "\n" if $txt_fh;
printf $htm_fh "\n" if $htm_fh;

if ($htm_fh)
{
  print $htm_fh "
\n"; print $htm_fh "
\n"; if ($HAVE_GD_Graph_linespoints && $charts && ($#chartdatavals > 0)) { # calculate the graph my @data = ( \@chartdatanames, \@chartdatavals ); my $graph = GD::Graph::linespoints->new(300, 300); $graph->set( x_label => 'Time', y_label => 'Amount', title => $text, x_labels_vertical => 1 ); my $pngname = "histogram_$text.png"; $pngname =~ s/[^\w\._]/_/g; my $gd = $graph->plot(\@data) or warn($graph->error); if ($gd) { open(IMG, ">$chartdir/$pngname") or die "Could not write $chartdir/$pngname: $!\n"; binmode IMG; print IMG $gd->png; close IMG; print $htm_fh ""; } } print $htm_fh "
\n"; } $col_hist++; # where to continue next times $row+=2; # leave some space after history block $run_hist=1; # we have done this once or more } ####################################################################### # print_league_table(); # # print_league_table($league_table_type,\%message_count,\%address_count,\%message_data,\%message_data_gigs, $spreadsheet, $row_sref); # # Given hashes of message count, address count, and message data, # which are keyed by the table type (eg by the sending host), print a # league table showing the top $topcount (defaults to 50). ####################################################################### sub print_league_table { my($text,$m_count,$a_count,$m_data,$m_data_gigs,$spreadsheet, $row_sref) = @_; my($name) = ($topcount == 1)? "$text" : "$topcount ${text}s"; my($title) = "Top $name by message count"; my(@chartdatanames) = (); my(@chartdatavals) = (); my $chartotherval = 0; $text = ucfirst($text); # Align non-local addresses to the right (so all the .com's line up). # Local addresses are aligned on the left as they are userids. my $align = ($text !~ /local/i) ? 'right' : 'left'; ################################################ # Generate the printf formats and table headers. ################################################ my(@headers) = ('Messages'); #push(@headers,'Addresses') if defined $a_count; push(@headers,'Addresses') if defined $a_count && %$a_count; push(@headers,'Bytes','Average') if defined $m_data; my $txt_format = "%10s " x @headers . " %s\n"; my $txt_col_headers = sprintf $txt_format, @headers, $text; my $htm_format = "" . '%s'x@headers . "%s\n"; my $htm_col_headers = sprintf $htm_format, @headers, $text; $htm_col_headers =~ s/(<\/?)td/$1th/g; #Convert 's to 's for the header. ################################################ # Write the table headers ################################################ printf $txt_fh ("%s\n%s\n%s", $title, "-" x length($title),$txt_col_headers) if $txt_fh; if ($htm_fh) { print $htm_fh <

$title

EoText print $htm_fh $htm_col_headers } if ($xls_fh) { $spreadsheet->write(${$row_sref}++, 0, $title, $f_header2); $spreadsheet->write(${$row_sref}++, 0, [@headers, $text], $f_headertab); } # write content foreach my $key (top_n_sort($topcount,$m_count,$m_data_gigs,$m_data)) { # When displaying the average figures, we calculate the average of # the rounded data, as the user would calculate it. This reduces # the accuracy slightly, but we have to do it this way otherwise # when using -merge to convert results from text to HTML and # vice-versa discrepencies would occur. my $messages = $$m_count{$key}; my @content = ($messages); push(@content, $$a_count{$key}) if defined $a_count; if (defined $m_data) { my $rounded_volume = volume_rounded($$m_data{$key},$$m_data_gigs{$key}); my($data,$gigs) = (0,0); un_round($rounded_volume,\$data,\$gigs); my $rounded_average = volume_rounded($data/$messages,$gigs/$messages); push(@content, $rounded_volume, $rounded_average); } # write content printf $txt_fh ($txt_format, @content, $key) if $txt_fh; if ($htm_fh) { my $htmlkey = $key; $htmlkey =~ s/>/\>\;/g; $htmlkey =~ s/write(${$row_sref}++, 0, [@content, $key], $f_default) if $xls_fh; if (scalar @chartdatanames < $ntopchart) { push(@chartdatanames, $key); push(@chartdatavals, $$m_count{$key}); } else { $chartotherval += $$m_count{$key}; } } push(@chartdatanames, "Other"); push(@chartdatavals, $chartotherval); print $txt_fh "\n" if $txt_fh; if ($htm_fh) { print $htm_fh "
\n"; print $htm_fh "
\n"; if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals > 0)) { # calculate the graph my @data = ( \@chartdatanames, \@chartdatavals ); my $graph = GD::Graph::pie->new(300, 300); $graph->set( x_label => 'Name', y_label => 'Amount', title => 'By count', ); my $gd = $graph->plot(\@data) or warn($graph->error); if ($gd) { my $temp = $text; $temp =~ s/ /_/g; open(IMG, ">$chartdir/${temp}_count.png") or die "Could not write $chartdir/${temp}_count.png: $!\n"; binmode IMG; print IMG $gd->png; close IMG; print $htm_fh ""; } } print $htm_fh "\n"; print $htm_fh "
\n\n"; } ++${$row_sref} if $xls_fh; if (defined $m_data) { # write header $title = "Top $name by volume"; printf $txt_fh ("%s\n%s\n%s", $title, "-" x length($title),$txt_col_headers) if $txt_fh; if ($htm_fh) { print $htm_fh <

$title

EoText print $htm_fh $htm_col_headers; } if ($xls_fh) { $spreadsheet->write(${$row_sref}++, 0, $title, $f_header2); $spreadsheet->write(${$row_sref}++, 0, [@headers, $text], $f_headertab); } @chartdatanames = (); @chartdatavals = (); $chartotherval = 0; my $use_gig = 0; foreach my $key (top_n_sort($topcount,$m_data_gigs,$m_data,$m_count)) { # The largest volume will be the first (top of the list). # If it has at least 1 gig, then just use gigabytes to avoid # risking an integer overflow when generating the pie charts. if ($$m_data_gigs{$key}) { $use_gig = 1; } my $messages = $$m_count{$key}; my @content = ($messages); push(@content, $$a_count{$key}) if defined $a_count; my $rounded_volume = volume_rounded($$m_data{$key},$$m_data_gigs{$key}); my($data ,$gigs) = (0,0); un_round($rounded_volume,\$data,\$gigs); my $rounded_average = volume_rounded($data/$messages,$gigs/$messages); push(@content, $rounded_volume, $rounded_average ); # write content printf $txt_fh ($txt_format, @content, $key) if $txt_fh; if ($htm_fh) { my $htmlkey = $key; $htmlkey =~ s/>/\>\;/g; $htmlkey =~ s/write(${$row_sref}++, 0, [@content, $key], $f_default) if $xls_fh; if (scalar @chartdatanames < $ntopchart) { if ($use_gig) { if ($$m_data_gigs{$key}) { push(@chartdatanames, $key); push(@chartdatavals, $$m_data_gigs{$key}); } } else { push(@chartdatanames, $key); push(@chartdatavals, $$m_data{$key}); } } else { $chartotherval += ($use_gig) ? $$m_data_gigs{$key} : $$m_data{$key}; } } push(@chartdatanames, "Other"); push(@chartdatavals, $chartotherval); print $txt_fh "\n" if $txt_fh; if ($htm_fh) { print $htm_fh "
\n"; print $htm_fh "
\n"; if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals > 0)) { # calculate the graph my @data = ( \@chartdatanames, \@chartdatavals ); my $graph = GD::Graph::pie->new(300, 300); $graph->set( x_label => 'Name', y_label => 'Volume' , title => 'By Volume', ); my $gd = $graph->plot(\@data) or warn($graph->error); if ($gd) { my $temp = $text; $temp =~ s/ /_/g; open(IMG, ">$chartdir/${temp}_volume.png") or die "Could not write $chartdir/${temp}_volume.png: $!\n"; binmode IMG; print IMG $gd->png; close IMG; print $htm_fh ""; } } print $htm_fh "\n"; print $htm_fh "
\n\n"; } ++${$row_sref} if $xls_fh; } } ####################################################################### # top_n_sort(); # # @sorted_keys = top_n_sort($n,$href1,$href2,$href3); # # Given a hash which has numerical values, return the sorted $n keys which # point to the top values. The second and third hashes are used as # tiebreakers. They all must have the same keys. # # The idea behind this routine is that when you only want to see the # top n members of a set, rather than sorting the entire set and then # plucking off the top n, sort through the stack as you go, discarding # any member which is lower than your current n'th highest member. # # This proves to be an order of magnitude faster for large hashes. # On 200,000 lines of mainlog it benchmarked 9 times faster. # On 700,000 lines of mainlog it benchmarked 13.8 times faster. # # We assume the values are > 0. ####################################################################### sub top_n_sort { my($n,$href1,$href2,$href3) = @_; # PH's original sort was: # # foreach $key (sort # { # $$m_count{$b} <=> $$m_count{$a} || # $$m_data_gigs{$b} <=> $$m_data_gigs{$a} || # $$m_data{$b} <=> $$m_data{$a} || # $a cmp $b # } # keys %{$m_count}) # #We use a key of '_' to represent non-existant values, as null keys are valid. #'_' is not a valid domain, edomain, host, or email. my(@top_n_keys) = ('_') x $n; my($minimum_value1,$minimum_value2,$minimum_value3) = (0,0,0); my $top_n_key = ''; my $n_minus_1 = $n - 1; my $n_minus_2 = $n - 2; # Create a dummy hash incase the user has not provided us with # tiebreaker hashes. my(%dummy_hash); $href2 = \%dummy_hash unless defined $href2; $href3 = \%dummy_hash unless defined $href3; # Pick out the top $n keys. my($key,$value1,$value2,$value3,$i,$comparison,$insert_position); while (($key,$value1) = each %$href1) { #print STDERR "key $key ($value1,",$href2->{$key},",",$href3->{$key},") <=> ($minimum_value1,$minimum_value2,$minimum_value3)\n"; # Check to see that the new value is bigger than the lowest of the # top n keys that we're keeping. We test the main key first, because # for the majority of cases we can skip creating dummy hash values # should the user have not provided real tie-breaking hashes. next unless $value1 >= $minimum_value1; # Create a dummy hash entry for the key if required. # Note that setting the dummy_hash value sets it for both href2 & # href3. Also note that currently we are guaranteed to have a real # value for href3 if a real value for href2 exists so don't need to # test for it as well. $dummy_hash{$key} = 0 unless exists $href2->{$key}; $comparison = $value1 <=> $minimum_value1 || $href2->{$key} <=> $minimum_value2 || $href3->{$key} <=> $minimum_value3 || $top_n_key cmp $key; next unless ($comparison == 1); # As we will be using these values a few times, extract them into scalars. $value2 = $href2->{$key}; $value3 = $href3->{$key}; # This key is bigger than the bottom n key, so the lowest position we # will insert it into is $n minus 1 (the bottom of the list). $insert_position = $n_minus_1; # Now go through the list, stopping when we find a key that we're # bigger than, or we come to the penultimate position - we've # already tested bigger than the last. # # Note: we go top down as the list starts off empty. # Note: stepping through the list in this way benchmarks nearly # three times faster than doing a sort() on the reduced list. # I assume this is because the list is already in order, and # we get a performance boost from not having to do hash lookups # on the new key. for ($i = 0; $i < $n_minus_1; $i++) { $top_n_key = $top_n_keys[$i]; if ( ($top_n_key eq '_') || ( ($value1 <=> $href1->{$top_n_key} || $value2 <=> $href2->{$top_n_key} || $value3 <=> $href3->{$top_n_key} || $top_n_key cmp $key) == 1 ) ) { $insert_position = $i; last; } } # Remove the last element, then insert the new one. $#top_n_keys = $n_minus_2; splice(@top_n_keys,$insert_position,0,$key); # Extract our new minimum values. $top_n_key = $top_n_keys[$n_minus_1]; if ($top_n_key ne '_') { $minimum_value1 = $href1->{$top_n_key}; $minimum_value2 = $href2->{$top_n_key}; $minimum_value3 = $href3->{$top_n_key}; } } # Return the top n list, grepping out non-existant values, just in case # we didn't have that many values. return(grep(!/^_$/,@top_n_keys)); } ####################################################################### # html_header(); # # $header = html_header($title); # # Print our HTML header and start the block. ####################################################################### sub html_header { my($title) = @_; my $text = << "EoText"; $title

$title

EoText return $text; } ####################################################################### # help(); # # help(); # # Display usage instructions and exit. ####################################################################### sub help { print << "EoText"; eximstats Version $VERSION Usage: eximstats [Output] [Options] mainlog1 mainlog2 ... eximstats -merge -html [Options] report.1.html ... > weekly_rep.html Examples: eximstats -html=eximstats.html mainlog1 mainlog2 ... eximstats mainlog1 mainlog2 ... > report.txt Parses exim mainlog or syslog files and generates a statistical analysis of the messages processed. Valid output types are: -txt[=] plain text (default unless no other type is specified) -html[=] HTML -xls[=] Excel With no type and file given, defaults to -txt and STDOUT. Valid options are: -h histogram divisions per hour. The default is 1, and 0 suppresses histograms. Other valid values are: 2, 3, 5, 10, 15, 20, 30 or 60. -ne don't display error information -nr don't display relaying information -nr/pattern/ don't display relaying information that matches -nt don't display transport information -nt/pattern/ don't display transport information that matches -nvr don't do volume rounding. Display in bytes, not KB/MB/GB. -t display top sources/destinations default is 50, 0 suppresses top listing -tnl omit local sources/destinations in top listing -t_remote_users show top user sources/destinations from non-local domains -q list of times for queuing information. -q0 suppresses. -show_rt Show the receipt times for all the messages. -show_dt Show the delivery times for all the messages. is an optional list of times in seconds. Eg -show_rt1,2,4,8. -include_original_destination show both the final and original destinations in the results rather than just the final ones. -byhost show results by sending host (default unless bydomain or byemail is specified) -bydomain show results by sending domain. -byemail show results by sender's email address -byedomain show results by sender's email domain -bylocaldomain show results by local domain -pattern "Description" /pattern/ Count lines matching specified patterns and show them in the results. It can be specified multiple times. Eg: -pattern 'Refused connections' '/refused connection/' -merge merge previously generated reports into a new report -charts Create charts (this requires the GD::Graph modules). Only valid with -html. -chartdir Create the charts' png files in the directory -chartrel Specify the relative directory for the "img src=" tags from where to include the charts in the html file -chartdir and -chartrel default to '.' -emptyok It is OK if there is no valid input, don't print an error. -d Debug mode - dump the eval'ed parser onto STDERR. EoText exit 1; } ####################################################################### # generate_parser(); # # $parser = generate_parser(); # # This subroutine generates the parsing routine which will be # used to parse the mainlog. We take the base operation, and remove bits not in use. # This improves performance depending on what bits you take out or add. # # I've tested using study(), but this does not improve performance. # # We store our parsing routing in a variable, and process it looking for #IFDEF (Expression) # or #IFNDEF (Expression) statements and corresponding #ENDIF (Expression) statements. If # the expression evaluates to true, then it is included/excluded accordingly. ####################################################################### sub generate_parser { my $parser = ' my($ip,$host,$email,$edomain,$domain,$thissize,$size,$old,$new); my($tod,$m_hour,$m_min,$id,$flag,$extra,$length); my($seconds,$queued,$rcpt_time,$local_domain); my $rej_id = 0; while (<$fh>) { # Convert syslog lines to mainlog format. if (! /^\\d{4}/) { next unless s/^.*? exim\\b.*?: //; } $length = length($_); next if ($length < 38); next unless /^ (\\d{4}\\-\\d\\d-\\d\\d\\s # 1: YYYYMMDD HHMMSS (\\d\\d) # 2: HH : (\\d\\d) # 3: MM :\\d\\d ) (\\.\\d+)? # 4: subseconds (\s[-+]\\d\\d\\d\\d)? # 5: tz-offset (\s\\[\\d+\\])? # 6: pid /ox; $tod = defined($5) ? $1 . $5 : $1; ($m_hour,$m_min) = ($2,$3); # PH - watch for GMT offsets in the timestamp. if (defined($5)) { $extra = 6; next if ($length < 44); } else { $extra = 0; } # watch for subsecond precision if (defined($4)) { $extra += length($4); next if ($length < 38 + $extra); } # PH - watch for PID added after the timestamp. if (defined($6)) { $extra += length($6); next if ($length < 38 + $extra); } # $id = substr($_, 20 + $extra, 16); # old ID was 16 chars $id = substr($_, 20 + $extra, 23); # new IS is 23 chars $id =~ s/(\S+).*/$1/; $extra += length($id) - 16; $flag = substr($_, 37 + $extra, 2); if ($flag !~ /^([<>=*-]+|SA)$/ && /rejected|refused|dropped/) { $flag = "Re"; $extra -= 3; } # Rejects can have no MSGID... if ($flag eq "Re" && $id !~ /^[-0-9a-zA-Z]+$/) { $id = "reject:" . ++$rej_id; $extra -= 17; } '; # Watch for user specified patterns. my $user_pattern_index = 0; foreach (@user_patterns) { $user_pattern_totals[$user_pattern_index] = 0; $parser .= " if ($_) {\n"; $parser .= " \$user_pattern_totals[$user_pattern_index]++;\n"; $parser .= " \$user_pattern_interval_count[$user_pattern_index][(\$m_hour*60 + \$m_min)/$hist_interval]++;\n" if ($hist_opt > 0); $parser .= " }\n"; $user_pattern_index++; } $parser .= ' next unless ($flag =~ /<=|=>|->|==|\\*\\*|Co|SA|Re/); #Strip away the timestamp, ID and flag to speed up later pattern matches. #The flags include Co (Completed), Re (Rejected), and SA (SpamAssassin). $_ = substr($_, 40 + $extra); # PH # Alias @message to the array of information about the message. # This minimises the number of calls to hash functions. $messages{$id} = [] unless exists $messages{$id}; *message = $messages{$id}; # JN - Skip over certain transports as specified via the "-nt/.../" command # line switch (where ... is a perl style regular expression). This is # required so that transports that skew stats such as SpamAssassin can be # ignored. #IFDEF ($transport_pattern) if (/\\sT=(\\S+)/) { next if ($1 =~ /$transport_pattern/o) ; } #ENDIF ($transport_pattern) # Do some pattern matches to get the host and IP address. # We expect lines to be of the form "H=[IpAddr]" or "H=Host [IpAddr]" or # "H=Host (UnverifiedHost) [IpAddr]" or "H=(UnverifiedHost) [IpAddr]". # We do 2 separate matches to keep the matches simple and fast. # Host is local unless otherwise specified. # Watch out for "H=([IpAddr])" in case they send "[IpAddr]" as their HELO! $ip = (/\\bH=(?:|.*? )(\\[[^]]+\\])/) ? $1 # 2008-03-31 06:25:22 Connection from [213.246.33.217]:39456 refused: too many connections from that IP address // .hs : (/Connection from (\[\S+\])/) ? $1 # 2008-03-31 06:52:40 SMTP call from mail.cacoshrf.com (ccsd02.ccsd.local) [69.24.118.229]:4511 dropped: too many nonmail commands (last was "RSET") // .hs : (/SMTP call from .*?(\[\S+\])/) ? $1 : "local"; $host = (/\\bH=(\\S+)/) ? $1 : "local"; $domain = "localdomain"; #Domain is localdomain unless otherwise specified. #IFDEF ($do_sender{Domain}) if ($host =~ /^\\[/ || $host =~ /^[\\d\\.]+$/) { # Host is just an IP address. $domain = $host; } elsif ($host =~ /^(\\(?)[^\\.]+\\.([^\\.]+\\..*)/) { # Remove the host portion from the DNS name. We ensure that we end up # with at least xxx.yyy. $host can be "(x.y.z)" or "x.y.z". $domain = lc("$1.$2"); $domain =~ s/^\\.//; #Remove preceding dot. } #ENDIF ($do_sender{Domain}) #IFDEF ($do_sender{Email}) #IFDEF ($include_original_destination) # Catch both "a@b.com " and "e@f.com" #$email = (/^(\S+) (<(\S*?)>)?/) ? $3 || $1 : ""; $email = (/^(\S+ (<[^@>]+@?[^>]*>)?)/) ? $1 : ""; chomp($email); #ENDIF ($include_original_destination) #IFNDEF ($include_original_destination) $email = (/^(\S+)/) ? $1 : ""; #ENDIF ($include_original_destination) #ENDIF ($do_sender{Email}) #IFDEF ($do_sender{Edomain}) if (/^(<>|blackhole)/) { $edomain = $1; } #IFDEF ($include_original_destination) elsif (/^(\S+ (<\S*?\\@(\S+?)>)?)/) { $edomain = $1; chomp($edomain); $edomain =~ s/@(\S+?)>/"@" . lc($1) . ">"/e; } #ENDIF ($include_original_destination) #IFNDEF ($include_original_destination) elsif (/^\S*?\\@(\S+)/) { $edomain = lc($1); } #ENDIF ($include_original_destination) else { $edomain = ""; } #ENDIF ($do_sender{Edomain}) if ($tod lt $begin) { $begin = $tod; } elsif ($tod gt $end) { $end = $tod; } if ($flag eq "<=") { $thissize = (/\\sS=(\\d+)( |$)/) ? $1 : 0; $message[$SIZE] = $thissize; $message[$PROTOCOL] = (/ P=(\S+)/) ? $1 : undef; #IFDEF ($show_relay) if ($host ne "local") { # Save incoming information in case it becomes interesting # later, when delivery lines are read. my($from) = /^(\\S+)/; $message[$FROM_HOST] = "$host$ip"; $message[$FROM_ADDRESS] = $from; } #ENDIF ($show_relay) #IFDEF ($local_league_table || $include_remote_users) if (/\sU=(\\S+)/) { my $user = $1; #IFDEF ($local_league_table && $include_remote_users) { #Store both local and remote users. #ENDIF ($local_league_table && $include_remote_users) #IFDEF ($local_league_table && ! $include_remote_users) if ($host eq "local") { #Store local users only. #ENDIF ($local_league_table && ! $include_remote_users) #IFDEF ($include_remote_users && ! $local_league_table) if ($host ne "local") { #Store remote users only. #ENDIF ($include_remote_users && ! $local_league_table) ++$received_count_user{$user}; add_volume(\\$received_data_user{$user},\\$received_data_gigs_user{$user},$thissize); } } #ENDIF ($local_league_table || $include_remote_users) #IFDEF ($do_sender{Host}) ++$received_count{Host}{$host}; add_volume(\\$received_data{Host}{$host},\\$received_data_gigs{Host}{$host},$thissize); #ENDIF ($do_sender{Host}) #IFDEF ($do_sender{Domain}) if ($domain) { ++$received_count{Domain}{$domain}; add_volume(\\$received_data{Domain}{$domain},\\$received_data_gigs{Domain}{$domain},$thissize); } #ENDIF ($do_sender{Domain}) #IFDEF ($do_sender{Email}) ++$received_count{Email}{$email}; add_volume(\\$received_data{Email}{$email},\\$received_data_gigs{Email}{$email},$thissize); #ENDIF ($do_sender{Email}) #IFDEF ($do_sender{Edomain}) ++$received_count{Edomain}{$edomain}; add_volume(\\$received_data{Edomain}{$edomain},\\$received_data_gigs{Edomain}{$edomain},$thissize); #ENDIF ($do_sender{Edomain}) ++$total_received_count; add_volume(\\$total_received_data,\\$total_received_data_gigs,$thissize); #IFDEF ($#queue_times >= 0 || $#rcpt_times >= 0) $message[$ARRIVAL_TIME] = $tod; #ENDIF ($#queue_times >= 0 || $#rcpt_times >= 0) #IFDEF ($hist_opt > 0) $received_interval_count[($m_hour*60 + $m_min)/$hist_interval]++; #ENDIF ($hist_opt > 0) } elsif ($flag eq "=>") { $size = $message[$SIZE] || 0; if ($host ne "local") { $message[$REMOTE_DELIVERED] = 1; #IFDEF ($show_relay) # Determine relaying address if either only one address listed, # or two the same. If they are different, it implies a forwarding # or aliasing, which is not relaying. Note that for multi-aliased # addresses, there may be a further address between the first # and last. if (defined $message[$FROM_HOST]) { if (/^(\\S+)(?:\\s+\\([^)]\\))?\\s+<([^>]+)>/) { ($old,$new) = ($1,$2); } else { $old = $new = ""; } if ("\\L$new" eq "\\L$old") { ($old) = /^(\\S+)/ if $old eq ""; my $key = "H=\\L$message[$FROM_HOST]\\E A=\\L$message[$FROM_ADDRESS]\\E => " . "H=\\L$host\\E$ip A=\\L$old\\E"; if (!defined $relay_pattern || $key !~ /$relay_pattern/o) { $relayed{$key} = 0 if !defined $relayed{$key}; ++$relayed{$key}; } else { ++$relayed_unshown; } } } #ENDIF ($show_relay) } #IFDEF ($local_league_table || $include_remote_users) #IFDEF ($local_league_table && $include_remote_users) { #Store both local and remote users. #ENDIF ($local_league_table && $include_remote_users) #IFDEF ($local_league_table && ! $include_remote_users) if ($host eq "local") { #Store local users only. #ENDIF ($local_league_table && ! $include_remote_users) #IFDEF ($include_remote_users && ! $local_league_table) if ($host ne "local") { #Store remote users only. #ENDIF ($include_remote_users && ! $local_league_table) if (my($user) = split((/\\s]*>)/; my($parent) = $_ =~ / (<.+?>) /; #DT 1.54 if (defined $parent) { $user = "$user $parent"; #IFDEF ($do_local_domain) if ($parent =~ /\\@(.+)>/) { $local_domain = lc($1); ++$delivered_messages_local_domain{$local_domain}; ++$delivered_addresses_local_domain{$local_domain}; add_volume(\\$delivered_data_local_domain{$local_domain},\\$delivered_data_gigs_local_domain{$local_domain},$size); } #ENDIF ($do_local_domain) } } ++$delivered_messages_user{$user}; ++$delivered_addresses_user{$user}; add_volume(\\$delivered_data_user{$user},\\$delivered_data_gigs_user{$user},$size); } } #ENDIF ($local_league_table || $include_remote_users) #IFDEF ($do_sender{Host}) $delivered_messages{Host}{$host}++; $delivered_addresses{Host}{$host}++; add_volume(\\$delivered_data{Host}{$host},\\$delivered_data_gigs{Host}{$host},$size); #ENDIF ($do_sender{Host}) #IFDEF ($do_sender{Domain}) if ($domain) { ++$delivered_messages{Domain}{$domain}; ++$delivered_addresses{Domain}{$domain}; add_volume(\\$delivered_data{Domain}{$domain},\\$delivered_data_gigs{Domain}{$domain},$size); } #ENDIF ($do_sender{Domain}) #IFDEF ($do_sender{Email}) ++$delivered_messages{Email}{$email}; ++$delivered_addresses{Email}{$email}; add_volume(\\$delivered_data{Email}{$email},\\$delivered_data_gigs{Email}{$email},$size); #ENDIF ($do_sender{Email}) #IFDEF ($do_sender{Edomain}) ++$delivered_messages{Edomain}{$edomain}; ++$delivered_addresses{Edomain}{$edomain}; add_volume(\\$delivered_data{Edomain}{$edomain},\\$delivered_data_gigs{Edomain}{$edomain},$size); #ENDIF ($do_sender{Edomain}) ++$total_delivered_messages; ++$total_delivered_addresses; add_volume(\\$total_delivered_data,\\$total_delivered_data_gigs,$size); #IFDEF ($show_transport) my $transport = (/\\sT=(\\S+)/) ? $1 : ":blackhole:"; ++$transported_count{$transport}; add_volume(\\$transported_data{$transport},\\$transported_data_gigs{$transport},$size); #ENDIF ($show_transport) #IFDEF ($hist_opt > 0) $delivered_interval_count[($m_hour*60 + $m_min)/$hist_interval]++; #ENDIF ($hist_opt > 0) #IFDEF ($#delivery_times > 0) if (/ DT=(\S+)/) { $seconds = wdhms_seconds($1); for ($i = 0; $i <= $#delivery_times; $i++) { if ($seconds < $delivery_times[$i]) { ++$dt_all_bin[$i]; ++$dt_remote_bin[$i] if $message[$REMOTE_DELIVERED]; last; } } if ($i > $#delivery_times) { ++$dt_all_overflow; ++$dt_remote_overflow if $message[$REMOTE_DELIVERED]; } } #ENDIF ($#delivery_times > 0) } elsif ($flag eq "->") { #IFDEF ($local_league_table || $include_remote_users) #IFDEF ($local_league_table && $include_remote_users) { #Store both local and remote users. #ENDIF ($local_league_table && $include_remote_users) #IFDEF ($local_league_table && ! $include_remote_users) if ($host eq "local") { #Store local users only. #ENDIF ($local_league_table && ! $include_remote_users) #IFDEF ($include_remote_users && ! $local_league_table) if ($host ne "local") { #Store remote users only. #ENDIF ($include_remote_users && ! $local_league_table) if (my($user) = split((/\\s]*>)/; my($parent) = $_ =~ / (<.+?>) /; #DT 1.54 $user = "$user $parent" if defined $parent; } ++$delivered_addresses_user{$user}; } } #ENDIF ($local_league_table || $include_remote_users) #IFDEF ($do_sender{Host}) $delivered_addresses{Host}{$host}++; #ENDIF ($do_sender{Host}) #IFDEF ($do_sender{Domain}) if ($domain) { ++$delivered_addresses{Domain}{$domain}; } #ENDIF ($do_sender{Domain}) #IFDEF ($do_sender{Email}) ++$delivered_addresses{Email}{$email}; #ENDIF ($do_sender{Email}) #IFDEF ($do_sender{Edomain}) ++$delivered_addresses{Edomain}{$edomain}; #ENDIF ($do_sender{Edomain}) ++$total_delivered_addresses; } elsif ($flag eq "==" && defined($message[$SIZE]) && !defined($message[$DELAYED])) { ++$delayed_count; $message[$DELAYED] = 1; } elsif ($flag eq "**") { if (defined ($message[$SIZE])) { unless (defined $message[$HAD_ERROR]) { ++$message_errors; $message[$HAD_ERROR] = 1; } } #IFDEF ($show_errors) ++$errors_count{$_}; #ENDIF ($show_errors) } elsif ($flag eq "Co") { #Completed? #IFDEF ($#queue_times >= 0) $queued = queue_time($tod, $message[$ARRIVAL_TIME], $id); for ($i = 0; $i <= $#queue_times; $i++) { if ($queued < $queue_times[$i]) { ++$qt_all_bin[$i]; ++$qt_remote_bin[$i] if $message[$REMOTE_DELIVERED]; last; } } if ($i > $#queue_times) { ++$qt_all_overflow; ++$qt_remote_overflow if $message[$REMOTE_DELIVERED]; } #ENDIF ($#queue_times >= 0) #IFDEF ($#rcpt_times >= 0) if (/ QT=(\S+)/) { $seconds = wdhms_seconds($1); #Calculate $queued if not previously calculated above. #IFNDEF ($#queue_times >= 0) $queued = queue_time($tod, $message[$ARRIVAL_TIME], $id); #ENDIF ($#queue_times >= 0) $rcpt_time = $seconds - $queued; my($protocol); if (defined $message[$PROTOCOL]) { $protocol = $message[$PROTOCOL]; # Create the bin if its not already defined. unless (exists $rcpt_times_bin{$protocol}) { initialise_rcpt_times($protocol); } } for ($i = 0; $i <= $#rcpt_times; ++$i) { if ($rcpt_time < $rcpt_times[$i]) { ++$rcpt_times_bin{all}[$i]; ++$rcpt_times_bin{$protocol}[$i] if defined $protocol; last; } } if ($i > $#rcpt_times) { ++$rcpt_times_overflow{all}; ++$rcpt_times_overflow{$protocol} if defined $protocol; } } #ENDIF ($#rcpt_times >= 0) delete($messages{$id}); } elsif ($flag eq "SA") { $ip = (/From.*?(\\[[^]]+\\])/ || /\\((local)\\)/) ? $1 : ""; #SpamAssassin message if (/Action: ((permanently|temporarily) rejected message|flagged as Spam but accepted): score=(\d+\.\d)/) { #add_volume(\\$spam_score,\\$spam_score_gigs,$3); ++$spam_count_by_ip{$ip}; } elsif (/Action: scanned but message isn\'t spam: score=(-?\d+\.\d)/) { #add_volume(\\$ham_score,\\$ham_score_gigs,$1); ++$ham_count_by_ip{$ip}; } elsif (/(Not running SA because SAEximRunCond expanded to false|check skipped due to message size)/) { ++$ham_count_by_ip{$ip}; } } # Look for Reject messages or blackholed messages (deliveries # without a transport) if ($flag eq "Re" || ($flag eq "=>" && ! /\\sT=\\S+/)) { # Correct the IP address for rejects: # rejected EHLO from my.test.net [10.0.0.5]: syntactically invalid argument(s): # rejected EHLO from [10.0.0.6]: syntactically invalid argument(s): $ip = $1 if ($ip eq "local" && /^rejected [HE][HE]LO from .*?(\[.+?\]):/); if (/SpamAssassin/) { ++$rejected_count_by_reason{"Rejected by SpamAssassin"}; ++$rejected_count_by_ip{$ip}; } elsif ( /(temporarily rejected [A-Z]*) .*?(: .*?)(:|\s*$)/ ) { ++$temporarily_rejected_count_by_reason{"\u$1$2"}; ++$temporarily_rejected_count_by_ip{$ip}; } elsif ( /(temporarily refused connection)/ ) { ++$temporarily_rejected_count_by_reason{"\u$1"}; ++$temporarily_rejected_count_by_ip{$ip}; } elsif ( /(listed at [^ ]+)/ || /(Forged IP detected in HELO)/ || /(Invalid domain or IP given in HELO\/EHLO)/ || /(unqualified recipient rejected)/ || /(closed connection (after|in response) .*?)\s*$/ || /(sender rejected)/ || # 2005-09-23 15:07:49 1EInHJ-0007Ex-Au H=(a.b.c) [10.0.0.1] F=<> rejected after DATA: This message contains a virus: (Eicar-Test-Signature) please scan your system. # 2005-10-06 10:50:07 1ENRS3-0000Nr-Kt => blackhole (DATA ACL discarded recipients): This message contains a virus: (Worm.SomeFool.P) please scan your system. / rejected after DATA: (.*)/ || / (rejected DATA: .*)/ || /.DATA ACL discarded recipients.: (.*)/ || /rejected after DATA: (unqualified address not permitted)/ || /(VRFY rejected)/ || # /(sender verify (defer|fail))/i || /(too many recipients)/ || /(refused relay.*?) to/ || /(rejected by non-SMTP ACL: .*)/ || /(rejected by local_scan.*)/ || # SMTP call from %s dropped: too many syntax or protocol errors (last command was "%s" # SMTP call from %s dropped: too many nonmail commands /(dropped: too many ((nonmail|unrecognized) commands|syntax or protocol errors))/ || # local_scan() function crashed with signal %d - message temporarily rejected # local_scan() function timed out - message temporarily rejected /(local_scan.. function .* - message temporarily rejected)/ || # SMTP protocol synchronization error (input sent without waiting for greeting): rejected connection from %s /(SMTP protocol .*?(error|violation))/ || /(message too big)/ ) { ++$rejected_count_by_reason{"\u$1"}; ++$rejected_count_by_ip{$ip}; } elsif (/rejected [HE][HE]LO from [^:]*: syntactically invalid argument/) { ++$rejected_count_by_reason{"Rejected HELO/EHLO: syntactically invalid argument"}; ++$rejected_count_by_ip{$ip}; } elsif (/response to "RCPT TO.*? was: (.*)/) { ++$rejected_count_by_reason{"Response to RCPT TO was: $1"}; ++$rejected_count_by_ip{$ip}; } elsif ( /(lookup of host )\S+ (failed)/ || # rejected from <%s>%s%s%s%s: message too big: /(rejected [A-Z]*) .*?(: .*?)(:|\s*$)/ || # refused connection from %s (host_reject_connection) # refused connection from %s (tcp wrappers) /(refused connection )from.*? (\(.*)/ || # error from remote mailer after RCPT TO:: host a.b.c [10.0.0.1]: 450 : Recipient address rejected: Greylisted for 60 seconds # error from remote mailer after MAIL FROM:<> SIZE=3468: host a.b.c [10.0.0.1]: 421 a.b.c has refused your connection because your server did not have a PTR record. /(error from remote mailer after .*?:).*(: .*?)(:|\s*$)/ || # a.b.c F= rejected after DATA: "@" or "." expected after "Undisclosed-Recipient": failing address in "To" header is: /rejected after DATA: ("." or "." expected).*?(: failing address in .*? header)/ || # connection from %s refused load average = %.2f /(Connection )from.*? (refused: load average)/ || # connection from %s refused (IP options) # Connection from %s refused: too many connections # connection from %s refused /([Cc]onnection )from.*? (refused.*)/ || # [10.0.0.1]: connection refused /: (Connection refused)()/ ) { ++$rejected_count_by_reason{"\u$1$2"}; ++$rejected_count_by_ip{$ip}; } elsif ( # 2008-03-31 06:25:22 H=mail.densitron.com [216.70.140.224]:45386 temporarily rejected connection in "connect" ACL: too fast reconnects // .hs # 2008-03-31 06:25:22 H=mail.densitron.com [216.70.140.224]:45386 temporarily rejected connection in "connect" ACL // .hs /(temporarily rejected connection in .*?ACL:?.*)/ ) { ++$temporarily_rejected_count_by_ip{$ip}; ++$temporarily_rejected_count_by_reason{"\u$1"}; } else { ++$rejected_count_by_reason{Unknown}; ++$rejected_count_by_ip{$ip}; print STDERR "Unknown rejection: $_" if $debug; } } }'; # We now do a 'C preprocessor style operation on our parser # to remove bits not in use. my(%defines_in_operation,$removing_lines,$processed_parser); foreach (split (/\n/,$parser)) { if ((/^\s*#\s*IFDEF\s*\((.*?)\)/i && ! eval $1) || (/^\s*#\s*IFNDEF\s*\((.*?)\)/i && eval $1) ) { $defines_in_operation{$1} = 1; $removing_lines = 1; } # Convert constants. while (/(\$[A-Z][A-Z_]*)\b/) { my $constant = eval $1; s/(\$[A-Z][A-Z_]*)\b/$constant/; } $processed_parser .= $_."\n" unless $removing_lines; if (/^\s*#\s*ENDIF\s*\((.*?)\)/i) { delete $defines_in_operation{$1}; unless (keys %defines_in_operation) { $removing_lines = 0; } } } print STDERR "# START OF PARSER:$processed_parser\n# END OF PARSER\n\n" if $debug; return $processed_parser; } ####################################################################### # parse(); # # parse($parser,\*FILEHANDLE); # # This subroutine accepts a parser and a filehandle from main and parses each # line. We store the results into global variables. ####################################################################### sub parse { my($parser,$fh) = @_; if ($merge_reports) { parse_old_eximstat_reports($fh); } else { eval $parser; die ($@) if $@; } } ####################################################################### # print_header(); # # print_header(); # # Print our headers and contents. ####################################################################### sub print_header { my $title = "Exim statistics from $begin to $end"; print $txt_fh "\n$title\n" if $txt_fh; if ($htm_fh) { print $htm_fh html_header($title); print $htm_fh "\n
\n"; } if ($xls_fh) { $ws_global->write($row++, $col+0, "Exim Statistics", $f_header1); &set_worksheet_line($ws_global, $row, $col, ["from:", $begin, "to:", $end], $f_default); $row+=2; } } ####################################################################### # print_grandtotals(); # # print_grandtotals(); # # Print the grand totals. ####################################################################### sub print_grandtotals { # Get the sender by headings and results. This is complicated as we can have # different numbers of columns. my($sender_txt_header,$sender_txt_format,$sender_html_format); my(@received_totals,@delivered_totals); my($row_tablehead, $row_max); my(@col_headers) = ('TOTAL', 'Volume', 'Messages', 'Addresses'); foreach ('Host','Domain','Email','Edomain') { next unless $do_sender{$_}; if ($merge_reports) { push(@received_totals, get_report_total($report_totals{Received},"${_}s")); push(@delivered_totals,get_report_total($report_totals{Delivered},"${_}s")); } else { push(@received_totals,scalar(keys %{$received_data{$_}})); push(@delivered_totals,scalar(keys %{$delivered_data{$_}})); } $sender_txt_header .= " " x ($COLUMN_WIDTHS - length($_)) . $_ . 's'; $sender_html_format .= "%s"; $sender_txt_format .= " " x ($COLUMN_WIDTHS - 5) . "%6s"; push(@col_headers,"${_}s"); } my $txt_format1 = " %-16s %9s %6d %6s $sender_txt_format"; my $txt_format2 = " %6d %4.1f%% %6d %4.1f%%", my $htm_format1 = "%s%s%s%s$sender_html_format"; my $htm_format2 = "%d%4.1f%%%d%4.1f%%"; if ($txt_fh) { my $sender_spaces = " " x length($sender_txt_header); print $txt_fh "\n"; print $txt_fh "Grand total summary\n"; print $txt_fh "-------------------\n"; print $txt_fh " $sender_spaces At least one address\n"; print $txt_fh " TOTAL Volume Messages Addresses $sender_txt_header Delayed Failed\n"; } if ($htm_fh) { print $htm_fh "\n"; print $htm_fh "

Grand total summary

\n"; print $htm_fh "\n"; print $htm_fh "\n"; } if ($xls_fh) { $ws_global->write($row++, 0, "Grand total summary", $f_header2); $ws_global->write($row, 0, \@col_headers, $f_header2); $ws_global->merge_range($row, scalar(@col_headers), $row, scalar(@col_headers)+1, "At least one addr Delayed", $f_header2_m); $ws_global->merge_range($row, scalar(@col_headers)+2, $row, scalar(@col_headers)+3, "At least one addr Failed", $f_header2_m); #$ws_global->write(++$row, scalar(@col_headers), ['Total','Percent','Total','Percent'], $f_header2); } my($volume,$failed_count); if ($merge_reports) { $volume = volume_rounded($report_totals{Received}{Volume}, $report_totals{Received}{'Volume-gigs'}); $total_received_count = get_report_total($report_totals{Received},'Messages'); $failed_count = get_report_total($report_totals{Received},'Failed'); $delayed_count = get_report_total($report_totals{Received},'Delayed'); } else { $volume = volume_rounded($total_received_data, $total_received_data_gigs); $failed_count = $message_errors; } { no integer; my @content=( $volume,$total_received_count,'', @received_totals, $delayed_count, ($total_received_count) ? ($delayed_count*100/$total_received_count) : 0, $failed_count, ($total_received_count) ? ($failed_count*100/$total_received_count) : 0 ); printf $txt_fh ("$txt_format1$txt_format2\n", 'Received', @content) if $txt_fh; printf $htm_fh ("$htm_format1$htm_format2\n", 'Received', @content) if $htm_fh; if ($xls_fh) { $ws_global->write(++$row, 0, 'Received', $f_default); for (my $i=0; $i < scalar(@content); $i++) { if ($i == 4 || $i == 6) { $ws_global->write($row, $i+1, $content[$i]/100, $f_percent); } else { $ws_global->write($row, $i+1, $content[$i], $f_default); } } } } if ($merge_reports) { $volume = volume_rounded($report_totals{Delivered}{Volume}, $report_totals{Delivered}{'Volume-gigs'}); $total_delivered_messages = get_report_total($report_totals{Delivered},'Messages'); $total_delivered_addresses = get_report_total($report_totals{Delivered},'Addresses'); } else { $volume = volume_rounded($total_delivered_data, $total_delivered_data_gigs); } my @content=($volume, $total_delivered_messages, $total_delivered_addresses, @delivered_totals); printf $txt_fh ("$txt_format1\n", 'Delivered', @content) if $txt_fh; printf $htm_fh ("$htm_format1\n", 'Delivered', @content) if $htm_fh; if ($xls_fh) { $ws_global->write(++$row, 0, 'Delivered', $f_default); for (my $i=0; $i < scalar(@content); $i++) { $ws_global->write($row, $i+1, $content[$i], $f_default); } } if ($merge_reports) { foreach ('Rejects', 'Temp Rejects', 'Ham', 'Spam') { my $messages = get_report_total($report_totals{$_},'Messages'); my $addresses = get_report_total($report_totals{$_},'Addresses'); if ($messages) { @content = ($_, '', $messages, ''); push(@content,get_report_total($report_totals{$_},'Hosts')) if $do_sender{Host}; #These rows do not have entries for the following columns (if specified) foreach ('Domain','Email','Edomain') { push(@content,'') if $do_sender{$_}; } printf $txt_fh ("$txt_format1\n", @content) if $txt_fh; printf $htm_fh ("$htm_format1\n", @content) if $htm_fh; $ws_global->write(++$row, 0, \@content) if $xls_fh; } } } else { foreach my $total_aref (['Rejects',\%rejected_count_by_ip], ['Temp Rejects',\%temporarily_rejected_count_by_ip], ['Ham',\%ham_count_by_ip], ['Spam',\%spam_count_by_ip]) { #Count the number of messages of this type. my $messages = 0; map {$messages += $_} values %{$total_aref->[1]}; if ($messages > 0) { @content = ($total_aref->[0], '', $messages, ''); #Count the number of distinct IPs for the Hosts column. push(@content,scalar(keys %{$total_aref->[1]})) if $do_sender{Host}; #These rows do not have entries for the following columns (if specified) foreach ('Domain','Email','Edomain') { push(@content,'') if $do_sender{$_}; } printf $txt_fh ("$txt_format1\n", @content) if $txt_fh; printf $htm_fh ("$htm_format1\n", @content) if $htm_fh; $ws_global->write(++$row, 0, \@content) if $xls_fh; } } } printf $txt_fh "\n" if $txt_fh; printf $htm_fh "
" . join('',@col_headers) . "At least one addr
Delayed
At least one addr
Failed
\n" if $htm_fh; ++$row; } ####################################################################### # print_user_patterns() # # print_user_patterns(); # # Print the counts of user specified patterns. ####################################################################### sub print_user_patterns { my $txt_format1 = " %-18s %6d"; my $htm_format1 = "%s%d"; if ($txt_fh) { print $txt_fh "User Specified Patterns\n"; print $txt_fh "-----------------------"; print $txt_fh "\n Total\n"; } if ($htm_fh) { print $htm_fh "

User Specified Patterns

\n"; print $htm_fh "\n"; print $htm_fh ""; if ($txt_fh) { print $txt_fh "Rejected mail by reason\n"; print $txt_fh "-----------------------"; print $txt_fh "\n Total\n"; } if ($htm_fh) { print $htm_fh "

Rejected mail by reason

\n"; print $htm_fh "
\n"; print $htm_fh "\n"; print $htm_fh "\n"; } if ($xls_fh) { $ws_global->write($row++, $col, "User Specified Patterns", $f_header2); &set_worksheet_line($ws_global, $row++, 1, ["Total"], $f_headertab); } my($key); if ($merge_reports) { # We are getting our data from previous reports. foreach $key (@user_descriptions) { my $count = get_report_total($report_totals{patterns}{$key},'Total'); printf $txt_fh ("$txt_format1\n",$key,$count) if $txt_fh; printf $htm_fh ("$htm_format1\n",$key,$count) if $htm_fh; if ($xls_fh) { &set_worksheet_line($ws_global, $row++, 0, [$key,$count], $f_default); } } } else { # We are getting our data from mainlog files. my $user_pattern_index = 0; foreach $key (@user_descriptions) { printf $txt_fh ("$txt_format1\n",$key,$user_pattern_totals[$user_pattern_index]) if $txt_fh; printf $htm_fh ("$htm_format1\n",$key,$user_pattern_totals[$user_pattern_index]) if $htm_fh; $ws_global->write($row++, 0, [$key,$user_pattern_totals[$user_pattern_index]]) if $xls_fh; $user_pattern_index++; } } print $txt_fh "\n" if $txt_fh; print $htm_fh "
 Total
\n\n" if $htm_fh; if ($xls_fh) { ++$row; } if ($hist_opt > 0) { my $user_pattern_index = 0; foreach $key (@user_descriptions) { print_histogram($key, 'occurence', @{$user_pattern_interval_count[$user_pattern_index]}); $user_pattern_index++; } } } ####################################################################### # print_rejects() # # print_rejects(); # # Print statistics about rejected mail. ####################################################################### sub print_rejects { my($format1,$reason); my $txt_format1 = " %-40s %6d"; my $htm_format1 = "
%s%d
\n"; print $htm_fh "\n"; } if ($xls_fh) { $ws_global->write($row++, $col, "Rejected mail by reason", $f_header2); &set_worksheet_line($ws_global, $row++, 1, ["Total"], $f_headertab); } my $href = ($merge_reports) ? $report_totals{rejected_mail_by_reason} : \%rejected_count_by_reason; my(@chartdatanames, @chartdatavals_count); foreach $reason (top_n_sort($topcount, $href, undef, undef)) { printf $txt_fh ("$txt_format1\n",$reason,$href->{$reason}) if $txt_fh; printf $htm_fh ("$htm_format1\n",$reason,$href->{$reason}) if $htm_fh; set_worksheet_line($ws_global, $row++, 0, [$reason,$href->{$reason}], $f_default) if $xls_fh; push(@chartdatanames, $reason); push(@chartdatavals_count, $href->{$reason}); } $row++ if $xls_fh; print $txt_fh "\n" if $txt_fh; if ($htm_fh) { print $htm_fh "
 Total
"; if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals_count > 0)) { # calculate the graph my @data = ( \@chartdatanames, \@chartdatavals_count ); my $graph = GD::Graph::pie->new(200, 200); $graph->set( x_label => 'Rejection Reasons', y_label => 'Messages', title => 'By count', ); my $gd = $graph->plot(\@data) or warn($graph->error); if ($gd) { open(IMG, ">$chartdir/rejections_count.png") or die "Could not write $chartdir/rejections_count.png: $!\n"; binmode IMG; print IMG $gd->png; close IMG; print $htm_fh ""; } } print $htm_fh "
\n\n"; } } ####################################################################### # print_transport(); # # print_transport(); # # Print totals by transport. ####################################################################### sub print_transport { my(@chartdatanames); my(@chartdatavals_count); my(@chartdatavals_vol); no integer; #Lose this for charting the data. my $txt_format1 = " %-18s %6s %6d"; my $htm_format1 = "%s%s%d"; if ($txt_fh) { print $txt_fh "Deliveries by transport\n"; print $txt_fh "-----------------------"; print $txt_fh "\n Volume Messages\n"; } if ($htm_fh) { print $htm_fh "

Deliveries by Transport

\n"; print $htm_fh "
\n"; print $htm_fh "\n"; } if ($xls_fh) { $ws_global->write(++$row, $col, "Deliveries by transport", $f_header2); $ws_global->write(++$row, 1, ["Volume", "Messages"], $f_headertab); } my($key); if ($merge_reports) { # We are getting our data from previous reports. foreach $key (sort keys %{$report_totals{transport}}) { my $count = get_report_total($report_totals{transport}{$key},'Messages'); my @content=($key, volume_rounded($report_totals{transport}{$key}{Volume}, $report_totals{transport}{$key}{'Volume-gigs'}), $count); push(@chartdatanames, $key); push(@chartdatavals_count, $count); push(@chartdatavals_vol, $report_totals{transport}{$key}{'Volume-gigs'}*$gig + $report_totals{transport}{$key}{Volume} ); printf $txt_fh ("$txt_format1\n", @content) if $txt_fh; printf $htm_fh ("$htm_format1\n", @content) if $htm_fh; $ws_global->write(++$row, 0, \@content) if $xls_fh; } } else { # We are getting our data from mainlog files. foreach $key (sort keys %transported_data) { my @content=($key, volume_rounded($transported_data{$key},$transported_data_gigs{$key}), $transported_count{$key}); push(@chartdatanames, $key); push(@chartdatavals_count, $transported_count{$key}); push(@chartdatavals_vol, $transported_data_gigs{$key}*$gig + $transported_data{$key}); printf $txt_fh ("$txt_format1\n", @content) if $txt_fh; printf $htm_fh ("$htm_format1\n", @content) if $htm_fh; $ws_global->write(++$row, 0, \@content) if $xls_fh; } } print $txt_fh "\n" if $txt_fh; if ($htm_fh) { print $htm_fh "
 VolumeMessages
"; if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals_count > 0)) { # calculate the graph my @data = ( \@chartdatanames, \@chartdatavals_count ); my $graph = GD::Graph::pie->new(200, 200); $graph->set( x_label => 'Transport', y_label => 'Messages', title => 'By count', ); my $gd = $graph->plot(\@data) or warn($graph->error); if ($gd) { open(IMG, ">$chartdir/transports_count.png") or die "Could not write $chartdir/transports_count.png: $!\n"; binmode IMG; print IMG $gd->png; close IMG; print $htm_fh ""; } } print $htm_fh ""; if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals_vol > 0)) { my @data = ( \@chartdatanames, \@chartdatavals_vol ); my $graph = GD::Graph::pie->new(200, 200); $graph->set( title => 'By volume', ); my $gd = $graph->plot(\@data) or warn($graph->error); if ($gd) { open(IMG, ">$chartdir/transports_vol.png") or die "Could not write $chartdir/transports_vol.png: $!\n"; binmode IMG; print IMG $gd->png; close IMG; print $htm_fh ""; } } print $htm_fh "
\n\n"; } } ####################################################################### # print_relay(); # # print_relay(); # # Print our totals by relay. ####################################################################### sub print_relay { my $row_print_relay=1; my $temp = "Relayed messages"; print $htm_fh "

$temp

\n" if $htm_fh; if (scalar(keys %relayed) > 0 || $relayed_unshown > 0) { my $shown = 0; my $spacing = ""; my $txt_format = "%7d %s\n => %s\n"; my $htm_format = "%d%s%s\n"; printf $txt_fh ("%s\n%s\n\n", $temp, "-" x length($temp)) if $txt_fh; if ($htm_fh) { print $htm_fh "\n"; print $htm_fh "\n"; } if ($xls_fh) { $ws_relayed->write($row_print_relay++, $col, $temp, $f_header2); &set_worksheet_line($ws_relayed, $row_print_relay++, 0, ["Count", "From", "To"], $f_headertab); } my($key); foreach $key (sort keys %relayed) { my $count = $relayed{$key}; $shown += $count; $key =~ s/[HA]=//g; my($one,$two) = split(/=> /, $key); my @content=($count, $one, $two); printf $txt_fh ($txt_format, @content) if $txt_fh; printf $htm_fh ($htm_format, @content) if $htm_fh; if ($xls_fh) { &set_worksheet_line($ws_relayed, $row_print_relay++, 0, \@content); } $spacing = "\n"; } print $htm_fh "
CountFromTo
\n

\n" if $htm_fh; print $txt_fh "${spacing}Total: $shown (plus $relayed_unshown unshown)\n\n" if $txt_fh; print $htm_fh "${spacing}Total: $shown (plus $relayed_unshown unshown)\n\n" if $htm_fh; if ($xls_fh) { &set_worksheet_line($ws_relayed, $row_print_relay++, 0, [$shown, "Sum of shown" ]); &set_worksheet_line($ws_relayed, $row_print_relay++, 0, [$relayed_unshown, "unshown"]); $row_print_relay++; } } else { print $txt_fh "No relayed messages\n-------------------\n\n" if $txt_fh; print $htm_fh "No relayed messages\n\n" if $htm_fh; if ($xls_fh) { $row_print_relay++; } } } ####################################################################### # print_errors(); # # print_errors(); # # Print our errors. In HTML, we display them as a list rather than a table - # Netscape doesn't like large tables! ####################################################################### sub print_errors { my $total_errors = 0; $row=1; if (scalar(keys %errors_count) != 0) { my $temp = "List of errors"; my $htm_format = "

  • %d - %s\n"; printf $txt_fh ("%s\n%s\n\n", $temp, "-" x length($temp)) if $txt_fh; if ($htm_fh) { print $htm_fh "

    $temp

    \n"; print $htm_fh "
    • Count - Error\n"; } if ($xls_fh) { $ws_errors->write($row++, 0, $temp, $f_header2); &set_worksheet_line($ws_errors, $row++, 0, ["Count", "Error"], $f_headertab); } my($key); foreach $key (sort keys %errors_count) { my $text = $key; chomp($text); $text =~ s/\s\s+/ /g; #Convert multiple spaces to a single space. $total_errors += $errors_count{$key}; if ($txt_fh) { printf $txt_fh ("%5d ", $errors_count{$key}); my $text_remaining = $text; while (length($text_remaining) > 65) { my($first,$rest) = $text_remaining =~ /(.{50}\S*)\s+(.+)/; last if !$first; printf $txt_fh ("%s\n\t ", $first); $text_remaining = $rest; } printf $txt_fh ("%s\n\n", $text_remaining); } if ($htm_fh) { #Translate HTML tag characters. Sergey Sholokh. $text =~ s/\/\>\;/g; printf $htm_fh ($htm_format,$errors_count{$key},$text); } if ($xls_fh) { &set_worksheet_line($ws_errors, $row++, 0, [$errors_count{$key},$text]); } } $temp = "Errors encountered: $total_errors"; if ($txt_fh) { print $txt_fh $temp, "\n"; print $txt_fh "-" x length($temp),"\n"; } if ($htm_fh) { print $htm_fh "
    \n

    \n"; print $htm_fh $temp, "\n"; } if ($xls_fh) { &set_worksheet_line($ws_errors, $row++, 0, [$total_errors, "Sum of Errors encountered"]); } } } ####################################################################### # parse_old_eximstat_reports(); # # parse_old_eximstat_reports($fh); # # Parse old eximstat output so we can merge daily stats to weekly stats and weekly to monthly etc. # # To test that the merging still works after changes, do something like the following. # All the diffs should produce no output. # # options='-bydomain -byemail -byhost -byedomain' # options="$options -show_rt1,2,4 -show_dt 1,2,4" # options="$options -pattern 'Completed Messages' /Completed/" # options="$options -pattern 'Received Messages' /<=/" # # ./eximstats $options mainlog > mainlog.txt # ./eximstats $options -merge mainlog.txt > mainlog.2.txt # diff mainlog.txt mainlog.2.txt # # ./eximstats $options -html mainlog > mainlog.html # ./eximstats $options -merge -html mainlog.txt > mainlog.2.html # diff mainlog.html mainlog.2.html # # ./eximstats $options -merge mainlog.html > mainlog.3.txt # diff mainlog.txt mainlog.3.txt # # ./eximstats $options -merge -html mainlog.html > mainlog.3.html # diff mainlog.html mainlog.3.html # # ./eximstats $options -nvr mainlog > mainlog.nvr.txt # ./eximstats $options -merge mainlog.nvr.txt > mainlog.4.txt # diff mainlog.txt mainlog.4.txt # # # double_mainlog.txt should have twice the values that mainlog.txt has. # ./eximstats $options mainlog mainlog > double_mainlog.txt ####################################################################### sub parse_old_eximstat_reports { my($fh) = @_; my(%league_table_value_entered, %league_table_value_was_zero, %table_order); my(%user_pattern_index); my $user_pattern_index = 0; map {$user_pattern_index{$_} = $user_pattern_index++} @user_descriptions; my $user_pattern_keys = join('|', @user_descriptions); while (<$fh>) { PARSE_OLD_REPORT_LINE: if (/Exim statistics from ([\d\-]+ [\d:]+(\s+[\+\-]\d+)?) to ([\d\-]+ [\d:]+(\s+[\+\-]\d+)?)/) { $begin = $1 if ($1 lt $begin); $end = $3 if ($3 gt $end); } elsif (/Grand total summary/) { # Fill in $report_totals{Received|Delivered}{Volume|Messages|Addresses|Hosts|Domains|...|Delayed|DelayedPercent|Failed|FailedPercent} my(@fields, @delivered_fields); my $doing_table = 0; while (<$fh>) { $_ = html2txt($_); #Convert general HTML markup to text. s/At least one addr//g; #Another part of the HTML output we don't want. # TOTAL Volume Messages Addresses Hosts Domains Delayed Failed # Received 26MB 237 177 23 8 3.4% 28 11.8% # Delivered 13MB 233 250 99 88 if (/TOTAL\s+(.*?)\s*$/) { $doing_table = 1; @delivered_fields = split(/\s+/,$1); #Delayed and Failed have two columns each, so add the extra field names in. splice(@delivered_fields,-1,1,'DelayedPercent','Failed','FailedPercent'); # Addresses only figure in the Delivered row, so remove them from the # normal fields. @fields = grep !/Addresses/, @delivered_fields; } elsif (/(Received)\s+(.*?)\s*$/) { print STDERR "Parsing $_" if $debug; add_to_totals($report_totals{$1},\@fields,$2); } elsif (/(Delivered)\s+(.*?)\s*$/) { print STDERR "Parsing $_" if $debug; add_to_totals($report_totals{$1},\@delivered_fields,$2); my $data = $2; # If we're merging an old report which doesn't include addresses, # then use the Messages field instead. unless (grep(/Addresses/, @delivered_fields)) { my %tmp; line_to_hash(\%tmp,\@delivered_fields,$data); add_to_totals($report_totals{Delivered},['Addresses'],$tmp{Messages}); } } elsif (/(Temp Rejects|Rejects|Ham|Spam)\s+(.*?)\s*$/) { print STDERR "Parsing $_" if $debug; add_to_totals($report_totals{$1},['Messages','Hosts'],$2); } else { last if $doing_table; } } } elsif (/User Specified Patterns/i) { #User Specified Patterns #----------------------- # Total # Description 85 while (<$fh>) { last if (/Total/); } #Wait until we get the table headers. while (<$fh>) { print STDERR "Parsing $_" if $debug; $_ = html2txt($_); #Convert general HTML markup to text. if (/^\s*(.*?)\s+(\d+)\s*$/) { $report_totals{patterns}{$1} = {} unless (defined $report_totals{patterns}{$1}); add_to_totals($report_totals{patterns}{$1},['Total'],$2); } last if (/^\s*$/); #Finished if we have a blank line. } } elsif (/(^|

    )($user_pattern_keys) per /o) { # Parse User defined pattern histograms if they exist. parse_histogram($fh, $user_pattern_interval_count[$user_pattern_index{$2}] ); } elsif (/Deliveries by transport/i) { #Deliveries by transport #----------------------- # Volume Messages # :blackhole: 70KB 51 # address_pipe 655KB 1 # smtp 11MB 151 while (<$fh>) { last if (/Volume/); } #Wait until we get the table headers. while (<$fh>) { print STDERR "Parsing $_" if $debug; $_ = html2txt($_); #Convert general HTML markup to text. if (/(\S+)\s+(\d+\S*\s+\d+)/) { $report_totals{transport}{$1} = {} unless (defined $report_totals{transport}{$1}); add_to_totals($report_totals{transport}{$1},['Volume','Messages'],$2); } last if (/^\s*$/); #Finished if we have a blank line. } } elsif (/Messages received per/) { parse_histogram($fh, \@received_interval_count); } elsif (/Deliveries per/) { parse_histogram($fh, \@delivered_interval_count); } #elsif (/Time spent on the queue: (all messages|messages with at least one remote delivery)/) { elsif (/(Time spent on the queue|Delivery times|Receipt times): ((\S+) messages|messages with at least one remote delivery)((<[^>]*>)*\s*)$/) { #Time spent on the queue: all messages #------------------------------------- # #Under 1m 217 91.9% 91.9% # 5m 2 0.8% 92.8% # 3h 8 3.4% 96.2% # 6h 7 3.0% 99.2% # 12h 2 0.8% 100.0% # Set a pointer to the queue bin so we can use the same code # block for both all messages and remote deliveries. #my $bin_aref = ($1 eq 'all messages') ? \@qt_all_bin : \@qt_remote_bin; my($bin_aref, $times_aref, $overflow_sref); if ($1 eq 'Time spent on the queue') { $times_aref = \@queue_times; if ($2 eq 'all messages') { $bin_aref = \@qt_all_bin; $overflow_sref = \$qt_all_overflow; } else { $bin_aref = \@qt_remote_bin; $overflow_sref = \$qt_remote_overflow; } } elsif ($1 eq 'Delivery times') { $times_aref = \@delivery_times; if ($2 eq 'all messages') { $bin_aref = \@dt_all_bin; $overflow_sref = \$dt_all_overflow; } else { $bin_aref = \@dt_remote_bin; $overflow_sref = \$dt_remote_overflow; } } else { unless (exists $rcpt_times_bin{$3}) { initialise_rcpt_times($3); } $bin_aref = $rcpt_times_bin{$3}; $times_aref = \@rcpt_times; $overflow_sref = \$rcpt_times_overflow{$3}; } my ($blank_lines, $reached_table) = (0,0); while (<$fh>) { $_ = html2txt($_); #Convert general HTML markup to text. # The table is preceded by one blank line, and has one blank line # following it. As the table may be empty, the best way to determine # that we've finished it is to look for the second blank line. ++$blank_lines if /^\s*$/; last if ($blank_lines >=2); #Finished the table ? $reached_table = 1 if (/\d/); next unless $reached_table; my $previous_seconds_on_queue = 0; if (/^\s*(Under|Over|)\s+(\d+[smhdw])\s+(\d+)/) { print STDERR "Parsing $_" if $debug; my($modifier,$formatted_time,$count) = ($1,$2,$3); my $seconds = unformat_time($formatted_time); my $time_on_queue = ($seconds + $previous_seconds_on_queue) / 2; $previous_seconds_on_queue = $seconds; $time_on_queue = $seconds * 2 if ($modifier eq 'Over'); my($i); for ($i = 0; $i <= $#$times_aref; $i++) { if ($time_on_queue < $times_aref->[$i]) { $$bin_aref[$i] += $count; last; } } $$overflow_sref += $count if ($i > $#$times_aref); } } } elsif (/Relayed messages/) { #Relayed messages #---------------- # # 1 addr.domain.com [1.2.3.4] a.user@domain.com # => addr2.domain2.com [5.6.7.8] a2.user2@domain2.com # #1addr.domain.com [1.2.3.4] a.user@domain.com addr2.domain2.com [5.6.7.8] a2.user2@domain2.com my $reached_table = 0; my($count,$sender); while (<$fh>) { unless ($reached_table) { last if (/No relayed messages/); $reached_table = 1 if (/^\s*\d/ || />\d+(\d+)<.td>(.*?) ?<.td>(.*?)\s+(.*?)\s*$/) { update_relayed($count,$sender,$1); } else { last; #Finished the table ? } } } elsif (/Top (.*?) by (message count|volume)/) { #Top 50 sending hosts by message count #------------------------------------- # # 48 1468KB local # Could also have average values for HTML output. # 48 1468KB 30KB local my($category,$by_count_or_volume) = ($1,$2); #As we show 2 views of each table (by count and by volume), #most (but not all) entries will appear in both tables. #Set up a hash to record which entries we have already seen #and one to record which ones we are seeing for the first time. if ($by_count_or_volume =~ /count/) { undef %league_table_value_entered; undef %league_table_value_was_zero; undef %table_order; } #As this section processes multiple different table categories, #set up pointers to the hashes to be updated. my($messages_href,$addresses_href,$data_href,$data_gigs_href); if ($category =~ /local sender/) { $messages_href = \%received_count_user; $addresses_href = undef; $data_href = \%received_data_user; $data_gigs_href = \%received_data_gigs_user; } elsif ($category =~ /sending (\S+?)s?\b/) { #Top 50 sending (host|domain|email|edomain)s #Top sending (host|domain|email|edomain) $messages_href = \%{$received_count{"\u$1"}}; $data_href = \%{$received_data{"\u$1"}}; $data_gigs_href = \%{$received_data_gigs{"\u$1"}}; } elsif ($category =~ /local destination/) { $messages_href = \%delivered_messages_user; $addresses_href = \%delivered_addresses_user; $data_href = \%delivered_data_user; $data_gigs_href = \%delivered_data_gigs_user; } elsif ($category =~ /local domain destination/) { $messages_href = \%delivered_messages_local_domain; $addresses_href = \%delivered_addresses_local_domain; $data_href = \%delivered_data_local_domain; $data_gigs_href = \%delivered_data_gigs_local_domain; } elsif ($category =~ /(\S+) destination/) { #Top 50 (host|domain|email|edomain) destinations #Top (host|domain|email|edomain) destination $messages_href = \%{$delivered_messages{"\u$1"}}; $addresses_href = \%{$delivered_addresses{"\u$1"}}; $data_href = \%{$delivered_data{"\u$1"}}; $data_gigs_href = \%{$delivered_data_gigs{"\u$1"}}; } elsif ($category =~ /temporarily rejected ips/) { $messages_href = \%temporarily_rejected_count_by_ip; } elsif ($category =~ /rejected ips/) { $messages_href = \%rejected_count_by_ip; } elsif ($category =~ /non-rejected spamming ips/) { $messages_href = \%spam_count_by_ip; } elsif ($category =~ /mail temporary rejection reasons/) { $messages_href = \%temporarily_rejected_count_by_reason; } elsif ($category =~ /mail rejection reasons/) { $messages_href = \%rejected_count_by_reason; } my $reached_table = 0; my $row_re; while (<$fh>) { # Watch out for empty tables. goto PARSE_OLD_REPORT_LINE if (/

    / or (/^\s*[a-zA-Z]/ && !/^\s*Messages/)); $_ = html2txt($_); #Convert general HTML markup to text. # Messages Addresses Bytes Average if (/^\s*Messages/) { my $pattern = '^\s*(\d+)'; $pattern .= (/Addresses/) ? '\s+(\d+)' : '()'; $pattern .= (/Bytes/) ? '\s+([\dKMGB]+)' : '()'; $pattern .= (/Average/) ? '\s+[\dKMGB]+' : ''; $pattern .= '\s+(.*?)\s*$'; $row_re = qr/$pattern/; $reached_table = 1; next; } next unless $reached_table; my($messages, $addresses, $rounded_volume, $entry); if (/$row_re/) { ($messages, $addresses, $rounded_volume, $entry) = ($1, $2, $3, $4); } else { #Else we have finished the table and we may need to do some #kludging to retain the order of the entries. if ($by_count_or_volume =~ /volume/) { #Add a few bytes to appropriate entries to preserve the order. foreach $rounded_volume (keys %table_order) { #For each rounded volume, we want to create a list which has things #ordered from the volume table at the front, and additional things #from the count table ordered at the back. @{$table_order{$rounded_volume}{volume}} = () unless defined $table_order{$rounded_volume}{volume}; @{$table_order{$rounded_volume}{'message count'}} = () unless defined $table_order{$rounded_volume}{'message count'}; my(@order,%mark); map {$mark{$_} = 1} @{$table_order{$rounded_volume}{volume}}; @order = @{$table_order{$rounded_volume}{volume}}; map {push(@order,$_)} grep(!$mark{$_},@{$table_order{$rounded_volume}{'message count'}}); my $bonus_bytes = $#order; $bonus_bytes = 511 if ($bonus_bytes > 511); #Don't go over the half-K boundary! while (@order and ($bonus_bytes > 0)) { my $entry = shift(@order); if ($league_table_value_was_zero{$entry}) { $$data_href{$entry} += $bonus_bytes; print STDERR "$category by $by_count_or_volume: added $bonus_bytes bonus bytes to $entry\n" if $debug; } $bonus_bytes--; } } } last; } # Store a new table entry. # Add the entry into the %table_order hash if it has a rounded # volume (KB/MB/GB). push(@{$table_order{$rounded_volume}{$by_count_or_volume}},$entry) if ($rounded_volume =~ /\D/); unless ($league_table_value_entered{$entry}) { $league_table_value_entered{$entry} = 1; unless ($$messages_href{$entry}) { $$messages_href{$entry} = 0; $$addresses_href{$entry} = 0; $$data_href{$entry} = 0; $$data_gigs_href{$entry} = 0; $league_table_value_was_zero{$entry} = 1; } $$messages_href{$entry} += $messages; # When adding the addresses, be aware that we could be merging # an old report which does not include addresses. In this case, # we add the messages instead. $$addresses_href{$entry} += ($addresses) ? $addresses : $messages; #Add the rounded value to the data and data_gigs hashes. un_round($rounded_volume,\$$data_href{$entry},\$$data_gigs_href{$entry}) if $rounded_volume; print STDERR "$category by $by_count_or_volume: added $messages,$rounded_volume to $entry\n" if $debug; } } } elsif (/List of errors/) { #List of errors #-------------- # # 1 07904931641@one2one.net R=external T=smtp: SMTP error # from remote mailer after RCPT TO:<07904931641@one2one.net>: # host mail.one2one.net [193.133.192.24]: 550 User unknown # #
  • 1 - ally.dufc@dunbar.org.uk R=external T=smtp: SMTP error from remote mailer after RCPT TO:: host mail.dunbar.org.uk [216.167.89.88]: 550 Unknown local part ally.dufc in my $reached_table = 0; my($count,$error,$blanks); while (<$fh>) { $reached_table = 1 if (/^( *|
  • )(\d+)/); next unless $reached_table; s/^
  • (\d+) -/$1/; #Convert an HTML line to a text line. $_ = html2txt($_); #Convert general HTML markup to text. if (/\t\s*(.*)/) { $error .= ' ' . $1; #Join a multiline error. } elsif (/^\s*(\d+)\s+(.*)/) { if ($error) { #Finished with a previous multiline error so save it. $errors_count{$error} = 0 unless $errors_count{$error}; $errors_count{$error} += $count; } ($count,$error) = ($1,$2); } elsif (/Errors encountered/) { if ($error) { #Finished the section, so save our stored last error. $errors_count{$error} = 0 unless $errors_count{$error}; $errors_count{$error} += $count; } last; } } } } } ####################################################################### # parse_histogram($fh, \@delivered_interval_count); # Parse a histogram into the provided array of counters. ####################################################################### sub parse_histogram { my($fh, $counters_aref) = @_; # Messages received per hour (each dot is 2 messages) #--------------------------------------------------- # #00-01 106 ..................................................... #01-02 103 ................................................... my $reached_table = 0; while (<$fh>) { $reached_table = 1 if (/^00/); next unless $reached_table; print STDERR "Parsing $_" if $debug; if (/^(\d+):(\d+)\s+(\d+)/) { #hh:mm start time format ? $$counters_aref[($1*60 + $2)/$hist_interval] += $3 if $hist_opt; } elsif (/^(\d+)-(\d+)\s+(\d+)/) { #hh-hh start-end time format ? $$counters_aref[($1*60)/$hist_interval] += $3 if $hist_opt; } else { #Finished the table ? last; } } } ####################################################################### # update_relayed(); # # update_relayed($count,$sender,$recipient); # # Adds an entry into the %relayed hash. Currently only used when # merging reports. ####################################################################### sub update_relayed { my($count,$sender,$recipient) = @_; #When generating the key, put in the 'H=' and 'A=' which can be used #in searches. my $key = "H=$sender => H=$recipient"; $key =~ s/ ([^=\s]+\@\S+|<>)/ A=$1/g; if (!defined $relay_pattern || $key !~ /$relay_pattern/o) { $relayed{$key} = 0 if !defined $relayed{$key}; $relayed{$key} += $count; } else { $relayed_unshown += $count; } } ####################################################################### # add_to_totals(); # # add_to_totals(\%totals,\@keys,$values); # # Given a line of space separated values, add them into the provided hash using @keys # as the hash keys. # # If the value contains a '%', then the value is set rather than added. Otherwise, we # convert the value to bytes and gigs. The gigs get added to I-gigs. ####################################################################### sub add_to_totals { my($totals_href,$keys_aref,$values) = @_; my(@values) = split(/\s+/,$values); for(my $i = 0; $i < @values && $i < @$keys_aref; ++$i) { my $key = $keys_aref->[$i]; if ($values[$i] =~ /%/) { $$totals_href{$key} = $values[$i]; } else { $$totals_href{$key} = 0 unless ($$totals_href{$key}); $$totals_href{"$key-gigs"} = 0 unless ($$totals_href{"$key-gigs"}); un_round($values[$i], \$$totals_href{$key}, \$$totals_href{"$key-gigs"}); print STDERR "Added $values[$i] to $key - $$totals_href{$key} , " . $$totals_href{"$key-gigs"} . "GB.\n" if $debug; } } } ####################################################################### # line_to_hash(); # # line_to_hash(\%hash,\@keys,$line); # # Given a line of space separated values, set them into the provided hash # using @keys as the hash keys. ####################################################################### sub line_to_hash { my($href,$keys_aref,$values) = @_; my(@values) = split(/\s+/,$values); for(my $i = 0; $i < @values && $i < @$keys_aref; ++$i) { $$href{$keys_aref->[$i]} = $values[$i]; } } ####################################################################### # get_report_total(); # # $total = get_report_total(\%hash,$key); # # If %hash contains values split into Units and Gigs, we calculate and return # # $hash{$key} + 1024*1024*1024 * $hash{"${key}-gigs"} ####################################################################### sub get_report_total { no integer; my($hash_ref,$key) = @_; if ($$hash_ref{"${key}-gigs"}) { return $$hash_ref{$key} + $gig * $$hash_ref{"${key}-gigs"}; } return $$hash_ref{$key} || 0; } ####################################################################### # html2txt(); # # $text_line = html2txt($html_line); # # Convert a line from html to text. Currently we just convert HTML tags to spaces # and convert >, <, and   tags back. ####################################################################### sub html2txt { ($_) = @_; # Convert HTML tags to spacing. Note that the reports may contain and # words, so explicitly specify the HTML tags we will remove # (the ones used by this program). If someone is careless enough to have their # Userid the same as an HTML tag, there's not much we can do about it. s/<\/?(html|head|title|body|h\d|ul|li|a\s+|table|tr|td|th|pre|hr|p|br)\b.*?>/ /g; s/\<\;/\/og; #Convert '>' to '>'. s/\ \;/ /og; #Convert ' ' to ' '. return($_); } ####################################################################### # get_next_arg(); # # $arg = get_next_arg(); # # Because eximstats arguments are often passed as variables, # we can't rely on shell parsing to deal with quotes. This # subroutine returns $ARGV[1] and does a shift. If $ARGV[1] # starts with a quote (' or "), and doesn't end in one, then # we append the next argument to it and shift again. We repeat # until we've got all of the argument. # # This isn't perfect as all white space gets reduced to one space, # but it's as good as we can get! If it's essential that spacing # be preserved precisely, then you get that by not using shell # variables. ####################################################################### sub get_next_arg { my $arg = ''; my $matched_pattern = 0; while ($ARGV[1]) { $arg .= ' ' if $arg; $arg .= $ARGV[1]; shift(@ARGV); if ($arg !~ /^['"]/) { $matched_pattern = 1; last; } if ($arg =~ s/^(['"])(.*)\1$/$2/) { $matched_pattern = 1; last; } } die "Mismatched argument quotes - <$arg>.\n" unless $matched_pattern; return $arg; } ####################################################################### # set_worksheet_line($ws_global, $startrow, $startcol, \@content, $format); # # set values to a sequence of cells in a row. # ####################################################################### sub set_worksheet_line { my ($worksheet, $row, $col, $content, $format) = @_; foreach my $token (@$content) { $worksheet->write($row, $col++, $token, $format ); } } ####################################################################### # @rcpt_times = parse_time_list($string); # # Parse a comma separated list of time values in seconds given by # the user and fill an array. # # Return a default list if $string is undefined. # Return () if $string eq '0'. ####################################################################### sub parse_time_list { my($string) = @_; if (! defined $string) { return(60, 5*60, 15*60, 30*60, 60*60, 3*60*60, 6*60*60, 12*60*60, 24*60*60); } my(@times) = split(/,/, $string); foreach my $q (@times) { $q = eval($q) + 0 } @times = sort { $a <=> $b } @times; @times = () if ($#times == 0 && $times[0] == 0); return(@times); } ####################################################################### # initialise_rcpt_times($protocol); # Initialise an array of rcpt_times to 0 for the specified protocol. ####################################################################### sub initialise_rcpt_times { my($protocol) = @_; for (my $i = 0; $i <= $#rcpt_times; ++$i) { $rcpt_times_bin{$protocol}[$i] = 0; } $rcpt_times_overflow{$protocol} = 0; } ################################################## # Main Program # ################################################## $last_timestamp = ''; $last_date = ''; $show_errors = 1; $show_relay = 1; $show_transport = 1; $topcount = 50; $local_league_table = 1; $include_remote_users = 0; $include_original_destination = 0; $hist_opt = 1; $volume_rounding = 1; $localtime_offset = calculate_localtime_offset(); # PH/FANF $charts = 0; $charts_option_specified = 0; $chartrel = "."; $chartdir = "."; @queue_times = parse_time_list(); @rcpt_times = (); @delivery_times = (); $last_offset = ''; $offset_seconds = 0; $row=1; $col=0; $col_hist=0; $run_hist=0; my(%output_files); # What output files have been specified? # Decode options while (@ARGV > 0 && substr($ARGV[0], 0, 1) eq '-') { if ($ARGV[0] =~ /^\-h(\d+)$/) { $hist_opt = $1 } elsif ($ARGV[0] =~ /^\-ne$/) { $show_errors = 0 } elsif ($ARGV[0] =~ /^\-nr(.?)(.*)\1$/) { if ($1 eq "") { $show_relay = 0 } else { $relay_pattern = $2 } } elsif ($ARGV[0] =~ /^\-q([,\d\+\-\*\/]+)$/) { @queue_times = parse_time_list($1) } elsif ($ARGV[0] =~ /^-nt$/) { $show_transport = 0 } elsif ($ARGV[0] =~ /^\-nt(.?)(.*)\1$/) { if ($1 eq "") { $show_transport = 0 } else { $transport_pattern = $2 } } elsif ($ARGV[0] =~ /^-t(\d+)$/) { $topcount = $1 } elsif ($ARGV[0] =~ /^-tnl$/) { $local_league_table = 0 } elsif ($ARGV[0] =~ /^-txt=?(\S*)$/) { $txt_fh = get_filehandle($1,\%output_files) } elsif ($ARGV[0] =~ /^-html=?(\S*)$/) { $htm_fh = get_filehandle($1,\%output_files) } elsif ($ARGV[0] =~ /^-xls=?(\S*)$/) { if ($HAVE_Spreadsheet_WriteExcel) { $xls_fh = get_filehandle($1,\%output_files); } else { warn "WARNING: CPAN Module Spreadsheet::WriteExcel not installed. Obtain from www.cpan.org\n"; } } elsif ($ARGV[0] =~ /^-merge$/) { $merge_reports = 1 } elsif ($ARGV[0] =~ /^-charts$/) { $charts = 1; warn "WARNING: CPAN Module GD::Graph::pie not installed. Obtain from www.cpan.org\n" unless $HAVE_GD_Graph_pie; warn "WARNING: CPAN Module GD::Graph::linespoints not installed. Obtain from www.cpan.org\n" unless $HAVE_GD_Graph_linespoints; } elsif ($ARGV[0] =~ /^-chartdir$/) { $chartdir = $ARGV[1]; shift; $charts_option_specified = 1; } elsif ($ARGV[0] =~ /^-chartrel$/) { $chartrel = $ARGV[1]; shift; $charts_option_specified = 1; } elsif ($ARGV[0] =~ /^-include_original_destination$/) { $include_original_destination = 1 } elsif ($ARGV[0] =~ /^-cache$/) { } #Not currently used. elsif ($ARGV[0] =~ /^-byhost$/) { $do_sender{Host} = 1 } elsif ($ARGV[0] =~ /^-bydomain$/) { $do_sender{Domain} = 1 } elsif ($ARGV[0] =~ /^-byemail$/) { $do_sender{Email} = 1 } elsif ($ARGV[0] =~ /^-byemaildomain$/) { $do_sender{Edomain} = 1 } elsif ($ARGV[0] =~ /^-byedomain$/) { $do_sender{Edomain} = 1 } elsif ($ARGV[0] =~ /^-bylocaldomain$/) { $do_local_domain = 1 } elsif ($ARGV[0] =~ /^-emptyok$/) { $emptyOK = 1 } elsif ($ARGV[0] =~ /^-nvr$/) { $volume_rounding = 0 } elsif ($ARGV[0] =~ /^-show_rt([,\d\+\-\*\/]+)?$/) { @rcpt_times = parse_time_list($1) } elsif ($ARGV[0] =~ /^-show_dt([,\d\+\-\*\/]+)?$/) { @delivery_times = parse_time_list($1) } elsif ($ARGV[0] =~ /^-d$/) { $debug = 1 } elsif ($ARGV[0] =~ /^--?h(elp)?$/){ help() } elsif ($ARGV[0] =~ /^-t_remote_users$/) { $include_remote_users = 1 } elsif ($ARGV[0] =~ /^-pattern$/) { push(@user_descriptions,get_next_arg()); push(@user_patterns,get_next_arg()); } elsif ($ARGV[0] =~ /^-utc$/) { # We don't need this value if the log is in UTC. $localtime_offset = undef; } else { print STDERR "Eximstats: Unknown or malformed option $ARGV[0]\n"; help(); } shift; } # keep old default behaviour if (! ($xls_fh or $htm_fh or $txt_fh)) { $txt_fh = \*STDOUT; } # Check that all the charts options are specified. warn "-charts option not specified. Use -help for help.\n" if ($charts_option_specified && ! $charts); # Default to display tables by sending Host. $do_sender{Host} = 1 unless ($do_sender{Domain} || $do_sender{Email} || $do_sender{Edomain}); # prepare xls Excel Workbook if (defined $xls_fh) { # Create a new Excel workbook $workbook = Spreadsheet::WriteExcel->new($xls_fh); # Add worksheets $ws_global = $workbook->addworksheet('Exim Statistik'); # show $ws_global as initial sheet $ws_global->set_first_sheet(); $ws_global->activate(); if ($show_relay) { $ws_relayed = $workbook->addworksheet('Relayed Messages'); $ws_relayed->set_column(1, 2, 80); } if ($show_errors) { $ws_errors = $workbook->addworksheet('Errors'); } # set column widths $ws_global->set_column(0, 2, 20); # Columns B-D width set to 30 $ws_global->set_column(3, 3, 15); # Columns B-D width set to 30 $ws_global->set_column(4, 4, 25); # Columns B-D width set to 30 # Define Formats $f_default = $workbook->add_format(); $f_header1 = $workbook->add_format(); $f_header1->set_bold(); #$f_header1->set_color('red'); $f_header1->set_size('15'); $f_header1->set_valign(); # $f_header1->set_align('center'); # $ws_global->write($row++, 2, "Testing Headers 1", $f_header1); $f_header2 = $workbook->add_format(); $f_header2->set_bold(); $f_header2->set_size('12'); $f_header2->set_valign(); # $ws_global->write($row++, 2, "Testing Headers 2", $f_header2); # Create another header2 for use in merged cells. $f_header2_m = $workbook->add_format(); $f_header2_m->set_bold(); $f_header2_m->set_size('8'); $f_header2_m->set_valign(); $f_header2_m->set_align('center'); $f_percent = $workbook->add_format(); $f_percent->set_num_format('0.0%'); $f_headertab = $workbook->add_format(); $f_headertab->set_bold(); $f_headertab->set_valign(); # $ws_global->write($row++, 2, "Testing Headers tab", $f_headertab); } # Initialise the queue/delivery/rcpt time counters. for (my $i = 0; $i <= $#queue_times; $i++) { $qt_all_bin[$i] = 0; $qt_remote_bin[$i] = 0; } for (my $i = 0; $i <= $#delivery_times; $i++) { $dt_all_bin[$i] = 0; $dt_remote_bin[$i] = 0; } initialise_rcpt_times('all'); # Compute the number of slots for the histogram if ($hist_opt > 0) { if ($hist_opt > 60 || 60 % $hist_opt != 0) { print STDERR "Eximstats: -h must specify a factor of 60\n"; exit 1; } $hist_interval = 60/$hist_opt; #Interval in minutes. $hist_number = (24*60)/$hist_interval; #Number of intervals per day. @received_interval_count = (0) x $hist_number; @delivered_interval_count = (0) x $hist_number; my $user_pattern_index = 0; for (my $user_pattern_index = 0; $user_pattern_index <= $#user_patterns; ++$user_pattern_index) { @{$user_pattern_interval_count[$user_pattern_index]} = (0) x $hist_number; } @dt_all_bin = (0) x $hist_number; @dt_remote_bin = (0) x $hist_number; } #$queue_unknown = 0; $total_received_data = 0; $total_received_data_gigs = 0; $total_received_count = 0; $total_delivered_data = 0; $total_delivered_data_gigs = 0; $total_delivered_messages = 0; $total_delivered_addresses = 0; $qt_all_overflow = 0; $qt_remote_overflow = 0; $dt_all_overflow = 0; $dt_remote_overflow = 0; $delayed_count = 0; $relayed_unshown = 0; $message_errors = 0; $begin = "9999-99-99 99:99:99"; $end = "0000-00-00 00:00:00"; my($section,$type); foreach $section ('Received','Delivered','Temp Rejects', 'Rejects','Ham','Spam') { foreach $type ('Volume','Messages','Delayed','Failed','Hosts','Domains','Emails','Edomains') { $report_totals{$section}{$type} = 0; } } # Generate our parser. my $parser = generate_parser(); if (@ARGV) { # Scan the input files and collect the data foreach my $file (@ARGV) { if ($file =~ /\.gz/) { unless (open(FILE,"gunzip -c $file |")) { print STDERR "Failed to gunzip -c $file: $!"; next; } } elsif ($file =~ /\.Z/) { unless (open(FILE,"uncompress -c $file |")) { print STDERR "Failed to uncompress -c $file: $!"; next; } } else { unless (open(FILE,$file)) { print STDERR "Failed to read $file: $!"; next; } } #Now parse the filehandle, updating the global variables. parse($parser,\*FILE); close FILE; } } else { #No files provided. Parse STDIN, updating the global variables. parse($parser,\*STDIN); } if ($begin eq "9999-99-99 99:99:99" && ! $emptyOK) { print STDERR "**** No valid log lines read\n"; exit 1; } # Output our results. print_header(); print_grandtotals(); # Print counts of user specified patterns if required. print_user_patterns() if @user_patterns; # Print rejection reasons. # print_rejects(); # Print totals by transport if required. print_transport() if $show_transport; # Print the deliveries per interval as a histogram, unless configured not to. # First find the maximum in one interval and scale accordingly. if ($hist_opt > 0) { print_histogram("Messages received", 'message', @received_interval_count); print_histogram("Deliveries", 'delivery', @delivered_interval_count); } # Print times on queue if required. if ($#queue_times >= 0) { print_duration_table("Time spent on the queue", "all messages", \@queue_times, \@qt_all_bin,$qt_all_overflow); print_duration_table("Time spent on the queue", "messages with at least one remote delivery", \@queue_times, \@qt_remote_bin,$qt_remote_overflow); } # Print delivery times if required. if ($#delivery_times >= 0) { print_duration_table("Delivery times", "all messages", \@delivery_times, \@dt_all_bin,$dt_all_overflow); print_duration_table("Delivery times", "messages with at least one remote delivery", \@delivery_times, \@dt_remote_bin,$dt_remote_overflow); } # Print rcpt times if required. if ($#rcpt_times >= 0) { foreach my $protocol ('all', grep(!/^all$/, sort keys %rcpt_times_bin)) { print_duration_table("Receipt times", "$protocol messages", \@rcpt_times, $rcpt_times_bin{$protocol}, $rcpt_times_overflow{$protocol}); } } # Print relay information if required. print_relay() if $show_relay; # Print the league tables, if topcount isn't zero. if ($topcount > 0) { my($ws_rej, $ws_top50, $ws_rej_row, $ws_top50_row, $ws_temp_rej, $ws_temp_rej_row); $ws_rej_row = $ws_temp_rej_row = $ws_top50_row = 0; if ($xls_fh) { $ws_top50 = $workbook->addworksheet('Deliveries'); $ws_rej = $workbook->addworksheet('Rejections') if (%rejected_count_by_reason || %rejected_count_by_ip || %spam_count_by_ip); $ws_temp_rej = $workbook->addworksheet('Temporary Rejections') if (%temporarily_rejected_count_by_reason || %temporarily_rejected_count_by_ip); } print_league_table("mail rejection reason", \%rejected_count_by_reason, undef, undef, undef, $ws_rej, \$ws_rej_row) if %rejected_count_by_reason; print_league_table("mail temporary rejection reason", \%temporarily_rejected_count_by_reason, undef, undef, undef, $ws_temp_rej, \$ws_temp_rej_row) if %temporarily_rejected_count_by_reason; foreach ('Host','Domain','Email','Edomain') { next unless $do_sender{$_}; print_league_table("sending \l$_", $received_count{$_}, undef, $received_data{$_},$received_data_gigs{$_}, $ws_top50, \$ws_top50_row); } print_league_table("local sender", \%received_count_user, undef, \%received_data_user,\%received_data_gigs_user, $ws_top50, \$ws_top50_row) if (($local_league_table || $include_remote_users) && %received_count_user); foreach ('Host','Domain','Email','Edomain') { next unless $do_sender{$_}; print_league_table("\l$_ destination", $delivered_messages{$_}, $delivered_addresses{$_}, $delivered_data{$_},$delivered_data_gigs{$_}, $ws_top50, \$ws_top50_row); } print_league_table("local destination", \%delivered_messages_user, \%delivered_addresses_user, \%delivered_data_user,\%delivered_data_gigs_user, $ws_top50, \$ws_top50_row) if (($local_league_table || $include_remote_users) && %delivered_messages_user); print_league_table("local domain destination", \%delivered_messages_local_domain, \%delivered_addresses_local_domain, \%delivered_data_local_domain,\%delivered_data_gigs_local_domain, $ws_top50, \$ws_top50_row) if (($local_league_table || $include_remote_users) && %delivered_messages_local_domain); print_league_table("rejected ip", \%rejected_count_by_ip, undef, undef, undef, $ws_rej, \$ws_rej_row) if %rejected_count_by_ip; print_league_table("temporarily rejected ip", \%temporarily_rejected_count_by_ip, undef, undef, undef, $ws_rej, \$ws_rej_row) if %temporarily_rejected_count_by_ip; print_league_table("non-rejected spamming ip", \%spam_count_by_ip, undef, undef, undef, $ws_rej, \$ws_rej_row) if %spam_count_by_ip; } # Print the error statistics if required. print_errors() if $show_errors; print $htm_fh "\n\n" if $htm_fh; $txt_fh->close if $txt_fh && ref $txt_fh; $htm_fh->close if $htm_fh; if ($xls_fh) { # close Excel Workbook $ws_global->set_first_sheet(); # FIXME: whyever - activate does not work :-/ $ws_global->activate(); $workbook->close(); } # End of eximstats