summaryrefslogtreecommitdiffstats
path: root/src/eximstats.src
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-27 09:44:07 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-27 09:44:07 +0000
commit39ce00b8d520cbecbd6af87257e8fb11df0ec273 (patch)
tree4c21a2674c19e5c44be3b3550b476b9e63d8ae3d /src/eximstats.src
parentInitial commit. (diff)
downloadexim4-upstream/4.94.2.tar.xz
exim4-upstream/4.94.2.zip
Adding upstream version 4.94.2.upstream/4.94.2upstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'src/eximstats.src')
-rw-r--r--src/eximstats.src4246
1 files changed, 4246 insertions, 0 deletions
diff --git a/src/eximstats.src b/src/eximstats.src
new file mode 100644
index 0000000..5e1a084
--- /dev/null
+++ b/src/eximstats.src
@@ -0,0 +1,4246 @@
+#!PERL_COMMAND
+
+# Copyright (c) 2001-2017 University of Cambridge.
+# See the file NOTICE for conditions of use and distribution.
+
+# Perl script to generate statistics from one or more Exim log files.
+
+# Usage: eximstats [<options>] <log file> <log file> ...
+
+# 1996-05-21: Ignore lines not starting with valid date/time, just in case
+# these get into a log file.
+# 1996-11-19: Add the -h option to control the size of the histogram,
+# and optionally turn it off.
+# Use some Perl 5 things; it should be everywhere by now.
+# Add the Perl -w option and rewrite so no warnings are given.
+# Add the -t option to control the length of the "top" listing.
+# Add the -ne, -nt options to turn off errors and transport
+# information.
+# Add information about length of time on queue, and -q<list> to
+# control the intervals and turn it off.
+# Add count and percentage of delayed messages to the Received
+# line.
+# Show total number of errors.
+# Add count and percentage of messages with errors to Received
+# line.
+# Add information about relaying and -nr to suppress it.
+# 1997-02-03 Merged in some of the things Nigel Metheringham had done:
+# Re-worded headings
+# Added received histogram as well as delivered
+# Added local senders' league table
+# Added local recipients' league table
+# 1997-03-10 Fixed typo "destinationss"
+# Allow for intermediate address between final and original
+# when testing for relaying
+# Give better message when no input
+# 1997-04-24 Fixed bug in layout of error listing that was depending on
+# text length (output line got repeated).
+# 1997-05-06 Bug in option decoding when only one option.
+# Overflow bug when handling very large volumes.
+# 1997-10-28 Updated to handle revised log format that might show
+# HELO name as well as host name before IP number
+# 1998-01-26 Bugs in the function for calculating the number of seconds
+# since 1970 from a log date
+# 1998-02-02 Delivery to :blackhole: doesn't have a T= entry in the log
+# line; cope with this, thereby avoiding undefined problems
+# Very short log line gave substring error
+# 1998-02-03 A routed delivery to a local transport may not have <> in the
+# log line; terminate the address at white space, not <
+# 1998-09-07 If first line of input was a => line, $thissize was undefined;
+# ensure it is zero.
+# 1998-12-21 Adding of $thissize from => line should have been adding $size.
+# Oops. Should have looked more closely when fixing the previous
+# bug!
+# 1999-11-12 Increased the field widths for printed integers; numbers are
+# bigger than originally envisaged.
+# 2001-03-21 Converted seconds() routine to use Time::Local, fixing a bug
+# whereby seconds($timestamp) - id_seconds($id) gave an
+# incorrect result.
+# Added POD documentation.
+# Moved usage instructions into help() subroutine.
+# Added 'use strict' and declared all global variables.
+# Added '-html' flag and resultant code.
+# Added '-cache' flag and resultant code.
+# Added add_volume() routine and converted all volume variables
+# to use it, fixing the overflow problems for individual hosts
+# on large sites.
+# Converted all volume output to GB/MB/KB as appropriate.
+# Don't store local user stats if -nfl is specified.
+# Modifications done by: Steve Campbell (<steve@computurn.com>)
+# 2001-04-02 Added the -t_remote_users flag. Steve Campbell.
+# 2001-10-15 Added the -domain flag. Steve Campbell.
+# 2001-10-16 Accept files on STDIN or on the command line. Steve Campbell.
+# 2001-10-21 Removed -domain flag and added -bydomain, -byhost, and -byemail.
+# We now generate our main parsing subroutine as an eval statement
+# which improves performance dramatically when not all the results
+# are required. We also cache the last timestamp to time conversion.
+#
+# NOTE: 'Top 50 destinations by (message count|volume)' lines are
+# now 'Top N (host|email|domain) destinations by (message count|volume)'
+# where N is the topcount. Steve Campbell.
+#
+# 2001-10-30 V1.16 Joachim Wieland.
+# Fixed minor bugs in add_volume() when taking over this version
+# for use in Exim 4: -w gave uninitialized value warnings in
+# two situations: for the first addition to a counter, and if
+# there were never any gigabytes, thereby leaving the $gigs
+# value unset.
+# Initialized $last_timestamp to stop a -w uninitialized warning.
+# Minor layout tweak for grand totals (nitpicking).
+# Put the IP addresses for relaying stats in [] and separated by
+# a space from the domain name.
+# Removed the IPv4-specific address test when picking out addresses
+# for relaying. Anything inside [] is OK.
+#
+# 2002-07-02 Philip Hazel
+# Fixed "uninitialized variable" message that occurred for relay
+# messages that arrived from H=[1.2.3.4] hosts (no name shown).
+# This bug didn't affect the output.
+#
+# 2002-04-15 V1.17 Joachim Wieland.
+# Added -charts, -chartdir. -chartrel options which use
+# GD::Graph modules to create graphical charts of the statistics.
+#
+# 2002-04-15 V1.18 Steve Campbell.
+# Added a check for $domain to to stop a -w uninitialized warning.
+# Added -byemaildomain option.
+# Only print HTML header links to included tables!
+#
+# 2002-08-02 V1.19 Steve Campbell.
+# Changed the debug mode to dump the parser onto STDERR rather
+# than STDOUT. Documented the -d flag into the help().
+# Rejoined the divergent 2002-04-15 and 2002-07-02 releases.
+#
+# 2002-08-21 V1.20 Steve Campbell.
+# Added the '-merge' option to allow merging of previous reports.
+# Fixed a missing semicolon when doing -bydomain.
+# Make volume charts plot the data gigs and bytes rather than just bytes.
+# Only process log lines with $flag =~ /<=|=>|->|==|\*\*|Co/
+# Converted Emaildomain to Edomain - the column header was too wide!
+# This changes the text output slightly. You can revert to the old
+# column widths by changing $COLUMN_WIDTHS to 7;
+#
+# 2002-09-04 V1.21 Andreas J Mueller
+# Local deliveries domain now defaults to 'localdomain'.
+# Don't match F=<From> when looking for the user.
+#
+# 2002-09-05 V1.22 Steve Campbell
+# Fixed a perl 5.005 incompatibility problem ('our' variables).
+#
+# 2002-09-11 V1.23 Steve Campbell
+# Stopped -charts option from throwing errors on null data.
+# Don't print out 'Errors encountered' unless there are any.
+
+# 2002-10-21 V1.23a Philip Hazel - patch from Tony Finch put in until
+# Steve's eximstats catches up.
+# Handle log files that include the timezone after the timestamp.
+# Switch to assuming that log timestamps are in local time, with
+# an option for UTC timestamps, as in Exim itself.
+#
+# 2003-02-05 V1.24 Steve Campbell
+# Added in Sergey Sholokh's code to convert '<' and '>' characters
+# in HTML output. Also added code to convert them back with -merge.
+# Fixed timestamp offsets to convert to seconds rather than minutes.
+# Updated -merge to work with output files using timezones.
+# Added caching to speed up the calculation of timezone offsets.
+#
+# 2003-02-07 V1.25 Steve Campbell
+# Optimised the usage of mktime() in the seconds subroutine.
+# Removed the now redundant '-cache' option.
+# html2txt() now explicitly matches HTML tags.
+# Implemented a new sorting algorithm - the top_n_sort() routine.
+# Added Danny Carroll's '-nvr' flag and code.
+#
+# 2003-03-13 V1.26 Steve Campbell
+# Implemented HTML compliance changes recommended by Bernard Massot.
+# Bug fix to allow top_n_sort() to handle null keys.
+# Convert all domains and edomains to lowercase.
+# Remove preceding dots from domains.
+#
+# 2003-03-13 V1.27 Steve Campbell
+# Replaced border attributes with 'border=1', as recommended by
+# Bernard Massot.
+#
+# 2003-06-03 V1.28 John Newman
+# Added in the ability to skip over the parsing and evaluation of
+# specific transports as passed to eximstats via the new "-nt/.../"
+# command line argument. This new switch allows the viewing of
+# not more accurate statistics but more applicable statistics when
+# special transports are in use (ie; SpamAssassin). We need to be
+# able to ignore transports such as this otherwise the resulting
+# local deliveries are significantly skewed (doubled)...
+#
+# 2003-11-06 V1.29 Steve Campbell
+# Added the '-pattern "Description" "/pattern/"' option.
+#
+# 2004-02-17 V1.30 Steve Campbell
+# Added warnings if required GD::Graph modules are not available or
+# insufficient -chart* options are specified.
+#
+# 2004-02-20 V1.31 Andrea Balzi
+# Only show the Local Sender/Destination links if the tables exist.
+#
+# 2004-07-05 V1.32 Steve Campbell
+# Fix '-merge -h0' divide by zero error.
+#
+# 2004-07-15 V1.33 Steve Campbell
+# Documentation update - I've converted the subroutine
+# documentation from POD to comments.
+#
+# 2004-12-10 V1.34 Steve Campbell
+# Eximstats can now parse syslog lines as well as mainlog lines.
+#
+# 2004-12-20 V1.35 Wouter Verhelst
+# Pie charts by volume were actually generated by count. Fixed.
+#
+# 2005-02-07 V1.36 Gregor Herrmann / Steve Campbell
+# Added average sizes to HTML Top tables.
+#
+# 2005-04-26 V1.37 Frank Heydlauf
+# Added -xls and the ability to specify output files.
+#
+# 2005-04-29 V1.38 Steve Campbell
+# Use FileHandles for outputting results.
+# Allow any combination of xls, txt, and html output.
+# Fixed display of large numbers with -nvr option
+# Fixed merging of reports with empty tables.
+#
+# 2005-05-27 V1.39 Steve Campbell
+# Added the -include_original_destination flag
+# Removed tabs and trailing whitespace.
+#
+# 2005-06-03 V1.40 Steve Campbell
+# Whilst parsing the mainlog(s), store information about
+# the messages in a hash of arrays rather than using
+# individual hashes. This is a bit cleaner and results in
+# dramatic memory savings, albeit at a slight CPU cost.
+#
+# 2005-06-15 V1.41 Steve Campbell
+# Added the -show_rt<list> flag.
+# Added the -show_dt<list> flag.
+#
+# 2005-06-24 V1.42 Steve Campbell
+# Added Histograms for user specified patterns.
+#
+# 2005-06-30 V1.43 Steve Campbell
+# Bug fix for V1.42 with -h0 specified. Spotted by Chris Lear.
+#
+# 2005-07-26 V1.44 Steve Campbell
+# Use a glob alias rather than an array ref in the generated
+# parser. This improves both readability and performance.
+#
+# 2005-09-30 V1.45 Marco Gaiarin / Steve Campbell
+# Collect SpamAssassin and rejection statistics.
+# Don't display local sender or destination tables unless
+# there is data to show.
+# Added average volumes into the top table text output.
+#
+# 2006-02-07 V1.46 Steve Campbell
+# Collect data on the number of addresses (recipients)
+# as well as the number of messages.
+#
+# 2006-05-05 V1.47 Steve Campbell
+# Added 'Message too big' to the list of mail rejection
+# reasons (thanks to Marco Gaiarin).
+#
+# 2006-06-05 V1.48 Steve Campbell
+# Mainlog lines which have GMT offsets and are too short to
+# have a flag are now skipped.
+#
+# 2006-11-10 V1.49 Alain Williams
+# Added the -emptyok flag.
+#
+# 2006-11-16 V1.50 Steve Campbell
+# Fixes for obtaining the IP address from reject messages.
+#
+# 2006-11-27 V1.51 Steve Campbell
+# Another update for obtaining the IP address from reject messages.
+#
+# 2006-11-27 V1.52 Steve Campbell
+# Tally any reject message containing SpamAssassin.
+#
+# 2007-01-31 V1.53 Philip Hazel
+# Allow for [pid] after date in log lines
+#
+# 2007-02-14 V1.54 Daniel Tiefnig
+# Improved the '($parent) =' pattern match.
+#
+# 2007-03-19 V1.55 Steve Campbell
+# Differentiate between permanent and temporary rejects.
+#
+# 2007-03-29 V1.56 Jez Hancock
+# Fixed some broken HTML links and added missing column headers.
+#
+# 2007-03-30 V1.57 Steve Campbell
+# Fixed Grand Total Summary Domains, Edomains, and Email columns
+# for Rejects, Temp Rejects, Ham, and Spam rows.
+#
+# 2007-04-11 V1.58 Steve Campbell
+# Fix to get <> and blackhole to show in edomain tables.
+#
+# 2007-09-20 V1.59 Steve Campbell
+# Added the -bylocaldomain option
+#
+# 2007-09-20 V1.60 Heiko Schlittermann
+# Fix for misinterpreted log lines
+#
+# 2013-01-14 V1.61 Steve Campbell
+# Watch out for senders sending "HELO [IpAddr]"
+#
+#
+# For documentation on the logfile format, see
+# http://www.exim.org/exim-html-4.50/doc/html/spec_48.html#IX2793
+
+=head1 NAME
+
+eximstats - generates statistics from Exim mainlog or syslog files.
+
+=head1 SYNOPSIS
+
+ eximstats [Output] [Options] mainlog1 mainlog2 ...
+ eximstats -merge [Options] report.1.txt report.2.txt ... > weekly_report.txt
+
+=head2 Output:
+
+=over 4
+
+=item B<-txt>
+
+Output the results in plain text to STDOUT.
+
+=item B<-txt>=I<filename>
+
+Output the results in plain text. Filename '-' for STDOUT is accepted.
+
+=item B<-html>
+
+Output the results in HTML to STDOUT.
+
+=item B<-html>=I<filename>
+
+Output the results in HTML. Filename '-' for STDOUT is accepted.
+
+=item B<-xls>
+
+Output the results in Excel compatible Format to STDOUT.
+Requires the Spreadsheet::WriteExcel CPAN module.
+
+=item B<-xls>=I<filename>
+
+Output the results in Excel compatible format. Filename '-' for STDOUT is accepted.
+
+
+=back
+
+=head2 Options:
+
+=over 4
+
+=item B<-h>I<number>
+
+histogram divisions per hour. The default is 1, and
+0 suppresses histograms. Valid values are:
+
+0, 1, 2, 3, 5, 10, 15, 20, 30 or 60.
+
+=item B<-ne>
+
+Don't display error information.
+
+=item B<-nr>
+
+Don't display relaying information.
+
+=item B<-nr>I</pattern/>
+
+Don't display relaying information that matches.
+
+=item B<-nt>
+
+Don't display transport information.
+
+=item B<-nt>I</pattern/>
+
+Don't display transport information that matches
+
+=item B<-q>I<list>
+
+List of times for queuing information single 0 item suppresses.
+
+=item B<-t>I<number>
+
+Display top <number> sources/destinations
+default is 50, 0 suppresses top listing.
+
+=item B<-tnl>
+
+Omit local sources/destinations in top listing.
+
+=item B<-t_remote_users>
+
+Include remote users in the top source/destination listings.
+
+=item B<-include_original_destination>
+
+Include the original destination email addresses rather than just
+using the final ones.
+Useful for finding out which of your mailing lists are receiving mail.
+
+=item B<-show_dt>I<list>
+
+Show the delivery times (B<DT>)for all the messages.
+
+Exim must have been configured to use the +deliver_time logging option
+for this option to work.
+
+I<list> is an optional list of times. Eg -show_dt1,2,4,8 will show
+the number of messages with delivery times under 1 second, 2 seconds, 4 seconds,
+8 seconds, and over 8 seconds.
+
+=item B<-show_rt>I<list>
+
+Show the receipt times for all the messages. The receipt time is
+defined as the Completed hh:mm:ss - queue_time_overall - the Receipt hh:mm:ss.
+These figures will be skewed by pipelined messages so might not be that useful.
+
+Exim must have been configured to use the +queue_time_overall logging option
+for this option to work.
+
+I<list> is an optional list of times. Eg -show_rt1,2,4,8 will show
+the number of messages with receipt times under 1 second, 2 seconds, 4 seconds,
+8 seconds, and over 8 seconds.
+
+=item B<-byhost>
+
+Show results by sending host. This may be combined with
+B<-bydomain> and/or B<-byemail> and/or B<-byedomain>. If none of these options
+are specified, then B<-byhost> is assumed as a default.
+
+=item B<-bydomain>
+
+Show results by sending domain.
+May be combined with B<-byhost> and/or B<-byemail> and/or B<-byedomain>.
+
+=item B<-byemail>
+
+Show results by sender's email address.
+May be combined with B<-byhost> and/or B<-bydomain> and/or B<-byedomain>.
+
+=item B<-byemaildomain> or B<-byedomain>
+
+Show results by sender's email domain.
+May be combined with B<-byhost> and/or B<-bydomain> and/or B<-byemail>.
+
+=item B<-pattern> I<Description> I</Pattern/>
+
+Look for the specified pattern and count the number of lines in which it appears.
+This option can be specified multiple times. Eg:
+
+ -pattern 'Refused connections' '/refused connection/'
+
+
+=item B<-merge>
+
+This option allows eximstats to merge old eximstat reports together. Eg:
+
+ eximstats mainlog.sun > report.sun.txt
+ eximstats mainlog.mon > report.mon.txt
+ eximstats mainlog.tue > report.tue.txt
+ eximstats mainlog.wed > report.web.txt
+ eximstats mainlog.thu > report.thu.txt
+ eximstats mainlog.fri > report.fri.txt
+ eximstats mainlog.sat > report.sat.txt
+ eximstats -merge report.*.txt > weekly_report.txt
+ eximstats -merge -html report.*.txt > weekly_report.html
+
+=over 4
+
+=item *
+
+You can merge text or html reports and output the results as text or html.
+
+=item *
+
+You can use all the normal eximstat output options, but only data
+included in the original reports can be shown!
+
+=item *
+
+When merging reports, some loss of accuracy may occur in the top I<n> lists.
+This will be towards the ends of the lists.
+
+=item *
+
+The order of items in the top I<n> lists may vary when the data volumes
+round to the same value.
+
+=back
+
+=item B<-charts>
+
+Create graphical charts to be displayed in HTML output.
+Only valid in combination with I<-html>.
+
+This requires the following modules which can be obtained
+from http://www.cpan.org/modules/01modules.index.html
+
+=over 4
+
+=item GD
+
+=item GDTextUtil
+
+=item GDGraph
+
+=back
+
+To install these, download and unpack them, then use the normal perl installation procedure:
+
+ perl Makefile.PL
+ make
+ make test
+ make install
+
+=item B<-chartdir>I <dir>
+
+Create the charts in the directory <dir>
+
+=item B<-chartrel>I <dir>
+
+Specify the relative directory for the "img src=" tags from where to include
+the charts
+
+=item B<-emptyok>
+
+Specify that it's OK to not find any valid log lines. Without this
+we will output an error message if we don't find any.
+
+=item B<-d>
+
+Debug flag. This outputs the eval()'d parser onto STDOUT which makes it
+easier to trap errors in the eval section. Remember to add 1 to the line numbers to allow for the
+title!
+
+=back
+
+=head1 DESCRIPTION
+
+Eximstats parses exim mainlog and syslog files to output a statistical
+analysis of the messages processed. By default, a text
+analysis is generated, but you can request other output formats
+using flags. See the help (B<-help>) to learn
+about how to create charts from the tables.
+
+=head1 AUTHOR
+
+There is a website at https://www.exim.org - this contains details of the
+mailing list exim-users@exim.org.
+
+=head1 TO DO
+
+This program does not perfectly handle messages whose received
+and delivered log lines are in different files, which can happen
+when you have multiple mail servers and a message cannot be
+immediately delivered. Fixing this could be tricky...
+
+Merging of xls files is not (yet) possible. Be free to implement :)
+
+=cut
+
+use warnings;
+use integer;
+BEGIN { pop @INC if $INC[-1] eq '.' };
+use strict;
+use IO::File;
+use File::Basename;
+
+# use Time::Local; # PH/FANF
+use POSIX;
+
+if (@ARGV and $ARGV[0] eq '--version') {
+ print basename($0) . ": $0\n",
+ "build: EXIM_RELEASE_VERSIONEXIM_VARIANT_VERSION\n",
+ "perl(runtime): $]\n";
+ exit 0;
+}
+
+use vars qw($HAVE_GD_Graph_pie $HAVE_GD_Graph_linespoints $HAVE_Spreadsheet_WriteExcel);
+eval { require GD::Graph::pie; };
+$HAVE_GD_Graph_pie = $@ ? 0 : 1;
+eval { require GD::Graph::linespoints; };
+$HAVE_GD_Graph_linespoints = $@ ? 0 : 1;
+eval { require Spreadsheet::WriteExcel; };
+$HAVE_Spreadsheet_WriteExcel = $@ ? 0 : 1;
+
+
+##################################################
+# Static data #
+##################################################
+# 'use vars' instead of 'our' as perl5.005 is still in use out there!
+use vars qw(@tab62 @days_per_month $gig);
+use vars qw($VERSION);
+use vars qw($COLUMN_WIDTHS);
+use vars qw($WEEK $DAY $HOUR $MINUTE);
+
+
+@tab62 =
+ (0,1,2,3,4,5,6,7,8,9,0,0,0,0,0,0, # 0-9
+ 0,10,11,12,13,14,15,16,17,18,19,20, # A-K
+ 21,22,23,24,25,26,27,28,29,30,31,32, # L-W
+ 33,34,35, 0, 0, 0, 0, 0, # X-Z
+ 0,36,37,38,39,40,41,42,43,44,45,46, # a-k
+ 47,48,49,50,51,52,53,54,55,56,57,58, # l-w
+ 59,60,61); # x-z
+
+@days_per_month = (0, 31, 59, 90, 120, 151, 181, 212, 243, 273, 304, 334);
+$gig = 1024 * 1024 * 1024;
+$VERSION = '1.61';
+
+# How much space do we allow for the Hosts/Domains/Emails/Edomains column headers?
+$COLUMN_WIDTHS = 8;
+
+$MINUTE = 60;
+$HOUR = 60 * $MINUTE;
+$DAY = 24 * $HOUR;
+$WEEK = 7 * $DAY;
+
+# Declare global variables.
+use vars qw($total_received_data $total_received_data_gigs $total_received_count);
+use vars qw($total_delivered_data $total_delivered_data_gigs $total_delivered_messages $total_delivered_addresses);
+use vars qw(%timestamp2time); #Hash of timestamp => time.
+use vars qw($last_timestamp $last_time); #The last time conversion done.
+use vars qw($last_date $date_seconds); #The last date conversion done.
+use vars qw($last_offset $offset_seconds); #The last time offset conversion done.
+use vars qw($localtime_offset);
+use vars qw($i); #General loop counter.
+use vars qw($debug); #Debug mode?
+use vars qw($ntopchart); #How many entries should make it into the chart?
+use vars qw($gddirectory); #Where to put files from GD::Graph
+
+# SpamAssassin variables
+use vars qw($spam_score $spam_score_gigs);
+use vars qw($ham_score $ham_score_gigs);
+use vars qw(%ham_count_by_ip %spam_count_by_ip);
+use vars qw(%rejected_count_by_ip %rejected_count_by_reason);
+use vars qw(%temporarily_rejected_count_by_ip %temporarily_rejected_count_by_reason);
+
+#For use in Spreadsheet::WriteExcel
+use vars qw($workbook $ws_global $ws_relayed $ws_errors);
+use vars qw($row $col $row_hist $col_hist);
+use vars qw($run_hist);
+use vars qw($f_default $f_header1 $f_header2 $f_header2_m $f_headertab $f_percent); #Format Header
+
+# Output FileHandles
+use vars qw($txt_fh $htm_fh $xls_fh);
+
+$ntopchart = 5;
+
+# The following are parameters whose values are
+# set by command line switches:
+use vars qw($show_errors $show_relay $show_transport $transport_pattern);
+use vars qw($topcount $local_league_table $include_remote_users $do_local_domain);
+use vars qw($hist_opt $hist_interval $hist_number $volume_rounding $emptyOK);
+use vars qw($relay_pattern @queue_times @user_patterns @user_descriptions);
+use vars qw(@rcpt_times @delivery_times);
+use vars qw($include_original_destination);
+use vars qw($txt_fh $htm_fh $xls_fh);
+
+use vars qw(%do_sender); #Do sender by Host, Domain, Email, and/or Edomain tables.
+use vars qw($charts $chartrel $chartdir $charts_option_specified);
+use vars qw($merge_reports); #Merge old reports ?
+
+# The following are modified in the parse() routine, and
+# referred to in the print_*() routines.
+use vars qw($delayed_count $relayed_unshown $begin $end);
+use vars qw(%messages @message);
+use vars qw(%received_count %received_data %received_data_gigs);
+use vars qw(%delivered_messages %delivered_data %delivered_data_gigs %delivered_addresses);
+use vars qw(%received_count_user %received_data_user %received_data_gigs_user);
+use vars qw(%delivered_messages_user %delivered_addresses_user %delivered_data_user %delivered_data_gigs_user);
+use vars qw(%delivered_messages_local_domain %delivered_addresses_local_domain %delivered_data_local_domain %delivered_data_gigs_local_domain);
+use vars qw(%transported_count %transported_data %transported_data_gigs);
+use vars qw(%relayed %errors_count $message_errors);
+use vars qw(@qt_all_bin @qt_remote_bin);
+use vars qw($qt_all_overflow $qt_remote_overflow);
+use vars qw(@dt_all_bin @dt_remote_bin %rcpt_times_bin);
+use vars qw($dt_all_overflow $dt_remote_overflow %rcpt_times_overflow);
+use vars qw(@received_interval_count @delivered_interval_count);
+use vars qw(@user_pattern_totals @user_pattern_interval_count);
+
+use vars qw(%report_totals);
+
+# Enumerations
+use vars qw($SIZE $FROM_HOST $FROM_ADDRESS $ARRIVAL_TIME $REMOTE_DELIVERED $PROTOCOL);
+use vars qw($DELAYED $HAD_ERROR);
+$SIZE = 0;
+$FROM_HOST = 1;
+$FROM_ADDRESS = 2;
+$ARRIVAL_TIME = 3;
+$REMOTE_DELIVERED = 4;
+$DELAYED = 5;
+$HAD_ERROR = 6;
+$PROTOCOL = 7;
+
+
+
+##################################################
+# Subroutines #
+##################################################
+
+#######################################################################
+# get_filehandle($file,\%output_files);
+# Return a filehandle writing to $file.
+#
+# If %output_files is defined, check that $output_files{$file}
+# doesn't exist and die if it does, or set it if it doesn't.
+#######################################################################
+sub get_filehandle {
+ my($file,$output_files_href) = @_;
+
+ $file = '-' if ($file eq '');
+
+ if (defined $output_files_href) {
+ die "You can only output to '$file' once! Use -h for help.\n" if exists $output_files_href->{$file};
+ $output_files_href->{$file} = 1;
+ }
+
+ if ($file eq '-') {
+ return \*STDOUT;
+ }
+
+ if (-e $file) {
+ unlink $file or die "Failed to rm $file: $!";
+ }
+
+ my $fh = new IO::File $file, O_WRONLY|O_CREAT|O_EXCL;
+ die "new IO::File $file failed: $!" unless (defined $fh);
+ return $fh;
+}
+
+
+#######################################################################
+# volume_rounded();
+#
+# $rounded_volume = volume_rounded($bytes,$gigabytes);
+#
+# Given a data size in bytes, round it to KB, MB, or GB
+# as appropriate.
+#
+# Eg 12000 => 12KB, 15000000 => 14GB, etc.
+#
+# Note: I've experimented with Math::BigInt and it results in a 33%
+# performance degredation as opposed to storing numbers split into
+# bytes and gigabytes.
+#######################################################################
+sub volume_rounded {
+ my($x,$g) = @_;
+ $x = 0 unless $x;
+ $g = 0 unless $g;
+ my($rounded);
+
+ while ($x > $gig) {
+ $g++;
+ $x -= $gig;
+ }
+
+ if ($volume_rounding) {
+ # Values < 1 GB
+ if ($g <= 0) {
+ if ($x < 10000) {
+ $rounded = sprintf("%6d", $x);
+ }
+ elsif ($x < 10000000) {
+ $rounded = sprintf("%4dKB", ($x + 512)/1024);
+ }
+ else {
+ $rounded = sprintf("%4dMB", ($x + 512*1024)/(1024*1024));
+ }
+ }
+ # Values between 1GB and 10GB are printed in MB
+ elsif ($g < 10) {
+ $rounded = sprintf("%4dMB", ($g * 1024) + ($x + 512*1024)/(1024*1024));
+ }
+ else {
+ # Handle values over 10GB
+ $rounded = sprintf("%4dGB", $g + ($x + $gig/2)/$gig);
+ }
+ }
+ else {
+ # We don't want any rounding to be done.
+ # and we don't need broken formatted output which on one hand avoids numbers from
+ # being interpreted as string by Spreadsheet Calculators, on the other hand
+ # breaks if more than 4 digits! -> flexible length instead of fixed length
+ # Format the return value at the output routine! -fh
+ #$rounded = sprintf("%d", ($g * $gig) + $x);
+ no integer;
+ $rounded = sprintf("%.0f", ($g * $gig) + $x);
+ }
+
+ return $rounded;
+}
+
+
+#######################################################################
+# un_round();
+#
+# un_round($rounded_volume,\$bytes,\$gigabytes);
+#
+# Given a volume in KB, MB or GB, as generated by volume_rounded(),
+# do the reverse transformation and convert it back into Bytes and Gigabytes.
+# These are added to the $bytes and $gigabytes parameters.
+#
+# Given a data size in bytes, round it to KB, MB, or GB
+# as appropriate.
+#
+# EG: 500 => (500,0), 14GB => (0,14), etc.
+#######################################################################
+sub un_round {
+ my($rounded,$bytes_sref,$gigabytes_sref) = @_;
+
+ if ($rounded =~ /(\d+)GB/) {
+ $$gigabytes_sref += $1;
+ }
+ elsif ($rounded =~ /(\d+)MB/) {
+ $$gigabytes_sref += $1 / 1024;
+ $$bytes_sref += (($1 % 1024 ) * 1024 * 1024);
+ }
+ elsif ($rounded =~ /(\d+)KB/) {
+ $$gigabytes_sref += $1 / (1024 * 1024);
+ $$bytes_sref += ($1 % (1024 * 1024) * 1024);
+ }
+ elsif ($rounded =~ /(\d+)/) {
+ # We need to turn off integer in case we are merging an -nvr report.
+ no integer;
+ $$gigabytes_sref += int($1 / $gig);
+ $$bytes_sref += $1 % $gig;
+ }
+
+ #Now reduce the bytes down to less than 1GB.
+ add_volume($bytes_sref,$gigabytes_sref,0) if ($$bytes_sref > $gig);
+}
+
+
+#######################################################################
+# add_volume();
+#
+# add_volume(\$bytes,\$gigs,$size);
+#
+# Add $size to $bytes/$gigs where this is a number split into
+# bytes ($bytes) and gigabytes ($gigs). This is significantly
+# faster than using Math::BigInt.
+#######################################################################
+sub add_volume {
+ my($bytes_ref,$gigs_ref,$size) = @_;
+ $$bytes_ref = 0 if ! defined $$bytes_ref;
+ $$gigs_ref = 0 if ! defined $$gigs_ref;
+ $$bytes_ref += $size;
+ while ($$bytes_ref > $gig) {
+ $$gigs_ref++;
+ $$bytes_ref -= $gig;
+ }
+}
+
+
+#######################################################################
+# format_time();
+#
+# $formatted_time = format_time($seconds);
+#
+# Given a time in seconds, break it down into
+# weeks, days, hours, minutes, and seconds.
+#
+# Eg 12005 => 3h20m5s
+#######################################################################
+sub format_time {
+my($t) = pop @_;
+my($s) = $t % 60;
+$t /= 60;
+my($m) = $t % 60;
+$t /= 60;
+my($h) = $t % 24;
+$t /= 24;
+my($d) = $t % 7;
+my($w) = $t/7;
+my($p) = "";
+$p .= "$w"."w" if $w > 0;
+$p .= "$d"."d" if $d > 0;
+$p .= "$h"."h" if $h > 0;
+$p .= "$m"."m" if $m > 0;
+$p .= "$s"."s" if $s > 0 || $p eq "";
+$p;
+}
+
+
+#######################################################################
+# unformat_time();
+#
+# $seconds = unformat_time($formatted_time);
+#
+# Given a time in weeks, days, hours, minutes, or seconds, convert it to seconds.
+#
+# Eg 3h20m5s => 12005
+#######################################################################
+sub unformat_time {
+ my($formatted_time) = pop @_;
+ my $time = 0;
+
+ while ($formatted_time =~ s/^(\d+)([wdhms]?)//) {
+ $time += $1 if ($2 eq '' || $2 eq 's');
+ $time += $1 * 60 if ($2 eq 'm');
+ $time += $1 * 60 * 60 if ($2 eq 'h');
+ $time += $1 * 60 * 60 * 24 if ($2 eq 'd');
+ $time += $1 * 60 * 60 * 24 * 7 if ($2 eq 'w');
+ }
+ $time;
+}
+
+
+#######################################################################
+# seconds();
+#
+# $time = seconds($timestamp);
+#
+# Given a time-of-day timestamp, convert it into a time() value using
+# POSIX::mktime. We expect the timestamp to be of the form
+# "$year-$mon-$day $hour:$min:$sec", with month going from 1 to 12,
+# and the year to be absolute (we do the necessary conversions). The
+# seconds value can be followed by decimals, which we ignore. The
+# timestamp may be followed with an offset from UTC like "+$hh$mm"; if the
+# offset is not present, and we have not been told that the log is in UTC
+# (with the -utc option), then we adjust the time by the current local
+# time offset so that it can be compared with the time recorded in message
+# IDs, which is UTC.
+#
+# To improve performance, we only use mktime on the date ($year-$mon-$day),
+# and only calculate it if the date is different to the previous time we
+# came here. We then add on seconds for the '$hour:$min:$sec'.
+#
+# We also store the results of the last conversion done, and only
+# recalculate if the date is different.
+#
+# We used to have the '-cache' flag which would store the results of the
+# mktime() call. However, the current way of just using mktime() on the
+# date obsoletes this.
+#######################################################################
+sub seconds {
+ my($timestamp) = @_;
+
+ # Is the timestamp the same as the last one?
+ return $last_time if ($last_timestamp eq $timestamp);
+
+ return 0 unless ($timestamp =~ /^((\d{4})\-(\d\d)-(\d\d))\s(\d\d):(\d\d):(\d\d)(?:\.\d+)?( ([+-])(\d\d)(\d\d))?/o);
+
+ unless ($last_date eq $1) {
+ $last_date = $1;
+ my(@timestamp) = (0,0,0,$4,$3,$2);
+ $timestamp[5] -= 1900;
+ $timestamp[4]--;
+ $date_seconds = mktime(@timestamp);
+ }
+ my $time = $date_seconds + ($5 * 3600) + ($6 * 60) + $7;
+
+ # SC. Use caching. Also note we want seconds not minutes.
+ #my($this_offset) = ($10 * 60 + $12) * ($9 . "1") if defined $8;
+ if (defined $8 && ($8 ne $last_offset)) {
+ $last_offset = $8;
+ $offset_seconds = ($10 * 60 + $11) * 60;
+ $offset_seconds = -$offset_seconds if ($9 eq '-');
+ }
+
+
+ if (defined $8) {
+ #$time -= $this_offset;
+ $time -= $offset_seconds;
+ } elsif (defined $localtime_offset) {
+ $time -= $localtime_offset;
+ }
+
+ # Store the last timestamp received.
+ $last_timestamp = $timestamp;
+ $last_time = $time;
+
+ $time;
+}
+
+
+#######################################################################
+# id_seconds();
+#
+# $time = id_seconds($message_id);
+#
+# Given a message ID, convert it into a time() value.
+#######################################################################
+sub id_seconds {
+my($sub_id) = substr((pop @_), 0, 6);
+my($s) = 0;
+my(@c) = split(//, $sub_id);
+while($#c >= 0) { $s = $s * 62 + $tab62[ord(shift @c) - ord('0')] }
+$s;
+}
+
+#######################################################################
+# wdhms_seconds();
+#
+# $seconds = wdhms_seconds($string);
+#
+# Convert a string in a week/day/hour/minute/second format (eg 4h10s)
+# into seconds.
+#######################################################################
+sub wdhms_seconds {
+ if ($_[0] =~ /^(?:(\d+)w)?(?:(\d+)d)?(?:(\d+)h)?(?:(\d+)m)?(?:(\d+)s)?/) {
+ return((($1||0) * $WEEK) + (($2||0) * $DAY) + (($3||0) * $HOUR) + (($4||0) * $MINUTE) + ($5||0));
+ }
+ return undef;
+}
+
+#######################################################################
+# queue_time();
+#
+# $queued = queue_time($completed_tod, $arrival_time, $id);
+#
+# Given the completed time of day and either the arrival time
+# (preferred), or the message ID, calculate how long the message has
+# been on the queue.
+#
+#######################################################################
+sub queue_time {
+ my($completed_tod, $arrival_time, $id) = @_;
+
+ # Note: id_seconds() benchmarks as 42% slower than seconds()
+ # and computing the time accounts for a significant portion of
+ # the run time.
+ if (defined $arrival_time) {
+ return(seconds($completed_tod) - seconds($arrival_time));
+ }
+ else {
+ return(seconds($completed_tod) - id_seconds($id));
+ }
+}
+
+
+#######################################################################
+# calculate_localtime_offset();
+#
+# $localtime_offset = calculate_localtime_offset();
+#
+# Calculate the the localtime offset from gmtime in seconds.
+#
+# $localtime = time() + $localtime_offset.
+#
+# These are the same semantics as ISO 8601 and RFC 2822 timezone offsets.
+# (West is negative, East is positive.)
+#######################################################################
+
+# $localtime = gmtime() + $localtime_offset. OLD COMMENT
+# This subroutine commented out as it's not currently in use.
+
+#sub calculate_localtime_offset {
+# # Pick an arbitrary date, convert it to localtime & gmtime, and return the difference.
+# my (@sample_date) = (0,0,0,5,5,100);
+# my $localtime = timelocal(@sample_date);
+# my $gmtime = timegm(@sample_date);
+# my $offset = $localtime - $gmtime;
+# return $offset;
+#}
+
+sub calculate_localtime_offset {
+ # Assume that the offset at the moment is valid across the whole
+ # period covered by the logs that we're analysing. This may not
+ # be true around the time the clocks change in spring or autumn.
+ my $utc = time;
+ # mktime works on local time and gmtime works in UTC
+ my $local = mktime(gmtime($utc));
+ return $local - $utc;
+}
+
+
+
+#######################################################################
+# print_duration_table();
+#
+# print_duration_table($title, $message_type, \@times, \@values, $overflow);
+#
+# Print a table showing how long a particular step took for
+# the messages. The parameters are:
+# $title Eg "Time spent on the queue"
+# $message_type Eg "Remote"
+# \@times The maximum time a message took for it to increment
+# the corresponding @values counter.
+# \@values An array of message counters.
+# $overflow The number of messages which exceeded the maximum
+# time.
+#######################################################################
+sub print_duration_table {
+no integer;
+my($title, $message_type, $times_aref, $values_aref, $overflow) = @_;
+my(@chartdatanames);
+my(@chartdatavals);
+
+my $printed_one = 0;
+my $cumulative_percent = 0;
+
+my $queue_total = $overflow;
+map {$queue_total += $_} @$values_aref;
+
+my $temp = "$title: $message_type";
+
+
+my $txt_format = "%5s %4s %6d %5.1f%% %5.1f%%\n";
+my $htm_format = "<tr><td align=\"right\">%s %s</td><td align=\"right\">%d</td><td align=\"right\">%5.1f%%</td><td align=\"right\">%5.1f%%</td>\n";
+
+# write header
+printf $txt_fh ("%s\n%s\n\n", $temp, "-" x length($temp)) if $txt_fh;
+if ($htm_fh) {
+ print $htm_fh "<hr><a name=\"$title $message_type\"></a><h2>$temp</h2>\n";
+ print $htm_fh "<table border=0 width=\"100%\"><tr><td><table border=1>\n";
+ print $htm_fh "<tr><th>Time</th><th>Messages</th><th>Percentage</th><th>Cumulative Percentage</th>\n";
+}
+if ($xls_fh) {
+ $ws_global->write($row++, $col, "$title: ".$message_type, $f_header2);
+ my @content=("Time", "Messages", "Percentage", "Cumulative Percentage");
+ &set_worksheet_line($ws_global, $row++, 1, \@content, $f_headertab);
+}
+
+
+for ($i = 0; $i <= $#$times_aref; ++$i) {
+ if ($$values_aref[$i] > 0)
+ {
+ my $percent = ($values_aref->[$i] * 100)/$queue_total;
+ $cumulative_percent += $percent;
+
+ my @content=($printed_one? " " : "Under",
+ format_time($times_aref->[$i]),
+ $values_aref->[$i], $percent, $cumulative_percent);
+
+ if ($htm_fh) {
+ printf $htm_fh ($htm_format, @content);
+ if (!defined($values_aref->[$i])) {
+ print $htm_fh "Not defined";
+ }
+ }
+ if ($txt_fh) {
+ printf $txt_fh ($txt_format, @content);
+ if (!defined($times_aref->[$i])) {
+ print $txt_fh "Not defined";
+ }
+ }
+ if ($xls_fh)
+ {
+ no integer;
+ &set_worksheet_line($ws_global, $row, 0, [@content[0,1,2]], $f_default);
+ &set_worksheet_line($ws_global, $row++, 3, [$content[3]/100,$content[4]/100], $f_percent);
+
+ if (!defined($times_aref->[$i])) {
+ $col=0;
+ $ws_global->write($row++, $col, "Not defined" );
+ }
+ }
+
+ push(@chartdatanames,
+ ($printed_one? "" : "Under") . format_time($times_aref->[$i]));
+ push(@chartdatavals, $$values_aref[$i]);
+ $printed_one = 1;
+ }
+}
+
+if ($overflow && $overflow > 0) {
+ my $percent = ($overflow * 100)/$queue_total;
+ $cumulative_percent += $percent;
+
+ my @content = ("Over ", format_time($times_aref->[-1]),
+ $overflow, $percent, $cumulative_percent);
+
+ printf $txt_fh ($txt_format, @content) if $txt_fh;
+ printf $htm_fh ($htm_format, @content) if $htm_fh;
+ if ($xls_fh)
+ {
+ &set_worksheet_line($ws_global, $row, 0, [@content[0,1,2]], $f_default);
+ &set_worksheet_line($ws_global, $row++, 3, [$content[3]/100,$content[4]/100], $f_percent);
+ }
+
+}
+
+push(@chartdatanames, "Over " . format_time($times_aref->[-1]));
+push(@chartdatavals, $overflow);
+
+#printf("Unknown %6d\n", $queue_unknown) if $queue_unknown > 0;
+if ($htm_fh) {
+ print $htm_fh "</table></td><td>";
+
+ if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals > 0)) {
+ my @data = (
+ \@chartdatanames,
+ \@chartdatavals
+ );
+ my $graph = GD::Graph::pie->new(200, 200);
+ my $pngname = "$title-$message_type.png";
+ $pngname =~ s/[^\w\-\.]/_/;
+
+ my $graph_title = "$title ($message_type)";
+ $graph->set(title => $graph_title) if (length($graph_title) < 21);
+
+ my $gd = $graph->plot(\@data) or warn($graph->error);
+ if ($gd) {
+ open(IMG, ">$chartdir/$pngname") or die "Could not write $chartdir/$pngname: $!\n";
+ binmode IMG;
+ print IMG $gd->png;
+ close IMG;
+ print $htm_fh "<img src=\"$chartrel/$pngname\">";
+ }
+ }
+ print $htm_fh "</td></tr></table>\n";
+}
+
+if ($xls_fh)
+{
+ $row++;
+}
+print $txt_fh "\n" if $txt_fh;
+print $htm_fh "\n" if $htm_fh;
+
+}
+
+
+#######################################################################
+# print_histogram();
+#
+# print_histogram('Deliveries|Messages received|$pattern', $unit, @interval_count);
+#
+# Print a histogram of the messages delivered/received per time slot
+# (hour by default).
+#######################################################################
+sub print_histogram {
+my($text, $unit, @interval_count) = @_;
+my(@chartdatanames);
+my(@chartdatavals);
+my($maxd) = 0;
+
+# save first row of print_histogram for xls output
+if (!$run_hist) {
+ $row_hist = $row;
+}
+else {
+ $row = $row_hist;
+}
+
+for ($i = 0; $i < $hist_number; $i++)
+ { $maxd = $interval_count[$i] if $interval_count[$i] > $maxd; }
+
+my $scale = int(($maxd + 25)/50);
+$scale = 1 if $scale == 0;
+
+if ($scale != 1) {
+ if ($unit !~ s/y$/ies/) {
+ $unit .= 's';
+ }
+}
+
+# make and output title
+my $title = sprintf("$text per %s",
+ ($hist_interval == 60)? "hour" :
+ ($hist_interval == 1)? "minute" : "$hist_interval minutes");
+
+my $txt_htm_title = $title . " (each dot is $scale $unit)";
+
+printf $txt_fh ("%s\n%s\n\n", $txt_htm_title, "-" x length($txt_htm_title)) if $txt_fh;
+
+if ($htm_fh) {
+ print $htm_fh "<hr><a name=\"$text\"></a><h2>$txt_htm_title</h2>\n";
+ print $htm_fh "<table border=0 width=\"100%\">\n";
+ print $htm_fh "<tr><td><pre>\n";
+}
+
+if ($xls_fh) {
+ $title =~ s/Messages/Msg/ ;
+ $row += 2;
+ $ws_global->write($row++, $col_hist+1, $title, $f_headertab);
+}
+
+
+my $hour = 0;
+my $minutes = 0;
+for ($i = 0; $i < $hist_number; $i++) {
+ my $c = $interval_count[$i];
+
+ # If the interval is an hour (the maximum) print the starting and
+ # ending hours as a label. Otherwise print the starting hour and
+ # minutes, which take up the same space.
+
+ my $temp;
+ if ($hist_opt == 1) {
+ $temp = sprintf("%02d-%02d", $hour, $hour + 1);
+
+ print $txt_fh $temp if $txt_fh;
+ print $htm_fh $temp if $htm_fh;
+
+ if ($xls_fh) {
+ if ($run_hist==0) {
+ # only on first run
+ $ws_global->write($row, 0, [$temp], $f_default);
+ }
+ }
+
+ push(@chartdatanames, $temp);
+ $hour++;
+ }
+ else {
+ if ($minutes == 0)
+ { $temp = sprintf("%02d:%02d", $hour, $minutes) }
+ else
+ { $temp = sprintf(" :%02d", $minutes) }
+
+ print $txt_fh $temp if $txt_fh;
+ print $htm_fh $temp if $htm_fh;
+ if (($xls_fh) and ($run_hist==0)) {
+ # only on first run
+ $temp = sprintf("%02d:%02d", $hour, $minutes);
+ $ws_global->write($row, 0, [$temp], $f_default);
+ }
+
+ push(@chartdatanames, $temp);
+ $minutes += $hist_interval;
+ if ($minutes >= 60) {
+ $minutes = 0;
+ $hour++;
+ }
+ }
+ push(@chartdatavals, $c);
+
+ printf $txt_fh (" %6d %s\n", $c, "." x ($c/$scale)) if $txt_fh;
+ printf $htm_fh (" %6d %s\n", $c, "." x ($c/$scale)) if $htm_fh;
+ $ws_global->write($row++, $col_hist+1, [$c], $f_default) if $xls_fh;
+
+} #end for
+
+printf $txt_fh "\n" if $txt_fh;
+printf $htm_fh "\n" if $htm_fh;
+
+if ($htm_fh)
+{
+ print $htm_fh "</pre>\n";
+ print $htm_fh "</td><td>\n";
+ if ($HAVE_GD_Graph_linespoints && $charts && ($#chartdatavals > 0)) {
+ # calculate the graph
+ my @data = (
+ \@chartdatanames,
+ \@chartdatavals
+ );
+ my $graph = GD::Graph::linespoints->new(300, 300);
+ $graph->set(
+ x_label => 'Time',
+ y_label => 'Amount',
+ title => $text,
+ x_labels_vertical => 1
+ );
+ my $pngname = "histogram_$text.png";
+ $pngname =~ s/[^\w\._]/_/g;
+
+ my $gd = $graph->plot(\@data) or warn($graph->error);
+ if ($gd) {
+ open(IMG, ">$chartdir/$pngname") or die "Could not write $chartdir/$pngname: $!\n";
+ binmode IMG;
+ print IMG $gd->png;
+ close IMG;
+ print $htm_fh "<img src=\"$chartrel/$pngname\">";
+ }
+ }
+ print $htm_fh "</td></tr></table>\n";
+}
+
+$col_hist++; # where to continue next times
+
+$row+=2; # leave some space after history block
+$run_hist=1; # we have done this once or more
+}
+
+
+
+#######################################################################
+# print_league_table();
+#
+# print_league_table($league_table_type,\%message_count,\%address_count,\%message_data,\%message_data_gigs, $spreadsheet, $row_sref);
+#
+# Given hashes of message count, address count, and message data,
+# which are keyed by the table type (eg by the sending host), print a
+# league table showing the top $topcount (defaults to 50).
+#######################################################################
+sub print_league_table {
+ my($text,$m_count,$a_count,$m_data,$m_data_gigs,$spreadsheet, $row_sref) = @_;
+ my($name) = ($topcount == 1)? "$text" : "$topcount ${text}s";
+ my($title) = "Top $name by message count";
+ my(@chartdatanames) = ();
+ my(@chartdatavals) = ();
+ my $chartotherval = 0;
+ $text = ucfirst($text);
+
+ # Align non-local addresses to the right (so all the .com's line up).
+ # Local addresses are aligned on the left as they are userids.
+ my $align = ($text !~ /local/i) ? 'right' : 'left';
+
+
+ ################################################
+ # Generate the printf formats and table headers.
+ ################################################
+ my(@headers) = ('Messages');
+ #push(@headers,'Addresses') if defined $a_count;
+ push(@headers,'Addresses') if defined $a_count && %$a_count;
+ push(@headers,'Bytes','Average') if defined $m_data;
+
+ my $txt_format = "%10s " x @headers . " %s\n";
+ my $txt_col_headers = sprintf $txt_format, @headers, $text;
+ my $htm_format = "<tr>" . '<td align="right">%s</td>'x@headers . "<td align=\"$align\" nowrap>%s</td></tr>\n";
+ my $htm_col_headers = sprintf $htm_format, @headers, $text;
+ $htm_col_headers =~ s/(<\/?)td/$1th/g; #Convert <td>'s to <th>'s for the header.
+
+
+ ################################################
+ # Write the table headers
+ ################################################
+ printf $txt_fh ("%s\n%s\n%s", $title, "-" x length($title),$txt_col_headers) if $txt_fh;
+
+ if ($htm_fh) {
+ print $htm_fh <<EoText;
+<hr><a name="$text count"></a><h2>$title</h2>
+<table border=0 width="100%">
+<tr><td>
+<table border=1>
+EoText
+ print $htm_fh $htm_col_headers
+ }
+
+ if ($xls_fh) {
+ $spreadsheet->write(${$row_sref}++, 0, $title, $f_header2);
+ $spreadsheet->write(${$row_sref}++, 0, [@headers, $text], $f_headertab);
+ }
+
+
+ # write content
+ foreach my $key (top_n_sort($topcount,$m_count,$m_data_gigs,$m_data)) {
+
+ # When displaying the average figures, we calculate the average of
+ # the rounded data, as the user would calculate it. This reduces
+ # the accuracy slightly, but we have to do it this way otherwise
+ # when using -merge to convert results from text to HTML and
+ # vice-versa discrepencies would occur.
+ my $messages = $$m_count{$key};
+ my @content = ($messages);
+ push(@content, $$a_count{$key}) if defined $a_count;
+ if (defined $m_data) {
+ my $rounded_volume = volume_rounded($$m_data{$key},$$m_data_gigs{$key});
+ my($data,$gigs) = (0,0);
+ un_round($rounded_volume,\$data,\$gigs);
+ my $rounded_average = volume_rounded($data/$messages,$gigs/$messages);
+ push(@content, $rounded_volume, $rounded_average);
+ }
+
+ # write content
+ printf $txt_fh ($txt_format, @content, $key) if $txt_fh;
+
+ if ($htm_fh) {
+ my $htmlkey = $key;
+ $htmlkey =~ s/>/\&gt\;/g;
+ $htmlkey =~ s/</\&lt\;/g;
+ printf $htm_fh ($htm_format, @content, $htmlkey);
+ }
+ $spreadsheet->write(${$row_sref}++, 0, [@content, $key], $f_default) if $xls_fh;
+
+ if (scalar @chartdatanames < $ntopchart) {
+ push(@chartdatanames, $key);
+ push(@chartdatavals, $$m_count{$key});
+ }
+ else {
+ $chartotherval += $$m_count{$key};
+ }
+ }
+
+ push(@chartdatanames, "Other");
+ push(@chartdatavals, $chartotherval);
+
+ print $txt_fh "\n" if $txt_fh;
+ if ($htm_fh) {
+ print $htm_fh "</table>\n";
+ print $htm_fh "</td><td>\n";
+ if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals > 0))
+ {
+ # calculate the graph
+ my @data = (
+ \@chartdatanames,
+ \@chartdatavals
+ );
+ my $graph = GD::Graph::pie->new(300, 300);
+ $graph->set(
+ x_label => 'Name',
+ y_label => 'Amount',
+ title => 'By count',
+ );
+ my $gd = $graph->plot(\@data) or warn($graph->error);
+ if ($gd) {
+ my $temp = $text;
+ $temp =~ s/ /_/g;
+ open(IMG, ">$chartdir/${temp}_count.png") or die "Could not write $chartdir/${temp}_count.png: $!\n";
+ binmode IMG;
+ print IMG $gd->png;
+ close IMG;
+ print $htm_fh "<img src=\"$chartrel/${temp}_count.png\">";
+ }
+ }
+ print $htm_fh "</td><td>\n";
+ print $htm_fh "</td></tr></table>\n\n";
+ }
+ ++${$row_sref} if $xls_fh;
+
+
+ if (defined $m_data) {
+ # write header
+
+ $title = "Top $name by volume";
+
+ printf $txt_fh ("%s\n%s\n%s", $title, "-" x length($title),$txt_col_headers) if $txt_fh;
+
+ if ($htm_fh) {
+ print $htm_fh <<EoText;
+<hr><a name="$text volume"></a><h2>$title</h2>
+<table border=0 width="100%">
+<tr><td>
+<table border=1>
+EoText
+ print $htm_fh $htm_col_headers;
+ }
+ if ($xls_fh) {
+ $spreadsheet->write(${$row_sref}++, 0, $title, $f_header2);
+ $spreadsheet->write(${$row_sref}++, 0, [@headers, $text], $f_headertab);
+ }
+
+ @chartdatanames = ();
+ @chartdatavals = ();
+ $chartotherval = 0;
+ my $use_gig = 0;
+ foreach my $key (top_n_sort($topcount,$m_data_gigs,$m_data,$m_count)) {
+ # The largest volume will be the first (top of the list).
+ # If it has at least 1 gig, then just use gigabytes to avoid
+ # risking an integer overflow when generating the pie charts.
+ if ($$m_data_gigs{$key}) {
+ $use_gig = 1;
+ }
+
+ my $messages = $$m_count{$key};
+ my @content = ($messages);
+ push(@content, $$a_count{$key}) if defined $a_count;
+ my $rounded_volume = volume_rounded($$m_data{$key},$$m_data_gigs{$key});
+ my($data ,$gigs) = (0,0);
+ un_round($rounded_volume,\$data,\$gigs);
+ my $rounded_average = volume_rounded($data/$messages,$gigs/$messages);
+ push(@content, $rounded_volume, $rounded_average );
+
+ # write content
+ printf $txt_fh ($txt_format, @content, $key) if $txt_fh;
+ if ($htm_fh) {
+ my $htmlkey = $key;
+ $htmlkey =~ s/>/\&gt\;/g;
+ $htmlkey =~ s/</\&lt\;/g;
+ printf $htm_fh ($htm_format, @content, $htmlkey);
+ }
+ $spreadsheet->write(${$row_sref}++, 0, [@content, $key], $f_default) if $xls_fh;
+
+
+ if (scalar @chartdatanames < $ntopchart) {
+ if ($use_gig) {
+ if ($$m_data_gigs{$key}) {
+ push(@chartdatanames, $key);
+ push(@chartdatavals, $$m_data_gigs{$key});
+ }
+ }
+ else {
+ push(@chartdatanames, $key);
+ push(@chartdatavals, $$m_data{$key});
+ }
+ }
+ else {
+ $chartotherval += ($use_gig) ? $$m_data_gigs{$key} : $$m_data{$key};
+ }
+ }
+ push(@chartdatanames, "Other");
+ push(@chartdatavals, $chartotherval);
+
+ print $txt_fh "\n" if $txt_fh;
+ if ($htm_fh) {
+ print $htm_fh "</table>\n";
+ print $htm_fh "</td><td>\n";
+ if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals > 0)) {
+ # calculate the graph
+ my @data = (
+ \@chartdatanames,
+ \@chartdatavals
+ );
+ my $graph = GD::Graph::pie->new(300, 300);
+ $graph->set(
+ x_label => 'Name',
+ y_label => 'Volume' ,
+ title => 'By Volume',
+ );
+ my $gd = $graph->plot(\@data) or warn($graph->error);
+ if ($gd) {
+ my $temp = $text;
+ $temp =~ s/ /_/g;
+ open(IMG, ">$chartdir/${temp}_volume.png") or die "Could not write $chartdir/${temp}_volume.png: $!\n";
+ binmode IMG;
+ print IMG $gd->png;
+ close IMG;
+ print $htm_fh "<img src=\"$chartrel/${temp}_volume.png\">";
+ }
+ }
+ print $htm_fh "</td><td>\n";
+ print $htm_fh "</td></tr></table>\n\n";
+ }
+
+ ++${$row_sref} if $xls_fh;
+ }
+}
+
+
+#######################################################################
+# top_n_sort();
+#
+# @sorted_keys = top_n_sort($n,$href1,$href2,$href3);
+#
+# Given a hash which has numerical values, return the sorted $n keys which
+# point to the top values. The second and third hashes are used as
+# tiebreakers. They all must have the same keys.
+#
+# The idea behind this routine is that when you only want to see the
+# top n members of a set, rather than sorting the entire set and then
+# plucking off the top n, sort through the stack as you go, discarding
+# any member which is lower than your current n'th highest member.
+#
+# This proves to be an order of magnitude faster for large hashes.
+# On 200,000 lines of mainlog it benchmarked 9 times faster.
+# On 700,000 lines of mainlog it benchmarked 13.8 times faster.
+#
+# We assume the values are > 0.
+#######################################################################
+sub top_n_sort {
+ my($n,$href1,$href2,$href3) = @_;
+
+ # PH's original sort was:
+ #
+ # foreach $key (sort
+ # {
+ # $$m_count{$b} <=> $$m_count{$a} ||
+ # $$m_data_gigs{$b} <=> $$m_data_gigs{$a} ||
+ # $$m_data{$b} <=> $$m_data{$a} ||
+ # $a cmp $b
+ # }
+ # keys %{$m_count})
+ #
+
+ #We use a key of '_' to represent non-existant values, as null keys are valid.
+ #'_' is not a valid domain, edomain, host, or email.
+ my(@top_n_keys) = ('_') x $n;
+ my($minimum_value1,$minimum_value2,$minimum_value3) = (0,0,0);
+ my $top_n_key = '';
+ my $n_minus_1 = $n - 1;
+ my $n_minus_2 = $n - 2;
+
+ # Create a dummy hash incase the user has not provided us with
+ # tiebreaker hashes.
+ my(%dummy_hash);
+ $href2 = \%dummy_hash unless defined $href2;
+ $href3 = \%dummy_hash unless defined $href3;
+
+ # Pick out the top $n keys.
+ my($key,$value1,$value2,$value3,$i,$comparison,$insert_position);
+ while (($key,$value1) = each %$href1) {
+
+ #print STDERR "key $key ($value1,",$href2->{$key},",",$href3->{$key},") <=> ($minimum_value1,$minimum_value2,$minimum_value3)\n";
+
+ # Check to see that the new value is bigger than the lowest of the
+ # top n keys that we're keeping. We test the main key first, because
+ # for the majority of cases we can skip creating dummy hash values
+ # should the user have not provided real tie-breaking hashes.
+ next unless $value1 >= $minimum_value1;
+
+ # Create a dummy hash entry for the key if required.
+ # Note that setting the dummy_hash value sets it for both href2 &
+ # href3. Also note that currently we are guaranteed to have a real
+ # value for href3 if a real value for href2 exists so don't need to
+ # test for it as well.
+ $dummy_hash{$key} = 0 unless exists $href2->{$key};
+
+ $comparison = $value1 <=> $minimum_value1 ||
+ $href2->{$key} <=> $minimum_value2 ||
+ $href3->{$key} <=> $minimum_value3 ||
+ $top_n_key cmp $key;
+ next unless ($comparison == 1);
+
+ # As we will be using these values a few times, extract them into scalars.
+ $value2 = $href2->{$key};
+ $value3 = $href3->{$key};
+
+ # This key is bigger than the bottom n key, so the lowest position we
+ # will insert it into is $n minus 1 (the bottom of the list).
+ $insert_position = $n_minus_1;
+
+ # Now go through the list, stopping when we find a key that we're
+ # bigger than, or we come to the penultimate position - we've
+ # already tested bigger than the last.
+ #
+ # Note: we go top down as the list starts off empty.
+ # Note: stepping through the list in this way benchmarks nearly
+ # three times faster than doing a sort() on the reduced list.
+ # I assume this is because the list is already in order, and
+ # we get a performance boost from not having to do hash lookups
+ # on the new key.
+ for ($i = 0; $i < $n_minus_1; $i++) {
+ $top_n_key = $top_n_keys[$i];
+ if ( ($top_n_key eq '_') ||
+ ( ($value1 <=> $href1->{$top_n_key} ||
+ $value2 <=> $href2->{$top_n_key} ||
+ $value3 <=> $href3->{$top_n_key} ||
+ $top_n_key cmp $key) == 1
+ )
+ ) {
+ $insert_position = $i;
+ last;
+ }
+ }
+
+ # Remove the last element, then insert the new one.
+ $#top_n_keys = $n_minus_2;
+ splice(@top_n_keys,$insert_position,0,$key);
+
+ # Extract our new minimum values.
+ $top_n_key = $top_n_keys[$n_minus_1];
+ if ($top_n_key ne '_') {
+ $minimum_value1 = $href1->{$top_n_key};
+ $minimum_value2 = $href2->{$top_n_key};
+ $minimum_value3 = $href3->{$top_n_key};
+ }
+ }
+
+ # Return the top n list, grepping out non-existant values, just in case
+ # we didn't have that many values.
+ return(grep(!/^_$/,@top_n_keys));
+}
+
+
+
+#######################################################################
+# html_header();
+#
+# $header = html_header($title);
+#
+# Print our HTML header and start the <body> block.
+#######################################################################
+sub html_header {
+ my($title) = @_;
+ my $text = << "EoText";
+<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">
+<html>
+<head>
+<meta http-equiv="Content-Type" content="text/html; charset=iso-8859-15">
+<title>$title</title>
+</head>
+<body bgcolor="white">
+<h1>$title</h1>
+EoText
+ return $text;
+}
+
+
+
+#######################################################################
+# help();
+#
+# help();
+#
+# Display usage instructions and exit.
+#######################################################################
+sub help {
+ print << "EoText";
+
+eximstats Version $VERSION
+
+Usage:
+ eximstats [Output] [Options] mainlog1 mainlog2 ...
+ eximstats -merge -html [Options] report.1.html ... > weekly_rep.html
+
+Examples:
+ eximstats -html=eximstats.html mainlog1 mainlog2 ...
+ eximstats mainlog1 mainlog2 ... > report.txt
+
+Parses exim mainlog or syslog files and generates a statistical analysis
+of the messages processed.
+
+Valid output types are:
+-txt[=<file>] plain text (default unless no other type is specified)
+-html[=<file>] HTML
+-xls[=<file>] Excel
+With no type and file given, defaults to -txt and STDOUT.
+
+Valid options are:
+-h<number> histogram divisions per hour. The default is 1, and
+ 0 suppresses histograms. Other valid values are:
+ 2, 3, 5, 10, 15, 20, 30 or 60.
+-ne don't display error information
+-nr don't display relaying information
+-nr/pattern/ don't display relaying information that matches
+-nt don't display transport information
+-nt/pattern/ don't display transport information that matches
+-nvr don't do volume rounding. Display in bytes, not KB/MB/GB.
+-t<number> display top <number> sources/destinations
+ default is 50, 0 suppresses top listing
+-tnl omit local sources/destinations in top listing
+-t_remote_users show top user sources/destinations from non-local domains
+-q<list> list of times for queuing information. -q0 suppresses.
+-show_rt<list> Show the receipt times for all the messages.
+-show_dt<list> Show the delivery times for all the messages.
+ <list> is an optional list of times in seconds.
+ Eg -show_rt1,2,4,8.
+
+-include_original_destination show both the final and original
+ destinations in the results rather than just the final ones.
+
+-byhost show results by sending host (default unless bydomain or
+ byemail is specified)
+-bydomain show results by sending domain.
+-byemail show results by sender's email address
+-byedomain show results by sender's email domain
+-bylocaldomain show results by local domain
+
+-pattern "Description" /pattern/
+ Count lines matching specified patterns and show them in
+ the results. It can be specified multiple times. Eg:
+ -pattern 'Refused connections' '/refused connection/'
+
+-merge merge previously generated reports into a new report
+
+-charts Create charts (this requires the GD::Graph modules).
+ Only valid with -html.
+-chartdir <dir> Create the charts' png files in the directory <dir>
+-chartrel <dir> Specify the relative directory for the "img src=" tags
+ from where to include the charts in the html file
+ -chartdir and -chartrel default to '.'
+
+-emptyok It is OK if there is no valid input, don't print an error.
+
+-d Debug mode - dump the eval'ed parser onto STDERR.
+
+EoText
+
+ exit 1;
+}
+
+
+
+#######################################################################
+# generate_parser();
+#
+# $parser = generate_parser();
+#
+# This subroutine generates the parsing routine which will be
+# used to parse the mainlog. We take the base operation, and remove bits not in use.
+# This improves performance depending on what bits you take out or add.
+#
+# I've tested using study(), but this does not improve performance.
+#
+# We store our parsing routing in a variable, and process it looking for #IFDEF (Expression)
+# or #IFNDEF (Expression) statements and corresponding #ENDIF (Expression) statements. If
+# the expression evaluates to true, then it is included/excluded accordingly.
+#######################################################################
+sub generate_parser {
+ my $parser = '
+ my($ip,$host,$email,$edomain,$domain,$thissize,$size,$old,$new);
+ my($tod,$m_hour,$m_min,$id,$flag,$extra,$length);
+ my($seconds,$queued,$rcpt_time,$local_domain);
+ my $rej_id = 0;
+ while (<$fh>) {
+
+ # Convert syslog lines to mainlog format.
+ if (! /^\\d{4}/) {
+ next unless s/^.*? exim\\b.*?: //;
+ }
+
+ $length = length($_);
+ next if ($length < 38);
+ next unless /^
+ (\\d{4}\\-\\d\\d-\\d\\d\\s # 1: YYYYMMDD HHMMSS
+ (\\d\\d) # 2: HH
+ :
+ (\\d\\d) # 3: MM
+ :\\d\\d
+ )
+ (\\.\\d+)? # 4: subseconds
+ (\s[-+]\\d\\d\\d\\d)? # 5: tz-offset
+ (\s\\[\\d+\\])? # 6: pid
+ /ox;
+
+ $tod = defined($5) ? $1 . $5 : $1;
+ ($m_hour,$m_min) = ($2,$3);
+
+ # PH - watch for GMT offsets in the timestamp.
+ if (defined($5)) {
+ $extra = 6;
+ next if ($length < 44);
+ }
+ else {
+ $extra = 0;
+ }
+
+ # watch for subsecond precision
+ if (defined($4)) {
+ $extra += length($4);
+ next if ($length < 38 + $extra);
+ }
+
+ # PH - watch for PID added after the timestamp.
+ if (defined($6)) {
+ $extra += length($6);
+ next if ($length < 38 + $extra);
+ }
+
+ $id = substr($_, 20 + $extra, 16);
+ $flag = substr($_, 37 + $extra, 2);
+
+ if ($flag !~ /^([<>=*-]+|SA)$/ && /rejected|refused|dropped/) {
+ $flag = "Re";
+ $extra -= 3;
+ }
+
+ # Rejects can have no MSGID...
+ if ($flag eq "Re" && $id !~ /^[-0-9a-zA-Z]+$/) {
+ $id = "reject:" . ++$rej_id;
+ $extra -= 17;
+ }
+';
+
+ # Watch for user specified patterns.
+ my $user_pattern_index = 0;
+ foreach (@user_patterns) {
+ $user_pattern_totals[$user_pattern_index] = 0;
+ $parser .= " if ($_) {\n";
+ $parser .= " \$user_pattern_totals[$user_pattern_index]++;\n";
+ $parser .= " \$user_pattern_interval_count[$user_pattern_index][(\$m_hour*60 + \$m_min)/$hist_interval]++;\n" if ($hist_opt > 0);
+ $parser .= " }\n";
+ $user_pattern_index++;
+ }
+
+ $parser .= '
+ next unless ($flag =~ /<=|=>|->|==|\\*\\*|Co|SA|Re/);
+
+ #Strip away the timestamp, ID and flag to speed up later pattern matches.
+ #The flags include Co (Completed), Re (Rejected), and SA (SpamAssassin).
+ $_ = substr($_, 40 + $extra); # PH
+
+ # Alias @message to the array of information about the message.
+ # This minimises the number of calls to hash functions.
+ $messages{$id} = [] unless exists $messages{$id};
+ *message = $messages{$id};
+
+
+ # JN - Skip over certain transports as specified via the "-nt/.../" command
+ # line switch (where ... is a perl style regular expression). This is
+ # required so that transports that skew stats such as SpamAssassin can be
+ # ignored.
+ #IFDEF ($transport_pattern)
+ if (/\\sT=(\\S+)/) {
+ next if ($1 =~ /$transport_pattern/o) ;
+ }
+ #ENDIF ($transport_pattern)
+
+
+
+ # Do some pattern matches to get the host and IP address.
+ # We expect lines to be of the form "H=[IpAddr]" or "H=Host [IpAddr]" or
+ # "H=Host (UnverifiedHost) [IpAddr]" or "H=(UnverifiedHost) [IpAddr]".
+ # We do 2 separate matches to keep the matches simple and fast.
+ # Host is local unless otherwise specified.
+ # Watch out for "H=([IpAddr])" in case they send "[IpAddr]" as their HELO!
+ $ip = (/\\bH=(?:|.*? )(\\[[^]]+\\])/) ? $1
+ # 2008-03-31 06:25:22 Connection from [213.246.33.217]:39456 refused: too many connections from that IP address // .hs
+ : (/Connection from (\[\S+\])/) ? $1
+ # 2008-03-31 06:52:40 SMTP call from mail.cacoshrf.com (ccsd02.ccsd.local) [69.24.118.229]:4511 dropped: too many nonmail commands (last was "RSET") // .hs
+ : (/SMTP call from .*?(\[\S+\])/) ? $1
+ : "local";
+ $host = (/\\bH=(\\S+)/) ? $1 : "local";
+
+ $domain = "localdomain"; #Domain is localdomain unless otherwise specified.
+
+ #IFDEF ($do_sender{Domain})
+ if ($host =~ /^\\[/ || $host =~ /^[\\d\\.]+$/) {
+ # Host is just an IP address.
+ $domain = $host;
+ }
+ elsif ($host =~ /^(\\(?)[^\\.]+\\.([^\\.]+\\..*)/) {
+ # Remove the host portion from the DNS name. We ensure that we end up
+ # with at least xxx.yyy. $host can be "(x.y.z)" or "x.y.z".
+ $domain = lc("$1.$2");
+ $domain =~ s/^\\.//; #Remove preceding dot.
+ }
+ #ENDIF ($do_sender{Domain})
+
+ #IFDEF ($do_sender{Email})
+ #IFDEF ($include_original_destination)
+ # Catch both "a@b.com <c@d.com>" and "e@f.com"
+ #$email = (/^(\S+) (<(\S*?)>)?/) ? $3 || $1 : "";
+ $email = (/^(\S+ (<[^@>]+@?[^>]*>)?)/) ? $1 : "";
+ chomp($email);
+ #ENDIF ($include_original_destination)
+
+ #IFNDEF ($include_original_destination)
+ $email = (/^(\S+)/) ? $1 : "";
+ #ENDIF ($include_original_destination)
+ #ENDIF ($do_sender{Email})
+
+ #IFDEF ($do_sender{Edomain})
+ if (/^(<>|blackhole)/) {
+ $edomain = $1;
+ }
+ #IFDEF ($include_original_destination)
+ elsif (/^(\S+ (<\S*?\\@(\S+?)>)?)/) {
+ $edomain = $1;
+ chomp($edomain);
+ $edomain =~ s/@(\S+?)>/"@" . lc($1) . ">"/e;
+ }
+ #ENDIF ($include_original_destination)
+ #IFNDEF ($include_original_destination)
+ elsif (/^\S*?\\@(\S+)/) {
+ $edomain = lc($1);
+ }
+ #ENDIF ($include_original_destination)
+ else {
+ $edomain = "";
+ }
+
+ #ENDIF ($do_sender{Edomain})
+
+ if ($tod lt $begin) {
+ $begin = $tod;
+ }
+ elsif ($tod gt $end) {
+ $end = $tod;
+ }
+
+
+ if ($flag eq "<=") {
+ $thissize = (/\\sS=(\\d+)( |$)/) ? $1 : 0;
+ $message[$SIZE] = $thissize;
+ $message[$PROTOCOL] = (/ P=(\S+)/) ? $1 : undef;
+
+ #IFDEF ($show_relay)
+ if ($host ne "local") {
+ # Save incoming information in case it becomes interesting
+ # later, when delivery lines are read.
+ my($from) = /^(\\S+)/;
+ $message[$FROM_HOST] = "$host$ip";
+ $message[$FROM_ADDRESS] = $from;
+ }
+ #ENDIF ($show_relay)
+
+ #IFDEF ($local_league_table || $include_remote_users)
+ if (/\sU=(\\S+)/) {
+ my $user = $1;
+
+ #IFDEF ($local_league_table && $include_remote_users)
+ { #Store both local and remote users.
+ #ENDIF ($local_league_table && $include_remote_users)
+
+ #IFDEF ($local_league_table && ! $include_remote_users)
+ if ($host eq "local") { #Store local users only.
+ #ENDIF ($local_league_table && ! $include_remote_users)
+
+ #IFDEF ($include_remote_users && ! $local_league_table)
+ if ($host ne "local") { #Store remote users only.
+ #ENDIF ($include_remote_users && ! $local_league_table)
+
+ ++$received_count_user{$user};
+ add_volume(\\$received_data_user{$user},\\$received_data_gigs_user{$user},$thissize);
+ }
+ }
+ #ENDIF ($local_league_table || $include_remote_users)
+
+ #IFDEF ($do_sender{Host})
+ ++$received_count{Host}{$host};
+ add_volume(\\$received_data{Host}{$host},\\$received_data_gigs{Host}{$host},$thissize);
+ #ENDIF ($do_sender{Host})
+
+ #IFDEF ($do_sender{Domain})
+ if ($domain) {
+ ++$received_count{Domain}{$domain};
+ add_volume(\\$received_data{Domain}{$domain},\\$received_data_gigs{Domain}{$domain},$thissize);
+ }
+ #ENDIF ($do_sender{Domain})
+
+ #IFDEF ($do_sender{Email})
+ ++$received_count{Email}{$email};
+ add_volume(\\$received_data{Email}{$email},\\$received_data_gigs{Email}{$email},$thissize);
+ #ENDIF ($do_sender{Email})
+
+ #IFDEF ($do_sender{Edomain})
+ ++$received_count{Edomain}{$edomain};
+ add_volume(\\$received_data{Edomain}{$edomain},\\$received_data_gigs{Edomain}{$edomain},$thissize);
+ #ENDIF ($do_sender{Edomain})
+
+ ++$total_received_count;
+ add_volume(\\$total_received_data,\\$total_received_data_gigs,$thissize);
+
+ #IFDEF ($#queue_times >= 0 || $#rcpt_times >= 0)
+ $message[$ARRIVAL_TIME] = $tod;
+ #ENDIF ($#queue_times >= 0 || $#rcpt_times >= 0)
+
+ #IFDEF ($hist_opt > 0)
+ $received_interval_count[($m_hour*60 + $m_min)/$hist_interval]++;
+ #ENDIF ($hist_opt > 0)
+ }
+
+ elsif ($flag eq "=>") {
+ $size = $message[$SIZE] || 0;
+ if ($host ne "local") {
+ $message[$REMOTE_DELIVERED] = 1;
+
+
+ #IFDEF ($show_relay)
+ # Determine relaying address if either only one address listed,
+ # or two the same. If they are different, it implies a forwarding
+ # or aliasing, which is not relaying. Note that for multi-aliased
+ # addresses, there may be a further address between the first
+ # and last.
+
+ if (defined $message[$FROM_HOST]) {
+ if (/^(\\S+)(?:\\s+\\([^)]\\))?\\s+<([^>]+)>/) {
+ ($old,$new) = ($1,$2);
+ }
+ else {
+ $old = $new = "";
+ }
+
+ if ("\\L$new" eq "\\L$old") {
+ ($old) = /^(\\S+)/ if $old eq "";
+ my $key = "H=\\L$message[$FROM_HOST]\\E A=\\L$message[$FROM_ADDRESS]\\E => " .
+ "H=\\L$host\\E$ip A=\\L$old\\E";
+ if (!defined $relay_pattern || $key !~ /$relay_pattern/o) {
+ $relayed{$key} = 0 if !defined $relayed{$key};
+ ++$relayed{$key};
+ }
+ else {
+ ++$relayed_unshown;
+ }
+ }
+ }
+ #ENDIF ($show_relay)
+
+ }
+
+ #IFDEF ($local_league_table || $include_remote_users)
+ #IFDEF ($local_league_table && $include_remote_users)
+ { #Store both local and remote users.
+ #ENDIF ($local_league_table && $include_remote_users)
+
+ #IFDEF ($local_league_table && ! $include_remote_users)
+ if ($host eq "local") { #Store local users only.
+ #ENDIF ($local_league_table && ! $include_remote_users)
+
+ #IFDEF ($include_remote_users && ! $local_league_table)
+ if ($host ne "local") { #Store remote users only.
+ #ENDIF ($include_remote_users && ! $local_league_table)
+
+ if (my($user) = split((/\\s</)? " <" : " ", $_)) {
+ #IFDEF ($include_original_destination)
+ {
+ #ENDIF ($include_original_destination)
+ #IFNDEF ($include_original_destination)
+ if ($user =~ /^[\\/|]/) {
+ #ENDIF ($include_original_destination)
+ #my($parent) = $_ =~ /(<[^@]+@?[^>]*>)/;
+ my($parent) = $_ =~ / (<.+?>) /; #DT 1.54
+ if (defined $parent) {
+ $user = "$user $parent";
+ #IFDEF ($do_local_domain)
+ if ($parent =~ /\\@(.+)>/) {
+ $local_domain = lc($1);
+ ++$delivered_messages_local_domain{$local_domain};
+ ++$delivered_addresses_local_domain{$local_domain};
+ add_volume(\\$delivered_data_local_domain{$local_domain},\\$delivered_data_gigs_local_domain{$local_domain},$size);
+ }
+ #ENDIF ($do_local_domain)
+ }
+ }
+ ++$delivered_messages_user{$user};
+ ++$delivered_addresses_user{$user};
+ add_volume(\\$delivered_data_user{$user},\\$delivered_data_gigs_user{$user},$size);
+ }
+ }
+ #ENDIF ($local_league_table || $include_remote_users)
+
+ #IFDEF ($do_sender{Host})
+ $delivered_messages{Host}{$host}++;
+ $delivered_addresses{Host}{$host}++;
+ add_volume(\\$delivered_data{Host}{$host},\\$delivered_data_gigs{Host}{$host},$size);
+ #ENDIF ($do_sender{Host})
+ #IFDEF ($do_sender{Domain})
+ if ($domain) {
+ ++$delivered_messages{Domain}{$domain};
+ ++$delivered_addresses{Domain}{$domain};
+ add_volume(\\$delivered_data{Domain}{$domain},\\$delivered_data_gigs{Domain}{$domain},$size);
+ }
+ #ENDIF ($do_sender{Domain})
+ #IFDEF ($do_sender{Email})
+ ++$delivered_messages{Email}{$email};
+ ++$delivered_addresses{Email}{$email};
+ add_volume(\\$delivered_data{Email}{$email},\\$delivered_data_gigs{Email}{$email},$size);
+ #ENDIF ($do_sender{Email})
+ #IFDEF ($do_sender{Edomain})
+ ++$delivered_messages{Edomain}{$edomain};
+ ++$delivered_addresses{Edomain}{$edomain};
+ add_volume(\\$delivered_data{Edomain}{$edomain},\\$delivered_data_gigs{Edomain}{$edomain},$size);
+ #ENDIF ($do_sender{Edomain})
+
+ ++$total_delivered_messages;
+ ++$total_delivered_addresses;
+ add_volume(\\$total_delivered_data,\\$total_delivered_data_gigs,$size);
+
+ #IFDEF ($show_transport)
+ my $transport = (/\\sT=(\\S+)/) ? $1 : ":blackhole:";
+ ++$transported_count{$transport};
+ add_volume(\\$transported_data{$transport},\\$transported_data_gigs{$transport},$size);
+ #ENDIF ($show_transport)
+
+ #IFDEF ($hist_opt > 0)
+ $delivered_interval_count[($m_hour*60 + $m_min)/$hist_interval]++;
+ #ENDIF ($hist_opt > 0)
+
+ #IFDEF ($#delivery_times > 0)
+ if (/ DT=(\S+)/) {
+ $seconds = wdhms_seconds($1);
+ for ($i = 0; $i <= $#delivery_times; $i++) {
+ if ($seconds < $delivery_times[$i]) {
+ ++$dt_all_bin[$i];
+ ++$dt_remote_bin[$i] if $message[$REMOTE_DELIVERED];
+ last;
+ }
+ }
+ if ($i > $#delivery_times) {
+ ++$dt_all_overflow;
+ ++$dt_remote_overflow if $message[$REMOTE_DELIVERED];
+ }
+ }
+ #ENDIF ($#delivery_times > 0)
+
+ }
+
+ elsif ($flag eq "->") {
+
+ #IFDEF ($local_league_table || $include_remote_users)
+ #IFDEF ($local_league_table && $include_remote_users)
+ { #Store both local and remote users.
+ #ENDIF ($local_league_table && $include_remote_users)
+
+ #IFDEF ($local_league_table && ! $include_remote_users)
+ if ($host eq "local") { #Store local users only.
+ #ENDIF ($local_league_table && ! $include_remote_users)
+
+ #IFDEF ($include_remote_users && ! $local_league_table)
+ if ($host ne "local") { #Store remote users only.
+ #ENDIF ($include_remote_users && ! $local_league_table)
+
+ if (my($user) = split((/\\s</)? " <" : " ", $_)) {
+ #IFDEF ($include_original_destination)
+ {
+ #ENDIF ($include_original_destination)
+ #IFNDEF ($include_original_destination)
+ if ($user =~ /^[\\/|]/) {
+ #ENDIF ($include_original_destination)
+ #my($parent) = $_ =~ /(<[^@]+@?[^>]*>)/;
+ my($parent) = $_ =~ / (<.+?>) /; #DT 1.54
+ $user = "$user $parent" if defined $parent;
+ }
+ ++$delivered_addresses_user{$user};
+ }
+ }
+ #ENDIF ($local_league_table || $include_remote_users)
+
+ #IFDEF ($do_sender{Host})
+ $delivered_addresses{Host}{$host}++;
+ #ENDIF ($do_sender{Host})
+ #IFDEF ($do_sender{Domain})
+ if ($domain) {
+ ++$delivered_addresses{Domain}{$domain};
+ }
+ #ENDIF ($do_sender{Domain})
+ #IFDEF ($do_sender{Email})
+ ++$delivered_addresses{Email}{$email};
+ #ENDIF ($do_sender{Email})
+ #IFDEF ($do_sender{Edomain})
+ ++$delivered_addresses{Edomain}{$edomain};
+ #ENDIF ($do_sender{Edomain})
+
+ ++$total_delivered_addresses;
+ }
+
+ elsif ($flag eq "==" && defined($message[$SIZE]) && !defined($message[$DELAYED])) {
+ ++$delayed_count;
+ $message[$DELAYED] = 1;
+ }
+
+ elsif ($flag eq "**") {
+ if (defined ($message[$SIZE])) {
+ unless (defined $message[$HAD_ERROR]) {
+ ++$message_errors;
+ $message[$HAD_ERROR] = 1;
+ }
+ }
+
+ #IFDEF ($show_errors)
+ ++$errors_count{$_};
+ #ENDIF ($show_errors)
+
+ }
+
+ elsif ($flag eq "Co") {
+ #Completed?
+ #IFDEF ($#queue_times >= 0)
+ $queued = queue_time($tod, $message[$ARRIVAL_TIME], $id);
+
+ for ($i = 0; $i <= $#queue_times; $i++) {
+ if ($queued < $queue_times[$i]) {
+ ++$qt_all_bin[$i];
+ ++$qt_remote_bin[$i] if $message[$REMOTE_DELIVERED];
+ last;
+ }
+ }
+ if ($i > $#queue_times) {
+ ++$qt_all_overflow;
+ ++$qt_remote_overflow if $message[$REMOTE_DELIVERED];
+ }
+ #ENDIF ($#queue_times >= 0)
+
+ #IFDEF ($#rcpt_times >= 0)
+ if (/ QT=(\S+)/) {
+ $seconds = wdhms_seconds($1);
+ #Calculate $queued if not previously calculated above.
+ #IFNDEF ($#queue_times >= 0)
+ $queued = queue_time($tod, $message[$ARRIVAL_TIME], $id);
+ #ENDIF ($#queue_times >= 0)
+ $rcpt_time = $seconds - $queued;
+ my($protocol);
+
+ if (defined $message[$PROTOCOL]) {
+ $protocol = $message[$PROTOCOL];
+
+ # Create the bin if its not already defined.
+ unless (exists $rcpt_times_bin{$protocol}) {
+ initialise_rcpt_times($protocol);
+ }
+ }
+
+
+ for ($i = 0; $i <= $#rcpt_times; ++$i) {
+ if ($rcpt_time < $rcpt_times[$i]) {
+ ++$rcpt_times_bin{all}[$i];
+ ++$rcpt_times_bin{$protocol}[$i] if defined $protocol;
+ last;
+ }
+ }
+
+ if ($i > $#rcpt_times) {
+ ++$rcpt_times_overflow{all};
+ ++$rcpt_times_overflow{$protocol} if defined $protocol;
+ }
+ }
+ #ENDIF ($#rcpt_times >= 0)
+
+ delete($messages{$id});
+ }
+ elsif ($flag eq "SA") {
+ $ip = (/From.*?(\\[[^]]+\\])/ || /\\((local)\\)/) ? $1 : "";
+ #SpamAssassin message
+ if (/Action: ((permanently|temporarily) rejected message|flagged as Spam but accepted): score=(\d+\.\d)/) {
+ #add_volume(\\$spam_score,\\$spam_score_gigs,$3);
+ ++$spam_count_by_ip{$ip};
+ } elsif (/Action: scanned but message isn\'t spam: score=(-?\d+\.\d)/) {
+ #add_volume(\\$ham_score,\\$ham_score_gigs,$1);
+ ++$ham_count_by_ip{$ip};
+ } elsif (/(Not running SA because SAEximRunCond expanded to false|check skipped due to message size)/) {
+ ++$ham_count_by_ip{$ip};
+ }
+ }
+
+ # Look for Reject messages or blackholed messages (deliveries
+ # without a transport)
+ if ($flag eq "Re" || ($flag eq "=>" && ! /\\sT=\\S+/)) {
+ # Correct the IP address for rejects:
+ # rejected EHLO from my.test.net [10.0.0.5]: syntactically invalid argument(s):
+ # rejected EHLO from [10.0.0.6]: syntactically invalid argument(s):
+ $ip = $1 if ($ip eq "local" && /^rejected [HE][HE]LO from .*?(\[.+?\]):/);
+ if (/SpamAssassin/) {
+ ++$rejected_count_by_reason{"Rejected by SpamAssassin"};
+ ++$rejected_count_by_ip{$ip};
+ }
+ elsif (
+ /(temporarily rejected [A-Z]*) .*?(: .*?)(:|\s*$)/
+ ) {
+ ++$temporarily_rejected_count_by_reason{"\u$1$2"};
+ ++$temporarily_rejected_count_by_ip{$ip};
+ }
+ elsif (
+ /(temporarily refused connection)/
+ ) {
+ ++$temporarily_rejected_count_by_reason{"\u$1"};
+ ++$temporarily_rejected_count_by_ip{$ip};
+ }
+ elsif (
+ /(listed at [^ ]+)/ ||
+ /(Forged IP detected in HELO)/ ||
+ /(Invalid domain or IP given in HELO\/EHLO)/ ||
+ /(unqualified recipient rejected)/ ||
+ /(closed connection (after|in response) .*?)\s*$/ ||
+ /(sender rejected)/ ||
+ # 2005-09-23 15:07:49 1EInHJ-0007Ex-Au H=(a.b.c) [10.0.0.1] F=<> rejected after DATA: This message contains a virus: (Eicar-Test-Signature) please scan your system.
+ # 2005-10-06 10:50:07 1ENRS3-0000Nr-Kt => blackhole (DATA ACL discarded recipients): This message contains a virus: (Worm.SomeFool.P) please scan your system.
+ / rejected after DATA: (.*)/ ||
+ / (rejected DATA: .*)/ ||
+ /.DATA ACL discarded recipients.: (.*)/ ||
+ /rejected after DATA: (unqualified address not permitted)/ ||
+ /(VRFY rejected)/ ||
+# /(sender verify (defer|fail))/i ||
+ /(too many recipients)/ ||
+ /(refused relay.*?) to/ ||
+ /(rejected by non-SMTP ACL: .*)/ ||
+ /(rejected by local_scan.*)/ ||
+ # SMTP call from %s dropped: too many syntax or protocol errors (last command was "%s"
+ # SMTP call from %s dropped: too many nonmail commands
+ /(dropped: too many ((nonmail|unrecognized) commands|syntax or protocol errors))/ ||
+
+ # local_scan() function crashed with signal %d - message temporarily rejected
+ # local_scan() function timed out - message temporarily rejected
+ /(local_scan.. function .* - message temporarily rejected)/ ||
+ # SMTP protocol synchronization error (input sent without waiting for greeting): rejected connection from %s
+ /(SMTP protocol .*?(error|violation))/ ||
+ /(message too big)/
+ ) {
+ ++$rejected_count_by_reason{"\u$1"};
+ ++$rejected_count_by_ip{$ip};
+ }
+ elsif (/rejected [HE][HE]LO from [^:]*: syntactically invalid argument/) {
+ ++$rejected_count_by_reason{"Rejected HELO/EHLO: syntactically invalid argument"};
+ ++$rejected_count_by_ip{$ip};
+ }
+ elsif (/response to "RCPT TO.*? was: (.*)/) {
+ ++$rejected_count_by_reason{"Response to RCPT TO was: $1"};
+ ++$rejected_count_by_ip{$ip};
+ }
+ elsif (
+ /(lookup of host )\S+ (failed)/ ||
+
+ # rejected from <%s>%s%s%s%s: message too big:
+ /(rejected [A-Z]*) .*?(: .*?)(:|\s*$)/ ||
+ # refused connection from %s (host_reject_connection)
+ # refused connection from %s (tcp wrappers)
+ /(refused connection )from.*? (\(.*)/ ||
+
+ # error from remote mailer after RCPT TO:<a@b.c>: host a.b.c [10.0.0.1]: 450 <a@b.c>: Recipient address rejected: Greylisted for 60 seconds
+ # error from remote mailer after MAIL FROM:<> SIZE=3468: host a.b.c [10.0.0.1]: 421 a.b.c has refused your connection because your server did not have a PTR record.
+ /(error from remote mailer after .*?:).*(: .*?)(:|\s*$)/ ||
+
+ # a.b.c F=<a@b.c> rejected after DATA: "@" or "." expected after "Undisclosed-Recipient": failing address in "To" header is: <Undisclosed-Recipient:;>
+ /rejected after DATA: ("." or "." expected).*?(: failing address in .*? header)/ ||
+
+ # connection from %s refused load average = %.2f
+ /(Connection )from.*? (refused: load average)/ ||
+ # connection from %s refused (IP options)
+ # Connection from %s refused: too many connections
+ # connection from %s refused
+ /([Cc]onnection )from.*? (refused.*)/ ||
+ # [10.0.0.1]: connection refused
+ /: (Connection refused)()/
+ ) {
+ ++$rejected_count_by_reason{"\u$1$2"};
+ ++$rejected_count_by_ip{$ip};
+ }
+ elsif (
+ # 2008-03-31 06:25:22 H=mail.densitron.com [216.70.140.224]:45386 temporarily rejected connection in "connect" ACL: too fast reconnects // .hs
+ # 2008-03-31 06:25:22 H=mail.densitron.com [216.70.140.224]:45386 temporarily rejected connection in "connect" ACL // .hs
+ /(temporarily rejected connection in .*?ACL:?.*)/
+ ) {
+ ++$temporarily_rejected_count_by_ip{$ip};
+ ++$temporarily_rejected_count_by_reason{"\u$1"};
+ }
+ else {
+ ++$rejected_count_by_reason{Unknown};
+ ++$rejected_count_by_ip{$ip};
+ print STDERR "Unknown rejection: $_" if $debug;
+ }
+ }
+ }';
+
+ # We now do a 'C preprocessor style operation on our parser
+ # to remove bits not in use.
+ my(%defines_in_operation,$removing_lines,$processed_parser);
+ foreach (split (/\n/,$parser)) {
+ if ((/^\s*#\s*IFDEF\s*\((.*?)\)/i && ! eval $1) ||
+ (/^\s*#\s*IFNDEF\s*\((.*?)\)/i && eval $1) ) {
+ $defines_in_operation{$1} = 1;
+ $removing_lines = 1;
+ }
+
+ # Convert constants.
+ while (/(\$[A-Z][A-Z_]*)\b/) {
+ my $constant = eval $1;
+ s/(\$[A-Z][A-Z_]*)\b/$constant/;
+ }
+
+ $processed_parser .= $_."\n" unless $removing_lines;
+
+ if (/^\s*#\s*ENDIF\s*\((.*?)\)/i) {
+ delete $defines_in_operation{$1};
+ unless (keys %defines_in_operation) {
+ $removing_lines = 0;
+ }
+ }
+ }
+ print STDERR "# START OF PARSER:$processed_parser\n# END OF PARSER\n\n" if $debug;
+
+ return $processed_parser;
+}
+
+
+
+#######################################################################
+# parse();
+#
+# parse($parser,\*FILEHANDLE);
+#
+# This subroutine accepts a parser and a filehandle from main and parses each
+# line. We store the results into global variables.
+#######################################################################
+sub parse {
+ my($parser,$fh) = @_;
+
+ if ($merge_reports) {
+ parse_old_eximstat_reports($fh);
+ }
+ else {
+ eval $parser;
+ die ($@) if $@;
+ }
+
+}
+
+
+
+#######################################################################
+# print_header();
+#
+# print_header();
+#
+# Print our headers and contents.
+#######################################################################
+sub print_header {
+
+
+ my $title = "Exim statistics from $begin to $end";
+
+ print $txt_fh "\n$title\n" if $txt_fh;
+ if ($htm_fh) {
+ print $htm_fh html_header($title);
+ print $htm_fh "<ul>\n";
+ print $htm_fh "<li><a href=\"#Grandtotal\">Grand total summary</a>\n";
+ print $htm_fh "<li><a href=\"#Patterns\">User Specified Patterns</a>\n" if @user_patterns;
+ print $htm_fh "<li><a href=\"#Transport\">Deliveries by Transport</a>\n" if $show_transport;
+ if ($hist_opt) {
+ print $htm_fh "<li><a href=\"#Messages received\">Messages received per hour</a>\n";
+ print $htm_fh "<li><a href=\"#Deliveries\">Deliveries per hour</a>\n";
+ }
+
+ if ($#queue_times >= 0) {
+ print $htm_fh "<li><a href=\"#Time spent on the queue all messages\">Time spent on the queue: all messages</a>\n";
+ print $htm_fh "<li><a href=\"#Time spent on the queue messages with at least one remote delivery\">Time spent on the queue: messages with at least one remote delivery</a>\n";
+ }
+
+ if ($#delivery_times >= 0) {
+ print $htm_fh "<li><a href=\"#Delivery times all messages\">Delivery times: all messages</a>\n";
+ print $htm_fh "<li><a href=\"#Delivery times messages with at least one remote delivery\">Delivery times: messages with at least one remote delivery</a>\n";
+ }
+
+ if ($#rcpt_times >= 0) {
+ print $htm_fh "<li><a href=\"#Receipt times all messages\">Receipt times</a>\n";
+ }
+
+ print $htm_fh "<li><a href=\"#Relayed messages\">Relayed messages</a>\n" if $show_relay;
+ if ($topcount) {
+ print $htm_fh "<li><a href=\"#Mail rejection reason count\">Top $topcount mail rejection reasons by message count</a>\n" if %rejected_count_by_reason;
+ foreach ('Host','Domain','Email','Edomain') {
+ next unless $do_sender{$_};
+ print $htm_fh "<li><a href=\"#Sending \l$_ count\">Top $topcount sending \l${_}s by message count</a>\n";
+ print $htm_fh "<li><a href=\"#Sending \l$_ volume\">Top $topcount sending \l${_}s by volume</a>\n";
+ }
+ if (($local_league_table || $include_remote_users) && %received_count_user) {
+ print $htm_fh "<li><a href=\"#Local sender count\">Top $topcount local senders by message count</a>\n";
+ print $htm_fh "<li><a href=\"#Local sender volume\">Top $topcount local senders by volume</a>\n";
+ }
+ foreach ('Host','Domain','Email','Edomain') {
+ next unless $do_sender{$_};
+ print $htm_fh "<li><a href=\"#$_ destination count\">Top $topcount \l$_ destinations by message count</a>\n";
+ print $htm_fh "<li><a href=\"#$_ destination volume\">Top $topcount \l$_ destinations by volume</a>\n";
+ }
+ if (($local_league_table || $include_remote_users) && %delivered_messages_user) {
+ print $htm_fh "<li><a href=\"#Local destination count\">Top $topcount local destinations by message count</a>\n";
+ print $htm_fh "<li><a href=\"#Local destination volume\">Top $topcount local destinations by volume</a>\n";
+ }
+ if (($local_league_table || $include_remote_users) && %delivered_messages_local_domain) {
+ print $htm_fh "<li><a href=\"#Local domain destination count\">Top $topcount local domain destinations by message count</a>\n";
+ print $htm_fh "<li><a href=\"#Local domain destination volume\">Top $topcount local domain destinations by volume</a>\n";
+ }
+
+ print $htm_fh "<li><a href=\"#Rejected ip count\">Top $topcount rejected ips by message count</a>\n" if %rejected_count_by_ip;
+ print $htm_fh "<li><a href=\"#Temporarily rejected ip count\">Top $topcount temporarily rejected ips by message count</a>\n" if %temporarily_rejected_count_by_ip;
+ print $htm_fh "<li><a href=\"#Non-rejected spamming ip count\">Top $topcount non-rejected spamming ips by message count</a>\n" if %spam_count_by_ip;
+
+ }
+ print $htm_fh "<li><a href=\"#errors\">List of errors</a>\n" if %errors_count;
+ print $htm_fh "</ul>\n<hr>\n";
+ }
+ if ($xls_fh)
+ {
+ $ws_global->write($row++, $col+0, "Exim Statistics", $f_header1);
+ &set_worksheet_line($ws_global, $row, $col, ["from:", $begin, "to:", $end], $f_default);
+ $row+=2;
+ }
+}
+
+
+#######################################################################
+# print_grandtotals();
+#
+# print_grandtotals();
+#
+# Print the grand totals.
+#######################################################################
+sub print_grandtotals {
+
+ # Get the sender by headings and results. This is complicated as we can have
+ # different numbers of columns.
+ my($sender_txt_header,$sender_txt_format,$sender_html_format);
+ my(@received_totals,@delivered_totals);
+ my($row_tablehead, $row_max);
+ my(@col_headers) = ('TOTAL', 'Volume', 'Messages', 'Addresses');
+
+ foreach ('Host','Domain','Email','Edomain') {
+ next unless $do_sender{$_};
+ if ($merge_reports) {
+ push(@received_totals, get_report_total($report_totals{Received},"${_}s"));
+ push(@delivered_totals,get_report_total($report_totals{Delivered},"${_}s"));
+ }
+ else {
+ push(@received_totals,scalar(keys %{$received_data{$_}}));
+ push(@delivered_totals,scalar(keys %{$delivered_data{$_}}));
+ }
+ $sender_txt_header .= " " x ($COLUMN_WIDTHS - length($_)) . $_ . 's';
+ $sender_html_format .= "<td align=\"right\">%s</td>";
+ $sender_txt_format .= " " x ($COLUMN_WIDTHS - 5) . "%6s";
+ push(@col_headers,"${_}s");
+ }
+
+ my $txt_format1 = " %-16s %9s %6d %6s $sender_txt_format";
+ my $txt_format2 = " %6d %4.1f%% %6d %4.1f%%",
+ my $htm_format1 = "<tr><td>%s</td><td align=\"right\">%s</td><td align=\"right\">%s</td><td align=\"right\">%s</td>$sender_html_format";
+ my $htm_format2 = "<td align=\"right\">%d</td><td align=\"right\">%4.1f%%</td><td align=\"right\">%d</td><td align=\"right\">%4.1f%%</td>";
+
+ if ($txt_fh) {
+ my $sender_spaces = " " x length($sender_txt_header);
+ print $txt_fh "\n";
+ print $txt_fh "Grand total summary\n";
+ print $txt_fh "-------------------\n";
+ print $txt_fh " $sender_spaces At least one address\n";
+ print $txt_fh " TOTAL Volume Messages Addresses $sender_txt_header Delayed Failed\n";
+ }
+ if ($htm_fh) {
+ print $htm_fh "<a name=\"Grandtotal\"></a>\n";
+ print $htm_fh "<h2>Grand total summary</h2>\n";
+ print $htm_fh "<table border=1>\n";
+ print $htm_fh "<tr><th>" . join('</th><th>',@col_headers) . "</th><th colspan=2>At least one addr<br>Delayed</th><th colspan=2>At least one addr<br>Failed</th>\n";
+ }
+ if ($xls_fh) {
+ $ws_global->write($row++, 0, "Grand total summary", $f_header2);
+ $ws_global->write($row, 0, \@col_headers, $f_header2);
+ $ws_global->merge_range($row, scalar(@col_headers), $row, scalar(@col_headers)+1, "At least one addr Delayed", $f_header2_m);
+ $ws_global->merge_range($row, scalar(@col_headers)+2, $row, scalar(@col_headers)+3, "At least one addr Failed", $f_header2_m);
+ #$ws_global->write(++$row, scalar(@col_headers), ['Total','Percent','Total','Percent'], $f_header2);
+ }
+
+
+ my($volume,$failed_count);
+ if ($merge_reports) {
+ $volume = volume_rounded($report_totals{Received}{Volume}, $report_totals{Received}{'Volume-gigs'});
+ $total_received_count = get_report_total($report_totals{Received},'Messages');
+ $failed_count = get_report_total($report_totals{Received},'Failed');
+ $delayed_count = get_report_total($report_totals{Received},'Delayed');
+ }
+ else {
+ $volume = volume_rounded($total_received_data, $total_received_data_gigs);
+ $failed_count = $message_errors;
+ }
+
+ {
+ no integer;
+
+ my @content=(
+ $volume,$total_received_count,'',
+ @received_totals,
+ $delayed_count,
+ ($total_received_count) ? ($delayed_count*100/$total_received_count) : 0,
+ $failed_count,
+ ($total_received_count) ? ($failed_count*100/$total_received_count) : 0
+ );
+
+ printf $txt_fh ("$txt_format1$txt_format2\n", 'Received', @content) if $txt_fh;
+ printf $htm_fh ("$htm_format1$htm_format2\n", 'Received', @content) if $htm_fh;
+ if ($xls_fh) {
+ $ws_global->write(++$row, 0, 'Received', $f_default);
+ for (my $i=0; $i < scalar(@content); $i++) {
+ if ($i == 4 || $i == 6) {
+ $ws_global->write($row, $i+1, $content[$i]/100, $f_percent);
+ }
+ else {
+ $ws_global->write($row, $i+1, $content[$i], $f_default);
+ }
+ }
+ }
+ }
+
+ if ($merge_reports) {
+ $volume = volume_rounded($report_totals{Delivered}{Volume}, $report_totals{Delivered}{'Volume-gigs'});
+ $total_delivered_messages = get_report_total($report_totals{Delivered},'Messages');
+ $total_delivered_addresses = get_report_total($report_totals{Delivered},'Addresses');
+ }
+ else {
+ $volume = volume_rounded($total_delivered_data, $total_delivered_data_gigs);
+ }
+
+ my @content=($volume, $total_delivered_messages, $total_delivered_addresses, @delivered_totals);
+ printf $txt_fh ("$txt_format1\n", 'Delivered', @content) if $txt_fh;
+ printf $htm_fh ("$htm_format1\n", 'Delivered', @content) if $htm_fh;
+
+ if ($xls_fh) {
+ $ws_global->write(++$row, 0, 'Delivered', $f_default);
+ for (my $i=0; $i < scalar(@content); $i++) {
+ $ws_global->write($row, $i+1, $content[$i], $f_default);
+ }
+ }
+
+ if ($merge_reports) {
+ foreach ('Rejects', 'Temp Rejects', 'Ham', 'Spam') {
+ my $messages = get_report_total($report_totals{$_},'Messages');
+ my $addresses = get_report_total($report_totals{$_},'Addresses');
+ if ($messages) {
+ @content = ($_, '', $messages, '');
+ push(@content,get_report_total($report_totals{$_},'Hosts')) if $do_sender{Host};
+ #These rows do not have entries for the following columns (if specified)
+ foreach ('Domain','Email','Edomain') {
+ push(@content,'') if $do_sender{$_};
+ }
+
+ printf $txt_fh ("$txt_format1\n", @content) if $txt_fh;
+ printf $htm_fh ("$htm_format1\n", @content) if $htm_fh;
+ $ws_global->write(++$row, 0, \@content) if $xls_fh;
+ }
+ }
+ }
+ else {
+ foreach my $total_aref (['Rejects',\%rejected_count_by_ip],
+ ['Temp Rejects',\%temporarily_rejected_count_by_ip],
+ ['Ham',\%ham_count_by_ip],
+ ['Spam',\%spam_count_by_ip]) {
+ #Count the number of messages of this type.
+ my $messages = 0;
+ map {$messages += $_} values %{$total_aref->[1]};
+
+ if ($messages > 0) {
+ @content = ($total_aref->[0], '', $messages, '');
+
+ #Count the number of distinct IPs for the Hosts column.
+ push(@content,scalar(keys %{$total_aref->[1]})) if $do_sender{Host};
+
+ #These rows do not have entries for the following columns (if specified)
+ foreach ('Domain','Email','Edomain') {
+ push(@content,'') if $do_sender{$_};
+ }
+
+ printf $txt_fh ("$txt_format1\n", @content) if $txt_fh;
+ printf $htm_fh ("$htm_format1\n", @content) if $htm_fh;
+ $ws_global->write(++$row, 0, \@content) if $xls_fh;
+ }
+ }
+ }
+
+ printf $txt_fh "\n" if $txt_fh;
+ printf $htm_fh "</table>\n" if $htm_fh;
+ ++$row;
+}
+
+
+#######################################################################
+# print_user_patterns()
+#
+# print_user_patterns();
+#
+# Print the counts of user specified patterns.
+#######################################################################
+sub print_user_patterns {
+ my $txt_format1 = " %-18s %6d";
+ my $htm_format1 = "<tr><td>%s</td><td align=\"right\">%d</td>";
+
+ if ($txt_fh) {
+ print $txt_fh "User Specified Patterns\n";
+ print $txt_fh "-----------------------";
+ print $txt_fh "\n Total\n";
+ }
+ if ($htm_fh) {
+ print $htm_fh "<hr><a name=\"Patterns\"></a><h2>User Specified Patterns</h2>\n";
+ print $htm_fh "<table border=0 width=\"100%\">\n";
+ print $htm_fh "<tr><td>\n";
+ print $htm_fh "<table border=1>\n";
+ print $htm_fh "<tr><th>&nbsp;</th><th>Total</th>\n";
+ }
+ if ($xls_fh) {
+ $ws_global->write($row++, $col, "User Specified Patterns", $f_header2);
+ &set_worksheet_line($ws_global, $row++, 1, ["Total"], $f_headertab);
+ }
+
+
+ my($key);
+ if ($merge_reports) {
+ # We are getting our data from previous reports.
+ foreach $key (@user_descriptions) {
+ my $count = get_report_total($report_totals{patterns}{$key},'Total');
+ printf $txt_fh ("$txt_format1\n",$key,$count) if $txt_fh;
+ printf $htm_fh ("$htm_format1\n",$key,$count) if $htm_fh;
+ if ($xls_fh)
+ {
+ &set_worksheet_line($ws_global, $row++, 0, [$key,$count], $f_default);
+ }
+ }
+ }
+ else {
+ # We are getting our data from mainlog files.
+ my $user_pattern_index = 0;
+ foreach $key (@user_descriptions) {
+ printf $txt_fh ("$txt_format1\n",$key,$user_pattern_totals[$user_pattern_index]) if $txt_fh;
+ printf $htm_fh ("$htm_format1\n",$key,$user_pattern_totals[$user_pattern_index]) if $htm_fh;
+ $ws_global->write($row++, 0, [$key,$user_pattern_totals[$user_pattern_index]]) if $xls_fh;
+ $user_pattern_index++;
+ }
+ }
+ print $txt_fh "\n" if $txt_fh;
+ print $htm_fh "</table>\n\n" if $htm_fh;
+ if ($xls_fh)
+ {
+ ++$row;
+ }
+
+ if ($hist_opt > 0) {
+ my $user_pattern_index = 0;
+ foreach $key (@user_descriptions) {
+ print_histogram($key, 'occurence', @{$user_pattern_interval_count[$user_pattern_index]});
+ $user_pattern_index++;
+ }
+ }
+}
+
+#######################################################################
+# print_rejects()
+#
+# print_rejects();
+#
+# Print statistics about rejected mail.
+#######################################################################
+sub print_rejects {
+ my($format1,$reason);
+
+ my $txt_format1 = " %-40s %6d";
+ my $htm_format1 = "<tr><td>%s</td><td align=\"right\">%d</td>";
+
+ if ($txt_fh) {
+ print $txt_fh "Rejected mail by reason\n";
+ print $txt_fh "-----------------------";
+ print $txt_fh "\n Total\n";
+ }
+ if ($htm_fh) {
+ print $htm_fh "<hr><a name=\"patterns\"></a><h2>Rejected mail by reason</h2>\n";
+ print $htm_fh "<table border=0 width=\"100%\"><tr><td><table border=1>\n";
+ print $htm_fh "<tr><th>&nbsp;</th><th>Total</th>\n";
+ }
+ if ($xls_fh) {
+ $ws_global->write($row++, $col, "Rejected mail by reason", $f_header2);
+ &set_worksheet_line($ws_global, $row++, 1, ["Total"], $f_headertab);
+ }
+
+
+ my $href = ($merge_reports) ? $report_totals{rejected_mail_by_reason} : \%rejected_count_by_reason;
+ my(@chartdatanames, @chartdatavals_count);
+
+ foreach $reason (top_n_sort($topcount, $href, undef, undef)) {
+ printf $txt_fh ("$txt_format1\n",$reason,$href->{$reason}) if $txt_fh;
+ printf $htm_fh ("$htm_format1\n",$reason,$href->{$reason}) if $htm_fh;
+ set_worksheet_line($ws_global, $row++, 0, [$reason,$href->{$reason}], $f_default) if $xls_fh;
+ push(@chartdatanames, $reason);
+ push(@chartdatavals_count, $href->{$reason});
+ }
+
+ $row++ if $xls_fh;
+ print $txt_fh "\n" if $txt_fh;
+
+ if ($htm_fh) {
+ print $htm_fh "</tr></table></td><td>";
+ if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals_count > 0)) {
+ # calculate the graph
+ my @data = (
+ \@chartdatanames,
+ \@chartdatavals_count
+ );
+ my $graph = GD::Graph::pie->new(200, 200);
+ $graph->set(
+ x_label => 'Rejection Reasons',
+ y_label => 'Messages',
+ title => 'By count',
+ );
+ my $gd = $graph->plot(\@data) or warn($graph->error);
+ if ($gd) {
+ open(IMG, ">$chartdir/rejections_count.png") or die "Could not write $chartdir/rejections_count.png: $!\n";
+ binmode IMG;
+ print IMG $gd->png;
+ close IMG;
+ print $htm_fh "<img src=\"$chartrel/rejections_count.png\">";
+ }
+ }
+ print $htm_fh "</td></tr></table>\n\n";
+ }
+}
+
+
+
+
+
+#######################################################################
+# print_transport();
+#
+# print_transport();
+#
+# Print totals by transport.
+#######################################################################
+sub print_transport {
+ my(@chartdatanames);
+ my(@chartdatavals_count);
+ my(@chartdatavals_vol);
+ no integer; #Lose this for charting the data.
+
+ my $txt_format1 = " %-18s %6s %6d";
+ my $htm_format1 = "<tr><td>%s</td><td align=\"right\">%s</td><td align=\"right\">%d</td>";
+
+ if ($txt_fh) {
+ print $txt_fh "Deliveries by transport\n";
+ print $txt_fh "-----------------------";
+ print $txt_fh "\n Volume Messages\n";
+ }
+ if ($htm_fh) {
+ print $htm_fh "<hr><a name=\"Transport\"></a><h2>Deliveries by Transport</h2>\n";
+ print $htm_fh "<table border=0 width=\"100%\"><tr><td><table border=1>\n";
+ print $htm_fh "<tr><th>&nbsp;</th><th>Volume</th><th>Messages</th>\n";
+ }
+ if ($xls_fh) {
+ $ws_global->write(++$row, $col, "Deliveries by transport", $f_header2);
+ $ws_global->write(++$row, 1, ["Volume", "Messages"], $f_headertab);
+ }
+
+ my($key);
+ if ($merge_reports) {
+ # We are getting our data from previous reports.
+ foreach $key (sort keys %{$report_totals{transport}}) {
+ my $count = get_report_total($report_totals{transport}{$key},'Messages');
+ my @content=($key, volume_rounded($report_totals{transport}{$key}{Volume},
+ $report_totals{transport}{$key}{'Volume-gigs'}), $count);
+ push(@chartdatanames, $key);
+ push(@chartdatavals_count, $count);
+ push(@chartdatavals_vol, $report_totals{transport}{$key}{'Volume-gigs'}*$gig + $report_totals{transport}{$key}{Volume} );
+ printf $txt_fh ("$txt_format1\n", @content) if $txt_fh;
+ printf $htm_fh ("$htm_format1\n", @content) if $htm_fh;
+ $ws_global->write(++$row, 0, \@content) if $xls_fh;
+ }
+ }
+ else {
+ # We are getting our data from mainlog files.
+ foreach $key (sort keys %transported_data) {
+ my @content=($key, volume_rounded($transported_data{$key},$transported_data_gigs{$key}),
+ $transported_count{$key});
+ push(@chartdatanames, $key);
+ push(@chartdatavals_count, $transported_count{$key});
+ push(@chartdatavals_vol, $transported_data_gigs{$key}*$gig + $transported_data{$key});
+ printf $txt_fh ("$txt_format1\n", @content) if $txt_fh;
+ printf $htm_fh ("$htm_format1\n", @content) if $htm_fh;
+ $ws_global->write(++$row, 0, \@content) if $xls_fh;
+ }
+ }
+ print $txt_fh "\n" if $txt_fh;
+ if ($htm_fh) {
+ print $htm_fh "</tr></table></td><td>";
+
+ if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals_count > 0))
+ {
+ # calculate the graph
+ my @data = (
+ \@chartdatanames,
+ \@chartdatavals_count
+ );
+ my $graph = GD::Graph::pie->new(200, 200);
+ $graph->set(
+ x_label => 'Transport',
+ y_label => 'Messages',
+ title => 'By count',
+ );
+ my $gd = $graph->plot(\@data) or warn($graph->error);
+ if ($gd) {
+ open(IMG, ">$chartdir/transports_count.png") or die "Could not write $chartdir/transports_count.png: $!\n";
+ binmode IMG;
+ print IMG $gd->png;
+ close IMG;
+ print $htm_fh "<img src=\"$chartrel/transports_count.png\">";
+ }
+ }
+ print $htm_fh "</td><td>";
+
+ if ($HAVE_GD_Graph_pie && $charts && ($#chartdatavals_vol > 0)) {
+ my @data = (
+ \@chartdatanames,
+ \@chartdatavals_vol
+ );
+ my $graph = GD::Graph::pie->new(200, 200);
+ $graph->set(
+ title => 'By volume',
+ );
+ my $gd = $graph->plot(\@data) or warn($graph->error);
+ if ($gd) {
+ open(IMG, ">$chartdir/transports_vol.png") or die "Could not write $chartdir/transports_vol.png: $!\n";
+ binmode IMG;
+ print IMG $gd->png;
+ close IMG;
+ print $htm_fh "<img src=\"$chartrel/transports_vol.png\">";
+ }
+ }
+
+ print $htm_fh "</td></tr></table>\n\n";
+ }
+}
+
+
+
+#######################################################################
+# print_relay();
+#
+# print_relay();
+#
+# Print our totals by relay.
+#######################################################################
+sub print_relay {
+ my $row_print_relay=1;
+ my $temp = "Relayed messages";
+ print $htm_fh "<hr><a name=\"$temp\"></a><h2>$temp</h2>\n" if $htm_fh;
+ if (scalar(keys %relayed) > 0 || $relayed_unshown > 0) {
+ my $shown = 0;
+ my $spacing = "";
+ my $txt_format = "%7d %s\n => %s\n";
+ my $htm_format = "<tr><td align=\"right\">%d</td><td>%s</td><td>%s</td>\n";
+
+ printf $txt_fh ("%s\n%s\n\n", $temp, "-" x length($temp)) if $txt_fh;
+ if ($htm_fh) {
+ print $htm_fh "<table border=1>\n";
+ print $htm_fh "<tr><th>Count</th><th>From</th><th>To</th>\n";
+ }
+ if ($xls_fh) {
+ $ws_relayed->write($row_print_relay++, $col, $temp, $f_header2);
+ &set_worksheet_line($ws_relayed, $row_print_relay++, 0, ["Count", "From", "To"], $f_headertab);
+ }
+
+
+ my($key);
+ foreach $key (sort keys %relayed) {
+ my $count = $relayed{$key};
+ $shown += $count;
+ $key =~ s/[HA]=//g;
+ my($one,$two) = split(/=> /, $key);
+ my @content=($count, $one, $two);
+ printf $txt_fh ($txt_format, @content) if $txt_fh;
+ printf $htm_fh ($htm_format, @content) if $htm_fh;
+ if ($xls_fh)
+ {
+ &set_worksheet_line($ws_relayed, $row_print_relay++, 0, \@content);
+ }
+ $spacing = "\n";
+ }
+
+ print $htm_fh "</table>\n<p>\n" if $htm_fh;
+ print $txt_fh "${spacing}Total: $shown (plus $relayed_unshown unshown)\n\n" if $txt_fh;
+ print $htm_fh "${spacing}Total: $shown (plus $relayed_unshown unshown)\n\n" if $htm_fh;
+ if ($xls_fh)
+ {
+ &set_worksheet_line($ws_relayed, $row_print_relay++, 0, [$shown, "Sum of shown" ]);
+ &set_worksheet_line($ws_relayed, $row_print_relay++, 0, [$relayed_unshown, "unshown"]);
+ $row_print_relay++;
+ }
+ }
+ else {
+ print $txt_fh "No relayed messages\n-------------------\n\n" if $txt_fh;
+ print $htm_fh "No relayed messages\n\n" if $htm_fh;
+ if ($xls_fh)
+ {
+ $row_print_relay++;
+ }
+ }
+}
+
+
+
+#######################################################################
+# print_errors();
+#
+# print_errors();
+#
+# Print our errors. In HTML, we display them as a list rather than a table -
+# Netscape doesn't like large tables!
+#######################################################################
+sub print_errors {
+ my $total_errors = 0;
+ $row=1;
+
+ if (scalar(keys %errors_count) != 0) {
+ my $temp = "List of errors";
+ my $htm_format = "<li>%d - %s\n";
+
+ printf $txt_fh ("%s\n%s\n\n", $temp, "-" x length($temp)) if $txt_fh;
+ if ($htm_fh) {
+ print $htm_fh "<hr><a name=\"errors\"></a><h2>$temp</h2>\n";
+ print $htm_fh "<ul><li><b>Count - Error</b>\n";
+ }
+ if ($xls_fh)
+ {
+ $ws_errors->write($row++, 0, $temp, $f_header2);
+ &set_worksheet_line($ws_errors, $row++, 0, ["Count", "Error"], $f_headertab);
+ }
+
+
+ my($key);
+ foreach $key (sort keys %errors_count) {
+ my $text = $key;
+ chomp($text);
+ $text =~ s/\s\s+/ /g; #Convert multiple spaces to a single space.
+ $total_errors += $errors_count{$key};
+
+ if ($txt_fh) {
+ printf $txt_fh ("%5d ", $errors_count{$key});
+ my $text_remaining = $text;
+ while (length($text_remaining) > 65) {
+ my($first,$rest) = $text_remaining =~ /(.{50}\S*)\s+(.+)/;
+ last if !$first;
+ printf $txt_fh ("%s\n\t ", $first);
+ $text_remaining = $rest;
+ }
+ printf $txt_fh ("%s\n\n", $text_remaining);
+ }
+
+ if ($htm_fh) {
+
+ #Translate HTML tag characters. Sergey Sholokh.
+ $text =~ s/\</\&lt\;/g;
+ $text =~ s/\>/\&gt\;/g;
+
+ printf $htm_fh ($htm_format,$errors_count{$key},$text);
+ }
+ if ($xls_fh)
+ {
+ &set_worksheet_line($ws_errors, $row++, 0, [$errors_count{$key},$text]);
+ }
+ }
+
+ $temp = "Errors encountered: $total_errors";
+
+ if ($txt_fh) {
+ print $txt_fh $temp, "\n";
+ print $txt_fh "-" x length($temp),"\n";
+ }
+ if ($htm_fh) {
+ print $htm_fh "</ul>\n<p>\n";
+ print $htm_fh $temp, "\n";
+ }
+ if ($xls_fh)
+ {
+ &set_worksheet_line($ws_errors, $row++, 0, [$total_errors, "Sum of Errors encountered"]);
+ }
+ }
+
+}
+
+
+#######################################################################
+# parse_old_eximstat_reports();
+#
+# parse_old_eximstat_reports($fh);
+#
+# Parse old eximstat output so we can merge daily stats to weekly stats and weekly to monthly etc.
+#
+# To test that the merging still works after changes, do something like the following.
+# All the diffs should produce no output.
+#
+# options='-bydomain -byemail -byhost -byedomain'
+# options="$options -show_rt1,2,4 -show_dt 1,2,4"
+# options="$options -pattern 'Completed Messages' /Completed/"
+# options="$options -pattern 'Received Messages' /<=/"
+#
+# ./eximstats $options mainlog > mainlog.txt
+# ./eximstats $options -merge mainlog.txt > mainlog.2.txt
+# diff mainlog.txt mainlog.2.txt
+#
+# ./eximstats $options -html mainlog > mainlog.html
+# ./eximstats $options -merge -html mainlog.txt > mainlog.2.html
+# diff mainlog.html mainlog.2.html
+#
+# ./eximstats $options -merge mainlog.html > mainlog.3.txt
+# diff mainlog.txt mainlog.3.txt
+#
+# ./eximstats $options -merge -html mainlog.html > mainlog.3.html
+# diff mainlog.html mainlog.3.html
+#
+# ./eximstats $options -nvr mainlog > mainlog.nvr.txt
+# ./eximstats $options -merge mainlog.nvr.txt > mainlog.4.txt
+# diff mainlog.txt mainlog.4.txt
+#
+# # double_mainlog.txt should have twice the values that mainlog.txt has.
+# ./eximstats $options mainlog mainlog > double_mainlog.txt
+#######################################################################
+sub parse_old_eximstat_reports {
+ my($fh) = @_;
+
+ my(%league_table_value_entered, %league_table_value_was_zero, %table_order);
+
+ my(%user_pattern_index);
+ my $user_pattern_index = 0;
+ map {$user_pattern_index{$_} = $user_pattern_index++} @user_descriptions;
+ my $user_pattern_keys = join('|', @user_descriptions);
+
+ while (<$fh>) {
+ PARSE_OLD_REPORT_LINE:
+ if (/Exim statistics from ([\d\-]+ [\d:]+(\s+[\+\-]\d+)?) to ([\d\-]+ [\d:]+(\s+[\+\-]\d+)?)/) {
+ $begin = $1 if ($1 lt $begin);
+ $end = $3 if ($3 gt $end);
+ }
+ elsif (/Grand total summary/) {
+ # Fill in $report_totals{Received|Delivered}{Volume|Messages|Addresses|Hosts|Domains|...|Delayed|DelayedPercent|Failed|FailedPercent}
+ my(@fields, @delivered_fields);
+ my $doing_table = 0;
+ while (<$fh>) {
+ $_ = html2txt($_); #Convert general HTML markup to text.
+ s/At least one addr//g; #Another part of the HTML output we don't want.
+
+# TOTAL Volume Messages Addresses Hosts Domains Delayed Failed
+# Received 26MB 237 177 23 8 3.4% 28 11.8%
+# Delivered 13MB 233 250 99 88
+ if (/TOTAL\s+(.*?)\s*$/) {
+ $doing_table = 1;
+ @delivered_fields = split(/\s+/,$1);
+
+ #Delayed and Failed have two columns each, so add the extra field names in.
+ splice(@delivered_fields,-1,1,'DelayedPercent','Failed','FailedPercent');
+
+ # Addresses only figure in the Delivered row, so remove them from the
+ # normal fields.
+ @fields = grep !/Addresses/, @delivered_fields;
+ }
+ elsif (/(Received)\s+(.*?)\s*$/) {
+ print STDERR "Parsing $_" if $debug;
+ add_to_totals($report_totals{$1},\@fields,$2);
+ }
+ elsif (/(Delivered)\s+(.*?)\s*$/) {
+ print STDERR "Parsing $_" if $debug;
+ add_to_totals($report_totals{$1},\@delivered_fields,$2);
+ my $data = $2;
+ # If we're merging an old report which doesn't include addresses,
+ # then use the Messages field instead.
+ unless (grep(/Addresses/, @delivered_fields)) {
+ my %tmp;
+ line_to_hash(\%tmp,\@delivered_fields,$data);
+ add_to_totals($report_totals{Delivered},['Addresses'],$tmp{Messages});
+ }
+ }
+ elsif (/(Temp Rejects|Rejects|Ham|Spam)\s+(.*?)\s*$/) {
+ print STDERR "Parsing $_" if $debug;
+ add_to_totals($report_totals{$1},['Messages','Hosts'],$2);
+ }
+ else {
+ last if $doing_table;
+ }
+ }
+ }
+
+ elsif (/User Specified Patterns/i) {
+#User Specified Patterns
+#-----------------------
+# Total
+# Description 85
+
+ while (<$fh>) { last if (/Total/); } #Wait until we get the table headers.
+ while (<$fh>) {
+ print STDERR "Parsing $_" if $debug;
+ $_ = html2txt($_); #Convert general HTML markup to text.
+ if (/^\s*(.*?)\s+(\d+)\s*$/) {
+ $report_totals{patterns}{$1} = {} unless (defined $report_totals{patterns}{$1});
+ add_to_totals($report_totals{patterns}{$1},['Total'],$2);
+ }
+ last if (/^\s*$/); #Finished if we have a blank line.
+ }
+ }
+
+ elsif (/(^|<h2>)($user_pattern_keys) per /o) {
+ # Parse User defined pattern histograms if they exist.
+ parse_histogram($fh, $user_pattern_interval_count[$user_pattern_index{$2}] );
+ }
+
+
+ elsif (/Deliveries by transport/i) {
+#Deliveries by transport
+#-----------------------
+# Volume Messages
+# :blackhole: 70KB 51
+# address_pipe 655KB 1
+# smtp 11MB 151
+
+ while (<$fh>) { last if (/Volume/); } #Wait until we get the table headers.
+ while (<$fh>) {
+ print STDERR "Parsing $_" if $debug;
+ $_ = html2txt($_); #Convert general HTML markup to text.
+ if (/(\S+)\s+(\d+\S*\s+\d+)/) {
+ $report_totals{transport}{$1} = {} unless (defined $report_totals{transport}{$1});
+ add_to_totals($report_totals{transport}{$1},['Volume','Messages'],$2);
+ }
+ last if (/^\s*$/); #Finished if we have a blank line.
+ }
+ }
+ elsif (/Messages received per/) {
+ parse_histogram($fh, \@received_interval_count);
+ }
+ elsif (/Deliveries per/) {
+ parse_histogram($fh, \@delivered_interval_count);
+ }
+
+ #elsif (/Time spent on the queue: (all messages|messages with at least one remote delivery)/) {
+ elsif (/(Time spent on the queue|Delivery times|Receipt times): ((\S+) messages|messages with at least one remote delivery)((<[^>]*>)*\s*)$/) {
+#Time spent on the queue: all messages
+#-------------------------------------
+#
+#Under 1m 217 91.9% 91.9%
+# 5m 2 0.8% 92.8%
+# 3h 8 3.4% 96.2%
+# 6h 7 3.0% 99.2%
+# 12h 2 0.8% 100.0%
+
+ # Set a pointer to the queue bin so we can use the same code
+ # block for both all messages and remote deliveries.
+ #my $bin_aref = ($1 eq 'all messages') ? \@qt_all_bin : \@qt_remote_bin;
+ my($bin_aref, $times_aref, $overflow_sref);
+ if ($1 eq 'Time spent on the queue') {
+ $times_aref = \@queue_times;
+ if ($2 eq 'all messages') {
+ $bin_aref = \@qt_all_bin;
+ $overflow_sref = \$qt_all_overflow;
+ }
+ else {
+ $bin_aref = \@qt_remote_bin;
+ $overflow_sref = \$qt_remote_overflow;
+ }
+ }
+ elsif ($1 eq 'Delivery times') {
+ $times_aref = \@delivery_times;
+ if ($2 eq 'all messages') {
+ $bin_aref = \@dt_all_bin;
+ $overflow_sref = \$dt_all_overflow;
+ }
+ else {
+ $bin_aref = \@dt_remote_bin;
+ $overflow_sref = \$dt_remote_overflow;
+ }
+ }
+ else {
+ unless (exists $rcpt_times_bin{$3}) {
+ initialise_rcpt_times($3);
+ }
+ $bin_aref = $rcpt_times_bin{$3};
+ $times_aref = \@rcpt_times;
+ $overflow_sref = \$rcpt_times_overflow{$3};
+ }
+
+
+ my ($blank_lines, $reached_table) = (0,0);
+ while (<$fh>) {
+ $_ = html2txt($_); #Convert general HTML markup to text.
+ # The table is preceded by one blank line, and has one blank line
+ # following it. As the table may be empty, the best way to determine
+ # that we've finished it is to look for the second blank line.
+ ++$blank_lines if /^\s*$/;
+ last if ($blank_lines >=2); #Finished the table ?
+ $reached_table = 1 if (/\d/);
+ next unless $reached_table;
+ my $previous_seconds_on_queue = 0;
+ if (/^\s*(Under|Over|)\s+(\d+[smhdw])\s+(\d+)/) {
+ print STDERR "Parsing $_" if $debug;
+ my($modifier,$formatted_time,$count) = ($1,$2,$3);
+ my $seconds = unformat_time($formatted_time);
+ my $time_on_queue = ($seconds + $previous_seconds_on_queue) / 2;
+ $previous_seconds_on_queue = $seconds;
+ $time_on_queue = $seconds * 2 if ($modifier eq 'Over');
+ my($i);
+ for ($i = 0; $i <= $#$times_aref; $i++) {
+ if ($time_on_queue < $times_aref->[$i]) {
+ $$bin_aref[$i] += $count;
+ last;
+ }
+ }
+ $$overflow_sref += $count if ($i > $#$times_aref);
+
+ }
+ }
+ }
+
+ elsif (/Relayed messages/) {
+#Relayed messages
+#----------------
+#
+# 1 addr.domain.com [1.2.3.4] a.user@domain.com
+# => addr2.domain2.com [5.6.7.8] a2.user2@domain2.com
+#
+#<tr><td align="right">1</td><td>addr.domain.com [1.2.3.4] a.user@domain.com </td><td>addr2.domain2.com [5.6.7.8] a2.user2@domain2.com</td>
+
+ my $reached_table = 0;
+ my($count,$sender);
+ while (<$fh>) {
+ unless ($reached_table) {
+ last if (/No relayed messages/);
+ $reached_table = 1 if (/^\s*\d/ || />\d+</);
+ next unless $reached_table;
+ }
+ if (/>(\d+)<.td><td>(.*?) ?<.td><td>(.*?)</) {
+ update_relayed($1,$2,$3);
+ }
+ elsif (/^\s*(\d+)\s+(.*?)\s*$/) {
+ ($count,$sender) = ($1,$2);
+ }
+ elsif (/=>\s+(.*?)\s*$/) {
+ update_relayed($count,$sender,$1);
+ }
+ else {
+ last; #Finished the table ?
+ }
+ }
+ }
+
+ elsif (/Top (.*?) by (message count|volume)/) {
+#Top 50 sending hosts by message count
+#-------------------------------------
+#
+# 48 1468KB local
+# Could also have average values for HTML output.
+# 48 1468KB 30KB local
+
+ my($category,$by_count_or_volume) = ($1,$2);
+
+ #As we show 2 views of each table (by count and by volume),
+ #most (but not all) entries will appear in both tables.
+ #Set up a hash to record which entries we have already seen
+ #and one to record which ones we are seeing for the first time.
+ if ($by_count_or_volume =~ /count/) {
+ undef %league_table_value_entered;
+ undef %league_table_value_was_zero;
+ undef %table_order;
+ }
+
+ #As this section processes multiple different table categories,
+ #set up pointers to the hashes to be updated.
+ my($messages_href,$addresses_href,$data_href,$data_gigs_href);
+ if ($category =~ /local sender/) {
+ $messages_href = \%received_count_user;
+ $addresses_href = undef;
+ $data_href = \%received_data_user;
+ $data_gigs_href = \%received_data_gigs_user;
+ }
+ elsif ($category =~ /sending (\S+?)s?\b/) {
+ #Top 50 sending (host|domain|email|edomain)s
+ #Top sending (host|domain|email|edomain)
+ $messages_href = \%{$received_count{"\u$1"}};
+ $data_href = \%{$received_data{"\u$1"}};
+ $data_gigs_href = \%{$received_data_gigs{"\u$1"}};
+ }
+ elsif ($category =~ /local destination/) {
+ $messages_href = \%delivered_messages_user;
+ $addresses_href = \%delivered_addresses_user;
+ $data_href = \%delivered_data_user;
+ $data_gigs_href = \%delivered_data_gigs_user;
+ }
+ elsif ($category =~ /local domain destination/) {
+ $messages_href = \%delivered_messages_local_domain;
+ $addresses_href = \%delivered_addresses_local_domain;
+ $data_href = \%delivered_data_local_domain;
+ $data_gigs_href = \%delivered_data_gigs_local_domain;
+ }
+ elsif ($category =~ /(\S+) destination/) {
+ #Top 50 (host|domain|email|edomain) destinations
+ #Top (host|domain|email|edomain) destination
+ $messages_href = \%{$delivered_messages{"\u$1"}};
+ $addresses_href = \%{$delivered_addresses{"\u$1"}};
+ $data_href = \%{$delivered_data{"\u$1"}};
+ $data_gigs_href = \%{$delivered_data_gigs{"\u$1"}};
+ }
+ elsif ($category =~ /temporarily rejected ips/) {
+ $messages_href = \%temporarily_rejected_count_by_ip;
+ }
+ elsif ($category =~ /rejected ips/) {
+ $messages_href = \%rejected_count_by_ip;
+ }
+ elsif ($category =~ /non-rejected spamming ips/) {
+ $messages_href = \%spam_count_by_ip;
+ }
+ elsif ($category =~ /mail temporary rejection reasons/) {
+ $messages_href = \%temporarily_rejected_count_by_reason;
+ }
+ elsif ($category =~ /mail rejection reasons/) {
+ $messages_href = \%rejected_count_by_reason;
+ }
+
+ my $reached_table = 0;
+ my $row_re;
+ while (<$fh>) {
+ # Watch out for empty tables.
+ goto PARSE_OLD_REPORT_LINE if (/<h2>/ or (/^\s*[a-zA-Z]/ && !/^\s*Messages/));
+
+ $_ = html2txt($_); #Convert general HTML markup to text.
+
+ # Messages Addresses Bytes Average
+ if (/^\s*Messages/) {
+ my $pattern = '^\s*(\d+)';
+ $pattern .= (/Addresses/) ? '\s+(\d+)' : '()';
+ $pattern .= (/Bytes/) ? '\s+([\dKMGB]+)' : '()';
+ $pattern .= (/Average/) ? '\s+[\dKMGB]+' : '';
+ $pattern .= '\s+(.*?)\s*$';
+ $row_re = qr/$pattern/;
+ $reached_table = 1;
+ next;
+ }
+ next unless $reached_table;
+
+ my($messages, $addresses, $rounded_volume, $entry);
+
+ if (/$row_re/) {
+ ($messages, $addresses, $rounded_volume, $entry) = ($1, $2, $3, $4);
+ }
+ else {
+ #Else we have finished the table and we may need to do some
+ #kludging to retain the order of the entries.
+
+ if ($by_count_or_volume =~ /volume/) {
+ #Add a few bytes to appropriate entries to preserve the order.
+ foreach $rounded_volume (keys %table_order) {
+ #For each rounded volume, we want to create a list which has things
+ #ordered from the volume table at the front, and additional things
+ #from the count table ordered at the back.
+ @{$table_order{$rounded_volume}{volume}} = () unless defined $table_order{$rounded_volume}{volume};
+ @{$table_order{$rounded_volume}{'message count'}} = () unless defined $table_order{$rounded_volume}{'message count'};
+ my(@order,%mark);
+ map {$mark{$_} = 1} @{$table_order{$rounded_volume}{volume}};
+ @order = @{$table_order{$rounded_volume}{volume}};
+ map {push(@order,$_)} grep(!$mark{$_},@{$table_order{$rounded_volume}{'message count'}});
+
+ my $bonus_bytes = $#order;
+ $bonus_bytes = 511 if ($bonus_bytes > 511); #Don't go over the half-K boundary!
+ while (@order and ($bonus_bytes > 0)) {
+ my $entry = shift(@order);
+ if ($league_table_value_was_zero{$entry}) {
+ $$data_href{$entry} += $bonus_bytes;
+ print STDERR "$category by $by_count_or_volume: added $bonus_bytes bonus bytes to $entry\n" if $debug;
+ }
+ $bonus_bytes--;
+ }
+ }
+ }
+ last;
+ }
+
+ # Store a new table entry.
+
+ # Add the entry into the %table_order hash if it has a rounded
+ # volume (KB/MB/GB).
+ push(@{$table_order{$rounded_volume}{$by_count_or_volume}},$entry) if ($rounded_volume =~ /\D/);
+
+ unless ($league_table_value_entered{$entry}) {
+ $league_table_value_entered{$entry} = 1;
+ unless ($$messages_href{$entry}) {
+ $$messages_href{$entry} = 0;
+ $$addresses_href{$entry} = 0;
+ $$data_href{$entry} = 0;
+ $$data_gigs_href{$entry} = 0;
+ $league_table_value_was_zero{$entry} = 1;
+ }
+
+ $$messages_href{$entry} += $messages;
+
+ # When adding the addresses, be aware that we could be merging
+ # an old report which does not include addresses. In this case,
+ # we add the messages instead.
+ $$addresses_href{$entry} += ($addresses) ? $addresses : $messages;
+
+ #Add the rounded value to the data and data_gigs hashes.
+ un_round($rounded_volume,\$$data_href{$entry},\$$data_gigs_href{$entry}) if $rounded_volume;
+ print STDERR "$category by $by_count_or_volume: added $messages,$rounded_volume to $entry\n" if $debug;
+ }
+
+ }
+ }
+ elsif (/List of errors/) {
+#List of errors
+#--------------
+#
+# 1 07904931641@one2one.net R=external T=smtp: SMTP error
+# from remote mailer after RCPT TO:<07904931641@one2one.net>:
+# host mail.one2one.net [193.133.192.24]: 550 User unknown
+#
+#<li>1 - ally.dufc@dunbar.org.uk R=external T=smtp: SMTP error from remote mailer after RCPT TO:<ally.dufc@dunbar.org.uk>: host mail.dunbar.org.uk [216.167.89.88]: 550 Unknown local part ally.dufc in <ally.dufc@dunbar.org.uk>
+
+
+ my $reached_table = 0;
+ my($count,$error,$blanks);
+ while (<$fh>) {
+ $reached_table = 1 if (/^( *|<li>)(\d+)/);
+ next unless $reached_table;
+
+ s/^<li>(\d+) -/$1/; #Convert an HTML line to a text line.
+ $_ = html2txt($_); #Convert general HTML markup to text.
+
+ if (/\t\s*(.*)/) {
+ $error .= ' ' . $1; #Join a multiline error.
+ }
+ elsif (/^\s*(\d+)\s+(.*)/) {
+ if ($error) {
+ #Finished with a previous multiline error so save it.
+ $errors_count{$error} = 0 unless $errors_count{$error};
+ $errors_count{$error} += $count;
+ }
+ ($count,$error) = ($1,$2);
+ }
+ elsif (/Errors encountered/) {
+ if ($error) {
+ #Finished the section, so save our stored last error.
+ $errors_count{$error} = 0 unless $errors_count{$error};
+ $errors_count{$error} += $count;
+ }
+ last;
+ }
+ }
+ }
+
+ }
+}
+
+#######################################################################
+# parse_histogram($fh, \@delivered_interval_count);
+# Parse a histogram into the provided array of counters.
+#######################################################################
+sub parse_histogram {
+ my($fh, $counters_aref) = @_;
+
+ # Messages received per hour (each dot is 2 messages)
+ #---------------------------------------------------
+ #
+ #00-01 106 .....................................................
+ #01-02 103 ...................................................
+
+ my $reached_table = 0;
+ while (<$fh>) {
+ $reached_table = 1 if (/^00/);
+ next unless $reached_table;
+ print STDERR "Parsing $_" if $debug;
+ if (/^(\d+):(\d+)\s+(\d+)/) { #hh:mm start time format ?
+ $$counters_aref[($1*60 + $2)/$hist_interval] += $3 if $hist_opt;
+ }
+ elsif (/^(\d+)-(\d+)\s+(\d+)/) { #hh-hh start-end time format ?
+ $$counters_aref[($1*60)/$hist_interval] += $3 if $hist_opt;
+ }
+ else { #Finished the table ?
+ last;
+ }
+ }
+}
+
+
+#######################################################################
+# update_relayed();
+#
+# update_relayed($count,$sender,$recipient);
+#
+# Adds an entry into the %relayed hash. Currently only used when
+# merging reports.
+#######################################################################
+sub update_relayed {
+ my($count,$sender,$recipient) = @_;
+
+ #When generating the key, put in the 'H=' and 'A=' which can be used
+ #in searches.
+ my $key = "H=$sender => H=$recipient";
+ $key =~ s/ ([^=\s]+\@\S+|<>)/ A=$1/g;
+ if (!defined $relay_pattern || $key !~ /$relay_pattern/o) {
+ $relayed{$key} = 0 if !defined $relayed{$key};
+ $relayed{$key} += $count;
+ }
+ else {
+ $relayed_unshown += $count;
+ }
+}
+
+
+#######################################################################
+# add_to_totals();
+#
+# add_to_totals(\%totals,\@keys,$values);
+#
+# Given a line of space separated values, add them into the provided hash using @keys
+# as the hash keys.
+#
+# If the value contains a '%', then the value is set rather than added. Otherwise, we
+# convert the value to bytes and gigs. The gigs get added to I<Key>-gigs.
+#######################################################################
+sub add_to_totals {
+ my($totals_href,$keys_aref,$values) = @_;
+ my(@values) = split(/\s+/,$values);
+
+ for(my $i = 0; $i < @values && $i < @$keys_aref; ++$i) {
+ my $key = $keys_aref->[$i];
+ if ($values[$i] =~ /%/) {
+ $$totals_href{$key} = $values[$i];
+ }
+ else {
+ $$totals_href{$key} = 0 unless ($$totals_href{$key});
+ $$totals_href{"$key-gigs"} = 0 unless ($$totals_href{"$key-gigs"});
+ un_round($values[$i], \$$totals_href{$key}, \$$totals_href{"$key-gigs"});
+ print STDERR "Added $values[$i] to $key - $$totals_href{$key} , " . $$totals_href{"$key-gigs"} . "GB.\n" if $debug;
+ }
+ }
+}
+
+
+#######################################################################
+# line_to_hash();
+#
+# line_to_hash(\%hash,\@keys,$line);
+#
+# Given a line of space separated values, set them into the provided hash
+# using @keys as the hash keys.
+#######################################################################
+sub line_to_hash {
+ my($href,$keys_aref,$values) = @_;
+ my(@values) = split(/\s+/,$values);
+ for(my $i = 0; $i < @values && $i < @$keys_aref; ++$i) {
+ $$href{$keys_aref->[$i]} = $values[$i];
+ }
+}
+
+
+#######################################################################
+# get_report_total();
+#
+# $total = get_report_total(\%hash,$key);
+#
+# If %hash contains values split into Units and Gigs, we calculate and return
+#
+# $hash{$key} + 1024*1024*1024 * $hash{"${key}-gigs"}
+#######################################################################
+sub get_report_total {
+ no integer;
+ my($hash_ref,$key) = @_;
+ if ($$hash_ref{"${key}-gigs"}) {
+ return $$hash_ref{$key} + $gig * $$hash_ref{"${key}-gigs"};
+ }
+ return $$hash_ref{$key} || 0;
+}
+
+#######################################################################
+# html2txt();
+#
+# $text_line = html2txt($html_line);
+#
+# Convert a line from html to text. Currently we just convert HTML tags to spaces
+# and convert &gt;, &lt;, and &nbsp; tags back.
+#######################################################################
+sub html2txt {
+ ($_) = @_;
+
+ # Convert HTML tags to spacing. Note that the reports may contain <Userid> and
+ # <Userid@Domain> words, so explicitly specify the HTML tags we will remove
+ # (the ones used by this program). If someone is careless enough to have their
+ # Userid the same as an HTML tag, there's not much we can do about it.
+ s/<\/?(html|head|title|body|h\d|ul|li|a\s+|table|tr|td|th|pre|hr|p|br)\b.*?>/ /g;
+
+ s/\&lt\;/\</og; #Convert '&lt;' to '<'.
+ s/\&gt\;/\>/og; #Convert '&gt;' to '>'.
+ s/\&nbsp\;/ /og; #Convert '&nbsp;' to ' '.
+ return($_);
+}
+
+#######################################################################
+# get_next_arg();
+#
+# $arg = get_next_arg();
+#
+# Because eximstats arguments are often passed as variables,
+# we can't rely on shell parsing to deal with quotes. This
+# subroutine returns $ARGV[1] and does a shift. If $ARGV[1]
+# starts with a quote (' or "), and doesn't end in one, then
+# we append the next argument to it and shift again. We repeat
+# until we've got all of the argument.
+#
+# This isn't perfect as all white space gets reduced to one space,
+# but it's as good as we can get! If it's essential that spacing
+# be preserved precisely, then you get that by not using shell
+# variables.
+#######################################################################
+sub get_next_arg {
+ my $arg = '';
+ my $matched_pattern = 0;
+ while ($ARGV[1]) {
+ $arg .= ' ' if $arg;
+ $arg .= $ARGV[1]; shift(@ARGV);
+ if ($arg !~ /^['"]/) {
+ $matched_pattern = 1;
+ last;
+ }
+ if ($arg =~ s/^(['"])(.*)\1$/$2/) {
+ $matched_pattern = 1;
+ last;
+ }
+ }
+ die "Mismatched argument quotes - <$arg>.\n" unless $matched_pattern;
+ return $arg;
+}
+
+#######################################################################
+# set_worksheet_line($ws_global, $startrow, $startcol, \@content, $format);
+#
+# set values to a sequence of cells in a row.
+#
+#######################################################################
+sub set_worksheet_line {
+ my ($worksheet, $row, $col, $content, $format) = @_;
+
+ foreach my $token (@$content)
+ {
+ $worksheet->write($row, $col++, $token, $format );
+ }
+
+}
+
+#######################################################################
+# @rcpt_times = parse_time_list($string);
+#
+# Parse a comma separated list of time values in seconds given by
+# the user and fill an array.
+#
+# Return a default list if $string is undefined.
+# Return () if $string eq '0'.
+#######################################################################
+sub parse_time_list {
+ my($string) = @_;
+ if (! defined $string) {
+ return(60, 5*60, 15*60, 30*60, 60*60, 3*60*60, 6*60*60, 12*60*60, 24*60*60);
+ }
+ my(@times) = split(/,/, $string);
+ foreach my $q (@times) { $q = eval($q) + 0 }
+ @times = sort { $a <=> $b } @times;
+ @times = () if ($#times == 0 && $times[0] == 0);
+ return(@times);
+}
+
+
+#######################################################################
+# initialise_rcpt_times($protocol);
+# Initialise an array of rcpt_times to 0 for the specified protocol.
+#######################################################################
+sub initialise_rcpt_times {
+ my($protocol) = @_;
+ for (my $i = 0; $i <= $#rcpt_times; ++$i) {
+ $rcpt_times_bin{$protocol}[$i] = 0;
+ }
+ $rcpt_times_overflow{$protocol} = 0;
+}
+
+
+##################################################
+# Main Program #
+##################################################
+
+
+$last_timestamp = '';
+$last_date = '';
+$show_errors = 1;
+$show_relay = 1;
+$show_transport = 1;
+$topcount = 50;
+$local_league_table = 1;
+$include_remote_users = 0;
+$include_original_destination = 0;
+$hist_opt = 1;
+$volume_rounding = 1;
+$localtime_offset = calculate_localtime_offset(); # PH/FANF
+
+$charts = 0;
+$charts_option_specified = 0;
+$chartrel = ".";
+$chartdir = ".";
+
+@queue_times = parse_time_list();
+@rcpt_times = ();
+@delivery_times = ();
+
+$last_offset = '';
+$offset_seconds = 0;
+
+$row=1;
+$col=0;
+$col_hist=0;
+$run_hist=0;
+my(%output_files); # What output files have been specified?
+
+# Decode options
+
+while (@ARGV > 0 && substr($ARGV[0], 0, 1) eq '-') {
+ if ($ARGV[0] =~ /^\-h(\d+)$/) { $hist_opt = $1 }
+ elsif ($ARGV[0] =~ /^\-ne$/) { $show_errors = 0 }
+ elsif ($ARGV[0] =~ /^\-nr(.?)(.*)\1$/) {
+ if ($1 eq "") { $show_relay = 0 } else { $relay_pattern = $2 }
+ }
+ elsif ($ARGV[0] =~ /^\-q([,\d\+\-\*\/]+)$/) { @queue_times = parse_time_list($1) }
+ elsif ($ARGV[0] =~ /^-nt$/) { $show_transport = 0 }
+ elsif ($ARGV[0] =~ /^\-nt(.?)(.*)\1$/)
+ {
+ if ($1 eq "") { $show_transport = 0 } else { $transport_pattern = $2 }
+ }
+ elsif ($ARGV[0] =~ /^-t(\d+)$/) { $topcount = $1 }
+ elsif ($ARGV[0] =~ /^-tnl$/) { $local_league_table = 0 }
+ elsif ($ARGV[0] =~ /^-txt=?(\S*)$/) { $txt_fh = get_filehandle($1,\%output_files) }
+ elsif ($ARGV[0] =~ /^-html=?(\S*)$/) { $htm_fh = get_filehandle($1,\%output_files) }
+ elsif ($ARGV[0] =~ /^-xls=?(\S*)$/) {
+ if ($HAVE_Spreadsheet_WriteExcel) {
+ $xls_fh = get_filehandle($1,\%output_files);
+ }
+ else {
+ warn "WARNING: CPAN Module Spreadsheet::WriteExcel not installed. Obtain from www.cpan.org\n";
+ }
+ }
+ elsif ($ARGV[0] =~ /^-merge$/) { $merge_reports = 1 }
+ elsif ($ARGV[0] =~ /^-charts$/) {
+ $charts = 1;
+ warn "WARNING: CPAN Module GD::Graph::pie not installed. Obtain from www.cpan.org\n" unless $HAVE_GD_Graph_pie;
+ warn "WARNING: CPAN Module GD::Graph::linespoints not installed. Obtain from www.cpan.org\n" unless $HAVE_GD_Graph_linespoints;
+ }
+ elsif ($ARGV[0] =~ /^-chartdir$/) { $chartdir = $ARGV[1]; shift; $charts_option_specified = 1; }
+ elsif ($ARGV[0] =~ /^-chartrel$/) { $chartrel = $ARGV[1]; shift; $charts_option_specified = 1; }
+ elsif ($ARGV[0] =~ /^-include_original_destination$/) { $include_original_destination = 1 }
+ elsif ($ARGV[0] =~ /^-cache$/) { } #Not currently used.
+ elsif ($ARGV[0] =~ /^-byhost$/) { $do_sender{Host} = 1 }
+ elsif ($ARGV[0] =~ /^-bydomain$/) { $do_sender{Domain} = 1 }
+ elsif ($ARGV[0] =~ /^-byemail$/) { $do_sender{Email} = 1 }
+ elsif ($ARGV[0] =~ /^-byemaildomain$/) { $do_sender{Edomain} = 1 }
+ elsif ($ARGV[0] =~ /^-byedomain$/) { $do_sender{Edomain} = 1 }
+ elsif ($ARGV[0] =~ /^-bylocaldomain$/) { $do_local_domain = 1 }
+ elsif ($ARGV[0] =~ /^-emptyok$/) { $emptyOK = 1 }
+ elsif ($ARGV[0] =~ /^-nvr$/) { $volume_rounding = 0 }
+ elsif ($ARGV[0] =~ /^-show_rt([,\d\+\-\*\/]+)?$/) { @rcpt_times = parse_time_list($1) }
+ elsif ($ARGV[0] =~ /^-show_dt([,\d\+\-\*\/]+)?$/) { @delivery_times = parse_time_list($1) }
+ elsif ($ARGV[0] =~ /^-d$/) { $debug = 1 }
+ elsif ($ARGV[0] =~ /^--?h(elp)?$/){ help() }
+ elsif ($ARGV[0] =~ /^-t_remote_users$/) { $include_remote_users = 1 }
+ elsif ($ARGV[0] =~ /^-pattern$/)
+ {
+ push(@user_descriptions,get_next_arg());
+ push(@user_patterns,get_next_arg());
+ }
+ elsif ($ARGV[0] =~ /^-utc$/)
+ {
+ # We don't need this value if the log is in UTC.
+ $localtime_offset = undef;
+ }
+ else
+ {
+ print STDERR "Eximstats: Unknown or malformed option $ARGV[0]\n";
+ help();
+ }
+ shift;
+ }
+
+ # keep old default behaviour
+ if (! ($xls_fh or $htm_fh or $txt_fh)) {
+ $txt_fh = \*STDOUT;
+ }
+
+ # Check that all the charts options are specified.
+ warn "-charts option not specified. Use -help for help.\n" if ($charts_option_specified && ! $charts);
+
+ # Default to display tables by sending Host.
+ $do_sender{Host} = 1 unless ($do_sender{Domain} || $do_sender{Email} || $do_sender{Edomain});
+
+ # prepare xls Excel Workbook
+ if (defined $xls_fh) {
+
+ # Create a new Excel workbook
+ $workbook = Spreadsheet::WriteExcel->new($xls_fh);
+
+ # Add worksheets
+ $ws_global = $workbook->addworksheet('Exim Statistik');
+ # show $ws_global as initial sheet
+ $ws_global->set_first_sheet();
+ $ws_global->activate();
+
+ if ($show_relay) {
+ $ws_relayed = $workbook->addworksheet('Relayed Messages');
+ $ws_relayed->set_column(1, 2, 80);
+ }
+ if ($show_errors) {
+ $ws_errors = $workbook->addworksheet('Errors');
+ }
+
+
+ # set column widths
+ $ws_global->set_column(0, 2, 20); # Columns B-D width set to 30
+ $ws_global->set_column(3, 3, 15); # Columns B-D width set to 30
+ $ws_global->set_column(4, 4, 25); # Columns B-D width set to 30
+
+ # Define Formats
+ $f_default = $workbook->add_format();
+
+ $f_header1 = $workbook->add_format();
+ $f_header1->set_bold();
+ #$f_header1->set_color('red');
+ $f_header1->set_size('15');
+ $f_header1->set_valign();
+ # $f_header1->set_align('center');
+ # $ws_global->write($row++, 2, "Testing Headers 1", $f_header1);
+
+ $f_header2 = $workbook->add_format();
+ $f_header2->set_bold();
+ $f_header2->set_size('12');
+ $f_header2->set_valign();
+ # $ws_global->write($row++, 2, "Testing Headers 2", $f_header2);
+
+ # Create another header2 for use in merged cells.
+ $f_header2_m = $workbook->add_format();
+ $f_header2_m->set_bold();
+ $f_header2_m->set_size('8');
+ $f_header2_m->set_valign();
+ $f_header2_m->set_align('center');
+
+ $f_percent = $workbook->add_format();
+ $f_percent->set_num_format('0.0%');
+
+ $f_headertab = $workbook->add_format();
+ $f_headertab->set_bold();
+ $f_headertab->set_valign();
+ # $ws_global->write($row++, 2, "Testing Headers tab", $f_headertab);
+
+ }
+
+
+# Initialise the queue/delivery/rcpt time counters.
+for (my $i = 0; $i <= $#queue_times; $i++) {
+ $qt_all_bin[$i] = 0;
+ $qt_remote_bin[$i] = 0;
+}
+for (my $i = 0; $i <= $#delivery_times; $i++) {
+ $dt_all_bin[$i] = 0;
+ $dt_remote_bin[$i] = 0;
+}
+initialise_rcpt_times('all');
+
+
+# Compute the number of slots for the histogram
+if ($hist_opt > 0)
+ {
+ if ($hist_opt > 60 || 60 % $hist_opt != 0)
+ {
+ print STDERR "Eximstats: -h must specify a factor of 60\n";
+ exit 1;
+ }
+ $hist_interval = 60/$hist_opt; #Interval in minutes.
+ $hist_number = (24*60)/$hist_interval; #Number of intervals per day.
+ @received_interval_count = (0) x $hist_number;
+ @delivered_interval_count = (0) x $hist_number;
+ my $user_pattern_index = 0;
+ for (my $user_pattern_index = 0; $user_pattern_index <= $#user_patterns; ++$user_pattern_index) {
+ @{$user_pattern_interval_count[$user_pattern_index]} = (0) x $hist_number;
+ }
+ @dt_all_bin = (0) x $hist_number;
+ @dt_remote_bin = (0) x $hist_number;
+}
+
+#$queue_unknown = 0;
+
+$total_received_data = 0;
+$total_received_data_gigs = 0;
+$total_received_count = 0;
+
+$total_delivered_data = 0;
+$total_delivered_data_gigs = 0;
+$total_delivered_messages = 0;
+$total_delivered_addresses = 0;
+
+$qt_all_overflow = 0;
+$qt_remote_overflow = 0;
+$dt_all_overflow = 0;
+$dt_remote_overflow = 0;
+$delayed_count = 0;
+$relayed_unshown = 0;
+$message_errors = 0;
+$begin = "9999-99-99 99:99:99";
+$end = "0000-00-00 00:00:00";
+my($section,$type);
+foreach $section ('Received','Delivered','Temp Rejects', 'Rejects','Ham','Spam') {
+ foreach $type ('Volume','Messages','Delayed','Failed','Hosts','Domains','Emails','Edomains') {
+ $report_totals{$section}{$type} = 0;
+ }
+}
+
+# Generate our parser.
+my $parser = generate_parser();
+
+
+
+if (@ARGV) {
+ # Scan the input files and collect the data
+ foreach my $file (@ARGV) {
+ if ($file =~ /\.gz/) {
+ unless (open(FILE,"gunzip -c $file |")) {
+ print STDERR "Failed to gunzip -c $file: $!";
+ next;
+ }
+ }
+ elsif ($file =~ /\.Z/) {
+ unless (open(FILE,"uncompress -c $file |")) {
+ print STDERR "Failed to uncompress -c $file: $!";
+ next;
+ }
+ }
+ else {
+ unless (open(FILE,$file)) {
+ print STDERR "Failed to read $file: $!";
+ next;
+ }
+ }
+ #Now parse the filehandle, updating the global variables.
+ parse($parser,\*FILE);
+ close FILE;
+ }
+}
+else {
+ #No files provided. Parse STDIN, updating the global variables.
+ parse($parser,\*STDIN);
+}
+
+
+if ($begin eq "9999-99-99 99:99:99" && ! $emptyOK) {
+ print STDERR "**** No valid log lines read\n";
+ exit 1;
+}
+
+# Output our results.
+print_header();
+print_grandtotals();
+
+# Print counts of user specified patterns if required.
+print_user_patterns() if @user_patterns;
+
+# Print rejection reasons.
+# print_rejects();
+
+# Print totals by transport if required.
+print_transport() if $show_transport;
+
+# Print the deliveries per interval as a histogram, unless configured not to.
+# First find the maximum in one interval and scale accordingly.
+if ($hist_opt > 0) {
+ print_histogram("Messages received", 'message', @received_interval_count);
+ print_histogram("Deliveries", 'delivery', @delivered_interval_count);
+}
+
+# Print times on queue if required.
+if ($#queue_times >= 0) {
+ print_duration_table("Time spent on the queue", "all messages", \@queue_times, \@qt_all_bin,$qt_all_overflow);
+ print_duration_table("Time spent on the queue", "messages with at least one remote delivery", \@queue_times, \@qt_remote_bin,$qt_remote_overflow);
+}
+
+# Print delivery times if required.
+if ($#delivery_times >= 0) {
+ print_duration_table("Delivery times", "all messages", \@delivery_times, \@dt_all_bin,$dt_all_overflow);
+ print_duration_table("Delivery times", "messages with at least one remote delivery", \@delivery_times, \@dt_remote_bin,$dt_remote_overflow);
+}
+
+# Print rcpt times if required.
+if ($#rcpt_times >= 0) {
+ foreach my $protocol ('all', grep(!/^all$/, sort keys %rcpt_times_bin)) {
+ print_duration_table("Receipt times", "$protocol messages", \@rcpt_times, $rcpt_times_bin{$protocol}, $rcpt_times_overflow{$protocol});
+ }
+}
+
+# Print relay information if required.
+print_relay() if $show_relay;
+
+# Print the league tables, if topcount isn't zero.
+if ($topcount > 0) {
+ my($ws_rej, $ws_top50, $ws_rej_row, $ws_top50_row, $ws_temp_rej, $ws_temp_rej_row);
+ $ws_rej_row = $ws_temp_rej_row = $ws_top50_row = 0;
+ if ($xls_fh) {
+ $ws_top50 = $workbook->addworksheet('Deliveries');
+ $ws_rej = $workbook->addworksheet('Rejections') if (%rejected_count_by_reason || %rejected_count_by_ip || %spam_count_by_ip);
+ $ws_temp_rej = $workbook->addworksheet('Temporary Rejections') if (%temporarily_rejected_count_by_reason || %temporarily_rejected_count_by_ip);
+ }
+
+ print_league_table("mail rejection reason", \%rejected_count_by_reason, undef, undef, undef, $ws_rej, \$ws_rej_row) if %rejected_count_by_reason;
+ print_league_table("mail temporary rejection reason", \%temporarily_rejected_count_by_reason, undef, undef, undef, $ws_temp_rej, \$ws_temp_rej_row) if %temporarily_rejected_count_by_reason;
+
+ foreach ('Host','Domain','Email','Edomain') {
+ next unless $do_sender{$_};
+ print_league_table("sending \l$_", $received_count{$_}, undef, $received_data{$_},$received_data_gigs{$_}, $ws_top50, \$ws_top50_row);
+ }
+
+ print_league_table("local sender", \%received_count_user, undef,
+ \%received_data_user,\%received_data_gigs_user, $ws_top50, \$ws_top50_row) if (($local_league_table || $include_remote_users) && %received_count_user);
+ foreach ('Host','Domain','Email','Edomain') {
+ next unless $do_sender{$_};
+ print_league_table("\l$_ destination", $delivered_messages{$_}, $delivered_addresses{$_}, $delivered_data{$_},$delivered_data_gigs{$_}, $ws_top50, \$ws_top50_row);
+ }
+ print_league_table("local destination", \%delivered_messages_user, \%delivered_addresses_user, \%delivered_data_user,\%delivered_data_gigs_user, $ws_top50, \$ws_top50_row) if (($local_league_table || $include_remote_users) && %delivered_messages_user);
+ print_league_table("local domain destination", \%delivered_messages_local_domain, \%delivered_addresses_local_domain, \%delivered_data_local_domain,\%delivered_data_gigs_local_domain, $ws_top50, \$ws_top50_row) if (($local_league_table || $include_remote_users) && %delivered_messages_local_domain);
+
+ print_league_table("rejected ip", \%rejected_count_by_ip, undef, undef, undef, $ws_rej, \$ws_rej_row) if %rejected_count_by_ip;
+ print_league_table("temporarily rejected ip", \%temporarily_rejected_count_by_ip, undef, undef, undef, $ws_rej, \$ws_rej_row) if %temporarily_rejected_count_by_ip;
+ print_league_table("non-rejected spamming ip", \%spam_count_by_ip, undef, undef, undef, $ws_rej, \$ws_rej_row) if %spam_count_by_ip;
+
+}
+
+# Print the error statistics if required.
+print_errors() if $show_errors;
+
+print $htm_fh "</body>\n</html>\n" if $htm_fh;
+
+
+$txt_fh->close if $txt_fh && ref $txt_fh;
+$htm_fh->close if $htm_fh;
+
+if ($xls_fh) {
+ # close Excel Workbook
+ $ws_global->set_first_sheet();
+ # FIXME: whyever - activate does not work :-/
+ $ws_global->activate();
+ $workbook->close();
+}
+
+
+# End of eximstats