summaryrefslogtreecommitdiffstats
path: root/samples
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-27 19:36:22 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-27 19:36:22 +0000
commitb799a1275c8e4bb4f1f07ce3e237708902cb8922 (patch)
treec75168eb236d7b16cda59ed6a27a937367d58dfe /samples
parentReleasing progress-linux version 2.9.0rel.0-2~progress7.99u1. (diff)
downloadlynx-b799a1275c8e4bb4f1f07ce3e237708902cb8922.tar.xz
lynx-b799a1275c8e4bb4f1f07ce3e237708902cb8922.zip
Merging upstream version 2.9.1.
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'samples')
-rwxr-xr-xsamples/mhtml-tool568
1 files changed, 568 insertions, 0 deletions
diff --git a/samples/mhtml-tool b/samples/mhtml-tool
new file mode 100755
index 0000000..aaf7f70
--- /dev/null
+++ b/samples/mhtml-tool
@@ -0,0 +1,568 @@
+#!/usr/bin/env perl
+
+use strict;
+use warnings;
+
+$| = 1;
+
+=pod
+
+=head1 NAME
+
+B<mhtml-tool> - Unpack a MIME HTML archive
+
+
+=head1 SYNOPSIS
+
+B<mhtml-tool> unpacks MIME HTML archives that some browsers (such as Opera)
+save by default. The file extensions of such archives are ".mht" or ".mhtml".
+
+The first HTML file in the archive is taken to be the primary web page, the
+other contained files for "page requisites" such as images or frames. The
+primary web page is written to the output directory (the current directory by
+default), the requisites to a subdirectory named after the primary HTML file
+name without extension, with "_files" appended. Link URLs in all HTML
+files referring to requisites are rewritten to point to the saved files.
+
+
+=head1 OPTIONS
+
+=over
+
+=item B<-e> program
+
+=item B<--exec> program
+
+Unpack the archive into a temporary directory and execute the given program
+in that directory, passing the primary HTML (first text/html entry) of the
+archive as a parameter.
+
+If B<-o> is given, use that directory location with the program.
+
+If the program is "lynx" or "(e)links(2)", this script uses the "-force-html"
+option to help when the entrypoint lacks a ".html" file-suffix.
+
+=item B<-h>
+
+=item B<-?>
+
+=item B<--help>
+
+Print a brief usage summary.
+
+=item B<-l>
+
+=item B<--list>
+
+List archive contents instead of unpacking. Four columns are output: file
+name, MIME type, size and URL. Unavailable entries are replaced by "(?)".
+
+=item B<-o> I<directory/ or name>
+
+=item B<--output> I<directory/ or name>
+
+If the argument ends in a slash or is an existing directory, unpack to that
+directory instead of current directory. Otherwise the argument is taken as a
+path to the file name to write the primary HTML file to. If the output
+directory does not exist, it is created.
+
+=back
+
+
+=head1 SEE ALSO
+
+https://github.com/ThomasDickey/mhtml-tool
+
+https://github.com/clapautius/mhtml-tool
+
+http://www.volkerschatz.com/unix/uware/unmht.html
+
+=head1 COPYLEFT
+
+B<unmht> is Copyright (c) 2012 Volker Schatz. It may be copied and/or
+modified under the same terms as Perl.
+
+B<mhtml-tool> is Copyright (c) 2018 Tudor M. Pristavu. It may be copied and/or
+modified under the same terms as Perl.
+
+B<mhtml-tool> is Copyright (c) 2024 Thomas E. Dickey. It may be copied and/or
+modified under the same terms as Perl.
+
+=cut
+
+use Cwd;
+use File::Path;
+use File::Copy;
+use File::Glob;
+use File::Temp qw/ tempdir /;
+use URI;
+use MIME::Base64;
+use MIME::QuotedPrint;
+use HTML::PullParser;
+use HTML::Tagset;
+use Getopt::Long;
+
+our $path_limit = 255;
+our %shorten_path;
+
+# RFC 7230 (obsoletes RFC 2616) indicates that URI length is limited by the
+# server line-length, which it suggests should be at least 8000 octets. The
+# problem is that components of pathnames might be limited to 255 bytes, while
+# the URI components can be much longer. As an additional complication, the
+# too-long components can be both directory- and file-names, and truncating
+# the components can produce collisions.
+sub shorten_path {
+ my $actual = $_[0];
+ if ( not defined $shorten_path{$actual} ) {
+ my @parts = split /\//, $actual;
+ for my $n ( 0 .. $#parts ) {
+ $parts[$n] = substr( $parts[$n], 0, $path_limit );
+ }
+ $shorten_path{$actual} = join "/", @parts;
+ }
+ return $shorten_path{$actual};
+}
+
+# URLs can include "&" and other characters which require quoting to use in
+# a shell command. We use system() rather than exec() to allow the temporary
+# directory to be cleaned up after running lynx.
+sub quoted {
+ my $param = $_[0];
+ $param =~ s/'/'\"'\"'/g;
+ return sprintf( "'%s'", $param );
+}
+
+# Add approriate ordinal suffix to a number.
+# -> Number
+# <- String of number with ordinal suffix
+sub ordinal {
+ return $_[0] . "th" if $_[0] > 3 && $_[0] < 20;
+ my $unitdig = $_[0] % 10;
+ return $_[0] . "st" if $unitdig == 1;
+ return $_[0] . "nd" if $unitdig == 2;
+ return $_[0] . "rd" if $unitdig == 3;
+ return $_[0] . "th";
+}
+
+{
+ my %taken;
+
+ # Find unique file name.
+ # -> Preferred file name, or undef
+ # MHT archive name (as a fallback if no name given)
+ # <- File name not conflicting with names returned by previous calls, but which
+ # may exist!
+ sub unique_name {
+ my ( $fname, $mhtname ) = @_;
+ my ( $trunc, $ext );
+
+ if ( defined $fname ) {
+ $fname =~ s/^\s+//;
+ $fname =~ s/\s+$//;
+ $taken{$fname} = 1, return $fname unless $taken{$fname};
+ ( $trunc, $ext ) = $fname =~ /^(.*?)(\.\w+)?$/;
+ $ext //= "";
+ }
+ else {
+ $trunc = $mhtname || "unpack";
+ $trunc =~ s/\.mht(?:ml?)?$//i;
+ $ext = "";
+ }
+ for my $suff ( 1 .. 9999 ) {
+ $fname = "${trunc}_$suff$ext";
+ $taken{$fname} = 1, return $fname unless $taken{$fname};
+ ++$suff;
+ }
+ return undef;
+ }
+
+}
+
+# Output error message and exit with return value 1.
+sub abort {
+ print STDERR "$_[0]\n";
+ exit 1;
+}
+
+# Generate output file directories and primary HTML file name depending on
+# --output option and primary HTML file name from archive. In case the primary
+# HTML file is not the first file in the archive, the secondary files directory
+# is renamed or the files moved on the second call, when the primary HTML file
+# name is known.
+# -> Value of --output option (or undef)
+# Primary HTML file name from MHT archive
+# Flag indicating if .._files subdirectory should be created
+# Hash reference to store resulting paths to
+sub mkfiledir {
+ my ( $outputopt, $firsthtmlname, $needfilesdir, $out ) = @_;
+
+ my $prevfilespath = $$out{filespath};
+ $firsthtmlname = "unpackmht-$$" unless defined $firsthtmlname;
+ if ( !defined $outputopt ) {
+ $$out{toppath} = ".";
+ }
+ elsif ( -d $outputopt || $outputopt =~ m!/$! ) {
+ $$out{toppath} = $outputopt;
+ }
+ else {
+ ( $$out{toppath}, $firsthtmlname ) = $outputopt =~ m!^(.*/)?([^/]+)$!;
+ abort "Empty output file name." unless defined $firsthtmlname;
+ $firsthtmlname .= ".html" unless $firsthtmlname =~ /\./;
+ $$out{toppath} = "." unless defined $$out{toppath};
+ }
+ $$out{toppath} =~ s!/$!!;
+ $$out{firstout} = "$$out{toppath}/$firsthtmlname";
+ $$out{filesdir} = $firsthtmlname;
+ $$out{filesdir} =~ s/\.[^.]+$//;
+ $$out{filesdir} = substr( $$out{filesdir}, 0, $path_limit - 6 );
+ $$out{filesdir} .= "_files";
+ $$out{filespath} = "$$out{toppath}/$$out{filesdir}";
+
+ if ( defined $prevfilespath ) {
+ return unless $prevfilespath ne $$out{filespath};
+ return unless -d $prevfilespath;
+ if ( !-d $$out{filespath} ) {
+ File::Copy::move( $prevfilespath, $$out{filespath} )
+ or abort "Could not rename secondary files directory.";
+ }
+ else {
+ for ( File::Glob::bsd_glob("$prevfilespath/*") ) {
+ File::Copy::move( $_, $$out{filespath} )
+ or abort "Could not move secondary files.";
+ }
+ }
+ }
+ else {
+ my $createall = $needfilesdir ? $$out{filespath} : $$out{toppath};
+ if ( !-d $createall ) {
+ File::Path::make_path($createall)
+ or abort "Could not create output directory $createall.";
+ }
+ }
+}
+
+my %opt;
+my @optdescr = ( 'exec|e=s', 'output|o=s', 'list|l!', 'help|h|?!', 'debug|d!' );
+my %config;
+
+my $status = GetOptions( \%opt, @optdescr );
+
+if ( !$status || $opt{help} ) {
+ print <<EOF;
+Usage: mhtml-tool [options] <MHT file>
+
+By default, mhtml-tool unpacks an MHT archive (an archive type saved by some
+browsers) to the current directory. The first HTML file in the archive is
+taken for the primary web page, and all other contained files are written to a
+directory named after that HTML file.
+
+Options:
+-e, --exec Execute the given program on a temporarily-unpacked archive
+-l, --list List archive contents (file name, MIME type, size and URL)
+-o, --output Unpack to directory <dir/> or to file <name>.html
+
+Use the command "pod2man mhtml-tool > mhtml-tool.1" or
+"pod2html mhtml-tool > mhtml-tool.html" to extract the manual.
+EOF
+ exit !$status;
+}
+
+my $origin = getcwd;
+my $tempdir = "";
+if ( $opt{exec} and not defined $opt{output} ) {
+ $tempdir = tempdir( CLEANUP => 1 );
+ $opt{output} = $tempdir . "/";
+}
+
+my $orig_sep = $/;
+my $crlf_file = 0;
+
+# Print a message to stdout (if debug is enabled).
+sub debug_msg {
+ if ( $opt{debug} ) {
+ print ":debug: $_[0]\n";
+ }
+}
+
+# Read next line. Remove line endings (both unix & windows style).
+sub read_next_line {
+ my $cur_line = <>;
+ chomp $cur_line;
+ if ( $cur_line =~ /\r$/ ) {
+ debug_msg("It's a CRLF file") if ( $crlf_file == 0 );
+ $crlf_file++;
+ }
+ $cur_line =~ s/\r//;
+ return $cur_line;
+}
+
+# Parse headers.
+# Handle multi-line headers (lines starting with spaces are part of a
+# multi-line header).
+# Read data from <> until the first empty line.
+# Set $crlf_file to 1 if it's a CRLF (windoze) file.
+sub parse_headers {
+ my %headers;
+ my $sep = $/;
+ $/ = $orig_sep;
+ my $cur_line = read_next_line;
+ my $full_line = $cur_line;
+
+ while ( not( $cur_line =~ /^$/ ) ) {
+ if ( $cur_line =~ /;$/ ) { # if a continued line...
+ while ( ( $cur_line = read_next_line ) =~ /^\h+/ ) {
+ $cur_line =~ s/^\h+//;
+ $full_line = $full_line . " " . $cur_line;
+ }
+ if ( $full_line =~ s/^([-\w]+): (.*)$// ) {
+ $headers{$1} = $2;
+ debug_msg("(1) New header $1 with value $2");
+ }
+ }
+ else {
+ if ( $full_line =~ s/^([-\w]+): (.*)$// ) {
+ $headers{$1} = $2;
+ debug_msg("(2) New header $1 with value $2");
+ }
+ }
+ $cur_line = read_next_line;
+ $full_line = $cur_line;
+ }
+ $/ = $sep;
+ return %headers;
+}
+
+my %global_headers = parse_headers();
+
+abort "Can't find Content-Type header - not a MIME HTML file?"
+ if ( !defined( $global_headers{'Content-Type'} ) );
+debug_msg("Global Content-Type: $global_headers{'Content-Type'}");
+
+my $boundary = '';
+my $endcode = '';
+
+# FIXME: - add other possible mime types
+# Types implemented so far are:
+# Content-Type: multipart/related; boundary="----=_NextPart_01D8BCC9.47C2B260"
+# Content-Type: text/html; charset="utf-8"
+if ( $global_headers{'Content-Type'} =~
+ m!multipart/related;.*\h+boundary="?([^"]*)"?$! )
+{
+ $endcode = $boundary = $1;
+ $endcode =~ s/\s+$//;
+ if ($crlf_file) {
+ $/ = "\r\n--$boundary\r\n";
+ }
+ else {
+ $/ = "\n--$boundary\n";
+ }
+ debug_msg("Boundary: $boundary");
+}
+elsif ( $global_headers{'Content-Type'} =~ m!text/html! ) {
+ $/ = '';
+}
+else {
+ die "Error: Unknown Content-Type: $global_headers{'Content-Type'}\n";
+}
+
+my %by_url;
+my @htmlfiles;
+my $fh;
+
+{
+ my $fileind = 1;
+ my $leading = "--$boundary\n";
+ while ( defined( my $data = <> ) ) {
+ chomp $data;
+ $data =~ s/\R/\n/g; # handle various other line-endings (windows, mac)
+ while ( index( $data, $leading ) == 0 ) {
+ $data = substr( $data, length($leading) );
+ }
+ my %headers;
+ while ( $data =~ s/^([-\w]+): (.*)\n// ) {
+ $headers{$1} = $2;
+ debug_msg("New header $1 with value $2");
+
+ # read (and ignore atm) lines starting with space (multi-line headers)
+ while ( $data =~ s/^\h+.*\n// ) {
+ debug_msg("empty line");
+ }
+ }
+ if ( scalar(%headers) == 0 ) {
+ %headers = %global_headers
+ ; # fallback to the global headers as needed, usually for "text"
+ }
+ $data =~ s/^\n//;
+ $data =~ s/\n--$endcode--\r?\n$/\n/s;
+ my ( $type, $origname );
+ if ( defined( $headers{"Content-Type"} ) ) {
+ ($type) = $headers{"Content-Type"} =~ /^(\w+\/\w+)\b/;
+ debug_msg("type=$type");
+ }
+ else {
+ print "Error: No Content-Type found, skipping chunk\n";
+
+ # we could print the bad chunk here...
+ # but it's probably just a "warning" about
+ # "if you see this your software isn't recognizing a web archive file"
+ next;
+ }
+ if ( defined( $headers{"Content-Type"} )
+ && $headers{"Content-Type"} =~ /\bname=([^;]*)/ )
+ {
+ $origname = $1;
+ ($type) = $headers{"Content-Type"} =~ /^(\w+\/\w+)\b/;
+ $type //= "";
+ }
+ elsif ( defined( $headers{"Content-Disposition"} )
+ && $headers{"Content-Disposition"} =~ /\bfilename=([^;]*)/ )
+ {
+ $origname = $1;
+ if ( !defined($type) ) {
+ $type = $origname =~ /\.html?$/i ? "text/html" : "";
+ }
+ }
+ elsif ( defined( $headers{"Content-Location"} ) ) {
+ $origname = $headers{"Content-Location"};
+
+ # for unknown reasons, files generated by IE11 may contain some
+ # local paths with '\' instead of '/'
+ if ( $origname =~ m!^file://.*\\! ) {
+ debug_msg("Windows-style path detected: $origname");
+ $origname =~ s!^.*\\!!;
+ }
+ $origname =~ s!^.*/!!;
+ if ( !defined($type) ) {
+ $type = "";
+ }
+ }
+
+ $origname = "noname" unless defined $origname;
+ $origname = "noname" if ( $origname eq "" );
+ my $fname = unique_name( $origname, $ARGV[0] );
+ if ( !defined( $headers{"Content-Transfer-Encoding"} ) ) {
+ print STDERR "Info: Encoding of ", ordinal($fileind),
+ " file not found - leaving as-is.\n";
+ }
+ elsif ( $headers{"Content-Transfer-Encoding"} =~ /\bbase64\b/i ) {
+ $data = MIME::Base64::decode($data);
+ }
+ elsif (
+ $headers{"Content-Transfer-Encoding"} =~ /\bquoted-printable\b/i )
+ {
+ $data = MIME::QuotedPrint::decode($data);
+ }
+ debug_msg("origname=$origname; fname=$fname; type=$type");
+ debug_msg( "Content-Type: " . $headers{"Content-Type"} );
+ debug_msg( "Data size: " . length($data) );
+ if ( $opt{list} ) {
+ $origname =~ s/\s+$// if defined $origname;
+ my $size = length($data);
+ print $fname // "(?)", "\t", $type || "(?)", "\t$size\t",
+ $headers{"Content-Location"} // "(?)", "\n";
+ next;
+ }
+ $headers{fname} = $fname;
+ if ( $headers{"Content-Location"} ) {
+ $headers{url} = $headers{"Content-Location"};
+ $headers{url} =~ s/\s+$//;
+ $by_url{ $headers{url} } = \%headers;
+ debug_msg("Content-Location: $headers{url}");
+ }
+ else {
+ debug_msg("?? no location");
+ }
+ if ( $type eq "text/html" ) {
+ $headers{data} = $data;
+ if ( scalar @htmlfiles == 0 ) { # first html file must have a name
+ if ( !$headers{fname} ) {
+ $headers{fname} = "index.html";
+ }
+ }
+ push @htmlfiles, \%headers;
+ debug_msg("New html file in list: $headers{fname}");
+ }
+ else {
+ mkfiledir( $opt{output}, $htmlfiles[0]->{fname}, 1, \%config );
+ $fname = shorten_path "$config{filespath}/$fname";
+ open $fh, ">$fname" or abort "Could not create file $fname.";
+ print $fh $data;
+ close $fh;
+ }
+ }
+ continue {
+ debug_msg("");
+ ++$fileind;
+ }
+}
+
+if ( $opt{list} ) {
+ exit(0);
+}
+
+mkfiledir( $opt{output}, $htmlfiles[0]->{fname}, 0, \%config );
+
+my $entrypoint = ".";
+my $filesprefix = $config{filesdir} . "/";
+my $outname = $config{firstout};
+print "primary html output name: $outname\n";
+
+for my $html (@htmlfiles) {
+ my $linksubst = "";
+ my $p = HTML::PullParser->new(
+ doc => \$html->{data},
+ "start" => 'text, attr, tagname',
+ "text" => 'text',
+ "end" => 'text'
+ );
+ while ( defined( my $tok = $p->get_token() ) ) {
+ my $linkary;
+ my @linkattrs;
+ if ( ref( $tok->[1] )
+ && ( $linkary = $HTML::Tagset::linkElements{ $tok->[2] } )
+ && ( @linkattrs = grep $tok->[1]->{$_}, @$linkary ) )
+ {
+ for my $attr (@linkattrs) {
+ my $uri = URI->new( $tok->[1]->{$attr} );
+ next unless defined $uri;
+ next unless defined $html->{url};
+ $uri = $uri->abs( $html->{url} );
+ $tok->[1]->{$attr} =
+ "$filesprefix" . $by_url{ $uri->as_string() }->{fname}
+ if $by_url{ $uri->as_string() };
+ }
+ delete $tok->[1]->{"/"};
+ $linksubst .=
+ "<$tok->[2] "
+ . join( " ",
+ map( "$_=\"$tok->[1]->{$_}\"", keys %{ $tok->[1] } ) )
+ . ">";
+ }
+ else {
+ $linksubst .= $tok->[0];
+ }
+ }
+ $outname = "$config{filespath}/$html->{fname}" unless defined $outname;
+ $outname = shorten_path $outname;
+ open $fh, ">$outname" or abort "Could not create file $outname.";
+ print $fh $linksubst;
+ close $fh;
+ $entrypoint = $outname if ( $entrypoint eq "." );
+
+ # for all except the first HTML file:
+ $filesprefix = "";
+ $outname = undef;
+}
+
+if ( $tempdir ne "" ) {
+ chdir $tempdir;
+ my $enforce = "";
+ $enforce = "-force-html"
+ if ( $entrypoint ne "."
+ and $entrypoint !~ /\.htm(l)?$/
+ and $opt{exec} =~ /(lynx|((e)?links(2)?))/ );
+ system( sprintf( "%s %s %s", $opt{exec}, $enforce, quoted $entrypoint ) );
+ chdir $origin;
+}
+
+1;