From 0d47952611198ef6b1163f366dc03922d20b1475 Mon Sep 17 00:00:00 2001 From: Daniel Baumann Date: Wed, 17 Apr 2024 09:42:04 +0200 Subject: Adding upstream version 7.94+git20230807.3be01efb1+dfsg. Signed-off-by: Daniel Baumann --- scripts/http-robots.txt.nse | 111 ++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 111 insertions(+) create mode 100644 scripts/http-robots.txt.nse (limited to 'scripts/http-robots.txt.nse') diff --git a/scripts/http-robots.txt.nse b/scripts/http-robots.txt.nse new file mode 100644 index 0000000..bc5632a --- /dev/null +++ b/scripts/http-robots.txt.nse @@ -0,0 +1,111 @@ +local http = require "http" +local nmap = require "nmap" +local shortport = require "shortport" +local strbuf = require "strbuf" +local table = require "table" + +description = [[ +Checks for disallowed entries in /robots.txt on a web server. + +The higher the verbosity or debug level, the more disallowed entries are shown. +]] + +--- +--@output +-- 80/tcp open http syn-ack +-- | http-robots.txt: 156 disallowed entries (40 shown) +-- | /news?output=xhtml& /search /groups /images /catalogs +-- | /catalogues /news /nwshp /news?btcid=*& /news?btaid=*& +-- | /setnewsprefs? /index.html? /? /addurl/image? /pagead/ /relpage/ +-- | /relcontent /sorry/ /imgres /keyword/ /u/ /univ/ /cobrand /custom +-- | /advanced_group_search /googlesite /preferences /setprefs /swr /url /default +-- | /m? /m/? /m/lcb /m/news? /m/setnewsprefs? /m/search? /wml? +-- |_ /wml/? /wml/search? + + + +author = "Eddie Bell" +license = "Same as Nmap--See https://nmap.org/book/man-legal.html" +categories = {"default", "discovery", "safe"} + +portrule = shortport.http +local last_len = 0 + +-- split the output in 50 character length lines +local function buildOutput(output, w) + local nl + + if w:len() == 0 then + return nil + end + + -- check for duplicates + for i,v in ipairs(output) do + if w == v or w == v:sub(2, v:len()) then + return nil + end + end + + -- format lines + if last_len == 0 or last_len + w:len() <= 50 then + last_len = last_len + w:len() + nl = '' + else + last_len = 0 + nl = '\n' + end + + output = output .. (nl .. w) +end + +-- parse all disallowed entries in body and add them to a strbuf +local function parse_robots(body, output) + for line in body:gmatch("[^\r\n]+") do + for w in line:gmatch('[Dd]isallow:%s*(.*)') do + w = w:gsub("%s*#.*", "") + buildOutput(output, w) + end + end + + return #output +end + +action = function(host, port) + local dis_count, noun + local answer = http.get(host, port, "/robots.txt" ) + + if answer.status ~= 200 then + return nil + end + + local v_level = nmap.verbosity() + (nmap.debugging()*2) + local output = strbuf.new() + local detail = 15 + + dis_count = parse_robots(answer.body, output) + + if dis_count == 0 then + return + end + + -- verbose/debug mode, print 50 entries + if v_level > 1 and v_level < 5 then + detail = 40 + -- double debug mode, print everything + elseif v_level >= 5 then + detail = dis_count + end + + -- check we have enough entries + if detail > dis_count then + detail = dis_count + end + + noun = dis_count == 1 and "entry " or "entries " + + local shown = (detail == 0 or detail == dis_count) + and "\n" or '(' .. detail .. ' shown)\n' + + return dis_count .. " disallowed " .. noun .. + shown .. table.concat(output, ' ', 1, detail) +end -- cgit v1.2.3