GitBucket
4.21.2
Toggle navigation
Snippets
Sign in
Files
Branches
1
Releases
Issues
Pull requests
Labels
Priorities
Milestones
Wiki
Forks
nigel.stanger
/
Digital_Repository
Browse code
- Fixed variable scoping problems.
master
1 parent
b998009
commit
5f735514ff284f19e04e9ff4ab3ce4938d25b760
nstanger
authored
on 24 Jul 2007
Patch
Showing
1 changed file
Repositories/statistics/scripts/eprints-usage_src.php
Ignore Space
Show notes
View
Repositories/statistics/scripts/eprints-usage_src.php
<?php /* NJS 2007-07-24 The database structure changed between versions 2.x and 3.x of EPrints, so we now need to check the major version number and alter the queries appropriately. Use only the MAJOR version number (i.e., 2 or 3, don't include the release number). */ $eprints_version = 3; /* NJS 2006-04-28 In earlier versions of this script, which eprints to count was determined by comparing the request date of the eprint against the "lastproc" date of this script (i.e., minimum time unit one day). This was fine if you only ran the script once per day, but if you ran it more than that, it counted multiple times requests whose $request_date == $lastproc. For example, if you ran this script five times per day, all the downloads that occurred during that day would be counted EVERY TIME this script ran, thus overinflating your stats by a factor of up to five :( The solution is to use the full time stamp for comparison rather than just the date. This timestamp MUST include time zone information so that things don't get screwed up by daylight saving time. As long as this is done consistently, there's no need to do things like convert to GMT, for example. The very first thing we need to do is grab the current time stamp with time zone, which will later be stored in the database as the "lastproc" time. This needs to happen first so that we don't "lose" any requests that occur while the script is running. */ $start_time = date('Y-m-d H:i:s O'); /* NJS 2007-01-30 A further twist! The original script ignored log lines that had a date falling before $lastproc, i.e., if log line date < $lastproc then it's already been dealt with. This is all fine. However, it didn't bother checking for log lines that were written after the script started running (i.e. log line date >= $start_time). Why is this a problem? We're reading the live Apache log file, so it's quite likely that new lines will be written to it after the script has started (i.e., after $start_time). Suppose $start_time is '2006-06-15 14:03:15 +1200', $lastproc is '2006-06-15 12:03:15 +1200' (i.e., the script is run every two hours) and the log file contains lines with the following dates: '2006-06-15 10:03:15 +1200' [1] <-- written before $lastproc '2006-06-15 12:03:14 +1200' [2] <-- written before $lastproc '2006-06-15 13:03:15 +1200' [3] <-- written before $start_time '2006-06-15 14:03:14 +1200' [4] <-- written before $start_time '2006-06-15 14:03:15 +1200' [5] <-- written at $start_time '2006-06-15 14:03:16 +1200' [6] <-- written after $start_time During this run, dates [1] and [2] are both < $lastproc and thus ignored. The remaining four dates ([4]--[6]) are >= $lastproc and thus processed. Two hours later, the script runs again, this time with $start_time set to '2006-06-15 16:03:15 +1200' and $lastproc to '2006-06-15 14:03:15 +1200'. Dates [1] through [4] are all < $lastproc and thus ignored. However, dates [5] and [6] are both >= $lastproc and are processed a second time, resulting in a duplicate entry in the database. The solution is to ignore any log line entries that occur at or after (>=) $start_time. In the example above, this would mean that in the first run, dates [1], [2], [5] and [6] would be ignored and dates [3] and [4] processed. In the second run, dates [1]--[4] would be ignored and dates [5] and [6] processed. */ $test_starttime = strtotime($start_time); // NJS 2005-12-09 Switched to GeoIP from GeoIP:IPfree. include("geoip.inc"); $gi = geoip_open("##GEOIP_DATABASE##",GEOIP_STANDARD); /* Apache log for ePrints uses this format: LogFormat "%h %l %u %t \"%r\" %>s %b \"%{Referer}i\" \"%{User-Agent}i\"" combined If the log format differs the regular expression matching would need to be adjusted. Parse: ip date YYYY MM DD archive ID */ // Web server log files $log_dir = '##APACHE_LOG_LOCATION##'; $log_file = array( 'otago_eprints' => '##APACHE_LOG_NAME_1##', 'cardrona' => '##APACHE_LOG_NAME_2##', ); // eprintstats db $sqlserver = 'localhost'; $sqluser = 'eprintstatspriv'; $sqlpass = 'AuldGrizzel'; $sqldatabase = 'eprintstats'; /* NJS 2006-05-26 SQL details of your ePrints installation(s). This has now been generalised to work with multiple archives. For each archive that you have, add an entry to this array in the following format: 'archive_name' => array( 'sqlserver' => 'db_host', 'username' => 'archive_name', 'password' => 'password', ), */ $eprintsdbs = array( 'otago_eprints' => array( 'sqlserver' => 'localhost', 'username' => 'otago_eprints', 'password' => 'DrSyntaxRidesAgain', ), 'cardrona' => array( 'sqlserver' => 'localhost', 'username' => 'cardrona', 'password' => 'chautquau', ), ); /* NJS 2005-12-16 IP address ranges for your local Intranet(s). You can have multiple ranges of IP addresses, each with a different "country name", so that they will appear as separate entries in the by country stats pages. You should use a different country code for each range (ISO 3166-1 specifies the range XA through XZ as "user-assignable", so you can use codes from there as necessary), and create flag icons as appropriate. Each address range key is the name that will appear in the statistics database (the "country name"), followed by a comma, followed by the appropriate ISO 3166-1 country code as noted above. Each entry in the range is either a single IP address, or an array specifying a lower and upper bound for a contiguous IP address range (see example below). All IP addresses must be converted to long values using the ip2long() function before being stored. Note that address ranges may overlap. The script will use the first range that matches a given IP, so list the ranges in the correct order of precedence for your needs. Example: $local_IPs = array( 'Repository Admin,XA' => array( ip2long('192.168.1.5'), ip2long('192.168.1.22'), array( lower => ip2long('192.168.1.30'), upper => ip2long('192.168.1.35'), ), ), 'Our Intranet,XI' => array( array( lower => ip2long('192.168.1.0'), upper => ip2long('192.168.255.255'), ), ), ); 'Repository Admin' covers the IP addresses 192.168.1.5, 192.168.1.22 and the range 192.168.1.30 to 192.168.1.35, inclusive. 'Our Intranet' covers the range 192.168.1.0 to 192.168.255.255, inclusive. A machine will only match the 'Our Intranet' range if it first fails to match the 'Repository Admin' range. */ $local_IPs = array( 'Repository Admin,XA' => array( ip2long('139.80.75.110'), // Nigel @ Uni ip2long('60.234.209.74'), // Nigel @ home ip2long('139.80.92.138'), // Monica & Jeremy ip2long('139.80.92.151'), // @ Uni ip2long('203.89.162.155'), // Monica @ home ip2long('139.80.81.50'), // eprints.otago.ac.nz ip2long('172.20.1.50'), // eprints.otago.ac.nz pre-switch ip2long('172.20.1.1'), // eprints.otago.ac.nz pre-switch ), 'Otago Intranet,XI' => array( array( 'lower' => ip2long('139.80.0.0'), 'upper' => ip2long('139.80.127.255'), ), ), ); /* NJS 2007-01-26 Patterns to match various search engine bots. Ideally, we'd use a similar mechanism to the $local_IPs variable above, but this isn't feasible because we'd need to know the IP ranges for the likes of Google, for example. This clearly isn't possible in practice. Fortunately, most search bots insert a readily identifiable string into the user-agent part of the HTTP response, which gets recorded in the Apache log file. We can look for these and re-code log entries as appropriate. The format of this list is similar to that of the $local_IPs variable. The key is the "country name" (in this case the name of the search engine) plus a non-standard four-character country code starting with "X@", separated by a comma. Each key value has an associated list of corresponding regular expressions that can occur in the user-agent part of the Apache log entry. If any one of these REs matches the user-agent part of the log entry, then we should re-code the country appropriately. A four-character code is used because that what the database allows, and it avoids having to reserve several of the "X" country codes for search engines. */ $bot_patterns = array( // Yahoo! (http://www.yahoo.com/) 'Yahoo!,X@YH' => array( '/yahoo! slurp/i', '/yahooseeker/i', ), // Windows Live Search (http://search.msn.com/) 'Windows Live Search,X@MS' => array( '/msnbot/i', ), // Google (http://www.google.com/) 'Google,X@GG' => array( '/googlebot/i', ), // Ask.com (http://www.ask.com/) 'Ask.com,X@AC' => array( '/ask jeeves\/teoma/i', ), // Everything else I could find in our log files :) 'Other search engine,X@OS' => array( // TAMU Internet Research Lab (http://irl.cs.tamu.edu/) '/http:\/\/irl\.cs\.tamu\.edu\/crawler/i', // Alexa web search (http://www.alexa.com/) '/ia_archiver/i', // TrueKnowledge for Web (http://www.authoritativeweb.com/) '/converacrawler/i', // Majestic 12 distributed search engine (http://www.majestic12.co.uk/) '/mj12bot/i', // Picsearch (http://www.picsearch.com/) '/psbot/i', // Exalead (http://www.exalead.com/search) '/exabot/i', // Cazoodle (note cazoodle.com doesn't exist) '/cazoodlebot crawler/i', '/mqbot@cazoodle\.com/i', // Gigablast (http://www.gigablast.com/) '/gigabot/i', // Houxou (http://www.houxou.com/) '/houxoucrawler/i', '/crawler at houxou dot com/i', // IBM Almaden Research Center Computer Science group (http://www.almaden.ibm.com/cs/) '/http:\/\/www\.almaden\.ibm\.com\/cs\/crawler/i', // Goo? (http://help.goo.ne.jp/) '/ichiro/i', // Daum Communications Corp (Korea) '/edacious & intelligent web robot/i', '/daum communications corp/i', '/daum web robot/i', '/msie is not me/i', '/daumoa/i', // Girafa (http://www.girafa.com/) '/girafabot/i', // The Generations Network (http://www.myfamilyinc.com/) '/myfamilybot/i', // Naver? (http://www.naver.com/) '/naverbot/i', // WiseNut (http://www.wisenutbot.com/) '/zyborg/i', '/wn-[0-9]+\.zyborg@looksmart\.net/i', // Accelobot (http://www.accelobot.com/) // This one seems particularly busy! '/heritrix/i', // Seeqpod (http://www.seeqpod.com/) '/seeqpod-vertical-crawler/i', // University of Illinois at Urbana-Champaign, Computer Science (http://www.cs.uiuc.edu/) '/mqbot crawler/i', '/mqbot@cs\.uiuc\.edu/i', // Microsoft Research (http://research.microsoft.com/) '/msrbot/i', // Nusearch '/nusearch spider/i', // SourceForge (http://www.sf.net/) '/nutch-agent@lists\.sourceforge\.net/i', // Lucene (http://lucene.apache.org/) '/nutch-agent@lucene\.apache\.org/i', '/raphael@unterreuth.de/i', // Computer Science, University of Washington (http://cs.washington.edu/) '/nutch running at uw/i', '/sycrawl@cs\.washington\.edu/i', // Chikayama & Taura Laboratory, University of Tokyo (http://www.logos.ic.i.u-tokyo.ac.jp/) '/shim-crawler/i', '/crawl@logos\.ic\.i\.u-tokyo\.ac\.jp/i', // Sproose (http://www.sproose.com/) '/sproose bot/i', '/crawler@sproose\.com/i', // Turnitin (http://www.turnitin.com/) '/turnitinbot/i', // WISH Project (http://wish.slis.tsukuba.ac.jp/) '/wish-project/i', // WWWster '/wwwster/i', '/gue@cis\.uni-muenchen\.de/i', // Forex Trading Network Organization (http://www.netforex.org/) '/forex trading network organization/i', '/info@netforex\.org/i', // FunnelBack (http://www.funnelback.com/) '/funnelback/i', // Baidu (http://www.baidu.com/) '/baiduspider/i', // Brandimensions (http://www.brandimensions.com/) '/bdfetch/i', // Blaiz Enterprises (http://www.blaiz.net/) '/blaiz-bee/i', // Boitho/SearchDaimon (http://www.boitho.com/ or http://www.searchdaimon.com/) '/boitho\.com-dc/i', // Celestial (OAI aggregator, see http://oai-perl.sourceforge.net/ for a little info) '/celestial/i', // Cipinet (http://www.cipinet.com/) '/cipinetbot/i', // iVia (http://ivia.ucr.edu/) '/crawlertest crawlertest/i', // Encyclopedia of Keywords (http://keywen.com/) '/easydl/i', // Everest-Vulcan Inc. (http://everest.vulcan.com/) '/everest-vulcan inc/i', // FactBites (http://www.factbites.com/) '/factbot/i', // Scirus (http://www.scirus.com/) '/scirus scirus-crawler@fast\.no/i', // UOL (http://www.uol.com.br/) '/uolcrawler/i', '/soscrawler@uol\.com\.br/i', // Always Updated (http://www.updated.com/) '/updated crawler/i', '/crawler@updated\.com/i', // FAST Enterprise Search (http://www.fast.no/) '/fast metaweb crawler/i', '/crawler@fast\.no/i', '/helpdesk at fastsearch dot com/i', // Deutsche Wortschatz Portal (http://wortschatz.uni-leipzig.de/) '/findlinks/i', // Gais (http://gais.cs.ccu.edu.tw/) '/gaisbot/i', '/robot[0-9]{2}@gais.cs.ccu.edu.tw/i', // http://ilse.net/ '/ingrid/i', // Krugle (http://corp.krugle.com/) '/krugle\/krugle/i', '/krugle web crawler/i', '/webcrawler@krugle\.com/i', // WebWobot (http://www.webwobot.com/) '/scollspider/i', // Omni-Explorer (http://www.omni-explorer.com/) '/omniexplorer_bot/i', '/worldindexer/i', // PageBull (http://www.pagebull.com/) '/pagebull http:\/\/www\.pagebull\.com\//i', // dir.com (http://dir.com/) '/pompos/i', // Sensis (http://sensis.com.au/) '/sensis web crawler/i', '/search_comments\\\\at\\\\sensis\\\\dot\\\\com\\\\dot\\\\au/i', // Shopwiki (http://www.shopwiki.com/) '/shopwiki/i', // Guruji (http://www.terrawiz.com/) '/terrawizbot/i', // Language Observatory Project (http://www.language-observatory.org/) '/ubicrawler/i', // MSIE offline bookmarks crawler '/msiecrawler/i', // Unidentified '/bot/i', '/crawler/i', '/spider/i', '/larbin/i', // also larbinSpider '/httrack/i', '/voyager/i', '/acadiauniversitywebcensusclient/i', '/feedchecker/i', '/knowitall\(knowitall@cs\.washington\.edu\)/i', '/mediapartners-google/i', '/psycheclone/i', '/topicblogs/i', '/nutch/i', ), ); ########################################### ## ## No configuration required below here. ## ########################################### $connect = mysql_pconnect ($sqlserver,$sqluser,$sqlpass); $db = mysql_select_db($sqldatabase,$connect) or die("Could not connect"); // First get the date of last update /* NJS 2006-04-28 Changed this from order by timeinsert to order by id. The ID is always guaranteed to increase temporally, but is otherwise time-independent and thus not affected by things like daylight savings. */ $query = "SELECT lastproc FROM lastproc ORDER BY id DESC LIMIT 1"; $result = mysql_query($query,$connect); $num_rows = mysql_num_rows($result); if ($num_rows > 0) { $row = mysql_fetch_assoc($result); $lastproc = $row["lastproc"]; // NJS 2007-01-30 Refactored $databaseA to more meaningful $test_lastproc. $test_lastproc = strtotime($lastproc); } else { $test_lastproc = 0; } // NJS 2006-06-14: Generalised connection list for multiple archives. $eprints_connections = array(); foreach ($eprintsdbs as $archive_name => $details) { $eprints_connections[$archive_name] = mysql_connect($details['sqlserver'],$details['username'],$details['password']); } $counter = 0; foreach($log_file as $archive_name=>$archive_log) { $logf = $log_dir . $archive_log; $handle = fopen($logf, "r"); while (!feof($handle)) { $buffer = fgets($handle, 4096); /* NJS 2007-01-26 Added user-agent match to all regexps to enable bot detection. NJS 2007-01-31 Refactored regexps from four down to one, after realising that (a) long EPrints URLs are a superset of the short ones, and (b) a regexp that matches domain names works just as well for IP addresses (the GeoIP lookup doesn't care which it gets). Also fixed the pattern so it can handle an arbitrary number of subdomains. Note that the latter would be the main argument for keeping a separate IP address pattern, as IP addresses always comprise exactly four parts. However, it's not really up to the script to verify IP addresses; Apache should be recording them correctly in the first place! The typical kinds of strings we are matching look something like this: fetch abstract (short, long): 168.192.1.1 - - [31/Jan/2007:09:15:36 +1300] "GET /1/ HTTP/1.1" 200 12345 "referer" "user-agent" 168.192.1.1 - - [31/Jan/2007:09:15:36 +1300] "GET /archive/00000001/ HTTP/1.1" 200 12345 "referer" "user-agent" download item (short, long): 168.192.1.1 - - [31/Jan/2007:09:15:37 +1300] "GET /1/01/foo.pdf HTTP/1.1" 200 12345 "referer" "user-agent" 168.192.1.1 - - [31/Jan/2007:09:15:37 +1300] "GET /archive/00000001/01/foo.pdf HTTP/1.1" 200 12345 "referer" "user-agent" Plus any of the above with a domain name substituted for the IP address (e.g., foo.bar.com instead of 168.192.1.1). */ if (preg_match("/^(\S+(?:\.\S+)+) - - \[(.*?)\] \"GET \/(?:archive\/0{1,8})?(\d{1,4}).*? HTTP\/1..\" 200 .*?(\"[^\"]+\")?$/i",$buffer,$matches)) { $counter++; $country_code = ''; $country_name = ''; $insertid = ''; $eprint_name = ''; $view_type = ''; $uniquebits = ''; /* NJS 2007-01-29 Moved date checking to the start of the loop, as there's no point in doing any of the regexp checks if we've already processed this log entry and will discard it anyway. */ $date = $matches[2]; /* NJS 2006-04-28 Switched to timestamp rather than date-based comparison. First, clean up the Apache request date into something that strtotime understands. Note that the Apache log dates include time zone info by default. */ $date = preg_replace("/:/"," ",$date,1); // Change first ":" to " ". $date = preg_replace("/\//", " ", $date); // Change all "/" to " ". // NJS 2007-01-30 Refactored $databaseB to more meaningful // $test_logdate. $test_logdate = strtotime($date); // NJS 2007-01-30 Added test for log dates >= $start_time. if ( ( $test_logdate < $test_lastproc ) || ( $test_logdate >= $test_starttime ) ) continue; // Convert to properly formatted date string. $request_date = date('Y-m-d H:i:s O', $test_logdate); /* NJS 2005-12-16 Determine country code and name. Check whether the IP number falls into any of the local intranet ranges. If so, then use that. */ $ip = $matches[1]; $ip_long = ip2long($ip); $found_country = FALSE; foreach ($local_IPs as $id => $addresses) { foreach ($addresses as $ip_range) { if (is_array($ip_range)) // check against lower/upper bounds { $found_country = (($ip_long >= $ip_range['lower']) && ($ip_long <= $ip_range['upper'])); break; } else if (is_long($ip_range)) // data type sanity check { $found_country = ($ip_long == $ip_range); break; } else // something is seriously broken, ignore this entry { print "Unsupported data type " . gettype($ip_range) . " (value " . $ip_range . ") in \$local_IPs (expected long).\n"; continue; } } if ($found_country) { list($country_name, $country_code) = explode(',', $id); break; } } // Otherwise, fall back to GeoIP. if (!$found_country) { $country_code = geoip_country_code_by_addr($gi, $ip); $country_name = geoip_country_name_by_addr($gi, $ip); } // end NJS 2005-12-16 /* NJS 2007-01-26 Check whether this is a bot reference. */ $user_agent = $matches[4]; $found_country = FALSE; foreach ($bot_patterns as $id => $patterns) { foreach ($patterns as $pat) { if (preg_match($pat, $user_agent)) { $found_country = TRUE; break; } } if ($found_country) { list($country_name, $country_code) = explode(',', $id); break; } } // end NJS 2007-01-26 // Now sort out the remaining bits and we're done. $eprint_id = $matches[3]; $uniquebits = $buffer; // NJS 2005-11-25 Added regexp for EPrints short URLs. // NJS 2007-01-31 Refactored into one regexp for both styles. if (preg_match("/GET \/(?:archive\/0{1,8})?\d{1,4}\/\d\d\//i",$buffer)) { $view_type = "download"; } else { $view_type = "abstract"; } if(isset($eprintname[$archive_name . $eprint_id])) { $eprint_name = $eprintname[$archive_name . $eprint_id]; } else { $eprint_name = getePrintName($eprints_connections[$archive_name],$archive_name,$eprint_id,$eprints_version); $eprintname[$archive_name . $eprint_id] = $eprint_name; } if($eprint_name=='') { // Do nothing. } else { $eprint_name = mysql_escape_string($eprint_name); /* NJS 2006-04-25 Requests containing apostrophes (') are dumped by MySQL unless we escape them. Looking in the GeoIP files I also see country names with apostrophes, so escape that as well. Everything else should be fine. */ $uniquebits = mysql_escape_string($uniquebits); $country_name = mysql_escape_string($country_name); // end NJS 2006-04-25 $query = " INSERT INTO view (uniquebits,archive_name,ip,request_date,archiveid,country_code,country_name,view_type,eprint_name) VALUES('".$uniquebits."','".$archive_name."','".$ip."','".$request_date."',".$eprint_id.",'".$country_code."','".$country_name."','".$view_type."','".$eprint_name."')"; $result = mysql_query($query,$connect); $insertid = mysql_insert_id($connect); } } else { // print "NO match" . "\n"; } } fclose($handle); } /* Keep track of where we are. Should avoid duplication of results if the script is run more than once on the same log file. */ // NJS 2006-04-28 Switched value inserted to $start_time instead of $request_date. $query = "INSERT into lastproc (lastproc) values('".$start_time."')"; $result = mysql_query($query,$connect); #print "Records counted: $counter\n"; #print "Last count: $request_date\n"; foreach ($eprints_connections as $connection) { mysql_close($connection); } mysql_close($connect); // Look up the title corresponding to the specified eprint id. function getePrintName($connection,$archive,$eprintid,$eprints_version) { // NJS 2006-06-14: DB connection now passed as an argument. $db = mysql_select_db($archive,$connection); // NJS 2007-07-24: Added check for EPrints version, as the // database structure changed between versions 2 and 3. if ( $eprints_version > 2 ) { $query3 = " SELECT title FROM eprint WHERE eprintid = $eprintid AND eprint_status = 'archive' "; } else { $query3 = " SELECT title FROM archive WHERE eprintid = $eprintid "; } $result3 = mysql_query($query3,$connection); $title = ''; $suffix = ''; // NJS 2006-04-25 Added check for empty result, probably a deleted item. // Look in the deleted items for details. if (mysql_num_rows($result3) == 0) { // NJS 2007-07-24: Added check for EPrints version, as the // database structure changed between versions 2 and 3. if ( $eprints_version > 2 ) { $query3 = " SELECT title FROM eprint WHERE eprintid = $eprintid AND eprint_status = 'deletion' "; } else { $query3 = " SELECT title FROM archive WHERE eprintid = $eprintid "; } $result3 = mysql_query($query3,$connection); // If it's not in deletion, then we have no clue what it is. if (mysql_num_rows($result3) == 0) { $title = "Unknown item [$eprintid]"; } else { $suffix = ' [deleted]'; } } if ($title == '') { $row = mysql_fetch_assoc($result3); $row["title"] = trim($row["title"]); $row["title"] = preg_replace("/\s+/"," ",$row["title"]); $title = $row["title"]; } return $title . $suffix; } ?>
<?php /* NJS 2007-07-24 The database structure changed between versions 2.x and 3.x of EPrints, so we now need to check the major version number and alter the queries appropriately. Use only the MAJOR version number (i.e., 2 or 3, don't include the release number). */ $eprints_version = 3; /* NJS 2006-04-28 In earlier versions of this script, which eprints to count was determined by comparing the request date of the eprint against the "lastproc" date of this script (i.e., minimum time unit one day). This was fine if you only ran the script once per day, but if you ran it more than that, it counted multiple times requests whose $request_date == $lastproc. For example, if you ran this script five times per day, all the downloads that occurred during that day would be counted EVERY TIME this script ran, thus overinflating your stats by a factor of up to five :( The solution is to use the full time stamp for comparison rather than just the date. This timestamp MUST include time zone information so that things don't get screwed up by daylight saving time. As long as this is done consistently, there's no need to do things like convert to GMT, for example. The very first thing we need to do is grab the current time stamp with time zone, which will later be stored in the database as the "lastproc" time. This needs to happen first so that we don't "lose" any requests that occur while the script is running. */ $start_time = date('Y-m-d H:i:s O'); /* NJS 2007-01-30 A further twist! The original script ignored log lines that had a date falling before $lastproc, i.e., if log line date < $lastproc then it's already been dealt with. This is all fine. However, it didn't bother checking for log lines that were written after the script started running (i.e. log line date >= $start_time). Why is this a problem? We're reading the live Apache log file, so it's quite likely that new lines will be written to it after the script has started (i.e., after $start_time). Suppose $start_time is '2006-06-15 14:03:15 +1200', $lastproc is '2006-06-15 12:03:15 +1200' (i.e., the script is run every two hours) and the log file contains lines with the following dates: '2006-06-15 10:03:15 +1200' [1] <-- written before $lastproc '2006-06-15 12:03:14 +1200' [2] <-- written before $lastproc '2006-06-15 13:03:15 +1200' [3] <-- written before $start_time '2006-06-15 14:03:14 +1200' [4] <-- written before $start_time '2006-06-15 14:03:15 +1200' [5] <-- written at $start_time '2006-06-15 14:03:16 +1200' [6] <-- written after $start_time During this run, dates [1] and [2] are both < $lastproc and thus ignored. The remaining four dates ([4]--[6]) are >= $lastproc and thus processed. Two hours later, the script runs again, this time with $start_time set to '2006-06-15 16:03:15 +1200' and $lastproc to '2006-06-15 14:03:15 +1200'. Dates [1] through [4] are all < $lastproc and thus ignored. However, dates [5] and [6] are both >= $lastproc and are processed a second time, resulting in a duplicate entry in the database. The solution is to ignore any log line entries that occur at or after (>=) $start_time. In the example above, this would mean that in the first run, dates [1], [2], [5] and [6] would be ignored and dates [3] and [4] processed. In the second run, dates [1]--[4] would be ignored and dates [5] and [6] processed. */ $test_starttime = strtotime($start_time); // NJS 2005-12-09 Switched to GeoIP from GeoIP:IPfree. include("geoip.inc"); $gi = geoip_open("##GEOIP_DATABASE##",GEOIP_STANDARD); /* Apache log for ePrints uses this format: LogFormat "%h %l %u %t \"%r\" %>s %b \"%{Referer}i\" \"%{User-Agent}i\"" combined If the log format differs the regular expression matching would need to be adjusted. Parse: ip date YYYY MM DD archive ID */ // Web server log files $log_dir = '##APACHE_LOG_LOCATION##'; $log_file = array( 'otago_eprints' => '##APACHE_LOG_NAME_1##', 'cardrona' => '##APACHE_LOG_NAME_2##', ); // eprintstats db $sqlserver = 'localhost'; $sqluser = 'eprintstatspriv'; $sqlpass = 'AuldGrizzel'; $sqldatabase = 'eprintstats'; /* NJS 2006-05-26 SQL details of your ePrints installation(s). This has now been generalised to work with multiple archives. For each archive that you have, add an entry to this array in the following format: 'archive_name' => array( 'sqlserver' => 'db_host', 'username' => 'archive_name', 'password' => 'password', ), */ $eprintsdbs = array( 'otago_eprints' => array( 'sqlserver' => 'localhost', 'username' => 'otago_eprints', 'password' => 'DrSyntaxRidesAgain', ), 'cardrona' => array( 'sqlserver' => 'localhost', 'username' => 'cardrona', 'password' => 'chautquau', ), ); /* NJS 2005-12-16 IP address ranges for your local Intranet(s). You can have multiple ranges of IP addresses, each with a different "country name", so that they will appear as separate entries in the by country stats pages. You should use a different country code for each range (ISO 3166-1 specifies the range XA through XZ as "user-assignable", so you can use codes from there as necessary), and create flag icons as appropriate. Each address range key is the name that will appear in the statistics database (the "country name"), followed by a comma, followed by the appropriate ISO 3166-1 country code as noted above. Each entry in the range is either a single IP address, or an array specifying a lower and upper bound for a contiguous IP address range (see example below). All IP addresses must be converted to long values using the ip2long() function before being stored. Note that address ranges may overlap. The script will use the first range that matches a given IP, so list the ranges in the correct order of precedence for your needs. Example: $local_IPs = array( 'Repository Admin,XA' => array( ip2long('192.168.1.5'), ip2long('192.168.1.22'), array( lower => ip2long('192.168.1.30'), upper => ip2long('192.168.1.35'), ), ), 'Our Intranet,XI' => array( array( lower => ip2long('192.168.1.0'), upper => ip2long('192.168.255.255'), ), ), ); 'Repository Admin' covers the IP addresses 192.168.1.5, 192.168.1.22 and the range 192.168.1.30 to 192.168.1.35, inclusive. 'Our Intranet' covers the range 192.168.1.0 to 192.168.255.255, inclusive. A machine will only match the 'Our Intranet' range if it first fails to match the 'Repository Admin' range. */ $local_IPs = array( 'Repository Admin,XA' => array( ip2long('139.80.75.110'), // Nigel @ Uni ip2long('60.234.209.74'), // Nigel @ home ip2long('139.80.92.138'), // Monica & Jeremy ip2long('139.80.92.151'), // @ Uni ip2long('203.89.162.155'), // Monica @ home ip2long('139.80.81.50'), // eprints.otago.ac.nz ip2long('172.20.1.50'), // eprints.otago.ac.nz pre-switch ip2long('172.20.1.1'), // eprints.otago.ac.nz pre-switch ), 'Otago Intranet,XI' => array( array( 'lower' => ip2long('139.80.0.0'), 'upper' => ip2long('139.80.127.255'), ), ), ); /* NJS 2007-01-26 Patterns to match various search engine bots. Ideally, we'd use a similar mechanism to the $local_IPs variable above, but this isn't feasible because we'd need to know the IP ranges for the likes of Google, for example. This clearly isn't possible in practice. Fortunately, most search bots insert a readily identifiable string into the user-agent part of the HTTP response, which gets recorded in the Apache log file. We can look for these and re-code log entries as appropriate. The format of this list is similar to that of the $local_IPs variable. The key is the "country name" (in this case the name of the search engine) plus a non-standard four-character country code starting with "X@", separated by a comma. Each key value has an associated list of corresponding regular expressions that can occur in the user-agent part of the Apache log entry. If any one of these REs matches the user-agent part of the log entry, then we should re-code the country appropriately. A four-character code is used because that what the database allows, and it avoids having to reserve several of the "X" country codes for search engines. */ $bot_patterns = array( // Yahoo! (http://www.yahoo.com/) 'Yahoo!,X@YH' => array( '/yahoo! slurp/i', '/yahooseeker/i', ), // Windows Live Search (http://search.msn.com/) 'Windows Live Search,X@MS' => array( '/msnbot/i', ), // Google (http://www.google.com/) 'Google,X@GG' => array( '/googlebot/i', ), // Ask.com (http://www.ask.com/) 'Ask.com,X@AC' => array( '/ask jeeves\/teoma/i', ), // Everything else I could find in our log files :) 'Other search engine,X@OS' => array( // TAMU Internet Research Lab (http://irl.cs.tamu.edu/) '/http:\/\/irl\.cs\.tamu\.edu\/crawler/i', // Alexa web search (http://www.alexa.com/) '/ia_archiver/i', // TrueKnowledge for Web (http://www.authoritativeweb.com/) '/converacrawler/i', // Majestic 12 distributed search engine (http://www.majestic12.co.uk/) '/mj12bot/i', // Picsearch (http://www.picsearch.com/) '/psbot/i', // Exalead (http://www.exalead.com/search) '/exabot/i', // Cazoodle (note cazoodle.com doesn't exist) '/cazoodlebot crawler/i', '/mqbot@cazoodle\.com/i', // Gigablast (http://www.gigablast.com/) '/gigabot/i', // Houxou (http://www.houxou.com/) '/houxoucrawler/i', '/crawler at houxou dot com/i', // IBM Almaden Research Center Computer Science group (http://www.almaden.ibm.com/cs/) '/http:\/\/www\.almaden\.ibm\.com\/cs\/crawler/i', // Goo? (http://help.goo.ne.jp/) '/ichiro/i', // Daum Communications Corp (Korea) '/edacious & intelligent web robot/i', '/daum communications corp/i', '/daum web robot/i', '/msie is not me/i', '/daumoa/i', // Girafa (http://www.girafa.com/) '/girafabot/i', // The Generations Network (http://www.myfamilyinc.com/) '/myfamilybot/i', // Naver? (http://www.naver.com/) '/naverbot/i', // WiseNut (http://www.wisenutbot.com/) '/zyborg/i', '/wn-[0-9]+\.zyborg@looksmart\.net/i', // Accelobot (http://www.accelobot.com/) // This one seems particularly busy! '/heritrix/i', // Seeqpod (http://www.seeqpod.com/) '/seeqpod-vertical-crawler/i', // University of Illinois at Urbana-Champaign, Computer Science (http://www.cs.uiuc.edu/) '/mqbot crawler/i', '/mqbot@cs\.uiuc\.edu/i', // Microsoft Research (http://research.microsoft.com/) '/msrbot/i', // Nusearch '/nusearch spider/i', // SourceForge (http://www.sf.net/) '/nutch-agent@lists\.sourceforge\.net/i', // Lucene (http://lucene.apache.org/) '/nutch-agent@lucene\.apache\.org/i', '/raphael@unterreuth.de/i', // Computer Science, University of Washington (http://cs.washington.edu/) '/nutch running at uw/i', '/sycrawl@cs\.washington\.edu/i', // Chikayama & Taura Laboratory, University of Tokyo (http://www.logos.ic.i.u-tokyo.ac.jp/) '/shim-crawler/i', '/crawl@logos\.ic\.i\.u-tokyo\.ac\.jp/i', // Sproose (http://www.sproose.com/) '/sproose bot/i', '/crawler@sproose\.com/i', // Turnitin (http://www.turnitin.com/) '/turnitinbot/i', // WISH Project (http://wish.slis.tsukuba.ac.jp/) '/wish-project/i', // WWWster '/wwwster/i', '/gue@cis\.uni-muenchen\.de/i', // Forex Trading Network Organization (http://www.netforex.org/) '/forex trading network organization/i', '/info@netforex\.org/i', // FunnelBack (http://www.funnelback.com/) '/funnelback/i', // Baidu (http://www.baidu.com/) '/baiduspider/i', // Brandimensions (http://www.brandimensions.com/) '/bdfetch/i', // Blaiz Enterprises (http://www.blaiz.net/) '/blaiz-bee/i', // Boitho/SearchDaimon (http://www.boitho.com/ or http://www.searchdaimon.com/) '/boitho\.com-dc/i', // Celestial (OAI aggregator, see http://oai-perl.sourceforge.net/ for a little info) '/celestial/i', // Cipinet (http://www.cipinet.com/) '/cipinetbot/i', // iVia (http://ivia.ucr.edu/) '/crawlertest crawlertest/i', // Encyclopedia of Keywords (http://keywen.com/) '/easydl/i', // Everest-Vulcan Inc. (http://everest.vulcan.com/) '/everest-vulcan inc/i', // FactBites (http://www.factbites.com/) '/factbot/i', // Scirus (http://www.scirus.com/) '/scirus scirus-crawler@fast\.no/i', // UOL (http://www.uol.com.br/) '/uolcrawler/i', '/soscrawler@uol\.com\.br/i', // Always Updated (http://www.updated.com/) '/updated crawler/i', '/crawler@updated\.com/i', // FAST Enterprise Search (http://www.fast.no/) '/fast metaweb crawler/i', '/crawler@fast\.no/i', '/helpdesk at fastsearch dot com/i', // Deutsche Wortschatz Portal (http://wortschatz.uni-leipzig.de/) '/findlinks/i', // Gais (http://gais.cs.ccu.edu.tw/) '/gaisbot/i', '/robot[0-9]{2}@gais.cs.ccu.edu.tw/i', // http://ilse.net/ '/ingrid/i', // Krugle (http://corp.krugle.com/) '/krugle\/krugle/i', '/krugle web crawler/i', '/webcrawler@krugle\.com/i', // WebWobot (http://www.webwobot.com/) '/scollspider/i', // Omni-Explorer (http://www.omni-explorer.com/) '/omniexplorer_bot/i', '/worldindexer/i', // PageBull (http://www.pagebull.com/) '/pagebull http:\/\/www\.pagebull\.com\//i', // dir.com (http://dir.com/) '/pompos/i', // Sensis (http://sensis.com.au/) '/sensis web crawler/i', '/search_comments\\\\at\\\\sensis\\\\dot\\\\com\\\\dot\\\\au/i', // Shopwiki (http://www.shopwiki.com/) '/shopwiki/i', // Guruji (http://www.terrawiz.com/) '/terrawizbot/i', // Language Observatory Project (http://www.language-observatory.org/) '/ubicrawler/i', // MSIE offline bookmarks crawler '/msiecrawler/i', // Unidentified '/bot/i', '/crawler/i', '/spider/i', '/larbin/i', // also larbinSpider '/httrack/i', '/voyager/i', '/acadiauniversitywebcensusclient/i', '/feedchecker/i', '/knowitall\(knowitall@cs\.washington\.edu\)/i', '/mediapartners-google/i', '/psycheclone/i', '/topicblogs/i', '/nutch/i', ), ); ########################################### ## ## No configuration required below here. ## ########################################### $connect = mysql_pconnect ($sqlserver,$sqluser,$sqlpass); $db = mysql_select_db($sqldatabase,$connect) or die("Could not connect"); // First get the date of last update /* NJS 2006-04-28 Changed this from order by timeinsert to order by id. The ID is always guaranteed to increase temporally, but is otherwise time-independent and thus not affected by things like daylight savings. */ $query = "SELECT lastproc FROM lastproc ORDER BY id DESC LIMIT 1"; $result = mysql_query($query,$connect); $num_rows = mysql_num_rows($result); if ($num_rows > 0) { $row = mysql_fetch_assoc($result); $lastproc = $row["lastproc"]; // NJS 2007-01-30 Refactored $databaseA to more meaningful $test_lastproc. $test_lastproc = strtotime($lastproc); } else { $test_lastproc = 0; } // NJS 2006-06-14: Generalised connection list for multiple archives. $eprints_connections = array(); foreach ($eprintsdbs as $archive_name => $details) { $eprints_connections[$archive_name] = mysql_connect($details['sqlserver'],$details['username'],$details['password']); } $counter = 0; foreach($log_file as $archive_name=>$archive_log) { $logf = $log_dir . $archive_log; $handle = fopen($logf, "r"); while (!feof($handle)) { $buffer = fgets($handle, 4096); /* NJS 2007-01-26 Added user-agent match to all regexps to enable bot detection. NJS 2007-01-31 Refactored regexps from four down to one, after realising that (a) long EPrints URLs are a superset of the short ones, and (b) a regexp that matches domain names works just as well for IP addresses (the GeoIP lookup doesn't care which it gets). Also fixed the pattern so it can handle an arbitrary number of subdomains. Note that the latter would be the main argument for keeping a separate IP address pattern, as IP addresses always comprise exactly four parts. However, it's not really up to the script to verify IP addresses; Apache should be recording them correctly in the first place! The typical kinds of strings we are matching look something like this: fetch abstract (short, long): 168.192.1.1 - - [31/Jan/2007:09:15:36 +1300] "GET /1/ HTTP/1.1" 200 12345 "referer" "user-agent" 168.192.1.1 - - [31/Jan/2007:09:15:36 +1300] "GET /archive/00000001/ HTTP/1.1" 200 12345 "referer" "user-agent" download item (short, long): 168.192.1.1 - - [31/Jan/2007:09:15:37 +1300] "GET /1/01/foo.pdf HTTP/1.1" 200 12345 "referer" "user-agent" 168.192.1.1 - - [31/Jan/2007:09:15:37 +1300] "GET /archive/00000001/01/foo.pdf HTTP/1.1" 200 12345 "referer" "user-agent" Plus any of the above with a domain name substituted for the IP address (e.g., foo.bar.com instead of 168.192.1.1). */ if (preg_match("/^(\S+(?:\.\S+)+) - - \[(.*?)\] \"GET \/(?:archive\/0{1,8})?(\d{1,4}).*? HTTP\/1..\" 200 .*?(\"[^\"]+\")?$/i",$buffer,$matches)) { $counter++; $country_code = ''; $country_name = ''; $insertid = ''; $eprint_name = ''; $view_type = ''; $uniquebits = ''; /* NJS 2007-01-29 Moved date checking to the start of the loop, as there's no point in doing any of the regexp checks if we've already processed this log entry and will discard it anyway. */ $date = $matches[2]; /* NJS 2006-04-28 Switched to timestamp rather than date-based comparison. First, clean up the Apache request date into something that strtotime understands. Note that the Apache log dates include time zone info by default. */ $date = preg_replace("/:/"," ",$date,1); // Change first ":" to " ". $date = preg_replace("/\//", " ", $date); // Change all "/" to " ". // NJS 2007-01-30 Refactored $databaseB to more meaningful // $test_logdate. $test_logdate = strtotime($date); // NJS 2007-01-30 Added test for log dates >= $start_time. if ( ( $test_logdate < $test_lastproc ) || ( $test_logdate >= $test_starttime ) ) continue; // Convert to properly formatted date string. $request_date = date('Y-m-d H:i:s O', $test_logdate); /* NJS 2005-12-16 Determine country code and name. Check whether the IP number falls into any of the local intranet ranges. If so, then use that. */ $ip = $matches[1]; $ip_long = ip2long($ip); $found_country = FALSE; foreach ($local_IPs as $id => $addresses) { foreach ($addresses as $ip_range) { if (is_array($ip_range)) // check against lower/upper bounds { $found_country = (($ip_long >= $ip_range['lower']) && ($ip_long <= $ip_range['upper'])); break; } else if (is_long($ip_range)) // data type sanity check { $found_country = ($ip_long == $ip_range); break; } else // something is seriously broken, ignore this entry { print "Unsupported data type " . gettype($ip_range) . " (value " . $ip_range . ") in \$local_IPs (expected long).\n"; continue; } } if ($found_country) { list($country_name, $country_code) = explode(',', $id); break; } } // Otherwise, fall back to GeoIP. if (!$found_country) { $country_code = geoip_country_code_by_addr($gi, $ip); $country_name = geoip_country_name_by_addr($gi, $ip); } // end NJS 2005-12-16 /* NJS 2007-01-26 Check whether this is a bot reference. */ $user_agent = $matches[4]; $found_country = FALSE; foreach ($bot_patterns as $id => $patterns) { foreach ($patterns as $pat) { if (preg_match($pat, $user_agent)) { $found_country = TRUE; break; } } if ($found_country) { list($country_name, $country_code) = explode(',', $id); break; } } // end NJS 2007-01-26 // Now sort out the remaining bits and we're done. $eprint_id = $matches[3]; $uniquebits = $buffer; // NJS 2005-11-25 Added regexp for EPrints short URLs. // NJS 2007-01-31 Refactored into one regexp for both styles. if (preg_match("/GET \/(?:archive\/0{1,8})?\d{1,4}\/\d\d\//i",$buffer)) { $view_type = "download"; } else { $view_type = "abstract"; } if(isset($eprintname[$archive_name . $eprint_id])) { $eprint_name = $eprintname[$archive_name . $eprint_id]; } else { $eprint_name = getePrintName($eprints_connections[$archive_name],$archive_name,$eprint_id); $eprintname[$archive_name . $eprint_id] = $eprint_name; } if($eprint_name=='') { // Do nothing. } else { $eprint_name = mysql_escape_string($eprint_name); /* NJS 2006-04-25 Requests containing apostrophes (') are dumped by MySQL unless we escape them. Looking in the GeoIP files I also see country names with apostrophes, so escape that as well. Everything else should be fine. */ $uniquebits = mysql_escape_string($uniquebits); $country_name = mysql_escape_string($country_name); // end NJS 2006-04-25 $query = " INSERT INTO view (uniquebits,archive_name,ip,request_date,archiveid,country_code,country_name,view_type,eprint_name) VALUES('".$uniquebits."','".$archive_name."','".$ip."','".$request_date."',".$eprint_id.",'".$country_code."','".$country_name."','".$view_type."','".$eprint_name."')"; $result = mysql_query($query,$connect); $insertid = mysql_insert_id($connect); } } else { // print "NO match" . "\n"; } } fclose($handle); } /* Keep track of where we are. Should avoid duplication of results if the script is run more than once on the same log file. */ // NJS 2006-04-28 Switched value inserted to $start_time instead of $request_date. $query = "INSERT into lastproc (lastproc) values('".$start_time."')"; $result = mysql_query($query,$connect); #print "Records counted: $counter\n"; #print "Last count: $request_date\n"; foreach ($eprints_connections as $connection) { mysql_close($connection); } mysql_close($connect); // Look up the title corresponding to the specified eprint id. function getePrintName($connection,$archive,$eprintid) { // NJS 2006-06-14: DB connection now passed as an argument. $db = mysql_select_db($archive,$connection); if ( $eprints_version > 2 ) { $query3 = " SELECT title FROM eprint WHERE eprintid = $eprintid AND eprint_status = 'archive' "; } else { $query3 = " SELECT title FROM archive WHERE eprintid = $eprintid "; } $result3 = mysql_query($query3,$connection); $title = ''; $suffix = ''; // NJS 2006-04-25 Added check for empty result, probably a deleted item. // Look in the deleted items for details. if (mysql_num_rows($result3) == 0) { if ( $eprints_version > 2 ) { $query3 = " SELECT title FROM eprint WHERE eprintid = $eprintid AND eprint_status = 'deletion' "; } else { $query3 = " SELECT title FROM archive WHERE eprintid = $eprintid "; } $result3 = mysql_query($query3,$connection); // If it's not in deletion, then we have no clue what it is. if (mysql_num_rows($result3) == 0) { $title = "Unknown item [$eprintid]"; } else { $suffix = ' [deleted]'; } } if ($title == '') { $row = mysql_fetch_assoc($result3); $row["title"] = trim($row["title"]); $row["title"] = preg_replace("/\s+/"," ",$row["title"]); $title = $row["title"]; } return $title . $suffix; } ?>
Show line notes below