Newer
Older
#!/usr/bin/python
# -*- coding: utf-8 -*-
import sys
if sys.version_info < (2, 6):
print >> sys.stderr, 'You must use Python 2.6 or greater'
sys.exit(1)
import base64
import ConfigParser
import datetime
import fnmatch
import gzip
import hashlib
import inspect
import itertools
import json
import logging
import optparse
import os
import os.path
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
import Queue
import re
import threading
import time
import urllib
import urllib2
##
## Constants.
##
_COMMON_LOG_FORMAT = (
'(?P<ip>\S+) \S+ \S+ \[(?P<date>.*?)\] '
'"\S+ (?P<path>.*?) \S+" (?P<status>\S+) (?P<length>\S+)'
)
_NSCA_EXTENDED_LOG_FORMAT = (
'(?P<ip>\S+) \S+ \S+ \[(?P<date>.*?)\] '
'"\S+ (?P<path>.*?) \S+" (?P<status>\S+) (?P<length>\S+) '
'"(?P<referrer>.*?)" "(?P<user_agent>.*?)"'
)
_COMMON_COMPLETE_LOG_FORMAT = (
'(?P<host>[\w\-\.]*)(?::\d+)? '
'(?P<ip>\S+) \S+ \S+ \[(?P<date>.*?)\] '
'"\S+ (?P<path>.*?) \S+" (?P<status>\S+) (?P<length>\S+) '
'"(?P<referrer>.*?)" "(?P<user_agent>.*?)"'
)
FORMATS = {
'common': _COMMON_LOG_FORMAT,
'common_vhost': '(?P<host>[\w\-\.]*)(?::\d+)? ' + _COMMON_LOG_FORMAT,
'nsca_extended': _NSCA_EXTENDED_LOG_FORMAT,
'common_complete': _COMMON_COMPLETE_LOG_FORMAT,
}
DATE_FORMAT = '%d/%b/%Y:%H:%M:%S'
EXCLUDED_EXTENSIONS = (
# Images
'.gif', '.jpg', '.jpeg', '.png', '.bmp', '.ico', '.svg',
# Fonts
'.ttf', '.eot', '.woff',
# Plugins
'.class', '.swf',
# Misc
'.css', '.js', '.xml', 'robots.txt',
)
DOWNLOAD_EXTENSIONS = (
'7z aac arc arj asf asx avi bin csv deb dmg doc exe flv gif gz gzip hqx '
'jar jpg jpeg js mpg mp2 mp3 mp4 mpeg mov movie msi msp odb odf odg odp '
'ods odt ogg ogv pdf phps png ppt qt qtm ra ram rar rpm sea sit tar tbz '
'bz2 tbz tgz torrent txt wav wma wmv wpd xls xml z zip'
).split()
# A good source is: http://phpbb-bots.blogspot.com/
EXCLUDED_USER_AGENTS = (
'AdsBot-Google',
'ia_archiver',
'Scooter/',
'Ask Jeeves',
'Baiduspider+(',
'Exabot',
'Googlebot',
'Mediapartners-Google',
'msnbot',
'Sosospider+',
'SurveyBot',
'Twiceler',
'VoilaBot',
'Yahoo',
'Yandex',
)
PIWIK_MAX_ATTEMPTS = 3
PIWIK_DELAY_AFTER_FAILURE = 2
PIWIK_EXPECTED_IMAGE = base64.b64decode(
'R0lGODlhAQABAIAAAAAAAAAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw=='
)
##
## Code.
##
class Configuration(object):
"""
Stores all the configuration options by reading sys.argv and parsing,
if needed, the config.inc.php.
It has 2 attributes: options and filenames.
"""
class Error(Exception):
pass
def __init__(self):
option_parser = optparse.OptionParser(
usage='Usage: %prog [options] log_file [ log_file [...] ]',
description="Import HTTP access logs to Piwik."
)
option_parser.add_option(
'-d', '--debug', dest='debug', action='count', default=0,
help="Enable debug output (specify multiple times for more verbose)",
)
option_parser.add_option(
'-n', '--dry-run', dest='dry_run',
action='store_true', default=False,
help="Perform a trial run with nothing being really inserted into Piwik",
)
option_parser.add_option(
'-u', '--url', dest='piwik_url',
help="Piwik install URL",
)
default_config = os.path.abspath(
os.path.join(os.path.dirname(__file__),
'../../config/config.ini.php'),
)
'-c', '--config', dest='config_file', default=default_config,
help=(
"Piwik configuration file (default: %default). This file is used to "
"get superuser credentials unless they are themselves given as "
"options"
)
)
option_parser.add_option(
'-l', '--login', dest='login',
help="Piwik superuser login"
)
option_parser.add_option(
'-p', '--password', dest='password',
help="Piwik superuser password"
)
option_parser.add_option(
'-t', '--token-auth', dest='piwik_token_auth',
help="Piwik token_auth",
)
option_parser.add_option(
'-f', '--format', dest='format', default=None,
help="Access log format. If not specified, the format will be autodetected",
)
option_parser.add_option(
'-i', '--idsite', dest='site_id',
help="Piwik site ID to use",
)
option_parser.add_option(
'--idsite-fallback', dest='site_id_fallback',
help="Default Piwik site ID to use if the hostname doesn't match any "
"known Piwik URL",
)
option_parser.add_option(
'--hostnames', dest='hostnames', action='append',
help="Accepted hostnames (others will be excluded)"
)
option_parser.add_option(
'-s', '--skip', dest='skip', default=0, type='int',
help="Skip the n first lines",
)
option_parser.add_option(
'-r', '--recorders', dest='recorders', default=1, type='int',
help="Number of simultaneous recorders (default: %default)",
)
option_parser.add_option(
'--add-sites-new-hosts', dest='add_sites_new_hosts',
action='store_true', default=False,
help="When a hostname is found in the log file, but not matched to any website "
"in Piwik, automatically create a new website in Piwik with this hostname to "
"import the logs"
)
option_parser.add_option(
'--useragent-exclude', dest='excluded_useragents',
action='append', default=[],
help="User agents to exclude (in addition to the standard excluded "
"user agents)",
)
option_parser.add_option(
'--show-progress', dest='show_progress',
action='store_true', default=os.isatty(sys.stdout.fileno()),
help="Print a progress report every second"
)
self.options, self.filenames = option_parser.parse_args(sys.argv[1:])
# Configure logging before calling logging.{debug,info}.
logging.basicConfig(
format='%(asctime)s: [%(levelname)s] %(message)s',
level=logging.DEBUG if self.options.debug >= 1 else logging.INFO,
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
)
if self.options.hostnames:
logging.debug('Accepted hostnames: %s', ', '.join(options.hostnames))
else:
logging.debug('Accepted hostnames: all')
if self.options.format:
try:
self.format_regexp = re.compile(FORMATS[self.options.format])
except KeyError:
fatal_error('invalid log format: %s' % self.options.format)
else:
self.format_regexp = None
if not self.options.piwik_url:
fatal_error('no URL given for Piwik')
if not (self.options.piwik_url.startswith('http://') or self.options.piwik_url.startswith('https://')):
self.options.piwik_url = 'http://' + self.options.piwik_url
logging.debug('Piwik URL is: %s', self.options.piwik_url)
if not self.options.piwik_token_auth:
self.options.piwik_token_auth = self._get_token_auth()
logging.debug('Authentication token is: %s', self.options.piwik_token_auth)
if self.options.recorders < 1:
self.options.recorders = 1
def _get_token_auth(self):
"""
If the token auth is not specified in the options, get it from Piwik.
"""
# Get superuser login/password from the options.
logging.debug('No token specified, getting it from Piwik')
piwik_login = self.options.login
piwik_password = hashlib.md5(self.options.password).hexdigest()
# Fallback to the given (or default) configuration file, then
# get the token from the API.
if not piwik_login or not piwik_password:
logging.debug(
'No credentials specified, reading them from "%s"',
self.options.config_file
)
config_file = ConfigParser.RawConfigParser()
success = len(config_file.read(self.options.config_file)) > 0
if not success:
fatal_error(
"couldn't open the configuration file, "
"required to get the authentication token"
)
piwik_login = config_file.get('superuser', 'login')
piwik_password = config_file.get('superuser', 'password')
logging.debug('Using credentials: (%s, %s)', piwik_login, piwik_password)
try:
api_result = piwik.call_api('UsersManager.getTokenAuth',
userLogin=piwik_login,
md5Password=piwik_password,
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
_token_auth='',
_url=self.options.piwik_url,
)
except urllib2.URLError, e:
fatal_error('error when getting authentication token: %s' % e)
try:
return api_result['value']
except KeyError:
# Happens when the credentials are invalid.
message = api_result.get('message')
fatal_error(
'error when getting authentication token%s' % (
': %s' % message if message else '')
)
def get_resolver(self):
if self.options.site_id:
logging.debug('Resolver: static')
return StaticResolver(self.options.site_id)
else:
logging.debug('Resolver: dynamic')
return DynamicResolver()
class Statistics(object):
"""
Store statistics about parsed logs and recorded entries.
Can optionally print statistics on standard output every second.
"""
class Counter(object):
"""
Simple integers cannot be used by multithreaded programs. See:
http://stackoverflow.com/questions/6320107/are-python-ints-thread-safe
"""
def __init__(self):
# itertools.count's implementation in C does not release the GIL and
# therefore is thread-safe.
self.counter = itertools.count(1)
self.value = 0
def increment(self):
self.value = self.counter.next()
def __str__(self):
return str(int(self.value))
def __init__(self):
self.time_start = None
self.time_stop = None
self.piwik_sites = set() # sites ID
self.piwik_sites_created = [] # (hostname, site ID)
self.piwik_sites_ignored = set() # hostname
self.count_lines_parsed = self.Counter()
self.count_lines_recorded = self.Counter()
# Do not match the regexp.
self.count_lines_invalid = self.Counter()
# No site ID found by the resolver.
self.count_lines_no_site = self.Counter()
# Hostname filtered by config.options.hostnames
self.count_lines_hostname_skipped = self.Counter()
# Static files.
self.count_lines_static = self.Counter()
# Ignored user-agents.
self.count_lines_skipped_user_agent = self.Counter()
# Downloads
self.count_lines_downloads = self.Counter()
# Misc
self.dates_recorded = set()
self.monitor_stop = False
def set_time_start(self):
self.time_start = time.time()
def set_time_stop(self):
self.time_stop = time.time()
def _compute_speed(self, value, start, end):
delta_time = end - start
if value == 0:
return 0
if delta_time == 0:
return 'very high!'
else:
return value / delta_time
def _round_value(self, value, base=100):
return round(value * base) / base
def _indent_text(self, lines, level=1):
"""
Return an indented text. 'lines' can be a list of lines or a single
line (as a string). One level of indentation is 4 spaces.
"""
prefix = ' ' * (4 * level)
if isinstance(lines, basestring):
return prefix + lines
else:
return '\n'.join(
prefix + line
for line in lines
)
def print_summary(self):
print '''
Logs import summary
-------------------
%(count_lines_recorded)d requests imported successfully
%(count_lines_downloads)d requests were downloads
%(total_lines_ignored)d requests ignored:
%(count_lines_invalid)d invalid log lines
%(count_lines_skipped_user_agent)d requests done by bots, search engines, ...
%(count_lines_static)d requests to static resources (images, stylesheets, ...)
%(count_lines_no_site)d requests did not match any known site
%(count_lines_hostname_skipped)d requests did not match any requested hostname
Website import summary
----------------------
%(count_lines_recorded)d requests imported to %(total_sites)d sites
%(total_sites_existing)d sites already existed
%(total_sites_created)d sites were created:
%(sites_created)s
%(total_sites_ignored)d distinct hostnames did not match any existing site:
%(sites_ignored)s
TIP: if one of these hosts is an alias host for one of the websites
in Piwik, you can add this host as an "Alias URL" in Settings > Websites.
TIP: use --add-sites-new-hosts if you wish to automatically create
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
one website for each of these hosts in Piwik rather than discarding
these requests.
Performance summary
-------------------
Total time: %(total_time)d seconds
Requests imported per second: %(speed_recording)s requests per second
''' % {
'count_lines_recorded': self.count_lines_recorded.value,
'count_lines_downloads': self.count_lines_downloads.value,
'total_lines_ignored': sum([
self.count_lines_invalid.value,
self.count_lines_skipped_user_agent.value,
self.count_lines_static.value,
self.count_lines_no_site.value,
self.count_lines_hostname_skipped.value,
]),
'count_lines_invalid': self.count_lines_invalid.value,
'count_lines_skipped_user_agent': self.count_lines_skipped_user_agent.value,
'count_lines_static': self.count_lines_static.value,
'count_lines_no_site': self.count_lines_no_site.value,
'count_lines_hostname_skipped': self.count_lines_hostname_skipped.value,
'total_sites': len(self.piwik_sites),
'total_sites_existing': len(self.piwik_sites - set(site_id for hostname, site_id in self.piwik_sites_created)),
'total_sites_created': len(self.piwik_sites_created),
'sites_created': self._indent_text(
['%s (ID: %d)' % (hostname, site_id) for hostname, site_id in self.piwik_sites_created],
level=3,
),
'total_sites_ignored': len(self.piwik_sites_ignored),
'sites_ignored': self._indent_text(
self.piwik_sites_ignored, level=3,
),
'total_time': self.time_stop - self.time_start,
'speed_recording': self._round_value(self._compute_speed(
self.count_lines_recorded.value,
self.time_start, self.time_stop,
)),
}
##
## The monitor is a thread that prints a short summary each second.
##
def _monitor(self):
latest_total_recorded = 0
while not self.monitor_stop:
current_total = stats.count_lines_recorded.value
print '%d lines parsed, %d lines recorded, %d records/sec' % (
stats.count_lines_parsed.value,
current_total,
current_total - latest_total_recorded,
)
latest_total_recorded = current_total
time.sleep(1)
def start_monitor(self):
t = threading.Thread(target=self._monitor)
t.daemon = True
t.start()
def stop_monitor(self):
self.monitor_stop = True
class Piwik(object):
"""
Make requests to Piwik.
"""
class Error(Exception):
pass
@staticmethod
def _call(path, args, headers=None, url=None):
"""
Make a request to the Piwik site. It is up to the caller to format
arguments, to embed authentication, etc.
"""
if url is None:
url = config.options.piwik_url
headers = headers or {}
# If Content-Type isn't defined, PHP do not parse the request's body.
headers['Content-type'] = 'application/x-www-form-urlencoded'
data = urllib.urlencode(args).encode('ascii', 'ignore')
request = urllib2.Request(url + path, data, headers)
response = urllib2.urlopen(request)
return response.read()
@staticmethod
def _call_api(method, **kwargs):
"""
Make a request to the Piwik API taking care of authentication, body
formatting, etc.
"""
args = {
'module' : 'API',
'format' : 'json',
'method' : method,
}
# token_auth, by default, is taken from config.
token_auth = kwargs.pop('_token_auth', None)
if token_auth is None:
token_auth = config.options.piwik_token_auth
if token_auth:
args['token_auth'] = token_auth
url = kwargs.pop('_url', None)
if kwargs:
args.update(kwargs)
# Convert lists into appropriate format.
# See: http://dev.piwik.org/trac/wiki/API/Reference#PassinganArrayParameter
# Warning: we have to pass the parameters in order: foo[0], foo[1], foo[2]
# and not foo[1], foo[0], foo[2] (it will break Piwik otherwise.)
final_args = []
for key, value in args.iteritems():
if isinstance(value, (list, tuple)):
for index, obj in enumerate(value):
final_args.append(('%s[%d]' % (key, index), obj))
else:
final_args.append((key, value))
res = Piwik._call('/', final_args, url=url)
try:
return json.loads(res)
except ValueError:
raise urllib2.URLError('Piwik returned an invalid response: ' + res)
def _call_wrapper(self, func, expected_response, *args, **kwargs):
"""
Try to make requests to Piwik at most PIWIK_FAILURE_MAX_RETRY times.
"""
errors = 0
while True:
try:
response = func(*args, **kwargs)
if expected_response is not None and response != expected_response:
raise urllib2.URLError("didn't receive the expected response")
return response
except (urllib2.URLError, ValueError), e:
logging.debug('Error when connecting to Piwik: %s', e)
errors += 1
if errors == PIWIK_MAX_ATTEMPTS:
raise Piwik.Error(str(e))
else:
time.sleep(PIWIK_DELAY_AFTER_FAILURE)
def call(self, path, args, expected_content=None, headers=None):
return self._call_wrapper(self._call, expected_content, path, args, headers)
def call_api(self, method, **kwargs):
return self._call_wrapper(self._call_api, None, method, **kwargs)
##
## Resolvers.
##
## A resolver is a class that turns a hostname into a Piwik site ID.
##
class StaticResolver(object):
"""
Always return the same site ID, specified in the configuration.
"""
def __init__(self, site_id):
self.site_id = site_id
# Go get the main URL
sites = piwik.call_api(
'SitesManager.getSiteFromId', idSite=self.site_id
)
try:
site = sites[0]
except (IndexError, KeyError):
fatal_error(
"cannot get the main URL of this site: invalid site ID: %s" % site_id
)
if site.get('result') == 'error':
fatal_error(
"cannot get the main URL of this site: %s" % site.get('message')
)
self._main_url = site['main_url']
stats.piwik_sites.add(self.site_id)
def resolve(self, hit):
return (self.site_id, self._main_url)
def check_format(self, format):
pass
class DynamicResolver(object):
"""
Use Piwik API to determine the site ID.
"""
def __init__(self):
self._cache = {}
def _resolve(self, hit):
main_url = 'http://' + hit.host
res = piwik.call_api(
'SitesManager.getSitesIdFromSiteUrl',
url=main_url,
)
if res:
# The site already exists.
site_id = res[0]['idsite']
else:
# The site doesn't exist.
logging.debug('No Piwik site found for the hostname: %s', hit.host)
if config.options.site_id_fallback is not None:
logging.debug('Using default site for hostname: %s', hit.host)
return config.options.site_id_fallback
elif config.options.add_sites_new_hosts:
if config.options.dry_run:
# Let's just return a fake ID.
site_id = 0
logging.debug('Creating a Piwik site for hostname %s', hit.host)
result = piwik.call_api(
'SitesManager.addSite',
siteName=hit.host,
urls=[main_url],
)
if result.get('result') == 'error':
logging.error("Couldn't create a Piwik site for host %s: %s",
hit.host, result.get('message'),
)
else:
site_id = result['value']
stats.piwik_sites_created.append((hit.host, site_id))
else:
# The site doesn't exist, we don't want to create new sites and
# there's no default site ID. We thus have to ignore this hit.
site_id = None
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
stats.piwik_sites.add(site_id)
def resolve(self, hit):
"""
Return the site ID from the cache if found, otherwise call _resolve.
"""
try:
site_id = self._cache[hit.host]
except KeyError:
logging.debug(
'Site ID for hostname %s not in cache', hit.host
)
site_id = self._resolve(hit)
logging.debug('Site ID for hostname %s: %s', hit.host, site_id)
self._cache[hit.host] = site_id
return (site_id, 'http://' + hit.host)
def check_format(self, format):
regexp = re.compile(format)
if 'host' not in regexp.groupindex:
fatal_error(
"the selected log format doesn't include the hostname: you must "
"specify the Piwik site ID with the -i argument"
)
class Recorder(object):
"""
A Recorder fetches hits from the Queue and inserts them into Piwik using
the API.
"""
recorders = []
def __init__(self):
self.queue = Queue.Queue(maxsize=10000)
@staticmethod
def launch(recorder_count):
"""
Launch a bunch of Recorder objects in a separate thread.
"""
for i in xrange(recorder_count):
recorder = Recorder()
Recorder.recorders.append(recorder)
t = threading.Thread(target=recorder._run)
t.daemon = True
t.start()
logging.debug('Launched recorder')
@staticmethod
def add_hit(hit):
"""
Add a hit in one of the recorders queue.
"""
# Get a queue so that one client IP will always use the same queue.
recorders = Recorder.recorders
queue = recorders[abs(hash(hit.ip)) % len(recorders)].queue
queue.put(hit)
@staticmethod
def wait_empty():
"""
Wait until all recorders have an empty queue.
"""
for recorder in Recorder.recorders:
recorder._wait_empty()
def _run(self):
while True:
hit = self.queue.get()
try:
self._record_hit(hit)
except Piwik.Error, e:
fatal_error(e, hit.filename, hit.lineno)
self.queue.task_done()
def _wait_empty(self):
"""
Wait until the queue is empty.
"""
while True:
if self.queue.empty():
# We still have to wait for the last queue item being processed
# (queue.empty() returns True before queue.task_done() is
# called).
self.queue.join()
return
time.sleep(1)
def date_to_piwik(self, date):
date, time = date.isoformat(sep=' ').split()
return '%s %s' % (date, time.replace('-', ':'))
def _record_hit(self, hit):
"""
Insert the hit into Piwik.
"""
site_id, main_url = resolver.resolve(hit)
if site_id is None:
# This hit doesn't match any known Piwik site.
stats.piwik_sites_ignored.add(hit.host)
stats.count_lines_no_site.increment()
return
stats.dates_recorded.add(hit.date.date())
args = {
'rec': '1',
'apiv': '1',
'url': main_url + hit.path[:1024],
'urlref': hit.referrer[:1024],
'cip': hit.ip,
'cdt': self.date_to_piwik(hit.date),
'idsite': site_id,
'dp': 1,
'token_auth': config.options.piwik_token_auth,
}
if hit.is_download:
args['download'] = args['url']
stats.count_lines_downloads.increment()
if hit.status == '404':
args['action_name'] = '404/URL = %s/From = %s' % (
urllib.quote(args['url']),
urllib.quote(args['urlref'])
)
if not config.options.dry_run:
piwik.call(
'/piwik.php', args,
expected_content=PIWIK_EXPECTED_IMAGE,
headers={'User-Agent' : hit.user_agent},
)
stats.count_lines_recorded.increment()
@staticmethod
def invalidate_reports():
if config.options.dry_run or not stats.dates_recorded:
return
dates = [date.strftime('%Y-%m-%d') for date in stats.dates_recorded]
print 'Purging Piwik archives for dates: %s' % dates
result = piwik.call_api(
'CoreAdminHome.invalidateArchivedReports',
dates=','.join(dates),
idSites=','.join(stats.piwik_sites),
)
class Hit(object):
"""
It's a simple container.
"""
def __init__(self, **kwargs):
for key, value in kwargs.iteritems():
setattr(self, key, value)
super(Hit, self).__init__()
class Parser(object):
"""
The Parser parses the lines in a specified file and inserts them into
a Queue.
"""
## All check_* methods are called for each hit and must return True if the
## hit can be imported, False otherwise.
def check_hostname(self, hit):
# Check against config.hostnames.
if not hasattr(hit, 'host') or not config.options.hostnames:
return True
# Accept the hostname only if it matches one pattern in the list.
result = any(
fnmatch.fnmatch(hit.host, pattern)
for pattern in config.options.hostnames
)
if not result:
stats.count_lines_hostname_skipped.increment()
return result
def check_extension(self, hit):
for extension in EXCLUDED_EXTENSIONS:
if hit.path.endswith(extension) and not hit.is_download:
stats.count_lines_static.increment()
return False
return True
def check_user_agent(self, hit):
for s in itertools.chain(EXCLUDED_USER_AGENTS, config.options.excluded_useragents):
if s in hit.user_agent:
stats.count_lines_skipped_user_agent.increment()
return False
return True
def parse(self, filename):
"""
Parse the specified filename and insert hits in the queue.
"""
def invalid_line(line):
stats.count_lines_invalid.increment()
if self.options.debug >= 2:
logging.debug('Invalid line detected: ' + line)
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
if config.options.show_progress:
print 'Parsing log %s...' % filename
if filename.endswith('.gz'):
open_func = gzip.open
else:
open_func = open
file = open_func(filename, 'r')
for lineno, line in enumerate(file):
# Guess the format if needed.
if not config.format_regexp:
logging.debug('Guessing the log format...')
for name, format in FORMATS.iteritems():
if re.match(format, line):
config.format = format
config.format_regexp = re.compile(format)
logging.debug('Format %s matches', name)
break
logging.debug('Format %s does not match', name)
if not config.format_regexp:
raise config.ConfigurationError(
'Cannot guess the logs format. Please give one using'
' the --format option'
)
# Make sure the format is compatible with the resolver.
resolver.check_format(format)
stats.count_lines_parsed.increment()
if stats.count_lines_parsed.value <= config.options.skip:
continue
match = config.format_regexp.match(line)
if not match:
invalid_line(line)
continue
hit = Hit(
filename=filename,
lineno=lineno,
status=match.group('status'),
full_path=match.group('path'),
)
# Strip query string
hit.path = hit.full_path.split('?', 1)[0]
# Parse date _with_ timezone to get an UTC timestamp.
date_string = match.group('date')
try:
tz = float(date_string[-5:])
hit.date = datetime.datetime.strptime(date_string[:-6], '%d/%b/%Y:%H:%M:%S')
except ValueError:
# Date format is incorrect, the line is probably badly formatted.
invalid_line(line)
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
continue
hit.date -= datetime.timedelta(hours=tz/100)
try:
hit.referrer = match.group('referrer')
except IndexError:
hit.referrer = ''
if hit.referrer == '-':
hit.referrer = ''
try:
hit.user_agent = match.group('user_agent')
except IndexError:
hit.user_agent = ''
hit.ip = match.group('ip')
try:
hit.length = int(match.group('length'))
except ValueError:
# Not all lines have a length (e.g. 304 redirects)
hit.length = 0
try:
hit.host = match.group('host')
except IndexError:
# Some formats have no host.
pass
hit.is_download = hit.path.rsplit('.', 1)[-1] in DOWNLOAD_EXTENSIONS
# Check if the hit must be excluded.
check_methods = inspect.getmembers(self, predicate=inspect.ismethod)
if all((method(hit) for name, method in check_methods if name.startswith('check_'))):
Recorder.add_hit(hit)
def main():
"""
Start the importing process.
"""
if config.options.show_progress:
stats.start_monitor()
stats.set_time_start()
recorders = Recorder.launch(config.options.recorders)
for filename in config.filenames:
if os.path.exists(filename):
parser.parse(filename)
else:
print >> sys.stderr, 'File %s does not exist' % filename
Recorder.wait_empty()
stats.set_time_stop()
if config.options.show_progress:
stats.stop_monitor()
try:
Recorder.invalidate_reports()
except Piwik.Error, e:
pass
stats.print_summary()
def fatal_error(error, filename=None, lineno=None):
print >> sys.stderr, 'Fatal error: %s' % error
if filename and lineno is not None:
print >> sys.stderr, (
'You can restart the import of "%s" from the point it failed by '
'specifying --skip=%d on the command line.\n' % (filename, lineno)
)