1
0
Fork 0
mirror of https://github.com/munin-monitoring/contrib.git synced 2025-07-22 02:51:03 +00:00
Munin-Contrib/plugins/moinmoin/moinmoin_pages

140 lines
3.7 KiB
Python
Executable file

#!/usr/bin/env python3
"""
=head1 NAME
moinmoin_pages - lists the number of pages (including ACL-protected pages) in all wikis of a MoinMoin wikifarm
=head1 CONFIGURATION
[moinmoin_*]
user www
=head1 IMPLEMENTATION NOTES
The plugin is quite koumbit-specific:
=over 4
=item 1. the wikifarm config is hardcoded
=item 2. "wikilist.py" is assumed to contain the list of wiki -> url patterns
=item 3. url patterns are assumed to be simple enough, that they are decodable into an url
=back
Also note that this plugin reuses code from MoinMoin/wikimacro.py's SystemInfo macro.
Finally, i tried using XMLRPC instead of native functions to fetch the data, but it ended up being
slower. For the record, here is what the getPageList() call would have looked like:
xmlrpclib.ServerProxy("http://wiki.koumbit.net/?action=xmlrpc2").getAllPages()
The quick benchmark I did yielded those results for the getAllPages() vs getPageList() calls:
xmlrpc: 2.35 real 0.12 user 0.04 sys
native: 1.44 real 1.07 user 0.35 sys
So the plugin is spending more time in the CPU (all time, actually), but it's doing in faster.
It is highly possible that the CPU time spared in XMLRPC is in fact used by the server.
=head1 AUTHORS
Copyleft 2007, The Anarcat <anarcat@koumbit.org>
=head1 LICENSE
Licensed under the GPLv2 or any later version.
SPDX-License-Identifier: GPL-2.0-or-later
=cut
"""
import operator
import os
from re import sub
import sys
os.chdir('/export/wiki/config')
sys.path.insert(0, '/export/wiki/config')
from MoinMoin import wikiutil
from MoinMoin.Page import Page
from farmconfig import wikis
import farmconfig
from MoinMoin.request import RequestCLI
def _formatInReadableUnits(size):
size = float(size)
unit = u' Byte'
if size > 9999:
unit = u' KiB'
size /= 1024
if size > 9999:
unit = u' MiB'
size /= 1024
if size > 9999:
unit = u' GiB'
size /= 1024
return u"%.1f %s" % (size, unit)
def _getDirectorySize(path):
try:
dirsize = 0
for root, dirs, files in os.walk(path):
dirsize += sum([os.path.getsize(os.path.join(root, name)) for name in files])
except EnvironmentError:
dirsize = -1
return dirsize
def main():
for wiki in wikis:
name = wiki[0]
url = wiki[1]
# XXX, hack: transform the regexp into a canonical url
# we need canonical urls in the config for this to be clean
# look for (foo|bar) and replace with foo
url = sub(r'\(([^\|]*)(\|[^\)]*\))+', r'\1', url)
# remove common regexp patterns and slap a protocol to make this a real url
url = sub(r'[\^\$]|(\.\*)', '', url)
mod = getattr(__import__(name), 'Config')
#print "Upgradeing wiki %s (%s)" % (getattr(mod, 'sitename'), url)
request = RequestCLI(url)
pagelist = request.rootpage.getPageList(user='')
systemPages = [page for page in pagelist if wikiutil.isSystemPage(request, page)]
print(name + '.value ' + str(len(pagelist) - len(systemPages)))
#totalsize = reduce(operator.add, [Page(request, name).size() for name in pagelist])
#print('Accumulated page sizes' + _formatInReadableUnits(totalsize))
def config():
print("""graph_title Wiki size
graph_vlabel Number of pages
graph_args --base 1000 -l 0
graph_scale no
graph_category wiki
graph_info The number of pages excludes system pages but includes ACL-protected pages.""")
for wiki in wikis:
name = wiki[0]
mod = getattr(__import__(name), 'Config')
print(name + '.label ' + getattr(mod, 'sitename'))
if __name__ == "__main__":
if len(sys.argv) > 1 and sys.argv[1] == 'config':
config()
else:
main()