X-Git-Url: https://projects.mako.cc/source/mw/blobdiff_plain/2c860f8d2f73c9d7322953b40b28927cef0187b9..2c712c501ee2e538d09ad94cf25c5749471c7bfd:/src/mw/metadir.py diff --git a/src/mw/metadir.py b/src/mw/metadir.py index e09c840..6c6bfeb 100644 --- a/src/mw/metadir.py +++ b/src/mw/metadir.py @@ -1,6 +1,6 @@ ### # mw - VCS-like nonsense for MediaWiki websites -# Copyright (C) 2009 Ian Weller +# Copyright (C) 2010 Ian Weller # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by @@ -16,9 +16,13 @@ # with this program. If not, see . ### +import bzrlib.diff +import codecs import ConfigParser import json +import mw.api import os +from StringIO import StringIO import sys @@ -45,7 +49,7 @@ class Metadir(object): else: self.config = None - def create(self, api_url): + def create(self, api_url, username=None): # create the directory if os.path.isdir(self.location): print '%s: you are already in a mw repo' % self.me @@ -60,6 +64,8 @@ class Metadir(object): self.config = ConfigParser.RawConfigParser() self.config.add_section('remote') self.config.set('remote', 'api_url', api_url) + if username != None: + self.config.set('remote', 'username', username) with open(self.config_loc, 'wb') as config_file: self.config.write(config_file) # create cache/ @@ -98,9 +104,10 @@ class Metadir(object): pagedata = json.loads(pagedata_raw) rvid = int(rv['revid']) pagedata[rvid] = { - 'user': rv['user'], 'timestamp': rv['timestamp'], - 'content': rv['*'], + 'user': rv['user'], 'timestamp': rv['timestamp'] } + if '*' in rv.keys(): + pagedata[rvid]['content'] = rv['*'] fd.seek(0) fd.write(json.dumps(pagedata)) fd.truncate() @@ -121,3 +128,60 @@ class Metadir(object): fd = file(pagefile, 'r') pagedata = json.loads(fd.read()) return pagedata[str(rvid)] + + def working_dir_status(self): + status = {} + check = [] + for root, dirs, files in os.walk(self.root): + if root == self.root: + dirs.remove('.mw') + for name in files: + check.append(os.path.join(root, name)) + check.sort() + for full in check: + name = os.path.split(full)[1] + if name[-5:] == '.wiki': + pagename = mw.api.filename_to_pagename(name[:-5]) + pageid = self.get_pageid_from_pagename(pagename) + if not pageid: + status[os.path.relpath(full, self.root)] = '?' + else: + rvid = self.pages_get_rv_list(pageid)[-1] + rv = self.pages_get_rv(pageid, rvid) + cur_content = codecs.open(full, 'r', 'utf-8').read() + if cur_content[-1] == '\n': + cur_content = cur_content[:-1] + if cur_content != rv['content']: + status[os.path.relpath(full, self.root)] = 'U' + return status + + def diff_rv_to_working(self, pagename, oldrvid=0, newrvid=0): + # oldrvid=0 means latest fetched revision + # newrvid=0 means working copy + filename = mw.api.pagename_to_filename(pagename) + '.wiki' + pageid = self.get_pageid_from_pagename(pagename) + if not pageid: + raise ValueError('page named %s has not been fetched' % pagename) + else: + if oldrvid == 0: + oldrvid = self.pages_get_rv_list(pageid)[-1] + oldrv = self.pages_get_rv(pageid, oldrvid) + oldname = 'a/%s (revision %i)' % (filename, oldrvid) + old = [i+'\n' for i in oldrv['content'].split('\n')] + if newrvid == 0: + cur_content = codecs.open(filename, 'r', 'utf-8').read() + if cur_content[-1] == '\n': + cur_content = cur_content[:-1] + newname = 'b/%s (working copy)' % filename + new = [i+'\n' for i in cur_content.split('\n')] + else: + newrv = self.pages_get_rv(pageid, newrvid) + newname = 'b/%s (revision %i)' % (filename, newrvid) + new = [i+'\n' for i in newrv['content'].split('\n')] + diff_fd = StringIO() + bzrlib.diff.internal_diff(oldname, old, newname, new, diff_fd) + diff_fd.seek(0) + diff = diff_fd.read() + if diff[-1] == '\n': + diff = diff[:-1] + return diff