# create cache/pages/
os.mkdir(os.path.join(self.location, 'cache', 'pages'), 0755)
+
+
+ def clean_page(self, pagename):
+ filename = mw.api.pagename_to_filename(pagename) + '.wiki'
+ cur_content = codecs.open(filename, 'r', 'utf-8').read()
+ if ( (len(cur_content) != 0) and (cur_content[-1] == '\n') ):
+ cur_content = cur_content[:-1]
+
+ fd = file(filename, 'w')
+ fd.write(cur_content.encode('utf-8'))
+ fd.close()
+
def pagedict_add(self, pagename, pageid, currentrv):
fd = file(os.path.join(self.location, 'cache', 'pagedict'), 'r+')
pagedict = json.loads(fd.read())
pagedata = json.loads(fd.read())
return pagedata[str(rvid)]
- def working_dir_status(self):
+ def working_dir_status(self, files=None):
status = {}
check = []
- for root, dirs, files in os.walk(self.root):
- if root == self.root:
- dirs.remove('.mw')
- for name in files:
- check.append(os.path.join(root, name))
+ if files == None or files == []:
+ for root, dirs, files in os.walk(self.root):
+ if root == self.root:
+ dirs.remove('.mw')
+ for name in files:
+ check.append(os.path.join(root, name))
+ else:
+ for file in files:
+ check.append(os.path.join(os.getcwd(), file))
check.sort()
for full in check:
name = os.path.split(full)[1]
rvid = self.pages_get_rv_list(pageid)[-1]
rv = self.pages_get_rv(pageid, rvid)
cur_content = codecs.open(full, 'r', 'utf-8').read()
- if cur_content[-1] == '\n':
+ if (len(cur_content) != 0) and (cur_content[-1] == '\n'):
cur_content = cur_content[:-1]
if cur_content != rv['content']:
status[os.path.relpath(full, self.root)] = 'U'
old = [i + '\n' for i in oldrv['content'].split('\n')]
if newrvid == 0:
cur_content = codecs.open(filename, 'r', 'utf-8').read()
- if cur_content[-1] == '\n':
+ if (len(cur_content) != 0) and (cur_content[-1] == '\n'):
cur_content = cur_content[:-1]
newname = 'b/%s (working copy)' % filename
new = [i + '\n' for i in cur_content.split('\n')]