Kiln » Kiln Extensions
Clone URL:  
Pushed to 2 repositories · View In Graph Contained in tip

Kiln Extensions as of Kiln 2.5b1

Changeset b4944cd04b48

Parent 154a3761de2e

by Profile picture of User 12Benjamin Pollack <benjamin@fogcreek.com>

Changes to 11 files · Browse files at b4944cd04b48 Showing diff from parent 154a3761de2e Diff from another changeset...

 
1
2
3
4
5
6
7
8
9
10
11
 
37
38
39
 
 
 
 
40
41
42
 
1
2
 
3
4
5
6
 
7
8
9
 
35
36
37
38
39
40
41
42
43
44
@@ -1,11 +1,9 @@
 '''Base class for store implementations and store-related utility code.'''   -import sys  import os  import tempfile  import binascii  import bfutil -import shutil    from mercurial import util, node, error, url as url_, hg  from mercurial.i18n import _ @@ -37,6 +35,10 @@
  '''Put source file into the store under <filename>/<hash>.'''   raise NotImplementedError('abstract method')   + def exists(self, hash): + '''Check to see if the store contains the given hash.''' + raise NotImplementedError('abstract method') +   def get(self, files):   '''Get the specified big files from the store and write to local   files under repo.root. files is a list of (filename, hash)
 
1
2
3
4
5
6
7
8
9
10
11
12
 
13
14
15
 
156
157
158
 
159
160
161
 
332
333
334
335
 
336
337
338
339
 
 
 
 
340
341
342
343
 
344
345
346
 
379
380
381
382
 
383
384
385
 
455
456
457
 
458
459
460
 
462
463
464
 
 
 
 
 
465
466
467
468
 
469
470
471
 
491
492
493
494
495
496
497
 
1
2
3
 
 
 
4
 
 
 
5
 
6
7
8
9
 
150
151
152
153
154
155
156
 
327
328
329
 
330
331
332
 
 
333
334
335
336
337
338
339
340
341
342
343
344
 
377
378
379
 
380
381
382
383
 
453
454
455
456
457
458
459
 
461
462
463
464
465
466
467
468
469
470
471
 
472
473
474
475
 
495
496
497
 
498
499
500
@@ -1,15 +1,9 @@
 '''High-level command functions: bfadd() et. al, plus the cmdtable.'''    import os -import re -import errno -import binascii  import shutil -import httplib -import posixpath -import BaseHTTPServer   -from mercurial import util, commands, match as match_, hg, node, context, error +from mercurial import util, match as match_, hg, node, context, error  from mercurial.i18n import _    import bfutil, basestore @@ -156,6 +150,7 @@
    hash = fctx.data().strip()   path = bfutil.find_file(rsrc, hash) + ### TODO: What if the file is not cached?   data = ''   with open(path, 'rb') as fd:   data = fd.read() @@ -332,15 +327,18 @@
    store = basestore._open_store(rsrc, rdst.path, put=True)   - at = 1 + at = 0   for hash in files:   ui.progress(_('Uploading bfiles'), at, unit='bfile', total=len(files)) - at += 2 - source = bfutil.find_file(rsrc, hash, False) + if store.exists(hash): + at += 1 + continue + source = bfutil.find_file(rsrc, hash)   if not source:   raise util.Abort(_('Missing bfile %s needs to be uploaded') % hash)   # XXX check for errors here   store.put(source, hash) + at += 1   ui.progress('Uploading bfiles', None)    def verify_bfiles(ui, repo, all=False, contents=False): @@ -379,7 +377,7 @@
  expectedhash = repo[None][bfutil.standin(bfile)].data().strip()   mode = os.stat(repo.wjoin(bfutil.standin(bfile))).st_mode   if not os.path.exists(repo.wjoin(bfile)) or expectedhash != bfutil.hashfile(repo.wjoin(bfile)): - path = bfutil.find_file(repo, expectedhash, False) + path = bfutil.find_file(repo, expectedhash)   if path is None:   toget.append((bfile, expectedhash))   else: @@ -455,6 +453,7 @@
  toget = []   at = 0   updated = 0 + removed = 0   printed = False   if bfiles:   ui.status(_('Getting changed bfiles\n')) @@ -462,10 +461,15 @@
    for bfile in bfiles:   at += 1 + if os.path.exists(repo.wjoin(bfile)) and not os.path.exists(repo.wjoin(bfutil.standin(bfile))): + os.unlink(repo.wjoin(bfile)) + removed += 1 + bfdirstate.forget(bfutil.unixpath(bfile)) + continue   expectedhash = repo[None][bfutil.standin(bfile)].data().strip()   mode = os.stat(repo.wjoin(bfutil.standin(bfile))).st_mode   if not os.path.exists(repo.wjoin(bfile)) or expectedhash != bfutil.hashfile(repo.wjoin(bfile)): - path = bfutil.find_file(repo, expectedhash, False) + path = bfutil.find_file(repo, expectedhash)   if not path:   toget.append((bfile, expectedhash))   else: @@ -491,7 +495,6 @@
  updated += 1   bfdirstate.normal(bfutil.unixpath(filename))   - removed = 0   for bfile in bfdirstate:   if bfile not in bfiles:   if os.path.exists(repo.wjoin(bfile)):
 
43
44
45
46
47
 
48
49
50
 
101
102
103
104
105
106
107
108
109
110
 
111
112
113
 
122
123
124
125
 
126
127
128
 
258
259
260
261
 
 
 
262
263
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
264
265
266
 
451
452
453
 
 
 
 
454
455
456
 
734
735
736
 
 
 
 
 
 
 
 
 
737
738
739
 
799
800
801
 
802
803
804
 
992
993
994
 
995
996
997
 
43
44
45
 
 
46
47
48
49
 
100
101
102
 
 
 
 
 
 
 
103
104
105
106
 
115
116
117
 
118
119
120
121
 
251
252
253
 
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
 
464
465
466
467
468
469
470
471
472
473
 
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
 
825
826
827
828
829
830
831
 
1019
1020
1021
1022
1023
1024
1025
@@ -43,8 +43,7 @@
  method = getattr(repo, name)   #if not (isinstance(method, types.MethodType) and   # method.im_func is repo.__class__.commitctx.im_func): - if (isinstance(method, types.FunctionType) and - method.func_name == 'wrap'): + if isinstance(method, types.FunctionType) and method.func_name == 'wrap':   ui.warn(_('kbfiles: repo method %r appears to have already been '   'wrapped by another extension: '   'kbfiles may behave incorrectly\n') @@ -101,13 +100,7 @@
    m = copy.copy(match)   m._files = [tostandin(f) for f in m._files] - orig_matchfn = m.matchfn - def matchfn(f): - if bfutil.is_standin(f): - return orig_matchfn(bfutil.split_standin(f)) - else: - return orig_matchfn(f) and not inctx(bfutil.standin(f), ctx2) - m.matchfn = matchfn +   # get ignored clean and unknown but remove them later if they were not asked for   try:   result = super(bfiles_repo, self).status(node1, node2, m, True, True, True, subrepos) @@ -122,7 +115,7 @@
  # was already computed using super's status.   bfdirstate = bfutil.open_bfdirstate(ui, self)   match._files = [f for f in match._files if f in bfdirstate] - s = bfdirstate.status(match, [], True, True, True) + s = bfdirstate.status(match, [], listignored, listclean, listunknown)   (unsure, modified, added, removed, missing, unknown, ignored, clean) = s   if parentworking:   for bfile in unsure: @@ -258,9 +251,29 @@
  bfiles = bfutil.list_bfiles(repo)   match = copy.copy(match)   orig_matchfn = match.matchfn - match._files = [f for f in match._files if f not in bfiles] + + # Check both the list of bfiles and the list of standins because if a bfile was removed, it + # won't be in the list of bfiles at this point   match._files += sorted(standins)   + actualfiles = [] + for f in match._files: + fstandin = bfutil.standin(f) + + # Ignore known bfiles and standins + if f in bfiles or fstandin in standins: + continue + + # Append directory separator to avoid collisions + if not fstandin.endswith('/'): + fstandin += '/' + + # Prevalidate matching standin directories + if any(st for st in match._files if st.startswith(fstandin)): + continue + actualfiles.append(f) + match._files = actualfiles +   def matchfn(f):   if orig_matchfn(f):   return f not in bfiles @@ -451,6 +464,10 @@
  for f in remove:   if not after:   os.unlink(repo.wjoin(f)) + currentdir = os.path.split(f)[0] + while currentdir and not os.listdir(repo.wjoin(currentdir)): + os.rmdir(repo.wjoin(currentdir)) + currentdir = os.path.split(currentdir)[0]   bfdirstate.remove(bfutil.unixpath(f))   bfdirstate.write()   @@ -734,6 +751,15 @@
  bfcommands.update_bfiles(repo.ui, repo)   return result   +# When we rebase a repository with remotely changed bfiles, we need +# to explicitly do a clean update so that the entries in .kbf are +# udpated and the new bfiles are pulled +def override_pull(orig, ui, repo, source="default", **opts): + result = orig(ui, repo, source, **opts) + if opts.get('rebase', False): + commands.update(repo.ui, repo, clean = True) + return result +  def override_archive(orig, repo, dest, node, kind, decode=True, matchfn=None,   prefix=None, mtime=None, subrepos=None):   # No need to lock because we are only reading history and bfile caches @@ -799,6 +825,7 @@
  getdata = ctx[f].data   if bfutil.is_standin(f):   path = bfutil.find_file(repo, getdata().strip()) + ### TODO: What if the file is not cached?   f = bfutil.split_standin(f)     def getdatafn(): @@ -992,6 +1019,7 @@
  entry[1].extend(summaryopt)     entry = extensions.wrapcommand(commands.table, 'update', override_update) + entry = extensions.wrapcommand(commands.table, 'pull', override_pull)   entry = extensions.wrapfunction(filemerge, 'filemerge', override_filemerge)   entry = extensions.wrapfunction(cmdutil, 'copy', override_copy)  
 
2
3
4
5
6
7
8
9
10
11
12
 
13
14
15
 
110
111
112
113
 
114
 
115
116
 
117
118
119
120
121
122
123
124
 
128
129
130
131
 
 
 
 
132
133
134
 
248
249
250
251
252
 
 
 
253
254
255
 
2
3
4
 
 
5
6
7
8
9
 
10
11
12
13
 
108
109
110
 
111
112
113
114
115
116
117
 
 
 
 
118
119
120
 
124
125
126
 
127
128
129
130
131
132
133
 
247
248
249
 
 
250
251
252
253
254
255
@@ -2,14 +2,12 @@
   import os  import errno -import binascii -import tempfile  import inspect  import shutil  import stat    from mercurial import \ - util, dirstate, context, cmdutil, error, match as match_, node + util, dirstate, cmdutil, match as match_  from mercurial.i18n import _    short_name = '.kbf' @@ -110,15 +108,13 @@
 def in_system_cache(ui, hash):   return os.path.exists(system_cache_path(ui, hash))   -def find_file(repo, hash, getfile=True): +def find_file(repo, hash):   if in_cache(repo, hash): + repo.ui.note(_('Found %s in cache\n') % hash)   return cache_path(repo, hash)   if in_system_cache(repo.ui, hash): + repo.ui.note(_('Found %s in system cache\n') % hash)   return system_cache_path(repo.ui, hash) - if getfile: - (success, failure) = basetore._open_store(repo).get([(f, hash)]) - if (f, hash) in success: - return cache_path(repo, hash)   return None    def open_bfdirstate(ui, repo): @@ -128,7 +124,10 @@
  '''   admin = repo.join(long_name)   opener = util.opener(admin) - bfdirstate = dirstate.dirstate(opener, ui, repo.root) + if hasattr(repo.dirstate, '_validate'): + bfdirstate = dirstate.dirstate(opener, ui, repo.root, repo.dirstate._validate) + else: + bfdirstate = dirstate.dirstate(opener, ui, repo.root)     # If the bfiles dirstate does not exist, populate and create it. This   # ensures that we create it on the first meaningful bfiles operation in @@ -248,8 +247,9 @@
   def compose_standin_matcher(repo, rmatcher):   '''Return a matcher that accepts standins corresponding to the files - accepted by rmatcher.''' - smatcher = get_standin_matcher(repo) + accepted by rmatcher. Pass the list of files in the matcher as the + paths specified by the user.''' + smatcher = get_standin_matcher(repo, rmatcher.files())   isstandin = smatcher.matchfn   def composed_matchfn(f):   return isstandin(f) and rmatcher.matchfn(split_standin(f))
 
21
22
23
 
 
 
24
25
26
27
28
29
 
30
31
32
33
 
 
34
35
36
 
39
40
41
42
 
43
44
45
 
21
22
23
24
25
26
27
28
29
30
31
32
33
34
 
 
 
35
36
37
38
39
 
42
43
44
 
45
46
47
48
@@ -21,16 +21,19 @@
  self.sendfile(source, hash)   self.ui.debug('put %s to remote store\n' % source)   + def exists(self, hash): + return self._verify(hash) +   def sendfile(self, filename, hash):   if self._verify(hash):   return     self.ui.debug('httpstore.sendfile(%s, %s)\n' % (filename, hash))   baseurl, authinfo = url_.getauthinfo(self.url) + fd = None   try: - fd = open(filename, 'rb') - request = urllib2.Request(bfutil.urljoin(baseurl, hash)) - request.add_data(fd.read()) + fd = url_.httpsendfile(filename, 'rb') + request = urllib2.Request(bfutil.urljoin(baseurl, hash), fd)   try:   url = self.opener.open(request)   self.ui.note(_('[OK] %s/%s\n') % (self.rawurl, url.geturl())) @@ -39,7 +42,7 @@
  except Exception, e:   raise util.Abort(_('%s') % e)   finally: - fd.close() + if fd: fd.close()     def _getfile(self, tmpfile, filename, hash):   (baseurl, authinfo) = url_.getauthinfo(self.url)
 
18
19
20
 
 
 
21
22
23
 
18
19
20
21
22
23
24
25
26
@@ -18,6 +18,9 @@
  '''Any file that is put must already be in the system wide cache so do nothing.'''   return   + def exists(self, hash): + return bfutil.in_system_cache(self.repo.ui, hash) +   def _getfile(self, tmpfile, filename, hash):   if bfutil.in_system_cache(self.ui, hash):   return bfutil.system_cache_path(self.ui, hash)
 
19
20
21
 
22
23
24
 
29
30
31
32
 
 
 
33
34
35
 
44
45
46
 
 
47
48
49
 
62
63
64
 
 
65
66
67
 
73
74
75
 
 
76
77
78
 
86
87
88
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
89
90
91
 
19
20
21
22
23
24
25
 
30
31
32
 
33
34
35
36
37
38
 
47
48
49
50
51
52
53
54
 
67
68
69
70
71
72
73
74
 
80
81
82
83
84
85
86
87
 
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
@@ -19,6 +19,7 @@
 hgt.hg(['init', '-q'])  hgt.writefile('normal1', 'foo')  os.mkdir('sub') +os.mkdir('sub2')  hgt.writefile('sub/normal2', 'bar')  hgt.writefile('sub/normal3.txt', 'bar2')  hgt.writefile('sub/normal4.txt', 'bar3') @@ -29,7 +30,9 @@
 hgt.writefile('big1', 'abc')  hgt.writefile('sub/big2', 'xyz')  hgt.writefile('sub/big3.txt', 'xyz') -hgt.hg(['add', '-q', '--bf', 'big1', rejoin('sub/big2'), rejoin('sub/big3.txt')]) +hgt.writefile('sub/big4', 'xyz') +hgt.writefile('sub2/big5', 'xyz') +hgt.hg(['add', '-q', '--bf', 'big1', rejoin('sub/big2'), rejoin('sub/big3.txt'), rejoin('sub/big4'), rejoin('sub2/big5')])  hgt.hg(['commit', '-m', 'added bfiles'])    hgt.announce('remove sub/*.txt') @@ -44,6 +47,8 @@
 hgt.asserttrue(os.path.exists('sub/normal2'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('big1'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('sub/big2'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub/big4'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub2/big5'), 'added file doesnt exist')  hgt.hg(['status'],   stdout=('R sub/big3.txt\n'   'R sub/normal3.txt\n' @@ -62,6 +67,8 @@
 hgt.asserttrue(os.path.exists('sub/normal2'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('big1'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('sub/big2'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub/big4'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub2/big5'), 'added file doesnt exist')  hgt.hg(['up'],   stdout=('0 files updated, 0 files merged, 3 files removed, 0 files unresolved\n'   'Getting changed bfiles\n' @@ -73,6 +80,8 @@
 hgt.asserttrue(os.path.exists('sub/normal2'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('big1'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('sub/big2'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub/big4'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub2/big5'), 'added file doesnt exist')    hgt.announce('remove single normal files and add')  hgt.hg(['remove', 'normal1', 'sub/normal2']) @@ -86,6 +95,23 @@
 hgt.asserttrue(os.path.exists('sub/normal2'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('big1'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('sub/big2'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub/big4'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub2/big5'), 'added file doesnt exist') +hgt.hg(['status']) + +hgt.announce('remove single bfile and commit with full path') +hgt.hg(['remove', 'sub/big4']) +hgt.hg(['status'],stdout=('R sub/big4\n')) +hgt.hg(['commit', '-m', 'removing big4', 'sub/big4']) +hgt.assertfalse(os.path.exists('sub/big4'), 'removed file exists') +hgt.hg(['status']) + +hgt.announce('remove single bfile and commit with partial path') +hgt.hg(['remove', 'sub2/big5']) +hgt.hg(['status'],stdout=('R sub2/big5\n')) +hgt.assertfalse(os.path.exists("sub2"), 'removed directory structure exists') +hgt.hg(['commit', '-m', 'removing big5', 'sub2']) +hgt.assertfalse(os.path.exists('sub2/big5'), 'removed file exists')  hgt.hg(['status'])    hgt.announce('remove single bfiles and add')
Change 1 of 1 Show Entire File big-push.py Stacked
 
75
76
77
 
78
79
80
 
81
82
83
84
85
86
87
 
 
88
89
 
90
91
 
92
93
94
 
75
76
77
78
79
80
 
81
82
83
84
85
86
 
87
88
89
90
 
91
92
 
93
94
95
96
@@ -75,20 +75,22 @@
  if not opts.get('force') and not opts.get('new_branch') and None == prepush(repo, other, False, revs)[0]:   return   try: + push_size = 1   while len(outgoing) > 0:   ui.debug('start: %d to push\n' % len(outgoing)) - current_push_size = min(max_push_size, len(outgoing)) + current_push_size = min(push_size, len(outgoing))   ui.debug('pushing: %d\n' % current_push_size)   # force the push, because we checked above that by the time the whole push is done, we'll have merged back to one head   remote_heads = repo.push(other, force=True, revs=outgoing[:current_push_size])   if remote_heads: # push succeeded   outgoing = outgoing[current_push_size:] - current_push_size = max_push_size   ui.debug('pushed %d ok\n' % current_push_size) + if push_size < max_push_size: + push_size *= 2   else: # push failed; try again with a smaller size - current_push_size /= 10 + push_size /= 2   ui.debug('failed, trying %d\n' % current_push_size) - if current_push_size == 0: + if push_size == 0:   raise UnpushableChangesetError   except UnpushableChangesetError:   ui.status(_('unable to push changeset %s\n') % outgoing[0])
Change 1 of 1 Show Entire File gestalt.py Stacked
 
160
161
162
163
 
 
164
165
166
 
160
161
162
 
163
164
165
166
167
@@ -160,7 +160,8 @@
 '''))   return True   - source, hashbranch = parseurl(ui.expandpath('default-push', ui.expandpath('default'))) + target = ui.config('paths', 'default-push') and ui.expandpath('default-push') or source + source, hashbranch = parseurl(source)   other = hg.repository(remoteui(repo, opts), source)   revs = addbranchrevs(repo, other, hashbranch)   ui.pushbuffer()
Change 1 of 6 Show Entire File kiln.py Stacked
 
1
 
2
3
4
 
38
39
40
 
41
42
43
44
45
 
 
46
47
48
 
66
67
68
69
70
71
72
73
 
174
175
176
177
 
178
179
180
 
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
202
203
204
205
 
 
 
206
207
208
 
556
557
558
559
560
561
 
562
563
564
 
569
570
571
572
 
573
574
575
 
 
1
2
3
4
 
38
39
40
41
42
43
44
45
 
46
47
48
49
50
 
68
69
70
 
 
71
72
73
 
174
175
176
 
177
178
 
 
179
180
181
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
 
 
 
 
208
209
210
211
212
213
 
561
562
563
 
 
 
564
565
566
567
 
572
573
574
 
575
576
577
578
@@ -1,4 +1,4 @@
-# Copyright (C) 2010 Fog Creek Software. All rights reserved. +# Copyright (C) 2011 Fog Creek Software. All rights reserved.  #  # To enable the "kiln" extension put these lines in your ~/.hgrc:  # [extensions] @@ -38,11 +38,13 @@
 import re  import urllib  import urllib2 +import subprocess  import sys    from cookielib import MozillaCookieJar  from hashlib import md5 -from mercurial import extensions, commands, demandimport, hg, util, httprepo, match +from mercurial import extensions, commands, demandimport, hg, util, httprepo, localrepo, match +from mercurial import ui as hgui  from mercurial import url as hgurl  from mercurial.error import RepoError  from mercurial.i18n import _ @@ -66,8 +68,6 @@
  win32api.ShellExecute(0, 'open', escape_reserved(url), None, None, 0)  demandimport.enable()   -KILN_CAPABILITY_PREFIX = 'kiln-' -KILN_CURRENT_VERSION = '1.0.0'  _did_version_check = False    class APIError(Exception): @@ -174,35 +174,40 @@
  ret = re.sub(r'([A-Za-z]):', r'\1:\\', ret)   return ret   -def _versioncheck(ui, repo, str): +def _upgradecheck(ui, repo):   global _did_version_check - m = re.match(KILN_CAPABILITY_PREFIX + '(?P<version>[0-9.]+).*', str) - if _did_version_check or not m: + if _did_version_check or not ui.configbool('kiln', 'autoupdate', True):   return   _did_version_check = True - version = m.group('version') - server_version = [int(s) for s in version.split('.')] - my_version = [int(s) for s in KILN_CURRENT_VERSION.split('.')] - ignore_version = [int(s) for s in ui.config('kiln', 'ignoreversion', '0.0.0').split('.')] - if server_version > my_version: - url = urljoin(repo.url()[:repo.url().lower().index('/repo')], 'Tools') - if server_version > ignore_version: - if ui.promptchoice(_('You are currently running Kiln client tools version %s. ' - 'Version %s is available.\nUpgrade now? (y/n)') % - (KILN_CURRENT_VERSION, version), ('&No', '&Yes'), default=0): - browse(url) - else: - if os.name == 'nt': - config_file = 'Mercurial.ini' - else: - config_file = '~/.hgrc' - ui.write(_('''If you'd like Kiln to stop prompting you about version %s and below, ''' - '''add ignoreversion=%s to the [kiln] section of your %s\n''') % (version, version, config_file)) + _upgrade(ui, repo) + +def _upgrade(ui, repo): + ext_dir = os.path.dirname(os.path.abspath(__file__)) + ui.debug('kiln: checking for extensions upgrade for %s\n' % ext_dir) + + try: + r = localrepo.localrepository(hgui.ui(), ext_dir) + except RepoError: + commands.init(hgui.ui(), dest=ext_dir) + r = localrepo.localrepository(hgui.ui(), ext_dir) + + r.ui.setconfig('kiln', 'autoupdate', False) + r.ui.pushbuffer() + try: + source = 'https://developers.kilnhg.com/Repo/Kiln/Group/Kiln-Extensions' + if commands.incoming(r.ui, r, bundle=None, force=False, source=source) != 0: + # no incoming changesets, or an error. Don't try to upgrade. + ui.debug('kiln: no extensions upgrade available\n') + return + ui.write(_('updating Kiln Extensions at %s... ') % ext_dir) + # pull and update return falsy values on success + if commands.pull(r.ui, r, source=source) or commands.update(r.ui, r, clean=True): + url = urljoin(repo.url()[:repo.url().lower().index('/repo')], 'Tools') + ui.write(_('unable to update\nvisit %s to download the newest extensions\n') % url)   else: - ui.write(_('You are currently running Kiln client tools version %s. ' - 'Version %s is available.\nVisit %s to download the new client tools.\n') % - (KILN_CURRENT_VERSION, version, url)) - ui.write('\n') + ui.write(_('complete\n')) + except Exception, e: + ui.debug(_('kiln: error updating Kiln Extensions: %s\n') % e)    def is_dest_a_path(ui, dest):   paths = ui.configitems('paths') @@ -556,9 +561,7 @@
   def reposetup(ui, repo):   if issubclass(repo.__class__, httprepo.httprepository): - for cap in repo.capabilities: - if cap.startswith(KILN_CAPABILITY_PREFIX): - _versioncheck(ui, repo, cap) + _upgradecheck(ui, repo)    cmdtable = {   'kiln': @@ -569,7 +572,7 @@
  ('l', 'filehistory', [], _('view the history of the file')),   ('o', 'outgoing', None, _('view the repository\'s outgoing tab')),   ('s', 'settings', None, _('view the repository\'s settings tab')), - ('p', 'path', '', _('override the default URL to use for Kiln')), + ('p', 'path', '', _('select which Kiln branch of the repository to use')),   ('r', 'rev', [], _('view the specified changeset in Kiln')),   ('t', 'targets', None, _('view the repository\'s targets')),   ('', 'logout', None, _('log out of Kiln sessions'))],
Change 1 of 3 Show Entire File kilnauth.py Stacked
 
36
37
38
 
39
 
40
41
42
 
49
50
51
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
53
54
 
69
70
71
72
 
73
74
75
 
36
37
38
39
40
41
42
43
44
 
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
 
102
103
104
 
105
106
107
108
@@ -36,7 +36,9 @@
 import re  from urllib2 import Request  from cookielib import MozillaCookieJar, Cookie +import shutil  import sys +import tempfile    try:   from hashlib import md5 @@ -49,6 +51,37 @@
   current_user = None   +class CookieJar(MozillaCookieJar, object): + def __init__(self, filename, *args, **kwargs): + self.__original_path = filename + tf = tempfile.NamedTemporaryFile(delete=False) + self.__temporary_path = tf.name + tf.close() + if os.path.exists(filename): + shutil.copyfile(filename, self.__temporary_path) + return super(CookieJar, self).__init__(self.__temporary_path, *args, **kwargs) + + def __enter__(self): + pass + + def __exit__(self, exc_type, exc_value, traceback): + os.unlink(self.__temporary_path) + self.__temporary_path = None + + def __del__(self): + try: + if self.__temporary_path: + os.unlink(self.__temporary_path) + except (OSError, IOError): + pass + + def save(self, *args, **kwargs): + super(CookieJar, self).save(*args, **kwargs) + try: + shutil.copyfile(self.__temporary_path, self.__original_path) + except IOError: + pass +  def get_cookiejar(ui):   global current_user   if os.name == 'nt': @@ -69,7 +102,7 @@
  cookie_path = re.sub(r'([A-Za-z]):', r'\1:\\', cookie_path)     try: - cj = MozillaCookieJar(cookie_path) + cj = CookieJar(cookie_path)   if not os.path.exists(cookie_path):   cj.save()   if os.name == 'posix':