Kiln » Kiln Extensions
Clone URL:  
Pushed to 2 repositories · View In Graph Contained in tip

Kiln Extensions as of Kiln 2.5b1

Changeset b4944cd04b48

Parent 154a3761de2e

by Profile picture of User 12Benjamin Pollack <benjamin@fogcreek.com>

Changes to 11 files · Browse files at b4944cd04b48 Showing diff from parent 154a3761de2e Diff from another changeset...

 
1
2
3
4
5
6
7
8
9
10
11
 
37
38
39
 
 
 
 
40
41
42
 
1
2
 
3
4
5
6
 
7
8
9
 
35
36
37
38
39
40
41
42
43
44
@@ -1,11 +1,9 @@
 '''Base class for store implementations and store-related utility code.'''   -import sys  import os  import tempfile  import binascii  import bfutil -import shutil    from mercurial import util, node, error, url as url_, hg  from mercurial.i18n import _ @@ -37,6 +35,10 @@
  '''Put source file into the store under <filename>/<hash>.'''   raise NotImplementedError('abstract method')   + def exists(self, hash): + '''Check to see if the store contains the given hash.''' + raise NotImplementedError('abstract method') +   def get(self, files):   '''Get the specified big files from the store and write to local   files under repo.root. files is a list of (filename, hash)
 
1
2
3
4
5
6
7
8
9
10
11
12
 
13
14
15
 
156
157
158
 
159
160
161
 
332
333
334
335
 
336
337
338
339
 
 
 
 
340
341
342
343
 
344
345
346
 
379
380
381
382
 
383
384
385
 
455
456
457
 
458
459
460
 
462
463
464
 
 
 
 
 
465
466
467
468
 
469
470
471
 
491
492
493
494
495
496
497
 
1
2
3
 
 
 
4
 
 
 
5
 
6
7
8
9
 
150
151
152
153
154
155
156
 
327
328
329
 
330
331
332
 
 
333
334
335
336
337
338
339
340
341
342
343
344
 
377
378
379
 
380
381
382
383
 
453
454
455
456
457
458
459
 
461
462
463
464
465
466
467
468
469
470
471
 
472
473
474
475
 
495
496
497
 
498
499
500
@@ -1,15 +1,9 @@
 '''High-level command functions: bfadd() et. al, plus the cmdtable.'''    import os -import re -import errno -import binascii  import shutil -import httplib -import posixpath -import BaseHTTPServer   -from mercurial import util, commands, match as match_, hg, node, context, error +from mercurial import util, match as match_, hg, node, context, error  from mercurial.i18n import _    import bfutil, basestore @@ -156,6 +150,7 @@
    hash = fctx.data().strip()   path = bfutil.find_file(rsrc, hash) + ### TODO: What if the file is not cached?   data = ''   with open(path, 'rb') as fd:   data = fd.read() @@ -332,15 +327,18 @@
    store = basestore._open_store(rsrc, rdst.path, put=True)   - at = 1 + at = 0   for hash in files:   ui.progress(_('Uploading bfiles'), at, unit='bfile', total=len(files)) - at += 2 - source = bfutil.find_file(rsrc, hash, False) + if store.exists(hash): + at += 1 + continue + source = bfutil.find_file(rsrc, hash)   if not source:   raise util.Abort(_('Missing bfile %s needs to be uploaded') % hash)   # XXX check for errors here   store.put(source, hash) + at += 1   ui.progress('Uploading bfiles', None)    def verify_bfiles(ui, repo, all=False, contents=False): @@ -379,7 +377,7 @@
  expectedhash = repo[None][bfutil.standin(bfile)].data().strip()   mode = os.stat(repo.wjoin(bfutil.standin(bfile))).st_mode   if not os.path.exists(repo.wjoin(bfile)) or expectedhash != bfutil.hashfile(repo.wjoin(bfile)): - path = bfutil.find_file(repo, expectedhash, False) + path = bfutil.find_file(repo, expectedhash)   if path is None:   toget.append((bfile, expectedhash))   else: @@ -455,6 +453,7 @@
  toget = []   at = 0   updated = 0 + removed = 0   printed = False   if bfiles:   ui.status(_('Getting changed bfiles\n')) @@ -462,10 +461,15 @@
    for bfile in bfiles:   at += 1 + if os.path.exists(repo.wjoin(bfile)) and not os.path.exists(repo.wjoin(bfutil.standin(bfile))): + os.unlink(repo.wjoin(bfile)) + removed += 1 + bfdirstate.forget(bfutil.unixpath(bfile)) + continue   expectedhash = repo[None][bfutil.standin(bfile)].data().strip()   mode = os.stat(repo.wjoin(bfutil.standin(bfile))).st_mode   if not os.path.exists(repo.wjoin(bfile)) or expectedhash != bfutil.hashfile(repo.wjoin(bfile)): - path = bfutil.find_file(repo, expectedhash, False) + path = bfutil.find_file(repo, expectedhash)   if not path:   toget.append((bfile, expectedhash))   else: @@ -491,7 +495,6 @@
  updated += 1   bfdirstate.normal(bfutil.unixpath(filename))   - removed = 0   for bfile in bfdirstate:   if bfile not in bfiles:   if os.path.exists(repo.wjoin(bfile)):
 
43
44
45
46
47
 
48
49
50
 
101
102
103
104
105
106
107
108
109
110
 
111
112
113
 
122
123
124
125
 
126
127
128
 
258
259
260
261
 
 
 
262
263
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
264
265
266
 
451
452
453
 
 
 
 
454
455
456
 
734
735
736
 
 
 
 
 
 
 
 
 
737
738
739
 
799
800
801
 
802
803
804
 
992
993
994
 
995
996
997
 
43
44
45
 
 
46
47
48
49
 
100
101
102
 
 
 
 
 
 
 
103
104
105
106
 
115
116
117
 
118
119
120
121
 
251
252
253
 
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
 
464
465
466
467
468
469
470
471
472
473
 
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
 
825
826
827
828
829
830
831
 
1019
1020
1021
1022
1023
1024
1025
@@ -43,8 +43,7 @@
  method = getattr(repo, name)   #if not (isinstance(method, types.MethodType) and   # method.im_func is repo.__class__.commitctx.im_func): - if (isinstance(method, types.FunctionType) and - method.func_name == 'wrap'): + if isinstance(method, types.FunctionType) and method.func_name == 'wrap':   ui.warn(_('kbfiles: repo method %r appears to have already been '   'wrapped by another extension: '   'kbfiles may behave incorrectly\n') @@ -101,13 +100,7 @@
    m = copy.copy(match)   m._files = [tostandin(f) for f in m._files] - orig_matchfn = m.matchfn - def matchfn(f): - if bfutil.is_standin(f): - return orig_matchfn(bfutil.split_standin(f)) - else: - return orig_matchfn(f) and not inctx(bfutil.standin(f), ctx2) - m.matchfn = matchfn +   # get ignored clean and unknown but remove them later if they were not asked for   try:   result = super(bfiles_repo, self).status(node1, node2, m, True, True, True, subrepos) @@ -122,7 +115,7 @@
  # was already computed using super's status.   bfdirstate = bfutil.open_bfdirstate(ui, self)   match._files = [f for f in match._files if f in bfdirstate] - s = bfdirstate.status(match, [], True, True, True) + s = bfdirstate.status(match, [], listignored, listclean, listunknown)   (unsure, modified, added, removed, missing, unknown, ignored, clean) = s   if parentworking:   for bfile in unsure: @@ -258,9 +251,29 @@
  bfiles = bfutil.list_bfiles(repo)   match = copy.copy(match)   orig_matchfn = match.matchfn - match._files = [f for f in match._files if f not in bfiles] + + # Check both the list of bfiles and the list of standins because if a bfile was removed, it + # won't be in the list of bfiles at this point   match._files += sorted(standins)   + actualfiles = [] + for f in match._files: + fstandin = bfutil.standin(f) + + # Ignore known bfiles and standins + if f in bfiles or fstandin in standins: + continue + + # Append directory separator to avoid collisions + if not fstandin.endswith('/'): + fstandin += '/' + + # Prevalidate matching standin directories + if any(st for st in match._files if st.startswith(fstandin)): + continue + actualfiles.append(f) + match._files = actualfiles +   def matchfn(f):   if orig_matchfn(f):   return f not in bfiles @@ -451,6 +464,10 @@
  for f in remove:   if not after:   os.unlink(repo.wjoin(f)) + currentdir = os.path.split(f)[0] + while currentdir and not os.listdir(repo.wjoin(currentdir)): + os.rmdir(repo.wjoin(currentdir)) + currentdir = os.path.split(currentdir)[0]   bfdirstate.remove(bfutil.unixpath(f))   bfdirstate.write()   @@ -734,6 +751,15 @@
  bfcommands.update_bfiles(repo.ui, repo)   return result   +# When we rebase a repository with remotely changed bfiles, we need +# to explicitly do a clean update so that the entries in .kbf are +# udpated and the new bfiles are pulled +def override_pull(orig, ui, repo, source="default", **opts): + result = orig(ui, repo, source, **opts) + if opts.get('rebase', False): + commands.update(repo.ui, repo, clean = True) + return result +  def override_archive(orig, repo, dest, node, kind, decode=True, matchfn=None,   prefix=None, mtime=None, subrepos=None):   # No need to lock because we are only reading history and bfile caches @@ -799,6 +825,7 @@
  getdata = ctx[f].data   if bfutil.is_standin(f):   path = bfutil.find_file(repo, getdata().strip()) + ### TODO: What if the file is not cached?   f = bfutil.split_standin(f)     def getdatafn(): @@ -992,6 +1019,7 @@
  entry[1].extend(summaryopt)     entry = extensions.wrapcommand(commands.table, 'update', override_update) + entry = extensions.wrapcommand(commands.table, 'pull', override_pull)   entry = extensions.wrapfunction(filemerge, 'filemerge', override_filemerge)   entry = extensions.wrapfunction(cmdutil, 'copy', override_copy)  
1
2
3
4
5
6
7
8
9
10
11
12
 
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
 
114
 
115
116
 
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
 
 
 
 
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
 
 
 
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
1
2
3
4
 
 
5
6
7
8
9
 
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
 
111
112
113
114
115
116
117
 
 
 
 
118
119
120
121
122
123
124
125
126
 
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
 
 
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
 '''bfiles utility code: must not import other modules in this package.'''    import os  import errno -import binascii -import tempfile  import inspect  import shutil  import stat    from mercurial import \ - util, dirstate, context, cmdutil, error, match as match_, node + util, dirstate, cmdutil, match as match_  from mercurial.i18n import _    short_name = '.kbf'  long_name = 'kilnbfiles'      # -- Portability wrappers ----------------------------------------------    if 'subrepos' in inspect.getargspec(dirstate.dirstate.status)[0]:   # for Mercurial >= 1.5   def dirstate_walk(dirstate, matcher, unknown=False, ignored=False):   return dirstate.walk(matcher, [], unknown, ignored)  else:   # for Mercurial <= 1.4   def dirstate_walk(dirstate, matcher, unknown=False, ignored=False):   return dirstate.walk(matcher, unknown, ignored)    def repo_add(repo, list):   try:   # Mercurial <= 1.5   add = repo.add   except AttributeError:   # Mercurial >= 1.6   add = repo[None].add   return add(list)    def repo_remove(repo, list, unlink=False):   try:   # Mercurial <= 1.5   remove = repo.remove   except AttributeError:   # Mercurial >= 1.6   remove = repo[None].remove   return remove(list, unlink=unlink)    def repo_forget(repo, list):   try:   # Mercurial <= 1.5   forget = repo.forget   except AttributeError:   # Mercurial >= 1.6   forget = repo[None].forget   return forget(list)    def dirstate_normaldirty(dirstate, file):   try:   normaldirty = dirstate.normaldirty   except AttributeError:   # Mercurial >= 1.6: HAAAACK: I should not be using normaldirty()   # (now called otherparent()), and dirstate in 1.6 prevents me   # from doing so. So reimplement it here until I figure out the   # right fix.   def normaldirty(f):   dirstate._dirty = True   dirstate._addpath(f)   dirstate._map[f] = ('n', 0, -2, -1)   if f in dirstate._copymap:   del dirstate._copymap[f]   normaldirty(file)    def findoutgoing(repo, remote, force):   # First attempt is for Mercurial <= 1.5 second is for >= 1.6   try:   return repo.findoutgoing(remote)   except AttributeError:   from mercurial import discovery   return discovery.findoutgoing(repo, remote, force=force)    # -- Private worker functions ------------------------------------------    if os.name == 'nt':   from mercurial import win32   linkfn = win32.os_link  else:   linkfn = os.link    def link(src, dest):   try:   linkfn(src, dest)   except OSError:   # If hardlinks fail fall back on copy   shutil.copyfile(src, dest)   os.chmod(dest, os.stat(src).st_mode)    def system_cache_path(ui, hash):   path = ui.config(long_name, 'systemcache', None)   if path:   path = os.path.join(path, hash)   else:   if os.name == 'nt':   path = os.path.join(os.getenv('LOCALAPPDATA') or os.getenv('APPDATA'), long_name, hash)   elif os.name == 'posix':   path = os.path.join(os.getenv('HOME'), '.' + long_name, hash)   else:   raise util.Abort(_('Unknown operating system: %s\n') % os.name)   return path    def in_system_cache(ui, hash):   return os.path.exists(system_cache_path(ui, hash))   -def find_file(repo, hash, getfile=True): +def find_file(repo, hash):   if in_cache(repo, hash): + repo.ui.note(_('Found %s in cache\n') % hash)   return cache_path(repo, hash)   if in_system_cache(repo.ui, hash): + repo.ui.note(_('Found %s in system cache\n') % hash)   return system_cache_path(repo.ui, hash) - if getfile: - (success, failure) = basetore._open_store(repo).get([(f, hash)]) - if (f, hash) in success: - return cache_path(repo, hash)   return None    def open_bfdirstate(ui, repo):   '''   Return a dirstate object that tracks big files: i.e. its root is the   repo root, but it is saved in .hg/bfiles/dirstate.   '''   admin = repo.join(long_name)   opener = util.opener(admin) - bfdirstate = dirstate.dirstate(opener, ui, repo.root) + if hasattr(repo.dirstate, '_validate'): + bfdirstate = dirstate.dirstate(opener, ui, repo.root, repo.dirstate._validate) + else: + bfdirstate = dirstate.dirstate(opener, ui, repo.root)     # If the bfiles dirstate does not exist, populate and create it. This   # ensures that we create it on the first meaningful bfiles operation in   # a new clone. It also gives us an easy way to forcibly rebuild bfiles   # state:   # rm .hg/bfiles/dirstate && hg bfstatus   # Or even, if things are really messed up:   # rm -rf .hg/bfiles && hg bfstatus   # (although that can lose data, e.g. pending big file revisions in   # .hg/bfiles/{pending,committed}).   if not os.path.exists(os.path.join(admin, 'dirstate')):   util.makedirs(admin)   matcher = get_standin_matcher(repo)   for standin in dirstate_walk(repo.dirstate, matcher):   bigfile = split_standin(standin)   hash = read_standin(repo, standin)   try:   curhash = hashfile(bigfile)   except IOError, err:   if err.errno == errno.ENOENT:   dirstate_normaldirty(bfdirstate, bigfile)   else:   raise   else:   if curhash == hash:   bfdirstate.normal(unixpath(bigfile))   else:   dirstate_normaldirty(bfdirstate, bigfile)     bfdirstate.write()     return bfdirstate    def bfdirstate_status(bfdirstate, repo, rev):   wlock = repo.wlock()   try:   match = match_.always(repo.root, repo.getcwd())   s = bfdirstate.status(match, [], False, False, False)   (unsure, modified, added, removed, missing, unknown, ignored, clean) = s   for bfile in unsure:   if repo[rev][standin(bfile)].data().strip() != hashfile(repo.wjoin(bfile)):   modified.append(bfile)   else:   clean.append(bfile)   bfdirstate.normal(unixpath(bfile))   bfdirstate.write()   finally:   wlock.release()   return (modified, added, removed, missing, unknown, ignored, clean)    def list_bfiles(repo, rev=None, matcher=None):   '''list big files in the working copy or specified changeset'''     if matcher is None:   matcher = get_standin_matcher(repo)     bfiles = []   if rev:   cctx = repo[rev]   for standin in cctx.walk(matcher):   filename = split_standin(standin)   bfiles.append(filename)   else:   for standin in sorted(dirstate_walk(repo.dirstate, matcher)):   filename = split_standin(standin)   bfiles.append(filename)   return bfiles    def in_cache(repo, hash):   return os.path.exists(cache_path(repo, hash))    def create_dir(dir):   if not os.path.exists(dir):   os.makedirs(dir)    def cache_path(repo, hash):   return repo.join(os.path.join(long_name, hash))    def copy_to_cache(repo, rev, file, uploaded=False):   hash = read_standin(repo, standin(file))   if in_cache(repo, hash):   return   create_dir(os.path.dirname(cache_path(repo, hash)))   if in_system_cache(repo.ui, hash):   link(system_cache_path(repo.ui, hash), cache_path(repo, hash))   else:   shutil.copyfile(repo.wjoin(file), cache_path(repo, hash))   os.chmod(cache_path(repo, hash), os.stat(repo.wjoin(file)).st_mode)   create_dir(os.path.dirname(system_cache_path(repo.ui, hash)))   link(cache_path(repo, hash), system_cache_path(repo.ui, hash))    def get_standin_matcher(repo, pats=[], opts={}):   '''Return a match object that applies pats to <repo>/.kbf.'''   standin_dir = repo.pathto(short_name)   if pats:   # patterns supplied: search .hgbfiles relative to current dir   cwd = repo.getcwd()   pats = [os.path.join(standin_dir, cwd, pat) for pat in pats]   elif os.path.isdir(standin_dir):   # no patterns: relative to repo root   pats = [standin_dir]   else:   # no patterns and no .hgbfiles dir: return matcher that matches nothing   match = match_.match(repo.root, None, [], exact=True)   match.matchfn = lambda f: False   return match   return get_matcher(repo, pats, opts, showbad=False)    def get_matcher(repo, pats=[], opts={}, showbad=True):   '''Wrapper around cmdutil.match() that adds showbad: if false, neuter   the match object\'s bad() method so it does not print any warnings   about missing files or directories.'''   match = cmdutil.match(repo, pats, opts)   if not showbad:   match.bad = lambda f, msg: None   return match    def compose_standin_matcher(repo, rmatcher):   '''Return a matcher that accepts standins corresponding to the files - accepted by rmatcher.''' - smatcher = get_standin_matcher(repo) + accepted by rmatcher. Pass the list of files in the matcher as the + paths specified by the user.''' + smatcher = get_standin_matcher(repo, rmatcher.files())   isstandin = smatcher.matchfn   def composed_matchfn(f):   return isstandin(f) and rmatcher.matchfn(split_standin(f))   smatcher.matchfn = composed_matchfn     return smatcher    def standin(filename):   '''Return the repo-relative path to the standin for the specified big   file.'''   # Notes:   # 1) Most callers want an absolute path, but _create_standin() needs   # it repo-relative so bfadd() can pass it to repo_add(). So leave   # it up to the caller to use repo.wjoin() to get an absolute path.   # 2) Join with '/' because that's what dirstate always uses, even on   # Windows. Change existing separator to '/' first in case we are   # passed filenames from an external source (like the command line).   return short_name + '/' + filename.replace(os.sep, '/')    def is_standin(filename):   '''Return true if filename is a big file standin. filename must   be in Mercurial\'s internal form (slash-separated).'''   return filename.startswith(short_name+'/')    def split_standin(filename):   # Split on / because that's what dirstate always uses, even on Windows.   # Change local separator to / first just in case we are passed filenames   # from an external source (like the command line).   bits = filename.replace(os.sep, '/').split('/', 1)   if len(bits) == 2 and bits[0] == short_name:   return bits[1]   else:   return None    def update_standin(repo, standin):   file = repo.wjoin(split_standin(standin))   hash = hashfile(file)   executable = get_executable(file)   write_standin(repo, standin, hash, executable)    def read_standin(repo, standin):   '''read hex hash from <repo.root>/<standin>'''   return read_hash(repo.wjoin(standin))    def write_standin(repo, standin, hash, executable):   '''write hhash to <repo.root>/<standin>'''   write_hash(hash, repo.wjoin(standin), executable)    def copy_and_hash(instream, outfile):   '''Read bytes from instream (iterable) and write them to outfile,   computing the SHA-1 hash of the data along the way. Close outfile   when done and return the binary hash.'''   hasher = util.sha1('')   for data in instream:   hasher.update(data)   outfile.write(data)     # Blecch: closing a file that somebody else opened is rude and   # wrong. But it's so darn convenient and practical! After all,   # outfile was opened just to copy and hash.   outfile.close()     return hasher.digest()    def hashrepofile(repo, file):   return hashfile(repo.wjoin(file))    def hashfile(file):   hasher = util.sha1('')   with open(file, 'rb') as fd:   for data in blockstream(fd):   hasher.update(data)   return hasher.hexdigest()    def blockstream(infile, blocksize=128*1024):   """Generator that yields blocks of data from infile and closes infile."""   while True:   data = infile.read(blocksize)   if not data:   break   yield data   # Same blecch as above.   infile.close()    def read_hash(filename):   rfile = open(filename, 'rb')   hash = rfile.read(40)   rfile.close()   if len(hash) < 40:   raise util.Abort(_('bad hash in \'%s\' (only %d bytes long)')   % (filename, len(hash)))   return hash    def write_hash(hash, filename, executable):   util.makedirs(os.path.dirname(filename))   if os.path.exists(filename):   os.unlink(filename)   if os.name == 'posix':   # Yuck: on Unix, go through open(2) to ensure that the caller's mode is   # filtered by umask() in the kernel, where it's supposed to be done.   wfile = os.fdopen(os.open(filename, os.O_WRONLY|os.O_CREAT, get_mode(executable)), 'wb')   else:   # But on Windows, use open() directly, since passing mode='wb' to os.fdopen()   # does not work. (Python bug?)   wfile = open(filename, 'wb')     try:   wfile.write(hash)   wfile.write('\n')   finally:   wfile.close()    def get_executable(filename):   mode = os.stat(filename).st_mode   return (mode & stat.S_IXUSR) and (mode & stat.S_IXGRP) and (mode & stat.S_IXOTH)    def get_mode(executable):   if executable:   return 0755   else:   return 0644    def urljoin(first, second, *arg):   def join(left, right):   if not left.endswith('/'):   left += '/'   if right.startswith('/'):   right = right[1:]   return left + right     url = join(first, second)   for a in arg:   url = join(url, a)   return url    # Convert a path to a unix style path. This is used to give a  # canonical path to the bfdirstate.  def unixpath(path):   return os.path.normpath(path).replace(os.sep, '/')
 
21
22
23
 
 
 
24
25
26
27
28
29
 
30
31
32
33
 
 
34
35
36
 
39
40
41
42
 
43
44
45
 
21
22
23
24
25
26
27
28
29
30
31
32
33
34
 
 
 
35
36
37
38
39
 
42
43
44
 
45
46
47
48
@@ -21,16 +21,19 @@
  self.sendfile(source, hash)   self.ui.debug('put %s to remote store\n' % source)   + def exists(self, hash): + return self._verify(hash) +   def sendfile(self, filename, hash):   if self._verify(hash):   return     self.ui.debug('httpstore.sendfile(%s, %s)\n' % (filename, hash))   baseurl, authinfo = url_.getauthinfo(self.url) + fd = None   try: - fd = open(filename, 'rb') - request = urllib2.Request(bfutil.urljoin(baseurl, hash)) - request.add_data(fd.read()) + fd = url_.httpsendfile(filename, 'rb') + request = urllib2.Request(bfutil.urljoin(baseurl, hash), fd)   try:   url = self.opener.open(request)   self.ui.note(_('[OK] %s/%s\n') % (self.rawurl, url.geturl())) @@ -39,7 +42,7 @@
  except Exception, e:   raise util.Abort(_('%s') % e)   finally: - fd.close() + if fd: fd.close()     def _getfile(self, tmpfile, filename, hash):   (baseurl, authinfo) = url_.getauthinfo(self.url)
 
18
19
20
 
 
 
21
22
23
 
18
19
20
21
22
23
24
25
26
@@ -18,6 +18,9 @@
  '''Any file that is put must already be in the system wide cache so do nothing.'''   return   + def exists(self, hash): + return bfutil.in_system_cache(self.repo.ui, hash) +   def _getfile(self, tmpfile, filename, hash):   if bfutil.in_system_cache(self.ui, hash):   return bfutil.system_cache_path(self.ui, hash)
 
19
20
21
 
22
23
24
 
29
30
31
32
 
 
 
33
34
35
 
44
45
46
 
 
47
48
49
 
62
63
64
 
 
65
66
67
 
73
74
75
 
 
76
77
78
 
86
87
88
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
89
90
91
 
19
20
21
22
23
24
25
 
30
31
32
 
33
34
35
36
37
38
 
47
48
49
50
51
52
53
54
 
67
68
69
70
71
72
73
74
 
80
81
82
83
84
85
86
87
 
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
@@ -19,6 +19,7 @@
 hgt.hg(['init', '-q'])  hgt.writefile('normal1', 'foo')  os.mkdir('sub') +os.mkdir('sub2')  hgt.writefile('sub/normal2', 'bar')  hgt.writefile('sub/normal3.txt', 'bar2')  hgt.writefile('sub/normal4.txt', 'bar3') @@ -29,7 +30,9 @@
 hgt.writefile('big1', 'abc')  hgt.writefile('sub/big2', 'xyz')  hgt.writefile('sub/big3.txt', 'xyz') -hgt.hg(['add', '-q', '--bf', 'big1', rejoin('sub/big2'), rejoin('sub/big3.txt')]) +hgt.writefile('sub/big4', 'xyz') +hgt.writefile('sub2/big5', 'xyz') +hgt.hg(['add', '-q', '--bf', 'big1', rejoin('sub/big2'), rejoin('sub/big3.txt'), rejoin('sub/big4'), rejoin('sub2/big5')])  hgt.hg(['commit', '-m', 'added bfiles'])    hgt.announce('remove sub/*.txt') @@ -44,6 +47,8 @@
 hgt.asserttrue(os.path.exists('sub/normal2'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('big1'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('sub/big2'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub/big4'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub2/big5'), 'added file doesnt exist')  hgt.hg(['status'],   stdout=('R sub/big3.txt\n'   'R sub/normal3.txt\n' @@ -62,6 +67,8 @@
 hgt.asserttrue(os.path.exists('sub/normal2'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('big1'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('sub/big2'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub/big4'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub2/big5'), 'added file doesnt exist')  hgt.hg(['up'],   stdout=('0 files updated, 0 files merged, 3 files removed, 0 files unresolved\n'   'Getting changed bfiles\n' @@ -73,6 +80,8 @@
 hgt.asserttrue(os.path.exists('sub/normal2'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('big1'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('sub/big2'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub/big4'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub2/big5'), 'added file doesnt exist')    hgt.announce('remove single normal files and add')  hgt.hg(['remove', 'normal1', 'sub/normal2']) @@ -86,6 +95,23 @@
 hgt.asserttrue(os.path.exists('sub/normal2'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('big1'), 'added file doesnt exist')  hgt.asserttrue(os.path.exists('sub/big2'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub/big4'), 'added file doesnt exist') +hgt.asserttrue(os.path.exists('sub2/big5'), 'added file doesnt exist') +hgt.hg(['status']) + +hgt.announce('remove single bfile and commit with full path') +hgt.hg(['remove', 'sub/big4']) +hgt.hg(['status'],stdout=('R sub/big4\n')) +hgt.hg(['commit', '-m', 'removing big4', 'sub/big4']) +hgt.assertfalse(os.path.exists('sub/big4'), 'removed file exists') +hgt.hg(['status']) + +hgt.announce('remove single bfile and commit with partial path') +hgt.hg(['remove', 'sub2/big5']) +hgt.hg(['status'],stdout=('R sub2/big5\n')) +hgt.assertfalse(os.path.exists("sub2"), 'removed directory structure exists') +hgt.hg(['commit', '-m', 'removing big5', 'sub2']) +hgt.assertfalse(os.path.exists('sub2/big5'), 'removed file exists')  hgt.hg(['status'])    hgt.announce('remove single bfiles and add')
Change 1 of 1 Show Entire File big-push.py Stacked
 
75
76
77
 
78
79
80
 
81
82
83
84
85
86
87
 
 
88
89
 
90
91
 
92
93
94
 
75
76
77
78
79
80
 
81
82
83
84
85
86
 
87
88
89
90
 
91
92
 
93
94
95
96
@@ -75,20 +75,22 @@
  if not opts.get('force') and not opts.get('new_branch') and None == prepush(repo, other, False, revs)[0]:   return   try: + push_size = 1   while len(outgoing) > 0:   ui.debug('start: %d to push\n' % len(outgoing)) - current_push_size = min(max_push_size, len(outgoing)) + current_push_size = min(push_size, len(outgoing))   ui.debug('pushing: %d\n' % current_push_size)   # force the push, because we checked above that by the time the whole push is done, we'll have merged back to one head   remote_heads = repo.push(other, force=True, revs=outgoing[:current_push_size])   if remote_heads: # push succeeded   outgoing = outgoing[current_push_size:] - current_push_size = max_push_size   ui.debug('pushed %d ok\n' % current_push_size) + if push_size < max_push_size: + push_size *= 2   else: # push failed; try again with a smaller size - current_push_size /= 10 + push_size /= 2   ui.debug('failed, trying %d\n' % current_push_size) - if current_push_size == 0: + if push_size == 0:   raise UnpushableChangesetError   except UnpushableChangesetError:   ui.status(_('unable to push changeset %s\n') % outgoing[0])
Change 1 of 1 Show Entire File gestalt.py Stacked
 
160
161
162
163
 
 
164
165
166
 
160
161
162
 
163
164
165
166
167
@@ -160,7 +160,8 @@
 '''))   return True   - source, hashbranch = parseurl(ui.expandpath('default-push', ui.expandpath('default'))) + target = ui.config('paths', 'default-push') and ui.expandpath('default-push') or source + source, hashbranch = parseurl(source)   other = hg.repository(remoteui(repo, opts), source)   revs = addbranchrevs(repo, other, hashbranch)   ui.pushbuffer()
Change 1 of 6 Show Entire File kiln.py Stacked
 
1
 
2
3
4
 
38
39
40
 
41
42
43
44
45
 
 
46
47
48
 
66
67
68
69
70
71
72
73
 
174
175
176
177
 
178
179
180
 
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
202
203
204
205
 
 
 
206
207
208
 
556
557
558
559
560
561
 
562
563
564
 
569
570
571
572
 
573
574
575
 
 
1
2
3
4
 
38
39
40
41
42
43
44
45
 
46
47
48
49
50
 
68
69
70
 
 
71
72
73
 
174
175
176
 
177
178
 
 
179
180
181
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
 
 
 
 
208
209
210
211
212
213
 
561
562
563
 
 
 
564
565
566
567
 
572
573
574
 
575
576
577
578
@@ -1,4 +1,4 @@
-# Copyright (C) 2010 Fog Creek Software. All rights reserved. +# Copyright (C) 2011 Fog Creek Software. All rights reserved.  #  # To enable the "kiln" extension put these lines in your ~/.hgrc:  # [extensions] @@ -38,11 +38,13 @@
 import re  import urllib  import urllib2 +import subprocess  import sys    from cookielib import MozillaCookieJar  from hashlib import md5 -from mercurial import extensions, commands, demandimport, hg, util, httprepo, match +from mercurial import extensions, commands, demandimport, hg, util, httprepo, localrepo, match +from mercurial import ui as hgui  from mercurial import url as hgurl  from mercurial.error import RepoError  from mercurial.i18n import _ @@ -66,8 +68,6 @@
  win32api.ShellExecute(0, 'open', escape_reserved(url), None, None, 0)  demandimport.enable()   -KILN_CAPABILITY_PREFIX = 'kiln-' -KILN_CURRENT_VERSION = '1.0.0'  _did_version_check = False    class APIError(Exception): @@ -174,35 +174,40 @@
  ret = re.sub(r'([A-Za-z]):', r'\1:\\', ret)   return ret   -def _versioncheck(ui, repo, str): +def _upgradecheck(ui, repo):   global _did_version_check - m = re.match(KILN_CAPABILITY_PREFIX + '(?P<version>[0-9.]+).*', str) - if _did_version_check or not m: + if _did_version_check or not ui.configbool('kiln', 'autoupdate', True):   return   _did_version_check = True - version = m.group('version') - server_version = [int(s) for s in version.split('.')] - my_version = [int(s) for s in KILN_CURRENT_VERSION.split('.')] - ignore_version = [int(s) for s in ui.config('kiln', 'ignoreversion', '0.0.0').split('.')] - if server_version > my_version: - url = urljoin(repo.url()[:repo.url().lower().index('/repo')], 'Tools') - if server_version > ignore_version: - if ui.promptchoice(_('You are currently running Kiln client tools version %s. ' - 'Version %s is available.\nUpgrade now? (y/n)') % - (KILN_CURRENT_VERSION, version), ('&No', '&Yes'), default=0): - browse(url) - else: - if os.name == 'nt': - config_file = 'Mercurial.ini' - else: - config_file = '~/.hgrc' - ui.write(_('''If you'd like Kiln to stop prompting you about version %s and below, ''' - '''add ignoreversion=%s to the [kiln] section of your %s\n''') % (version, version, config_file)) + _upgrade(ui, repo) + +def _upgrade(ui, repo): + ext_dir = os.path.dirname(os.path.abspath(__file__)) + ui.debug('kiln: checking for extensions upgrade for %s\n' % ext_dir) + + try: + r = localrepo.localrepository(hgui.ui(), ext_dir) + except RepoError: + commands.init(hgui.ui(), dest=ext_dir) + r = localrepo.localrepository(hgui.ui(), ext_dir) + + r.ui.setconfig('kiln', 'autoupdate', False) + r.ui.pushbuffer() + try: + source = 'https://developers.kilnhg.com/Repo/Kiln/Group/Kiln-Extensions' + if commands.incoming(r.ui, r, bundle=None, force=False, source=source) != 0: + # no incoming changesets, or an error. Don't try to upgrade. + ui.debug('kiln: no extensions upgrade available\n') + return + ui.write(_('updating Kiln Extensions at %s... ') % ext_dir) + # pull and update return falsy values on success + if commands.pull(r.ui, r, source=source) or commands.update(r.ui, r, clean=True): + url = urljoin(repo.url()[:repo.url().lower().index('/repo')], 'Tools') + ui.write(_('unable to update\nvisit %s to download the newest extensions\n') % url)   else: - ui.write(_('You are currently running Kiln client tools version %s. ' - 'Version %s is available.\nVisit %s to download the new client tools.\n') % - (KILN_CURRENT_VERSION, version, url)) - ui.write('\n') + ui.write(_('complete\n')) + except Exception, e: + ui.debug(_('kiln: error updating Kiln Extensions: %s\n') % e)    def is_dest_a_path(ui, dest):   paths = ui.configitems('paths') @@ -556,9 +561,7 @@
   def reposetup(ui, repo):   if issubclass(repo.__class__, httprepo.httprepository): - for cap in repo.capabilities: - if cap.startswith(KILN_CAPABILITY_PREFIX): - _versioncheck(ui, repo, cap) + _upgradecheck(ui, repo)    cmdtable = {   'kiln': @@ -569,7 +572,7 @@
  ('l', 'filehistory', [], _('view the history of the file')),   ('o', 'outgoing', None, _('view the repository\'s outgoing tab')),   ('s', 'settings', None, _('view the repository\'s settings tab')), - ('p', 'path', '', _('override the default URL to use for Kiln')), + ('p', 'path', '', _('select which Kiln branch of the repository to use')),   ('r', 'rev', [], _('view the specified changeset in Kiln')),   ('t', 'targets', None, _('view the repository\'s targets')),   ('', 'logout', None, _('log out of Kiln sessions'))],
Change 1 of 3 Show Entire File kilnauth.py Stacked
 
36
37
38
 
39
 
40
41
42
 
49
50
51
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
53
54
 
69
70
71
72
 
73
74
75
 
36
37
38
39
40
41
42
43
44
 
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
 
102
103
104
 
105
106
107
108
@@ -36,7 +36,9 @@
 import re  from urllib2 import Request  from cookielib import MozillaCookieJar, Cookie +import shutil  import sys +import tempfile    try:   from hashlib import md5 @@ -49,6 +51,37 @@
   current_user = None   +class CookieJar(MozillaCookieJar, object): + def __init__(self, filename, *args, **kwargs): + self.__original_path = filename + tf = tempfile.NamedTemporaryFile(delete=False) + self.__temporary_path = tf.name + tf.close() + if os.path.exists(filename): + shutil.copyfile(filename, self.__temporary_path) + return super(CookieJar, self).__init__(self.__temporary_path, *args, **kwargs) + + def __enter__(self): + pass + + def __exit__(self, exc_type, exc_value, traceback): + os.unlink(self.__temporary_path) + self.__temporary_path = None + + def __del__(self): + try: + if self.__temporary_path: + os.unlink(self.__temporary_path) + except (OSError, IOError): + pass + + def save(self, *args, **kwargs): + super(CookieJar, self).save(*args, **kwargs) + try: + shutil.copyfile(self.__temporary_path, self.__original_path) + except IOError: + pass +  def get_cookiejar(ui):   global current_user   if os.name == 'nt': @@ -69,7 +102,7 @@
  cookie_path = re.sub(r'([A-Za-z]):', r'\1:\\', cookie_path)     try: - cj = MozillaCookieJar(cookie_path) + cj = CookieJar(cookie_path)   if not os.path.exists(cookie_path):   cj.save()   if os.name == 'posix':