mirror of
https://github.com/newren/git-filter-repo.git
synced 2024-07-06 18:32:14 +02:00
filter-repo (python3): error messages should be strings instead of bytes
Signed-off-by: Elijah Newren <newren@gmail.com>
This commit is contained in:
parent
9b3134b68c
commit
0279e3882d
@ -265,7 +265,7 @@ class MailmapInfo(object):
|
|||||||
name_and_email_re = re.compile(r'(.*?)\s*<([^>]+)>\s*')
|
name_and_email_re = re.compile(r'(.*?)\s*<([^>]+)>\s*')
|
||||||
comment_re = re.compile(r'\s*#.*')
|
comment_re = re.compile(r'\s*#.*')
|
||||||
if not os.access(filename, os.R_OK):
|
if not os.access(filename, os.R_OK):
|
||||||
raise SystemExit(_("Cannot read %s") % filename)
|
raise SystemExit(_("Cannot read %s") % decode(filename))
|
||||||
with open(filename, 'br') as f:
|
with open(filename, 'br') as f:
|
||||||
count = 0
|
count = 0
|
||||||
for line in f:
|
for line in f:
|
||||||
@ -2450,8 +2450,8 @@ class RepoAnalyze(object):
|
|||||||
if equiv_group in seen:
|
if equiv_group in seen:
|
||||||
continue
|
continue
|
||||||
seen.add(equiv_group)
|
seen.add(equiv_group)
|
||||||
f.write("{} ->\n ".format(equiv_group[0]) +
|
f.write("{} ->\n ".format(decode(equiv_group[0])) +
|
||||||
"\n ".join(equiv_group[1:]) +
|
"\n ".join(decode(x) for x in equiv_group[1:]) +
|
||||||
"\n")
|
"\n")
|
||||||
|
|
||||||
# List directories in reverse sorted order of unpacked size
|
# List directories in reverse sorted order of unpacked size
|
||||||
@ -2565,7 +2565,7 @@ class RepoAnalyze(object):
|
|||||||
stats = RepoAnalyze.gather_data(args)
|
stats = RepoAnalyze.gather_data(args)
|
||||||
|
|
||||||
# Write the reports
|
# Write the reports
|
||||||
sys.stdout.write(_("Writing reports to %s...") % reportdir)
|
sys.stdout.write(_("Writing reports to %s...") % decode(reportdir))
|
||||||
sys.stdout.flush()
|
sys.stdout.flush()
|
||||||
RepoAnalyze.write_report(reportdir, stats)
|
RepoAnalyze.write_report(reportdir, stats)
|
||||||
sys.stdout.write(_("done.\n"))
|
sys.stdout.write(_("done.\n"))
|
||||||
@ -2739,7 +2739,7 @@ class RepoFilter(object):
|
|||||||
if len(f.read().splitlines()) > 1:
|
if len(f.read().splitlines()) > 1:
|
||||||
shortpath = pathname[len(reflog_dir)+1:]
|
shortpath = pathname[len(reflog_dir)+1:]
|
||||||
abort(_("expected at most one entry in the reflog for %s") %
|
abort(_("expected at most one entry in the reflog for %s") %
|
||||||
shortpath)
|
decode(shortpath))
|
||||||
|
|
||||||
# Make sure there are no stashed changes
|
# Make sure there are no stashed changes
|
||||||
if 'refs/stash' in refs:
|
if 'refs/stash' in refs:
|
||||||
@ -2761,9 +2761,11 @@ class RepoFilter(object):
|
|||||||
continue
|
continue
|
||||||
origin_ref = refname.replace('refs/heads/', 'refs/remotes/origin/')
|
origin_ref = refname.replace('refs/heads/', 'refs/remotes/origin/')
|
||||||
if origin_ref not in refs:
|
if origin_ref not in refs:
|
||||||
abort(_('%s exists, but %s not found') % (refname, origin_ref))
|
abort(_('%s exists, but %s not found') % (decode(refname),
|
||||||
|
decode(origin_ref)))
|
||||||
if rev != refs[origin_ref]:
|
if rev != refs[origin_ref]:
|
||||||
abort(_('%s does not match %s') % (refname, origin_ref))
|
abort(_('%s does not match %s') % (decode(refname),
|
||||||
|
decode(origin_ref)))
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def tweak_blob(args, blob):
|
def tweak_blob(args, blob):
|
||||||
@ -2974,7 +2976,8 @@ class RepoFilter(object):
|
|||||||
self._input = InputFileBackup(self._input, output)
|
self._input = InputFileBackup(self._input, output)
|
||||||
if self._args.debug:
|
if self._args.debug:
|
||||||
print("[DEBUG] Running: {}".format(' '.join(fep_cmd)))
|
print("[DEBUG] Running: {}".format(' '.join(fep_cmd)))
|
||||||
print(" (saving a copy of the output at {})".format(self._fe_orig))
|
print(" (saving a copy of the output at {})"
|
||||||
|
.format(decode(self._fe_orig)))
|
||||||
|
|
||||||
def _setup_output(self):
|
def _setup_output(self):
|
||||||
if not self._args.dry_run:
|
if not self._args.dry_run:
|
||||||
@ -2994,7 +2997,8 @@ class RepoFilter(object):
|
|||||||
if self._args.debug:
|
if self._args.debug:
|
||||||
self._output = DualFileWriter(self._fip.stdin, self._output)
|
self._output = DualFileWriter(self._fip.stdin, self._output)
|
||||||
print("[DEBUG] Running: {}".format(' '.join(fip_cmd)))
|
print("[DEBUG] Running: {}".format(' '.join(fip_cmd)))
|
||||||
print(" (using the following file as input: {})".format(self._fe_filt))
|
print(" (using the following file as input: {})"
|
||||||
|
.format(decode(self._fe_filt)))
|
||||||
|
|
||||||
def _migrate_origin_to_heads(self):
|
def _migrate_origin_to_heads(self):
|
||||||
if self._args.dry_run:
|
if self._args.dry_run:
|
||||||
@ -3099,10 +3103,10 @@ class RepoFilter(object):
|
|||||||
print(_("NOTE: Not running fast-import or cleaning up; --dry-run passed."))
|
print(_("NOTE: Not running fast-import or cleaning up; --dry-run passed."))
|
||||||
if self._fe_orig:
|
if self._fe_orig:
|
||||||
print(_(" Requested filtering can be seen by comparing:"))
|
print(_(" Requested filtering can be seen by comparing:"))
|
||||||
print(" " + self._fe_orig)
|
print(" " + decode(self._fe_orig))
|
||||||
else:
|
else:
|
||||||
print(_(" Requested filtering can be seen at:"))
|
print(_(" Requested filtering can be seen at:"))
|
||||||
print(" " + self._fe_filt)
|
print(" " + decode(self._fe_filt))
|
||||||
return
|
return
|
||||||
|
|
||||||
target_working_dir = self._args.target or '.'
|
target_working_dir = self._args.target or '.'
|
||||||
@ -3112,7 +3116,7 @@ class RepoFilter(object):
|
|||||||
if refs_to_nuke:
|
if refs_to_nuke:
|
||||||
if self._args.debug:
|
if self._args.debug:
|
||||||
print("[DEBUG] Deleting the following refs:\n "+
|
print("[DEBUG] Deleting the following refs:\n "+
|
||||||
"\n ".join(refs_to_nuke))
|
decode("\n ".join(refs_to_nuke)))
|
||||||
p = subprocess.Popen('git update-ref --stdin'.split(),
|
p = subprocess.Popen('git update-ref --stdin'.split(),
|
||||||
stdin=subprocess.PIPE,
|
stdin=subprocess.PIPE,
|
||||||
cwd=target_working_dir)
|
cwd=target_working_dir)
|
||||||
|
Loading…
Reference in New Issue
Block a user