aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rwxr-xr-xcreate-git.sh2
-rwxr-xr-xprocess_directory.sh2
-rwxr-xr-xrewrite-commit-dump.py11
3 files changed, 7 insertions, 8 deletions
diff --git a/create-git.sh b/create-git.sh
index dab679e..97448ff 100755
--- a/create-git.sh
+++ b/create-git.sh
@@ -48,7 +48,7 @@ time {
( cd "${root}"; ./rewrite-commit-dump.py; ) | \
( read line; { echo "$line"; cat; } | \
tee ../export-stream-rewritten |\
- time git fast-import
+ git fast-import
)
} 2>&1 > >(tee git-creation.log)
ret=$?
diff --git a/process_directory.sh b/process_directory.sh
index cb1041b..9684f1e 100755
--- a/process_directory.sh
+++ b/process_directory.sh
@@ -13,7 +13,7 @@ f() {
# Note- this must be canonical path, else it screws up our $Header rewriting.
cd "$(readlink -f "${output}" )"
export PYTHONPATH="${output}${PYTHONPATH:+:${PYTHONPATH}}"
- time cvs2git --options config -vv
+ time cvs2git --options config -v
cd git
git init --bare
# Note we're only pull in blob data here; this intentional- we need to
diff --git a/rewrite-commit-dump.py b/rewrite-commit-dump.py
index 607c7a5..3744fb3 100755
--- a/rewrite-commit-dump.py
+++ b/rewrite-commit-dump.py
@@ -201,7 +201,7 @@ def serialize_records(records, handle, target='refs/heads/master', progress=100)
for idx, record in enumerate(records, 1):
if idx % progress_interval == 0:
write('progress %s%%: %s of %i commits\n'
- % (str((100 * float(idx))/total).rjust(2), str(idx).rjust(total_len), total))
+ % (str(int(100 * (float(idx)/total))).rjust(2), str(idx).rjust(total_len), total))
write('commit %s\n' % target)
write('mark :%i\n' % idx)
# fields = ('mark', 'author', 'committer', 'msg', 'files')
@@ -295,7 +295,7 @@ import traceback
def process_record(data):
try:
return _process_record(data)
- except Exception, e:
+ except Exception:
return traceback.format_exc()
def _process_record(data):
@@ -343,10 +343,8 @@ def thin_manifest_conversion(records, processing_pool):
potentials.append((idx, manifests, record))
rewrites = deletes = 0
- processed = 0
for result in processing_pool.imap_unordered(
process_record, potentials, chunksize=30):
- processed += 1
if result is not None:
if not isinstance(result, tuple):
raise Exception(result)
@@ -369,8 +367,9 @@ def process_directory(paths):
deserialize_records(data, deserialize_blob_map(idx_path))))
def main(argv):
- # allocate the pool now, before we start getting memory abusive
- clean_pool = multiprocessing.Pool()
+ # allocate the pool now, before we start getting memory abusive; this is
+ # used for thin-manifest conversion if active/enabled.
+ #clean_pool = multiprocessing.Pool()
# Be careful here to just iterate over source; doing so allows this script
# to do basic processing as it goes (specifically while it's being fed from