mirror of
https://github.com/open-source-parsers/jsoncpp.git
synced 2024-12-12 18:10:27 +01:00
64ba062076
- makerelease now decompress the tarball, download and install scons, run scons check on provided platforms, decompress doc tarball and upload the doc on the project web
345 lines
13 KiB
Python
345 lines
13 KiB
Python
"""Tag the sandbox for release, make source and doc tarballs.
|
|
|
|
Requires Python 2.6
|
|
|
|
Example of invocation (use to test the script):
|
|
python makerelease.py --force --retag --platform=msvc6,msvc71,msvc80,mingw -ublep 0.5.0 0.6.0-dev
|
|
|
|
Example of invocation when doing a release:
|
|
python makerelease.py 0.5.0 0.6.0-dev
|
|
"""
|
|
import os.path
|
|
import subprocess
|
|
import sys
|
|
import doxybuild
|
|
import subprocess
|
|
import xml.etree.ElementTree as ElementTree
|
|
import shutil
|
|
import urllib2
|
|
import tempfile
|
|
import os
|
|
import time
|
|
from devtools import antglob, fixeol, tarball
|
|
|
|
SVN_ROOT = 'https://jsoncpp.svn.sourceforge.net/svnroot/jsoncpp/'
|
|
SVN_TAG_ROOT = SVN_ROOT + 'tags/jsoncpp'
|
|
SCONS_LOCAL_URL = 'http://sourceforge.net/projects/scons/files/scons-local/1.2.0/scons-local-1.2.0.tar.gz/download'
|
|
SOURCEFORGE_PROJECT = 'jsoncpp'
|
|
|
|
def set_version( version ):
|
|
with open('version','wb') as f:
|
|
f.write( version.strip() )
|
|
|
|
def rmdir_if_exist( dir_path ):
|
|
if os.path.isdir( dir_path ):
|
|
shutil.rmtree( dir_path )
|
|
|
|
class SVNError(Exception):
|
|
pass
|
|
|
|
def svn_command( command, *args ):
|
|
cmd = ['svn', '--non-interactive', command] + list(args)
|
|
print 'Running:', ' '.join( cmd )
|
|
process = subprocess.Popen( cmd,
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.STDOUT )
|
|
stdout = process.communicate()[0]
|
|
if process.returncode:
|
|
error = SVNError( 'SVN command failed:\n' + stdout )
|
|
error.returncode = process.returncode
|
|
raise error
|
|
return stdout
|
|
|
|
def check_no_pending_commit():
|
|
"""Checks that there is no pending commit in the sandbox."""
|
|
stdout = svn_command( 'status', '--xml' )
|
|
etree = ElementTree.fromstring( stdout )
|
|
msg = []
|
|
for entry in etree.getiterator( 'entry' ):
|
|
path = entry.get('path')
|
|
status = entry.find('wc-status').get('item')
|
|
if status != 'unversioned':
|
|
msg.append( 'File "%s" has pending change (status="%s")' % (path, status) )
|
|
if msg:
|
|
msg.insert(0, 'Pending change to commit found in sandbox. Commit them first!' )
|
|
return '\n'.join( msg )
|
|
|
|
def svn_join_url( base_url, suffix ):
|
|
if not base_url.endswith('/'):
|
|
base_url += '/'
|
|
if suffix.startswith('/'):
|
|
suffix = suffix[1:]
|
|
return base_url + suffix
|
|
|
|
def svn_check_if_tag_exist( tag_url ):
|
|
"""Checks if a tag exist.
|
|
Returns: True if the tag exist, False otherwise.
|
|
"""
|
|
try:
|
|
list_stdout = svn_command( 'list', tag_url )
|
|
except SVNError, e:
|
|
if e.returncode != 1 or not str(e).find('tag_url'):
|
|
raise e
|
|
# otherwise ignore error, meaning tag does not exist
|
|
return False
|
|
return True
|
|
|
|
def svn_tag_sandbox( tag_url, message ):
|
|
"""Makes a tag based on the sandbox revisions.
|
|
"""
|
|
svn_command( 'copy', '-m', message, '.', tag_url )
|
|
|
|
def svn_remove_tag( tag_url, message ):
|
|
"""Removes an existing tag.
|
|
"""
|
|
svn_command( 'delete', '-m', message, tag_url )
|
|
|
|
def svn_export( tag_url, export_dir ):
|
|
"""Exports the tag_url revision to export_dir.
|
|
Target directory, including its parent is created if it does not exist.
|
|
If the directory export_dir exist, it is deleted before export proceed.
|
|
"""
|
|
rmdir_if_exist( export_dir )
|
|
svn_command( 'export', tag_url, export_dir )
|
|
|
|
def fix_sources_eol( dist_dir ):
|
|
"""Set file EOL for tarball distribution.
|
|
"""
|
|
print 'Preparing exported source file EOL for distribution...'
|
|
prune_dirs = antglob.prune_dirs + 'scons-local* ./build* ./libs ./dist'
|
|
win_sources = antglob.glob( dist_dir,
|
|
includes = '**/*.sln **/*.vcproj',
|
|
prune_dirs = prune_dirs )
|
|
unix_sources = antglob.glob( dist_dir,
|
|
includes = '''**/*.h **/*.cpp **/*.inl **/*.txt **/*.dox **/*.py **/*.html **/*.in
|
|
sconscript *.json *.expected AUTHORS LICENSE''',
|
|
excludes = antglob.default_excludes + 'scons.py sconsign.py scons-*',
|
|
prune_dirs = prune_dirs )
|
|
for path in win_sources:
|
|
fixeol.fix_source_eol( path, is_dry_run = False, verbose = True, eol = '\r\n' )
|
|
for path in unix_sources:
|
|
fixeol.fix_source_eol( path, is_dry_run = False, verbose = True, eol = '\n' )
|
|
|
|
def download( url, target_path ):
|
|
"""Download file represented by url to target_path.
|
|
"""
|
|
f = urllib2.urlopen( url )
|
|
try:
|
|
data = f.read()
|
|
finally:
|
|
f.close()
|
|
fout = open( target_path, 'wb' )
|
|
try:
|
|
fout.write( data )
|
|
finally:
|
|
fout.close()
|
|
|
|
def check_compile( distcheck_top_dir, platform ):
|
|
cmd = [sys.executable, 'scons.py', 'platform=%s' % platform, 'check']
|
|
print 'Running:', ' '.join( cmd )
|
|
log_path = os.path.join( distcheck_top_dir, 'build-%s.log' % platform )
|
|
flog = open( log_path, 'wb' )
|
|
try:
|
|
process = subprocess.Popen( cmd,
|
|
stdout=flog,
|
|
stderr=subprocess.STDOUT,
|
|
cwd=distcheck_top_dir )
|
|
stdout = process.communicate()[0]
|
|
status = (process.returncode == 0)
|
|
finally:
|
|
flog.close()
|
|
return (status, log_path)
|
|
|
|
def write_tempfile( content, **kwargs ):
|
|
fd, path = tempfile.mkstemp( **kwargs )
|
|
f = os.fdopen( fd, 'wt' )
|
|
try:
|
|
f.write( content )
|
|
finally:
|
|
f.close()
|
|
return path
|
|
|
|
class SFTPError(Exception):
|
|
pass
|
|
|
|
def run_sftp_batch( userhost, sftp, batch, retry=0 ):
|
|
path = write_tempfile( batch, suffix='.sftp', text=True )
|
|
# psftp -agent -C blep,jsoncpp@web.sourceforge.net -batch -b batch.sftp -bc
|
|
cmd = [sftp, '-agent', '-C', '-batch', '-b', path, '-bc', userhost]
|
|
error = None
|
|
for retry_index in xrange(0, max(1,retry)):
|
|
heading = retry_index == 0 and 'Running:' or 'Retrying:'
|
|
print heading, ' '.join( cmd )
|
|
process = subprocess.Popen( cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT )
|
|
stdout = process.communicate()[0]
|
|
if process.returncode != 0:
|
|
error = SFTPError( 'SFTP batch failed:\n' + stdout )
|
|
else:
|
|
break
|
|
if error:
|
|
raise error
|
|
return stdout
|
|
|
|
def sourceforge_web_synchro( sourceforge_project, doc_dir,
|
|
user=None, sftp='sftp' ):
|
|
"""Notes: does not synchronize sub-directory of doc-dir.
|
|
"""
|
|
userhost = '%s,%s@web.sourceforge.net' % (user, sourceforge_project)
|
|
stdout = run_sftp_batch( userhost, sftp, """
|
|
cd htdocs
|
|
dir
|
|
exit
|
|
""" )
|
|
existing_paths = set()
|
|
collect = 0
|
|
for line in stdout.split('\n'):
|
|
line = line.strip()
|
|
if not collect and line.endswith('> dir'):
|
|
collect = True
|
|
elif collect and line.endswith('> exit'):
|
|
break
|
|
elif collect == 1:
|
|
collect = 2
|
|
elif collect == 2:
|
|
path = line.strip().split()[-1:]
|
|
if path and path[0] not in ('.', '..'):
|
|
existing_paths.add( path[0] )
|
|
upload_paths = set( [os.path.basename(p) for p in antglob.glob( doc_dir )] )
|
|
paths_to_remove = existing_paths - upload_paths
|
|
if paths_to_remove:
|
|
print 'Removing the following file from web:'
|
|
print '\n'.join( paths_to_remove )
|
|
stdout = run_sftp_batch( userhost, sftp, """cd htdocs
|
|
rm %s
|
|
exit""" % ' '.join(paths_to_remove) )
|
|
print 'Uploading %d files:' % len(upload_paths)
|
|
batch_size = 10
|
|
upload_paths = list(upload_paths)
|
|
start_time = time.time()
|
|
for index in xrange(0,len(upload_paths),batch_size):
|
|
paths = upload_paths[index:index+batch_size]
|
|
file_per_sec = (time.time() - start_time) / (index+1)
|
|
remaining_files = len(upload_paths) - index
|
|
remaining_sec = file_per_sec * remaining_files
|
|
print '%d/%d, ETA=%.1fs' % (index+1, len(upload_paths), remaining_sec)
|
|
run_sftp_batch( userhost, sftp, """cd htdocs
|
|
lcd %s
|
|
mput %s
|
|
exit""" % (doc_dir, ' '.join(paths) ), retry=3 )
|
|
|
|
|
|
def main():
|
|
usage = """%prog release_version next_dev_version
|
|
Update 'version' file to release_version and commit.
|
|
Generates the document tarball.
|
|
Tags the sandbox revision with release_version.
|
|
Update 'version' file to next_dev_version and commit.
|
|
|
|
Performs an svn export of tag release version, and build a source tarball.
|
|
|
|
Must be started in the project top directory.
|
|
|
|
Warning: --force should only be used when developping/testing the release script.
|
|
"""
|
|
from optparse import OptionParser
|
|
parser = OptionParser(usage=usage)
|
|
parser.allow_interspersed_args = False
|
|
parser.add_option('--dot', dest="dot_path", action='store', default=doxybuild.find_program('dot'),
|
|
help="""Path to GraphViz dot tool. Must be full qualified path. [Default: %default]""")
|
|
parser.add_option('--doxygen', dest="doxygen_path", action='store', default=doxybuild.find_program('doxygen'),
|
|
help="""Path to Doxygen tool. [Default: %default]""")
|
|
parser.add_option('--force', dest="ignore_pending_commit", action='store_true', default=False,
|
|
help="""Ignore pending commit. [Default: %default]""")
|
|
parser.add_option('--retag', dest="retag_release", action='store_true', default=False,
|
|
help="""Overwrite release existing tag if it exist. [Default: %default]""")
|
|
parser.add_option('-p', '--platforms', dest="platforms", action='store', default='',
|
|
help="""Comma separated list of platform passed to scons for build check.""")
|
|
parser.add_option('--no-test', dest="no_test", action='store', default=False,
|
|
help="""Skips build check.""")
|
|
parser.add_option('-u', '--upload-user', dest="user", action='store',
|
|
help="""Sourceforge user for SFTP documentation upload.""")
|
|
parser.add_option('--sftp', dest='sftp', action='store', default=doxybuild.find_program('psftp', 'sftp'),
|
|
help="""Path of the SFTP compatible binary used to upload the documentation.""")
|
|
parser.enable_interspersed_args()
|
|
options, args = parser.parse_args()
|
|
|
|
if len(args) < 1:
|
|
parser.error( 'release_version missing on command-line.' )
|
|
release_version = args[0]
|
|
|
|
if not options.platforms and not options.no_test:
|
|
parser.error( 'You must specify either --platform or --no-test option.' )
|
|
|
|
if options.ignore_pending_commit:
|
|
msg = ''
|
|
else:
|
|
msg = check_no_pending_commit()
|
|
if not msg:
|
|
print 'Setting version to', release_version
|
|
set_version( release_version )
|
|
tag_url = svn_join_url( SVN_TAG_ROOT, release_version )
|
|
if svn_check_if_tag_exist( tag_url ):
|
|
if options.retag_release:
|
|
svn_remove_tag( tag_url, 'Overwriting previous tag' )
|
|
else:
|
|
print 'Aborting, tag %s already exist. Use --retag to overwrite it!' % tag_url
|
|
sys.exit( 1 )
|
|
svn_tag_sandbox( tag_url, 'Release ' + release_version )
|
|
|
|
print 'Generated doxygen document...'
|
|
## doc_dirname = r'jsoncpp-api-html-0.5.0'
|
|
## doc_tarball_path = r'e:\prg\vc\Lib\jsoncpp-trunk\dist\jsoncpp-api-html-0.5.0.tar.gz'
|
|
doc_tarball_path, doc_dirname = doxybuild.build_doc( options, make_release=True )
|
|
doc_distcheck_dir = 'dist/doccheck'
|
|
tarball.decompress( doc_tarball_path, doc_distcheck_dir )
|
|
doc_distcheck_top_dir = os.path.join( doc_distcheck_dir, doc_dirname )
|
|
|
|
export_dir = 'dist/export'
|
|
svn_export( tag_url, export_dir )
|
|
fix_sources_eol( export_dir )
|
|
|
|
source_dir = 'jsoncpp-src-' + release_version
|
|
source_tarball_path = 'dist/%s.tar.gz' % source_dir
|
|
print 'Generating source tarball to', source_tarball_path
|
|
tarball.make_tarball( source_tarball_path, [export_dir], export_dir, prefix_dir=source_dir )
|
|
|
|
# Decompress source tarball, download and install scons-local
|
|
distcheck_dir = 'dist/distcheck'
|
|
distcheck_top_dir = distcheck_dir + '/' + source_dir
|
|
print 'Decompressing source tarball to', distcheck_dir
|
|
rmdir_if_exist( distcheck_dir )
|
|
tarball.decompress( source_tarball_path, distcheck_dir )
|
|
scons_local_path = 'dist/scons-local.tar.gz'
|
|
print 'Downloading scons-local to', scons_local_path
|
|
download( SCONS_LOCAL_URL, scons_local_path )
|
|
print 'Decompressing scons-local to', distcheck_top_dir
|
|
tarball.decompress( scons_local_path, distcheck_top_dir )
|
|
|
|
# Run compilation
|
|
print 'Compiling decompressed tarball'
|
|
all_build_status = True
|
|
for platform in options.platforms.split(','):
|
|
print 'Testing platform:', platform
|
|
build_status, log_path = check_compile( distcheck_top_dir, platform )
|
|
print 'see build log:', log_path
|
|
print build_status and '=> ok' or '=> FAILED'
|
|
all_build_status = all_build_status and build_status
|
|
if not build_status:
|
|
print 'Testing failed on at least one platform, aborting...'
|
|
svn_remove_tag( tag_url, 'Removing tag due to failed testing' )
|
|
sys.exit(1)
|
|
if options.user:
|
|
print 'Uploading documentation using user', options.user
|
|
sourceforge_web_synchro( SOURCEFORGE_PROJECT, doc_distcheck_top_dir, user=options.user, sftp=options.sftp )
|
|
print 'Completed documentatio upload'
|
|
else:
|
|
print 'No upload user specified. Documentation was not upload.'
|
|
print 'Tarball can be found at:', doc_tarball_path
|
|
#@todo:
|
|
#upload source & doc tarballs
|
|
else:
|
|
sys.stderr.write( msg + '\n' )
|
|
|
|
if __name__ == '__main__':
|
|
main()
|