[OE-core] [PATCH 1/6] archiver.bbclass: indent fixes

Robert Yang liezhi.yang at windriver.com
Sun Aug 26 06:51:56 UTC 2012


Several fixes:
* It uses mixed tab and whitespace as the indent in one function,
  Fix them to use "4 spaces" as the indent.

* Remove the unwanted blank, for example, more than one blank lines
  appeared together.

* Remove the tail whitespace.

[YOCTO #2619]

Signed-off-by: Robert Yang <liezhi.yang at windriver.com>
---
 meta/classes/archiver.bbclass    | 757 ++++++++++++++++++++-------------------
 meta/classes/package_rpm.bbclass |   1 -
 2 files changed, 379 insertions(+), 379 deletions(-)

diff --git a/meta/classes/archiver.bbclass b/meta/classes/archiver.bbclass
index 8e97e44..117ad02 100644
--- a/meta/classes/archiver.bbclass
+++ b/meta/classes/archiver.bbclass
@@ -1,6 +1,10 @@
+# ex:ts=4:sw=4:sts=4:et
+# -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*-
+#
 # This file is used for archiving sources ,patches,and logs to tarball.
-# It also output building environment to xxx.dump.data and create xxx.diff.gz to record 
-# all content in ${S} to a diff file. 
+# It also output building environment to xxx.dump.data and create xxx.diff.gz to record
+# all content in ${S} to a diff file.
+#
 
 ARCHIVE_EXCLUDE_FROM ?= ".pc autom4te.cache"
 ARCHIVE_TYPE ?= "TAR SRPM"
@@ -67,454 +71,451 @@ def tar_filter(d):
     if d.getVar('FILTER', True).upper() == "YES":
         included, reason = copyleft_should_include(d)
         if not included:
-                return False
+            return False
         else:
-                return True
+            return True
     else:
         return False
 
 def get_bb_inc(d):
-	'''create a directory "script-logs" including .bb and .inc file in ${WORKDIR}'''
-	import re
-	import shutil
-	
-	bbinc = []
-	pat=re.compile('require\s*([^\s]*\.*)(.*)')
-	work_dir = d.getVar('WORKDIR', True)
-	bbfile = d.getVar('FILE', True)
-	bbdir = os.path.dirname(bbfile)
-	script_logs = os.path.join(work_dir,'script-logs')
-	bb_inc = os.path.join(script_logs,'bb_inc')
-	bb.mkdirhier(script_logs)
-	bb.mkdirhier(bb_inc)
-	
-	def find_file(dir,file):
-		for root, dirs, files in os.walk(dir):
-				if file in files:
-					return os.path.join(root,file) 
-	
-	def get_inc (file):
-		f = open(file,'r')
-		for line in f.readlines():
-			if 'require' not  in line:
-				bbinc.append(file)
-			else:
-				try:
-					incfile = pat.match(line).group(1)
-					incfile = bb.data.expand(os.path.basename(incfile),d)
-					abs_incfile = find_file(bbdir,incfile)
-					if abs_incfile:
-						bbinc.append(abs_incfile)
-						get_inc(abs_incfile)
-				except AttributeError:
-					pass
-	get_inc(bbfile)
-	bbinc = list(set(bbinc))
-	for bbincfile in bbinc:
-		shutil.copy(bbincfile,bb_inc)
-
-	try:
-		bb.mkdirhier(os.path.join(script_logs,'temp'))
-		oe.path.copytree(os.path.join(work_dir,'temp'), os.path.join(script_logs,'temp'))
-	except (IOError,AttributeError):
-		pass
-	return script_logs
-
-def get_series(d):
-	'''copy patches and series file to a pointed directory which will be archived to tarball in ${WORKDIR}'''
-	import shutil
-	
-	src_patches=[]
-	pf = d.getVar('PF', True)
-	work_dir = d.getVar('WORKDIR', True)
-	s = d.getVar('S',True)
-	dest = os.path.join(work_dir, pf + '-series') 
-	shutil.rmtree(dest, ignore_errors=True)
-	bb.mkdirhier(dest)
-	
-	src_uri = d.getVar('SRC_URI', True).split()
-	fetch = bb.fetch2.Fetch(src_uri, d)
-	locals = (fetch.localpath(url) for url in fetch.urls)
-	for local in locals:
-		src_patches.append(local)
-	if not cmp(work_dir,s):
-		tmp_list = src_patches
-	else:
-		tmp_list = src_patches[1:]
-	
-	for patch in tmp_list:
-		try:
-			shutil.copy(patch,dest)
-		except IOError:
-			if os.path.isdir(patch):
-				bb.mkdirhier(os.path.join(dest,patch))
-				oe.path.copytree(patch, os.path.join(dest,patch))
-	return dest
-
-def get_applying_patches(d):
-	"""only copy applying patches to a pointed directory which will be archived to tarball"""
-	import shutil
+    '''create a directory "script-logs" including .bb and .inc file in ${WORKDIR}'''
+    import re
+    import shutil
+
+    bbinc = []
+    pat=re.compile('require\s*([^\s]*\.*)(.*)')
+    work_dir = d.getVar('WORKDIR', True)
+    bbfile = d.getVar('FILE', True)
+    bbdir = os.path.dirname(bbfile)
+    script_logs = os.path.join(work_dir,'script-logs')
+    bb_inc = os.path.join(script_logs,'bb_inc')
+    bb.mkdirhier(script_logs)
+    bb.mkdirhier(bb_inc)
+
+    def find_file(dir,file):
+        for root, dirs, files in os.walk(dir):
+            if file in files:
+                return os.path.join(root,file)
+
+    def get_inc (file):
+        f = open(file,'r')
+        for line in f.readlines():
+            if 'require' not  in line:
+                bbinc.append(file)
+            else:
+                try:
+                    incfile = pat.match(line).group(1)
+                    incfile = bb.data.expand(os.path.basename(incfile),d)
+                    abs_incfile = find_file(bbdir,incfile)
+                    if abs_incfile:
+                        bbinc.append(abs_incfile)
+                        get_inc(abs_incfile)
+                except AttributeError:
+                    pass
+    get_inc(bbfile)
+    bbinc = list(set(bbinc))
+    for bbincfile in bbinc:
+        shutil.copy(bbincfile,bb_inc)
 
+    try:
+        bb.mkdirhier(os.path.join(script_logs,'temp'))
+        oe.path.copytree(os.path.join(work_dir,'temp'), os.path.join(script_logs,'temp'))
+    except (IOError,AttributeError):
+        pass
+    return script_logs
 
-	pf = d.getVar('PF', True)
-	work_dir = d.getVar('WORKDIR', True)
-	dest = os.path.join(work_dir, pf + '-patches') 
-	shutil.rmtree(dest, ignore_errors=True)
-	bb.mkdirhier(dest)
+def get_series(d):
+    '''copy patches and series file to a pointed directory which will be archived to tarball in ${WORKDIR}'''
+    import shutil
+
+    src_patches=[]
+    pf = d.getVar('PF', True)
+    work_dir = d.getVar('WORKDIR', True)
+    s = d.getVar('S',True)
+    dest = os.path.join(work_dir, pf + '-series')
+    shutil.rmtree(dest, ignore_errors=True)
+    bb.mkdirhier(dest)
+
+    src_uri = d.getVar('SRC_URI', True).split()
+    fetch = bb.fetch2.Fetch(src_uri, d)
+    locals = (fetch.localpath(url) for url in fetch.urls)
+    for local in locals:
+        src_patches.append(local)
+    if not cmp(work_dir,s):
+        tmp_list = src_patches
+    else:
+        tmp_list = src_patches[1:]
 
+    for patch in tmp_list:
+        try:
+            shutil.copy(patch,dest)
+        except IOError:
+            if os.path.isdir(patch):
+                bb.mkdirhier(os.path.join(dest,patch))
+                oe.path.copytree(patch, os.path.join(dest,patch))
+    return dest
 
-	patches = src_patches(d)
-	for patch in patches:
-		_, _, local, _, _, parm = bb.decodeurl(patch)
-		if local:
-			 shutil.copy(local,dest)
-	return dest
+def get_applying_patches(d):
+    """only copy applying patches to a pointed directory which will be archived to tarball"""
+    import shutil
+
+    pf = d.getVar('PF', True)
+    work_dir = d.getVar('WORKDIR', True)
+    dest = os.path.join(work_dir, pf + '-patches')
+    shutil.rmtree(dest, ignore_errors=True)
+    bb.mkdirhier(dest)
+
+    patches = src_patches(d)
+    for patch in patches:
+        _, _, local, _, _, parm = bb.decodeurl(patch)
+        if local:
+             shutil.copy(local,dest)
+    return dest
 
 def not_tarball(d):
-	'''packages including key words 'work-shared','native', 'task-' will be passed'''
+    '''packages including key words 'work-shared','native', 'task-' will be passed'''
 
-	workdir = d.getVar('WORKDIR',True)
-	s = d.getVar('S',True)
-	if 'work-shared' in s or 'task-' in workdir or 'native' in workdir:
-		return True
-	else:
-		return False
+    workdir = d.getVar('WORKDIR',True)
+    s = d.getVar('S',True)
+    if 'work-shared' in s or 'task-' in workdir or 'native' in workdir:
+        return True
+    else:
+        return False
 
 def get_source_from_downloads(d,stage_name):
-	'''copy tarball of $P to $WORKDIR when this tarball exists in $DL_DIR'''
-	if stage_name in 'patched' 'configured':
-		return
-	pf = d.getVar('PF', True)
-	dl_dir = d.getVar('DL_DIR',True)
-	try:
-		source = os.path.join(dl_dir,os.path.basename(d.getVar('SRC_URI', True).split()[0]))
-		if os.path.exists(source) and not os.path.isdir(source):
-			return source
-	except (IndexError, OSError):
-		pass
-	return ''
+    '''copy tarball of $P to $WORKDIR when this tarball exists in $DL_DIR'''
+    if stage_name in 'patched' 'configured':
+        return
+    pf = d.getVar('PF', True)
+    dl_dir = d.getVar('DL_DIR',True)
+    try:
+        source = os.path.join(dl_dir,os.path.basename(d.getVar('SRC_URI', True).split()[0]))
+        if os.path.exists(source) and not os.path.isdir(source):
+            return source
+    except (IndexError, OSError):
+        pass
+    return ''
 
 def do_tarball(workdir,srcdir,tarname):
-	'''tar "srcdir" under "workdir" to "tarname"'''
-	import tarfile
-	
-	sav_dir = os.getcwd()
-	os.chdir(workdir)
-	if (len(os.listdir(srcdir))) != 0:
-		tar = tarfile.open(tarname, "w:gz")
-		tar.add(srcdir)
-		tar.close()
-	else:
-		tarname = ''
-	os.chdir(sav_dir)
-	return tarname
-
-def archive_sources_from_directory(d,stage_name): 
-	'''archive sources codes tree to tarball when tarball of $P doesn't exist in $DL_DIR'''
-	import shutil
-	
-	s = d.getVar('S',True)
-	work_dir=d.getVar('WORKDIR', True)
-	PF = d.getVar('PF',True)
-	tarname = PF + '-' + stage_name + ".tar.gz"
-
-	if os.path.exists(s) and work_dir in s:
-		try:
-			source_dir = os.path.join(work_dir,[ i for i in s.replace(work_dir,'').split('/') if i][0])
-		except IndexError:
-			if not cmp(s,work_dir):
-				return ''
-	else:
-		return ''
-	source = os.path.basename(source_dir)
-	return do_tarball(work_dir,source,tarname)
+    '''tar "srcdir" under "workdir" to "tarname"'''
+    import tarfile
+
+    sav_dir = os.getcwd()
+    os.chdir(workdir)
+    if (len(os.listdir(srcdir))) != 0:
+        tar = tarfile.open(tarname, "w:gz")
+        tar.add(srcdir)
+        tar.close()
+    else:
+        tarname = ''
+    os.chdir(sav_dir)
+    return tarname
+
+def archive_sources_from_directory(d,stage_name):
+    '''archive sources codes tree to tarball when tarball of $P doesn't exist in $DL_DIR'''
+    import shutil
+
+    s = d.getVar('S',True)
+    work_dir=d.getVar('WORKDIR', True)
+    PF = d.getVar('PF',True)
+    tarname = PF + '-' + stage_name + ".tar.gz"
+
+    if os.path.exists(s) and work_dir in s:
+        try:
+            source_dir = os.path.join(work_dir,[ i for i in s.replace(work_dir,'').split('/') if i][0])
+        except IndexError:
+            if not cmp(s,work_dir):
+                return ''
+    else:
+        return ''
+    source = os.path.basename(source_dir)
+    return do_tarball(work_dir,source,tarname)
 
 def archive_sources(d,stage_name):
-	'''copy tarball from $DL_DIR to $WORKDIR if have tarball, archive source codes tree in $WORKDIR if $P is directory instead of tarball'''
-	import shutil
-	work_dir = d.getVar('WORKDIR',True)
-	file = get_source_from_downloads(d,stage_name)
-	if file:
-		shutil.copy(file,work_dir)
-		file = os.path.basename(file)
-	else:
-		file = archive_sources_from_directory(d,stage_name)
-	return file
-
+    '''copy tarball from $DL_DIR to $WORKDIR if have tarball, archive source codes tree in $WORKDIR if $P is directory instead of tarball'''
+    import shutil
+    work_dir = d.getVar('WORKDIR',True)
+    file = get_source_from_downloads(d,stage_name)
+    if file:
+        shutil.copy(file,work_dir)
+        file = os.path.basename(file)
+    else:
+        file = archive_sources_from_directory(d,stage_name)
+    return file
 
 def archive_patches(d,patchdir,series):
-	'''archive patches to tarball and also include series files if 'series' is True'''
-	import shutil
-
-	s = d.getVar('S',True)
-	work_dir = d.getVar('WORKDIR', True)
-	patch_dir = os.path.basename(patchdir)
-	tarname = patch_dir + ".tar.gz"
-	if series  == 'all' and os.path.exists(os.path.join(s,'patches/series')):
-			shutil.copy(os.path.join(s,'patches/series'),patchdir)
-	tarname = do_tarball(work_dir,patch_dir,tarname)
-	shutil.rmtree(patchdir, ignore_errors=True)
-	return tarname
+    '''archive patches to tarball and also include series files if 'series' is True'''
+    import shutil
+
+    s = d.getVar('S',True)
+    work_dir = d.getVar('WORKDIR', True)
+    patch_dir = os.path.basename(patchdir)
+    tarname = patch_dir + ".tar.gz"
+    if series  == 'all' and os.path.exists(os.path.join(s,'patches/series')):
+        shutil.copy(os.path.join(s,'patches/series'),patchdir)
+    tarname = do_tarball(work_dir,patch_dir,tarname)
+    shutil.rmtree(patchdir, ignore_errors=True)
+    return tarname
 
 def select_archive_patches(d,option):
-	'''select to archive all patches including non-applying and series or applying patches '''
-	if option == "all":
-		patchdir = get_series(d)
-	elif option == "applying":
-		patchdir = get_applying_patches(d)
-	try:
-		os.rmdir(patchdir)
-	except OSError:
-			tarpatch = archive_patches(d,patchdir,option)
-			return tarpatch
-	return
+    '''select to archive all patches including non-applying and series or applying patches '''
+    if option == "all":
+        patchdir = get_series(d)
+    elif option == "applying":
+        patchdir = get_applying_patches(d)
+    try:
+        os.rmdir(patchdir)
+    except OSError:
+            tarpatch = archive_patches(d,patchdir,option)
+            return tarpatch
+    return
 
 def archive_logs(d,logdir,bbinc=False):
-	'''archive logs in temp to tarball and .bb and .inc files if bbinc is True '''
-	import shutil
-
-	pf = d.getVar('PF',True)
-	work_dir = d.getVar('WORKDIR',True)
-	log_dir =  os.path.basename(logdir)
-	tarname = pf + '-' + log_dir + ".tar.gz"
-	tarname = do_tarball(work_dir,log_dir,tarname)
-	if bbinc:
-		shutil.rmtree(logdir, ignore_errors=True)
-	return tarname 
+    '''archive logs in temp to tarball and .bb and .inc files if bbinc is True '''
+    import shutil
+
+    pf = d.getVar('PF',True)
+    work_dir = d.getVar('WORKDIR',True)
+    log_dir =  os.path.basename(logdir)
+    tarname = pf + '-' + log_dir + ".tar.gz"
+    tarname = do_tarball(work_dir,log_dir,tarname)
+    if bbinc:
+        shutil.rmtree(logdir, ignore_errors=True)
+    return tarname
 
 def get_licenses(d):
-	'''get licenses for running .bb file'''
-	import oe.license
-
-	licenses_type = d.getVar('LICENSE', True) or ""
-	lics = oe.license.is_included(licenses_type)[1:][0]
-	lice = ''
-	for lic in lics:
-		licens = d.getVarFlag('SPDXLICENSEMAP', lic)
-		if licens != None:
-			lice += licens
-		else:
-			lice += lic
-	return lice
-	
+    '''get licenses for running .bb file'''
+    import oe.license
+
+    licenses_type = d.getVar('LICENSE', True) or ""
+    lics = oe.license.is_included(licenses_type)[1:][0]
+    lice = ''
+    for lic in lics:
+        licens = d.getVarFlag('SPDXLICENSEMAP', lic)
+        if licens != None:
+            lice += licens
+        else:
+            lice += lic
+    return lice
+
 
 def move_tarball_deploy(d,tarball_list):
-	'''move tarball in location to ${DEPLOY_DIR}/sources'''
-	import shutil
-	
-	if tarball_list is []:
-		return
-	target_sys = d.getVar('TARGET_SYS', True)
-	pf = d.getVar('PF', True)
-	licenses = get_licenses(d)
-	work_dir = d.getVar('WORKDIR',True)
-	tar_sources = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf
-	if not os.path.exists(tar_sources):
-		bb.mkdirhier(tar_sources)
-	for source in tarball_list:
-		if source:
-			if os.path.exists(os.path.join(tar_sources, source)):
-				os.remove(os.path.join(tar_sources,source))
-			shutil.move(os.path.join(work_dir,source),tar_sources)
+    '''move tarball in location to ${DEPLOY_DIR}/sources'''
+    import shutil
+
+    if tarball_list is []:
+        return
+    target_sys = d.getVar('TARGET_SYS', True)
+    pf = d.getVar('PF', True)
+    licenses = get_licenses(d)
+    work_dir = d.getVar('WORKDIR',True)
+    tar_sources = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf
+    if not os.path.exists(tar_sources):
+        bb.mkdirhier(tar_sources)
+    for source in tarball_list:
+        if source:
+            if os.path.exists(os.path.join(tar_sources, source)):
+                os.remove(os.path.join(tar_sources,source))
+            shutil.move(os.path.join(work_dir,source),tar_sources)
 
 def check_archiving_type(d):
-	'''check the type for archiving package('tar' or 'srpm')'''
-	try:
-		if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in d.getVar('ARCHIVE_TYPE', True).split():
-			raise AttributeError	
-	except AttributeError:
-			bb.fatal("\"SOURCE_ARCHIVE_PACKAGE_TYPE\" is \'tar\' or \'srpm\', no other types") 
+    '''check the type for archiving package('tar' or 'srpm')'''
+    try:
+        if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in d.getVar('ARCHIVE_TYPE', True).split():
+            raise AttributeError
+    except AttributeError:
+        bb.fatal("\"SOURCE_ARCHIVE_PACKAGE_TYPE\" is \'tar\' or \'srpm\', no other types")
 
 def store_package(d,package_name):
-	'''store tarbablls name to file "tar-package"'''
-	try:
-		f = open(os.path.join(d.getVar('WORKDIR',True),'tar-package'),'a')
-		f.write(package_name + ' ')
-		f.close()
-	except IOError:
-		pass
+    '''store tarbablls name to file "tar-package"'''
+    try:
+        f = open(os.path.join(d.getVar('WORKDIR',True),'tar-package'),'a')
+        f.write(package_name + ' ')
+        f.close()
+    except IOError:
+        pass
 
 def get_package(d):
-	'''get tarballs name from "tar-package"'''
-	work_dir = (d.getVar('WORKDIR', True))
-	tarpackage = os.path.join(work_dir,'tar-package')
-	try:
-		f = open(tarpackage,'r')
-		line = list(set(f.readline().replace('\n','').split()))
-	except UnboundLocalError,IOError:
-		pass
-	f.close()
-	return line
+    '''get tarballs name from "tar-package"'''
+    work_dir = (d.getVar('WORKDIR', True))
+    tarpackage = os.path.join(work_dir,'tar-package')
+    try:
+        f = open(tarpackage,'r')
+        line = list(set(f.readline().replace('\n','').split()))
+    except UnboundLocalError,IOError:
+        pass
+    f.close()
+    return line
 
 
 def archive_sources_patches(d,stage_name):
-	'''archive sources and patches to tarball. stage_name will append strings ${stage_name} to ${PR} as middle name. for example, zlib-1.4.6-prepatch(stage_name).tar.gz '''
-	import shutil
-
-	check_archiving_type(d)	
-	if not_tarball(d) or tar_filter(d):
-		return
-	
-	source_tar_name = archive_sources(d,stage_name)
-	if stage_name == "prepatch":
-		if d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'TRUE':
-			patch_tar_name = select_archive_patches(d,"all")
-		elif d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'FALSE':
-			patch_tar_name = select_archive_patches(d,"applying")
-		else:
-			bb.fatal("Please define 'PATCHES_ARCHIVE_WITH_SERIES' is strings 'True' or 'False' ")
-	else:
-		patch_tar_name = ''
-	
-	if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM':
-		move_tarball_deploy(d,[source_tar_name,patch_tar_name])
-	else:
-		tarpackage = os.path.join(d.getVar('WORKDIR', True),'tar-package')
-		if os.path.exists(tarpackage):
-			os.remove(tarpackage)
-		for package in os.path.basename(source_tar_name), patch_tar_name:
-			if package:
-				store_package(d,str(package) + ' ')
+    '''archive sources and patches to tarball. stage_name will append strings ${stage_name} to ${PR} as middle name. for example, zlib-1.4.6-prepatch(stage_name).tar.gz '''
+    import shutil
+
+    check_archiving_type(d)
+    if not_tarball(d) or tar_filter(d):
+        return
+
+    source_tar_name = archive_sources(d,stage_name)
+    if stage_name == "prepatch":
+        if d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'TRUE':
+            patch_tar_name = select_archive_patches(d,"all")
+        elif d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'FALSE':
+            patch_tar_name = select_archive_patches(d,"applying")
+        else:
+            bb.fatal("Please define 'PATCHES_ARCHIVE_WITH_SERIES' is strings 'True' or 'False' ")
+    else:
+        patch_tar_name = ''
+
+    if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM':
+        move_tarball_deploy(d,[source_tar_name,patch_tar_name])
+    else:
+        tarpackage = os.path.join(d.getVar('WORKDIR', True),'tar-package')
+        if os.path.exists(tarpackage):
+            os.remove(tarpackage)
+        for package in os.path.basename(source_tar_name), patch_tar_name:
+            if package:
+                store_package(d,str(package) + ' ')
 
 def archive_scripts_logs(d):
-	'''archive scripts and logs. scripts include .bb and .inc files and logs include stuff in "temp".'''
-
-        if tar_filter(d):
-                return
-	work_dir = d.getVar('WORKDIR', True)
-	temp_dir = os.path.join(work_dir,'temp')
-	source_archive_log_with_scripts = d.getVar('SOURCE_ARCHIVE_LOG_WITH_SCRIPTS', True)
-	if source_archive_log_with_scripts == 'logs_with_scripts':
-		logdir = get_bb_inc(d)
-		tarlog = archive_logs(d,logdir,True)
-	elif source_archive_log_with_scripts == 'logs':
-		if os.path.exists(temp_dir):
-			tarlog = archive_logs(d,temp_dir,False)
-	else:
-		return
-		
-	if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM':
-		move_tarball_deploy(d,[tarlog])
-
-	else:
-		store_package(d,tarlog)
+    '''archive scripts and logs. scripts include .bb and .inc files and logs include stuff in "temp".'''
+
+    if tar_filter(d):
+        return
+    work_dir = d.getVar('WORKDIR', True)
+    temp_dir = os.path.join(work_dir,'temp')
+    source_archive_log_with_scripts = d.getVar('SOURCE_ARCHIVE_LOG_WITH_SCRIPTS', True)
+    if source_archive_log_with_scripts == 'logs_with_scripts':
+        logdir = get_bb_inc(d)
+        tarlog = archive_logs(d,logdir,True)
+    elif source_archive_log_with_scripts == 'logs':
+        if os.path.exists(temp_dir):
+            tarlog = archive_logs(d,temp_dir,False)
+    else:
+        return
+
+    if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM':
+        move_tarball_deploy(d,[tarlog])
+
+    else:
+        store_package(d,tarlog)
 
 def dumpdata(d):
-	'''dump environment to "${P}-${PR}.showdata.dump" including all kinds of variables and functions when running a task'''
-
-        if tar_filter(d):
-                return
-	workdir = bb.data.getVar('WORKDIR', d, 1)
-	distro = bb.data.getVar('DISTRO', d, 1)
-	s = d.getVar('S', True)
-	pf = d.getVar('PF', True)
-	target_sys = d.getVar('TARGET_SYS', True)
-	licenses = get_licenses(d)
-	dumpdir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf
-	if not os.path.exists(dumpdir):
-		bb.mkdirhier(dumpdir)
-
-	dumpfile = os.path.join(dumpdir, bb.data.expand("${P}-${PR}.showdata.dump",d))
-
-	bb.note("Dumping metadata into '%s'" % dumpfile)
-	f = open(dumpfile, "w")
-	# emit variables and shell functions
-        bb.data.emit_env(f, d, True)
-	# emit the metadata which isnt valid shell
-	for e in d.keys():
-		if bb.data.getVarFlag(e, 'python', d):
-			f.write("\npython %s () {\n%s}\n" % (e, bb.data.getVar(e, d, 1)))
-	f.close()
+    '''dump environment to "${P}-${PR}.showdata.dump" including all kinds of variables and functions when running a task'''
+
+    if tar_filter(d):
+        return
+    workdir = bb.data.getVar('WORKDIR', d, 1)
+    distro = bb.data.getVar('DISTRO', d, 1)
+    s = d.getVar('S', True)
+    pf = d.getVar('PF', True)
+    target_sys = d.getVar('TARGET_SYS', True)
+    licenses = get_licenses(d)
+    dumpdir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf
+    if not os.path.exists(dumpdir):
+        bb.mkdirhier(dumpdir)
+
+    dumpfile = os.path.join(dumpdir, bb.data.expand("${P}-${PR}.showdata.dump",d))
+
+    bb.note("Dumping metadata into '%s'" % dumpfile)
+    f = open(dumpfile, "w")
+    # emit variables and shell functions
+    bb.data.emit_env(f, d, True)
+    # emit the metadata which isnt valid shell
+    for e in d.keys():
+        if bb.data.getVarFlag(e, 'python', d):
+            f.write("\npython %s () {\n%s}\n" % (e, bb.data.getVar(e, d, 1)))
+    f.close()
 
 def create_diff_gz(d):
-	'''creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.g gz for mapping all content in 's' including patches to  xxx.diff.gz'''
-	import shutil
-	import subprocess
-
-        if tar_filter(d):
-                return
-	work_dir = d.getVar('WORKDIR', True)
-	exclude_from = d.getVar('ARCHIVE_EXCLUDE_FROM', True).split()
-	pf = d.getVar('PF', True)
-	licenses = get_licenses(d)
-	target_sys = d.getVar('TARGET_SYS', True)
-	diff_dir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf
-	diff_file = os.path.join(diff_dir, bb.data.expand("${P}-${PR}.diff.gz",d))
-
-	f = open(os.path.join(work_dir,'temp/exclude-from-file'), 'a')
-	for i in exclude_from:
-		f.write(i)
-		f.write("\n")
-	f.close()
-
-	s=d.getVar('S', True)
-	distro = d.getVar('DISTRO',True)
-	dest = s + '/' + distro + '/files'
-	if not os.path.exists(dest):
-		bb.mkdirhier(dest)
-	for i in os.listdir(os.getcwd()):
-		if os.path.isfile(i):
-			try:
-				shutil.copy(i, dest)
-			except IOError:
-				subprocess.call('fakeroot cp -rf ' + i + " " + dest, shell=True)
-	
-	bb.note("Creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.gz")
-	cmd = "LC_ALL=C TZ=UTC0 diff --exclude-from=" + work_dir + "/temp/exclude-from-file -Naur " + s + '.org' + ' ' +  s + " | gzip -c > " + diff_file
-	d.setVar('DIFF', cmd + "\n")
-	d.setVarFlag('DIFF', 'func', '1') 
-	bb.build.exec_func('DIFF', d)
-	shutil.rmtree(s + '.org', ignore_errors=True)
+    '''creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.g gz for mapping all content in 's' including patches to  xxx.diff.gz'''
+    import shutil
+    import subprocess
+
+    if tar_filter(d):
+        return
+    work_dir = d.getVar('WORKDIR', True)
+    exclude_from = d.getVar('ARCHIVE_EXCLUDE_FROM', True).split()
+    pf = d.getVar('PF', True)
+    licenses = get_licenses(d)
+    target_sys = d.getVar('TARGET_SYS', True)
+    diff_dir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf
+    diff_file = os.path.join(diff_dir, bb.data.expand("${P}-${PR}.diff.gz",d))
+
+    f = open(os.path.join(work_dir,'temp/exclude-from-file'), 'a')
+    for i in exclude_from:
+        f.write(i)
+        f.write("\n")
+    f.close()
+
+    s=d.getVar('S', True)
+    distro = d.getVar('DISTRO',True)
+    dest = s + '/' + distro + '/files'
+    if not os.path.exists(dest):
+        bb.mkdirhier(dest)
+    for i in os.listdir(os.getcwd()):
+        if os.path.isfile(i):
+            try:
+                shutil.copy(i, dest)
+            except IOError:
+                subprocess.call('fakeroot cp -rf ' + i + " " + dest, shell=True)
+
+    bb.note("Creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.gz")
+    cmd = "LC_ALL=C TZ=UTC0 diff --exclude-from=" + work_dir + "/temp/exclude-from-file -Naur " + s + '.org' + ' ' +  s + " | gzip -c > " + diff_file
+    d.setVar('DIFF', cmd + "\n")
+    d.setVarFlag('DIFF', 'func', '1')
+    bb.build.exec_func('DIFF', d)
+    shutil.rmtree(s + '.org', ignore_errors=True)
 
 # This function will run when user want to get tarball for sources and patches after do_unpack
 python do_archive_original_sources_patches(){
-	archive_sources_patches(d,'prepatch')
+    archive_sources_patches(d,'prepatch')
 }
 
 # This function will run when user want to get tarball for patched sources after do_patch
 python do_archive_patched_sources(){
-	archive_sources_patches(d,'patched')
+    archive_sources_patches(d,'patched')
 }
 
 # This function will run when user want to get tarball for configured sources after do_configure
 python do_archive_configured_sources(){
-	archive_sources_patches(d,'configured')
+    archive_sources_patches(d,'configured')
 }
 
 # This function will run when user want to get tarball for logs or both logs and scripts(.bb and .inc files)
 python do_archive_scripts_logs(){
-	archive_scripts_logs(d)
+    archive_scripts_logs(d)
 }
 
-# This function will run when user want to know what variable and functions in a running task are and also can get a diff file including 
+# This function will run when user want to know what variable and functions in a running task are and also can get a diff file including
 # all content a package should include.
 python do_dumpdata_create_diff_gz(){
-	dumpdata(d)
-	create_diff_gz(d)
+    dumpdata(d)
+    create_diff_gz(d)
 }
 
 # This functions prepare for archiving "linux-yocto" because this package create directory 's' before do_patch instead of after do_unpack.
 # This is special control for archiving linux-yocto only.
 python do_archive_linux_yocto(){
-	s = d.getVar('S', True)
-	if 'linux-yocto' in s:
-		source_tar_name = archive_sources(d,'')
-	if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM':
-		move_tarball_deploy(d,[source_tar_name,''])
+    s = d.getVar('S', True)
+    if 'linux-yocto' in s:
+        source_tar_name = archive_sources(d,'')
+    if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM':
+        move_tarball_deploy(d,[source_tar_name,''])
 }
 do_kernel_checkout[postfuncs] += "do_archive_linux_yocto "
 
 # remove tarball for sources, patches and logs after creating srpm.
 python do_remove_tarball(){
-	if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() == 'SRPM':
-		work_dir = d.getVar('WORKDIR', True)
-		try:
-			for file in os.listdir(os.getcwd()):
-				if file in get_package(d):
-					os.remove(file)
-			os.remove(os.path.join(work_dir,'tar-package'))
-		except (TypeError,OSError):
-				pass
+    if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() == 'SRPM':
+        work_dir = d.getVar('WORKDIR', True)
+        try:
+            for file in os.listdir(os.getcwd()):
+                if file in get_package(d):
+                    os.remove(file)
+            os.remove(os.path.join(work_dir,'tar-package'))
+        except (TypeError,OSError):
+            pass
 }
 do_remove_taball[deptask] = "do_archive_scripts_logs"
 do_package_write_rpm[postfuncs] += "do_remove_tarball "
diff --git a/meta/classes/package_rpm.bbclass b/meta/classes/package_rpm.bbclass
index 9abad5e..58a9aac 100644
--- a/meta/classes/package_rpm.bbclass
+++ b/meta/classes/package_rpm.bbclass
@@ -1152,7 +1152,6 @@ python do_package_rpm () {
         d.setVarFlag('SBUILDSPEC', 'func', '1')
         bb.build.exec_func('SBUILDSPEC', d)
 
-
     # Build the rpm package!
     d.setVar('BUILDSPEC', cmd + "\n")
     d.setVarFlag('BUILDSPEC', 'func', '1')
-- 
1.7.11.2





More information about the Openembedded-core mailing list