[OE-core] [PATCH] distrodata: replace Tab with Space

Saul Wold sgw at linux.intel.com
Mon Sep 10 06:53:46 UTC 2012


Signed-off-by: Saul Wold <sgw at linux.intel.com>
---
 meta/classes/distrodata.bbclass | 1090 +++++++++++++++++++-------------------
 1 files changed, 545 insertions(+), 545 deletions(-)

diff --git a/meta/classes/distrodata.bbclass b/meta/classes/distrodata.bbclass
index 0da10a1..945ff53 100644
--- a/meta/classes/distrodata.bbclass
+++ b/meta/classes/distrodata.bbclass
@@ -3,11 +3,11 @@ addhandler distro_eventhandler
 python distro_eventhandler() {
 
     if bb.event.getName(e) == "BuildStarted":
-	import oe.distro_check as dc
-	logfile = dc.create_log_file(e.data, "distrodata.csv")
-	lf = bb.utils.lockfile("%s.lock" % logfile)
-	f = open(logfile, "a")
-	f.write("Package,Description,Owner,License,VerMatch,Version,Upsteam,Reason,Recipe Status,Distro 1,Distro 2,Distro 3\n")
+        import oe.distro_check as dc
+        logfile = dc.create_log_file(e.data, "distrodata.csv")
+        lf = bb.utils.lockfile("%s.lock" % logfile)
+        f = open(logfile, "a")
+        f.write("Package,Description,Owner,License,VerMatch,Version,Upsteam,Reason,Recipe Status,Distro 1,Distro 2,Distro 3\n")
         f.close()
         bb.utils.unlockfile(lf)
 
@@ -17,7 +17,7 @@ python distro_eventhandler() {
 addtask distrodata_np
 do_distrodata_np[nostamp] = "1"
 python do_distrodata_np() {
-	localdata = bb.data.createCopy(d)
+        localdata = bb.data.createCopy(d)
         pn = d.getVar("PN", True)
         bb.note("Package Name: %s" % pn)
 
@@ -27,69 +27,69 @@ python do_distrodata_np() {
         datetime = localdata.getVar('DATETIME', True)
         dist_check.update_distro_data(distro_check_dir, datetime)
 
-	if pn.find("-native") != -1:
-	    pnstripped = pn.split("-native")
-	    bb.note("Native Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	if pn.find("nativesdk-") != -1:
-	    pnstripped = pn.replace("nativesdk-", "")
-	    bb.note("Native Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	if pn.find("-cross") != -1:
-	    pnstripped = pn.split("-cross")
-	    bb.note("cross Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	if pn.find("-crosssdk") != -1:
-	    pnstripped = pn.split("-crosssdk")
-	    bb.note("cross Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	if pn.find("-initial") != -1:
-	    pnstripped = pn.split("-initial")
-	    bb.note("initial Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	"""generate package information from .bb file"""
-	pname = localdata.getVar('PN', True)
-	pcurver = localdata.getVar('PV', True)
-	pdesc = localdata.getVar('DESCRIPTION', True)
+        if pn.find("-native") != -1:
+            pnstripped = pn.split("-native")
+            bb.note("Native Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        if pn.find("nativesdk-") != -1:
+            pnstripped = pn.replace("nativesdk-", "")
+            bb.note("Native Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        if pn.find("-cross") != -1:
+            pnstripped = pn.split("-cross")
+            bb.note("cross Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        if pn.find("-crosssdk") != -1:
+            pnstripped = pn.split("-crosssdk")
+            bb.note("cross Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        if pn.find("-initial") != -1:
+            pnstripped = pn.split("-initial")
+            bb.note("initial Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        """generate package information from .bb file"""
+        pname = localdata.getVar('PN', True)
+        pcurver = localdata.getVar('PV', True)
+        pdesc = localdata.getVar('DESCRIPTION', True)
         if pdesc is not None:
                 pdesc = pdesc.replace(',','')
                 pdesc = pdesc.replace('\n','')
 
-	pgrp = localdata.getVar('SECTION', True)
-	plicense = localdata.getVar('LICENSE', True).replace(',','_')
+        pgrp = localdata.getVar('SECTION', True)
+        plicense = localdata.getVar('LICENSE', True).replace(',','_')
 
-	rstatus = localdata.getVar('RECIPE_COLOR', True)
+        rstatus = localdata.getVar('RECIPE_COLOR', True)
         if rstatus is not None:
                 rstatus = rstatus.replace(',','')
-		
-	pupver = localdata.getVar('RECIPE_UPSTREAM_VERSION', True)
-	if pcurver == pupver:
-		vermatch="1"
-	else:
-		vermatch="0"
-	noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True)
-	if noupdate_reason is None:
-		noupdate="0"
-	else:
-		noupdate="1"
+                
+        pupver = localdata.getVar('RECIPE_UPSTREAM_VERSION', True)
+        if pcurver == pupver:
+                vermatch="1"
+        else:
+                vermatch="0"
+        noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True)
+        if noupdate_reason is None:
+                noupdate="0"
+        else:
+                noupdate="1"
                 noupdate_reason = noupdate_reason.replace(',','')
 
-	maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
-	rlrd = localdata.getVar('RECIPE_UPSTREAM_DATE', True)
+        maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
+        rlrd = localdata.getVar('RECIPE_UPSTREAM_DATE', True)
         result = dist_check.compare_in_distro_packages_list(distro_check_dir, localdata)
 
-	bb.note("DISTRO: %s,%s,%s,%s,%s,%s,%s,%s,%s\n" % \
-		  (pname, pdesc, maintainer, plicense, vermatch, pcurver, pupver, noupdate_reason, rstatus))
+        bb.note("DISTRO: %s,%s,%s,%s,%s,%s,%s,%s,%s\n" % \
+                  (pname, pdesc, maintainer, plicense, vermatch, pcurver, pupver, noupdate_reason, rstatus))
         line = pn
         for i in result:
             line = line + "," + i
@@ -99,12 +99,12 @@ python do_distrodata_np() {
 addtask distrodata
 do_distrodata[nostamp] = "1"
 python do_distrodata() {
-	logpath = d.getVar('LOG_DIR', True)
-	bb.utils.mkdirhier(logpath)
-	logfile = os.path.join(logpath, "distrodata.csv")
+        logpath = d.getVar('LOG_DIR', True)
+        bb.utils.mkdirhier(logpath)
+        logfile = os.path.join(logpath, "distrodata.csv")
 
         import oe.distro_check as dist_check
-	localdata = bb.data.createCopy(d)
+        localdata = bb.data.createCopy(d)
         tmpdir = d.getVar('TMPDIR', True)
         distro_check_dir = os.path.join(tmpdir, "distro_check")
         datetime = localdata.getVar('DATETIME', True)
@@ -113,61 +113,61 @@ python do_distrodata() {
         pn = d.getVar("PN", True)
         bb.note("Package Name: %s" % pn)
 
-	if pn.find("-native") != -1:
-	    pnstripped = pn.split("-native")
-	    bb.note("Native Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	if pn.find("-cross") != -1:
-	    pnstripped = pn.split("-cross")
-	    bb.note("cross Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	if pn.find("-initial") != -1:
-	    pnstripped = pn.split("-initial")
-	    bb.note("initial Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	"""generate package information from .bb file"""
-	pname = localdata.getVar('PN', True)
-	pcurver = localdata.getVar('PV', True)
-	pdesc = localdata.getVar('DESCRIPTION', True)
+        if pn.find("-native") != -1:
+            pnstripped = pn.split("-native")
+            bb.note("Native Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        if pn.find("-cross") != -1:
+            pnstripped = pn.split("-cross")
+            bb.note("cross Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        if pn.find("-initial") != -1:
+            pnstripped = pn.split("-initial")
+            bb.note("initial Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        """generate package information from .bb file"""
+        pname = localdata.getVar('PN', True)
+        pcurver = localdata.getVar('PV', True)
+        pdesc = localdata.getVar('DESCRIPTION', True)
         if pdesc is not None:
                 pdesc = pdesc.replace(',','')
                 pdesc = pdesc.replace('\n','')
 
-	pgrp = localdata.getVar('SECTION', True)
-	plicense = localdata.getVar('LICENSE', True).replace(',','_')
+        pgrp = localdata.getVar('SECTION', True)
+        plicense = localdata.getVar('LICENSE', True).replace(',','_')
 
-	rstatus = localdata.getVar('RECIPE_COLOR', True)
+        rstatus = localdata.getVar('RECIPE_COLOR', True)
         if rstatus is not None:
                 rstatus = rstatus.replace(',','')
-		
-	pupver = localdata.getVar('RECIPE_UPSTREAM_VERSION', True)
-	if pcurver == pupver:
-		vermatch="1"
-	else:
-		vermatch="0"
-
-	noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True)
-	if noupdate_reason is None:
-		noupdate="0"
-	else:
-		noupdate="1"
+                
+        pupver = localdata.getVar('RECIPE_UPSTREAM_VERSION', True)
+        if pcurver == pupver:
+                vermatch="1"
+        else:
+                vermatch="0"
+
+        noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True)
+        if noupdate_reason is None:
+                noupdate="0"
+        else:
+                noupdate="1"
                 noupdate_reason = noupdate_reason.replace(',','')
 
-	maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
-	rlrd = localdata.getVar('RECIPE_UPSTREAM_DATE', True)
+        maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
+        rlrd = localdata.getVar('RECIPE_UPSTREAM_DATE', True)
         # do the comparison
         result = dist_check.compare_in_distro_packages_list(distro_check_dir, localdata)
 
-	lf = bb.utils.lockfile("%s.lock" % logfile)
-	f = open(logfile, "a")
-	f.write("%s,%s,%s,%s,%s,%s,%s,%s,%s" % \
-		  (pname, pdesc, maintainer, plicense, vermatch, pcurver, pupver, noupdate_reason, rstatus))
+        lf = bb.utils.lockfile("%s.lock" % logfile)
+        f = open(logfile, "a")
+        f.write("%s,%s,%s,%s,%s,%s,%s,%s,%s" % \
+                  (pname, pdesc, maintainer, plicense, vermatch, pcurver, pupver, noupdate_reason, rstatus))
         line = ""
         for i in result:
             line = line + "," + i
@@ -180,18 +180,18 @@ addtask distrodataall after do_distrodata
 do_distrodataall[recrdeptask] = "do_distrodataall do_distrodata"
 do_distrodataall[nostamp] = "1"
 do_distrodataall() {
-	:
+        :
 }
 
 addhandler checkpkg_eventhandler
 python checkpkg_eventhandler() {
     if bb.event.getName(e) == "BuildStarted":
-	import oe.distro_check as dc
-	logfile = dc.create_log_file(e.data, "checkpkg.csv")
+        import oe.distro_check as dc
+        logfile = dc.create_log_file(e.data, "checkpkg.csv")
 
-	lf = bb.utils.lockfile("%s.lock" % logfile)
-	f = open(logfile, "a")
-	f.write("Package\tVersion\tUpver\tLicense\tSection\tHome\tRelease\tDepends\tBugTracker\tPE\tDescription\tStatus\tTracking\tURI\tMAINTAINER\n")
+        lf = bb.utils.lockfile("%s.lock" % logfile)
+        f = open(logfile, "a")
+        f.write("Package\tVersion\tUpver\tLicense\tSection\tHome\tRelease\tDepends\tBugTracker\tPE\tDescription\tStatus\tTracking\tURI\tMAINTAINER\n")
         f.close()
         bb.utils.unlockfile(lf)
     return
@@ -200,444 +200,444 @@ python checkpkg_eventhandler() {
 addtask checkpkg
 do_checkpkg[nostamp] = "1"
 python do_checkpkg() {
-	localdata = bb.data.createCopy(d)
-	import sys
-	import re
-	import tempfile
-	import subprocess
-
-	"""
-	sanity check to ensure same name and type. Match as many patterns as possible
-	such as:
-		gnome-common-2.20.0.tar.gz (most common format)
-		gtk+-2.90.1.tar.gz
-		xf86-input-synaptics-12.6.9.tar.gz
-		dri2proto-2.3.tar.gz
-		blktool_4.orig.tar.gz
-		libid3tag-0.15.1b.tar.gz
-		unzip552.tar.gz
-		icu4c-3_6-src.tgz
-		genext2fs_1.3.orig.tar.gz
-		gst-fluendo-mp3
-	"""
-	prefix1 = "[a-zA-Z][a-zA-Z0-9]*([\-_][a-zA-Z]\w+)*[\-_]"	# match most patterns which uses "-" as separator to version digits
-	prefix2 = "[a-zA-Z]+"			# a loose pattern such as for unzip552.tar.gz
-	prefix3 = "[0-9a-zA-Z]+"			# a loose pattern such as for 80325-quicky-0.4.tar.gz
-	prefix = "(%s|%s|%s)" % (prefix1, prefix2, prefix3)
-	suffix = "(tar\.gz|tgz|tar\.bz2|zip|xz|rpm)"
-	suffixtuple = ("tar.gz", "tgz", "zip", "tar.bz2", "tar.xz", "src.rpm")
-
-	sinterstr = "(?P<name>%s?)(?P<ver>.*)" % prefix
-	sdirstr = "(?P<name>%s)(?P<ver>.*)\.(?P<type>%s$)" % (prefix, suffix)
-
-	def parse_inter(s):
-		m = re.search(sinterstr, s)
-		if not m:
-			return None
-		else:
-			return (m.group('name'), m.group('ver'), "")
-
-	def parse_dir(s):
-		m = re.search(sdirstr, s)
-		if not m:
-			return None
-		else:
-			return (m.group('name'), m.group('ver'), m.group('type'))
-
-	"""
-	Check whether 'new' is newer than 'old' version. We use existing vercmp() for the
-	purpose. PE is cleared in comparison as it's not for build, and PV is cleared too
-	for simplicity as it's somehow difficult to get from various upstream format
-	"""
-	def __vercmp(old, new):
-		(on, ov, ot) = old
-		(en, ev, et) = new
-		if on != en or (et and et not in suffixtuple):
-			return 0
-		ov = re.search("[\d|\.]+[^a-zA-Z]+", ov).group()
-		ev = re.search("[\d|\.]+[^a-zA-Z]+", ev).group()
-		return bb.utils.vercmp(("0", ov, ""), ("0", ev, ""))
-
-	"""
-	wrapper for fetch upstream directory info
-		'url'	- upstream link customized by regular expression
-		'd'	- database
-		'tmpf'	- tmpfile for fetcher output
-	We don't want to exit whole build due to one recipe error. So handle all exceptions 
-	gracefully w/o leaking to outer. 
-	"""
-	def internal_fetch_wget(url, d, tmpf):
-		status = "ErrFetchUnknown"
-		"""
-		Clear internal url cache as it's a temporary check. Not doing so will have 
-		bitbake check url multiple times when looping through a single url
-		"""
-		fn = d.getVar('FILE', True)
-		bb.fetch2.urldata_cache[fn] = {}
-
-		"""
-		To avoid impacting bitbake build engine, this trick is required for reusing bitbake
-		interfaces. bb.fetch.go() is not appliable as it checks downloaded content in ${DL_DIR}
-		while we don't want to pollute that place. So bb.fetch2.checkstatus() is borrowed here
-		which is designed for check purpose but we override check command for our own purpose
-		"""
-		ld = bb.data.createCopy(d)
-		d.setVar('CHECKCOMMAND_wget', "/usr/bin/env wget -t 1 --passive-ftp -O %s --user-agent=\"Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.2.12) Gecko/20101027 Ubuntu/9.10 (karmic) Firefox/3.6.12\" '${URI}'" \
-					% tmpf.name)
-		bb.data.update_data(ld)
-
-		try:
-			fetcher = bb.fetch2.Fetch([url], ld)
-			fetcher.checkstatus()
-			status = "SUCC"
-		except bb.fetch2.BBFetchException, e:
-			status = "ErrFetch"
-
-		return status
-
-	"""
-	Check on middle version directory such as "2.4/" in "http://xxx/2.4/pkg-2.4.1.tar.gz", 
-		'url'	- upstream link customized by regular expression
-		'd'	- database
-		'curver' - current version
-	Return new version if success, or else error in "Errxxxx" style
-	"""
-	def check_new_dir(url, curver, d):
-		pn = d.getVar('PN', True)
-		f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-1-" % pn)
-		status = internal_fetch_wget(url, d, f)
-		fhtml = f.read()
-		if status == "SUCC" and len(fhtml):
-			newver = parse_inter(curver)
-
-			"""
-			match "*4.1/">*4.1/ where '*' matches chars
-			N.B. add package name, only match for digits
-			"""
-			m = re.search("^%s" % prefix, curver)
-			if m:
-				s = "%s[^\d\"]*?(\d+[\.\-_])+\d+/?" % m.group()
-			else:
-				s = "(\d+[\.\-_])+\d+/?"
-				
-			searchstr = "[hH][rR][eE][fF]=\"%s\">" % s
-			reg = re.compile(searchstr)
-
-			valid = 0
-			for line in fhtml.split("\n"):
-				if line.find(curver) >= 0:
-					valid = 1
-				m = reg.search(line)
-				if m:
-					ver = m.group().split("\"")[1]
-					ver = ver.strip("/")
-					ver = parse_inter(ver)
-					if ver and __vercmp(newver, ver) < 0:
-						newver = ver
-
-			"""Expect a match for curver in directory list, or else it indicates unknown format"""
-			if not valid:
-				status = "ErrParseInterDir"
-			else:
-				"""rejoin the path name"""
-				status = newver[0] + newver[1]
-		elif not len(fhtml):
-			status = "ErrHostNoDir"
-
-		f.close()
-		if status != "ErrHostNoDir" and re.match("Err", status):
-			logpath = d.getVar('LOG_DIR', True)
-			subprocess.call("cp %s %s/" % (f.name, logpath), shell=True)
-		os.unlink(f.name)
-		return status
-
-	"""
-	Check on the last directory to search '2.4.1' in "http://xxx/2.4/pkg-2.4.1.tar.gz", 
-		'url'	- upstream link customized by regular expression
-		'd'	- database
-		'curname' - current package name
-	Return new version if success, or else error in "Errxxxx" style
-	"""
-	def check_new_version(url, curname, d):
-		"""possible to have no version in pkg name, such as spectrum-fw"""
-		if not re.search("\d+", curname):
-			return pcurver
-		pn = d.getVar('PN', True)
-		f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-2-" % pn)
-		status = internal_fetch_wget(url, d, f)
-		fhtml = f.read()
-
-		if status == "SUCC" and len(fhtml):
-			newver = parse_dir(curname)
-
-			"""match "{PN}-5.21.1.tar.gz">{PN}-5.21.1.tar.gz """
-			pn1 = re.search("^%s" % prefix, curname).group()
-			
-			s = "[^\"]*%s[^\d\"]*?(\d+[\.\-_])+[^\"]*" % pn1
-			searchstr = "[hH][rR][eE][fF]=\"%s\".*[>\"]" % s
-			reg = re.compile(searchstr)
-	
-			valid = 0
-			for line in fhtml.split("\n"):
-				m = reg.search(line)
-				if m:
-					valid = 1
-					ver = m.group().split("\"")[1].split("/")[-1]
-					if ver == "download":
-						ver = m.group().split("\"")[1].split("/")[-2]
-					ver = parse_dir(ver)
-					if ver and __vercmp(newver, ver) < 0:
-						newver = ver
-	
-			"""Expect a match for curver in directory list, or else it indicates unknown format"""
-			if not valid:
-				status = "ErrParseDir"
-			else:
-				"""newver still contains a full package name string"""
-				status = re.search("(\d+[\.\-_])*(\d+[0-9a-zA-Z]*)", newver[1]).group()
-				if "_" in status:
-					status = re.sub("_",".",status)
-				elif "-" in status:
-					status = re.sub("-",".",status)
-		elif not len(fhtml):
-			status = "ErrHostNoDir"
-
-		f.close()
-		"""if host hasn't directory information, no need to save tmp file"""
-		if status != "ErrHostNoDir" and re.match("Err", status):
-			logpath = d.getVar('LOG_DIR', True)
-			subprocess.call("cp %s %s/" % (f.name, logpath), shell=True)
-		os.unlink(f.name)
-		return status
-
-	"""first check whether a uri is provided"""
-	src_uri = d.getVar('SRC_URI', True)
-	if not src_uri:
-		return
-
-	"""initialize log files."""
-	logpath = d.getVar('LOG_DIR', True)
-	bb.utils.mkdirhier(logpath)
-	logfile = os.path.join(logpath, "checkpkg.csv")
-
-	"""generate package information from .bb file"""
-	pname = d.getVar('PN', True)
-
-	if pname.find("-native") != -1:
-	    pnstripped = pname.split("-native")
-	    bb.note("Native Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	if pname.find("-cross") != -1:
-	    pnstripped = pname.split("-cross")
-	    bb.note("cross Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	if pname.find("-initial") != -1:
-	    pnstripped = pname.split("-initial")
-	    bb.note("initial Split: %s" % pnstripped)
-	    localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
-	    bb.data.update_data(localdata)
-
-	pdesc = localdata.getVar('DESCRIPTION', True)
-	pgrp = localdata.getVar('SECTION', True)
-	pversion = localdata.getVar('PV', True)
-	plicense = localdata.getVar('LICENSE', True)
-	psection = localdata.getVar('SECTION', True)
-	phome = localdata.getVar('HOMEPAGE', True)
-	prelease = localdata.getVar('PR', True)
-	pdepends = localdata.getVar('DEPENDS', True)
-	pbugtracker = localdata.getVar('BUGTRACKER', True)
-	ppe = localdata.getVar('PE', True)
-	psrcuri = localdata.getVar('SRC_URI', True)
-	maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
-
-	found = 0
-	for uri in src_uri.split():
-		m = re.compile('(?P<type>[^:]*)').match(uri)
-		if not m:
-			raise MalformedUrl(uri)
-		elif m.group('type') in ('http', 'https', 'ftp', 'cvs', 'svn', 'git'):
-			found = 1
-			pproto = m.group('type')
-			break
-	if not found:
-		pproto = "file"
-	pupver = "N/A"
-	pstatus = "ErrUnknown"
-
-	(type, host, path, user, pswd, parm) = bb.decodeurl(uri)
-	if type in ['http', 'https', 'ftp']:
-		pcurver = d.getVar('PV', True)
-	else:
-		pcurver = d.getVar("SRCREV", True)
-
-	if type in ['http', 'https', 'ftp']:
-		newver = pcurver
-		altpath = path
-		dirver = "-"
-		curname = "-"
-	
-		"""
-		match version number amid the path, such as "5.7" in:
-			http://download.gnome.org/sources/${PN}/5.7/${PN}-${PV}.tar.gz	
-		N.B. how about sth. like "../5.7/5.8/..."? Not find such example so far :-P
-		"""
-		m = re.search(r"[^/]*(\d+\.)+\d+([\-_]r\d+)*/", path)
-		if m:
-			altpath = path.split(m.group())[0]
-			dirver = m.group().strip("/")
-	
-			"""use new path and remove param. for wget only param is md5sum"""
-			alturi = bb.encodeurl([type, host, altpath, user, pswd, {}])
-	
-			newver = check_new_dir(alturi, dirver, d)
-			altpath = path
-			if not re.match("Err", newver) and dirver != newver:
-				altpath = altpath.replace(dirver, newver, True)
-				
-		"""Now try to acquire all remote files in current directory"""
-		if not re.match("Err", newver):
-			curname = altpath.split("/")[-1]
-	
-			"""get remote name by skipping pacakge name"""
-			m = re.search(r"/.*/", altpath)
-			if not m:
-				altpath = "/"
-			else:
-				altpath = m.group()
-	
-			alturi = bb.encodeurl([type, host, altpath, user, pswd, {}])
-			newver = check_new_version(alturi, curname, d)
-			while(newver == "ErrHostNoDir"):
-				if alturi == "/download":
-					break
-				else:
-					alturi = "/".join(alturi.split("/")[0:-2]) + "/download"
-					newver = check_new_version(alturi, curname, d)
-			if not re.match("Err", newver):
-				pupver = newver
-				if pupver != pcurver:
-					pstatus = "UPDATE"
-				else:
-					pstatus = "MATCH"
-	
-		if re.match("Err", newver):
-			pstatus = newver + ":" + altpath + ":" + dirver + ":" + curname
-	elif type == 'git':
-		if user:
-			gituser = user + '@'
-		else:
-			gituser = ""
-
-		if 'protocol' in parm:
-			gitproto = parm['protocol']
-		else:
-			gitproto = "git"
-		gitcmd = "git ls-remote %s://%s%s%s *tag* 2>&1" % (gitproto, gituser, host, path)
-		gitcmd2 = "git ls-remote %s://%s%s%s HEAD 2>&1" % (gitproto, gituser, host, path)
-		tmp = os.popen(gitcmd).read()
-		tmp2 = os.popen(gitcmd2).read()
-		#This is for those repo have tag like: refs/tags/1.2.2
-		if tmp:
-			tmpline = tmp.split("\n")
-			verflag = 0
-			for line in tmpline:
-				if len(line)==0:
-					break;
-				puptag = line.split("/")[-1]
-				puptag = re.search("[0-9][0-9|\.|_]+[0-9]", puptag)
-				if puptag == None:
-					continue;
-				puptag = puptag.group()
-				puptag = re.sub("_",".",puptag)
-				plocaltag = pversion.split("+")[0]
-				if "git" in plocaltag:
-					plocaltag = plocaltag.split("-")[0]
-				result = bb.utils.vercmp(("0", puptag, ""), ("0", plocaltag, ""))
-				if result > 0:
-					verflag = 1
-					pstatus = "UPDATE"
-					pupver = puptag
-				elif verflag == 0 :
-					pupver = plocaltag
-					pstatus = "MATCH"
-		#This is for those no tag repo
-		elif tmp2:
-			pupver = tmp2.split("\t")[0]
-			if pupver in pversion:
-				pstatus = "MATCH"
-			else:
-				pstatus = "UPDATE"
-		else:
-			pstatus = "ErrGitAccess"
-	elif type == 'svn':
-		options = []
-		if user:
-			options.append("--username %s" % user)
-		if pswd:
-			options.append("--password %s" % pswd)
-		svnproto = 'svn'
-		if 'proto' in parm:
-			svnproto = parm['proto']
-		if 'rev' in parm:
-			pcurver = parm['rev']
-
-		svncmd = "svn info %s %s://%s%s/%s/ 2>&1" % (" ".join(options), svnproto, host, path, parm["module"])
-		print svncmd
-		svninfo = os.popen(svncmd).read()
-		for line in svninfo.split("\n"):
-			if re.search("^Last Changed Rev:", line):
-				pupver = line.split(" ")[-1]
-				if pupver in pversion:
-					pstatus = "MATCH"
-				else:
-					pstatus = "UPDATE"
-
-		if re.match("Err", pstatus):
-			pstatus = "ErrSvnAccess"
-	elif type == 'cvs':
-		pupver = "HEAD"
-		pstatus = "UPDATE"
-	elif type == 'file':
-		"""local file is always up-to-date"""
-		pupver = pcurver
-		pstatus = "MATCH"
-	else:
-		pstatus = "ErrUnsupportedProto"
-
-	if re.match("Err", pstatus):
-		pstatus += ":%s%s" % (host, path)
-
-	"""Read from manual distro tracking fields as alternative"""
-	pmver = d.getVar("RECIPE_UPSTREAM_VERSION", True)
-	if not pmver:
-		pmver = "N/A"
-		pmstatus = "ErrNoRecipeData"
-	else:
-		if pmver == pcurver:
-			pmstatus = "MATCH"
-		else:
-			pmstatus = "UPDATE"
-	
-	psrcuri = psrcuri.split()[0]
-	pdepends = "".join(pdepends.split("\t"))
-	pdesc = "".join(pdesc.split("\t"))
-	lf = bb.utils.lockfile("%s.lock" % logfile)
-	f = open(logfile, "a")
-	f.write("%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\n" % \
-		  (pname,pversion,pupver,plicense,psection, phome,prelease, pdepends,pbugtracker,ppe,pdesc,pstatus,pmver,psrcuri,maintainer))
-	f.close()
-	bb.utils.unlockfile(lf)
+        localdata = bb.data.createCopy(d)
+        import sys
+        import re
+        import tempfile
+        import subprocess
+
+        """
+        sanity check to ensure same name and type. Match as many patterns as possible
+        such as:
+                gnome-common-2.20.0.tar.gz (most common format)
+                gtk+-2.90.1.tar.gz
+                xf86-input-synaptics-12.6.9.tar.gz
+                dri2proto-2.3.tar.gz
+                blktool_4.orig.tar.gz
+                libid3tag-0.15.1b.tar.gz
+                unzip552.tar.gz
+                icu4c-3_6-src.tgz
+                genext2fs_1.3.orig.tar.gz
+                gst-fluendo-mp3
+        """
+        prefix1 = "[a-zA-Z][a-zA-Z0-9]*([\-_][a-zA-Z]\w+)*[\-_]"        # match most patterns which uses "-" as separator to version digits
+        prefix2 = "[a-zA-Z]+"                        # a loose pattern such as for unzip552.tar.gz
+        prefix3 = "[0-9a-zA-Z]+"                        # a loose pattern such as for 80325-quicky-0.4.tar.gz
+        prefix = "(%s|%s|%s)" % (prefix1, prefix2, prefix3)
+        suffix = "(tar\.gz|tgz|tar\.bz2|zip|xz|rpm)"
+        suffixtuple = ("tar.gz", "tgz", "zip", "tar.bz2", "tar.xz", "src.rpm")
+
+        sinterstr = "(?P<name>%s?)(?P<ver>.*)" % prefix
+        sdirstr = "(?P<name>%s)(?P<ver>.*)\.(?P<type>%s$)" % (prefix, suffix)
+
+        def parse_inter(s):
+                m = re.search(sinterstr, s)
+                if not m:
+                        return None
+                else:
+                        return (m.group('name'), m.group('ver'), "")
+
+        def parse_dir(s):
+                m = re.search(sdirstr, s)
+                if not m:
+                        return None
+                else:
+                        return (m.group('name'), m.group('ver'), m.group('type'))
+
+        """
+        Check whether 'new' is newer than 'old' version. We use existing vercmp() for the
+        purpose. PE is cleared in comparison as it's not for build, and PV is cleared too
+        for simplicity as it's somehow difficult to get from various upstream format
+        """
+        def __vercmp(old, new):
+                (on, ov, ot) = old
+                (en, ev, et) = new
+                if on != en or (et and et not in suffixtuple):
+                        return 0
+                ov = re.search("[\d|\.]+[^a-zA-Z]+", ov).group()
+                ev = re.search("[\d|\.]+[^a-zA-Z]+", ev).group()
+                return bb.utils.vercmp(("0", ov, ""), ("0", ev, ""))
+
+        """
+        wrapper for fetch upstream directory info
+                'url'        - upstream link customized by regular expression
+                'd'        - database
+                'tmpf'        - tmpfile for fetcher output
+        We don't want to exit whole build due to one recipe error. So handle all exceptions 
+        gracefully w/o leaking to outer. 
+        """
+        def internal_fetch_wget(url, d, tmpf):
+                status = "ErrFetchUnknown"
+                """
+                Clear internal url cache as it's a temporary check. Not doing so will have 
+                bitbake check url multiple times when looping through a single url
+                """
+                fn = d.getVar('FILE', True)
+                bb.fetch2.urldata_cache[fn] = {}
+
+                """
+                To avoid impacting bitbake build engine, this trick is required for reusing bitbake
+                interfaces. bb.fetch.go() is not appliable as it checks downloaded content in ${DL_DIR}
+                while we don't want to pollute that place. So bb.fetch2.checkstatus() is borrowed here
+                which is designed for check purpose but we override check command for our own purpose
+                """
+                ld = bb.data.createCopy(d)
+                d.setVar('CHECKCOMMAND_wget', "/usr/bin/env wget -t 1 --passive-ftp -O %s --user-agent=\"Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.2.12) Gecko/20101027 Ubuntu/9.10 (karmic) Firefox/3.6.12\" '${URI}'" \
+                                        % tmpf.name)
+                bb.data.update_data(ld)
+
+                try:
+                        fetcher = bb.fetch2.Fetch([url], ld)
+                        fetcher.checkstatus()
+                        status = "SUCC"
+                except bb.fetch2.BBFetchException, e:
+                        status = "ErrFetch"
+
+                return status
+
+        """
+        Check on middle version directory such as "2.4/" in "http://xxx/2.4/pkg-2.4.1.tar.gz", 
+                'url'        - upstream link customized by regular expression
+                'd'        - database
+                'curver' - current version
+        Return new version if success, or else error in "Errxxxx" style
+        """
+        def check_new_dir(url, curver, d):
+                pn = d.getVar('PN', True)
+                f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-1-" % pn)
+                status = internal_fetch_wget(url, d, f)
+                fhtml = f.read()
+                if status == "SUCC" and len(fhtml):
+                        newver = parse_inter(curver)
+
+                        """
+                        match "*4.1/">*4.1/ where '*' matches chars
+                        N.B. add package name, only match for digits
+                        """
+                        m = re.search("^%s" % prefix, curver)
+                        if m:
+                                s = "%s[^\d\"]*?(\d+[\.\-_])+\d+/?" % m.group()
+                        else:
+                                s = "(\d+[\.\-_])+\d+/?"
+                                
+                        searchstr = "[hH][rR][eE][fF]=\"%s\">" % s
+                        reg = re.compile(searchstr)
+
+                        valid = 0
+                        for line in fhtml.split("\n"):
+                                if line.find(curver) >= 0:
+                                        valid = 1
+                                m = reg.search(line)
+                                if m:
+                                        ver = m.group().split("\"")[1]
+                                        ver = ver.strip("/")
+                                        ver = parse_inter(ver)
+                                        if ver and __vercmp(newver, ver) < 0:
+                                                newver = ver
+
+                        """Expect a match for curver in directory list, or else it indicates unknown format"""
+                        if not valid:
+                                status = "ErrParseInterDir"
+                        else:
+                                """rejoin the path name"""
+                                status = newver[0] + newver[1]
+                elif not len(fhtml):
+                        status = "ErrHostNoDir"
+
+                f.close()
+                if status != "ErrHostNoDir" and re.match("Err", status):
+                        logpath = d.getVar('LOG_DIR', True)
+                        subprocess.call("cp %s %s/" % (f.name, logpath), shell=True)
+                os.unlink(f.name)
+                return status
+
+        """
+        Check on the last directory to search '2.4.1' in "http://xxx/2.4/pkg-2.4.1.tar.gz", 
+                'url'        - upstream link customized by regular expression
+                'd'        - database
+                'curname' - current package name
+        Return new version if success, or else error in "Errxxxx" style
+        """
+        def check_new_version(url, curname, d):
+                """possible to have no version in pkg name, such as spectrum-fw"""
+                if not re.search("\d+", curname):
+                        return pcurver
+                pn = d.getVar('PN', True)
+                f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-2-" % pn)
+                status = internal_fetch_wget(url, d, f)
+                fhtml = f.read()
+
+                if status == "SUCC" and len(fhtml):
+                        newver = parse_dir(curname)
+
+                        """match "{PN}-5.21.1.tar.gz">{PN}-5.21.1.tar.gz """
+                        pn1 = re.search("^%s" % prefix, curname).group()
+                        
+                        s = "[^\"]*%s[^\d\"]*?(\d+[\.\-_])+[^\"]*" % pn1
+                        searchstr = "[hH][rR][eE][fF]=\"%s\".*[>\"]" % s
+                        reg = re.compile(searchstr)
+        
+                        valid = 0
+                        for line in fhtml.split("\n"):
+                                m = reg.search(line)
+                                if m:
+                                        valid = 1
+                                        ver = m.group().split("\"")[1].split("/")[-1]
+                                        if ver == "download":
+                                                ver = m.group().split("\"")[1].split("/")[-2]
+                                        ver = parse_dir(ver)
+                                        if ver and __vercmp(newver, ver) < 0:
+                                                newver = ver
+        
+                        """Expect a match for curver in directory list, or else it indicates unknown format"""
+                        if not valid:
+                                status = "ErrParseDir"
+                        else:
+                                """newver still contains a full package name string"""
+                                status = re.search("(\d+[\.\-_])*(\d+[0-9a-zA-Z]*)", newver[1]).group()
+                                if "_" in status:
+                                        status = re.sub("_",".",status)
+                                elif "-" in status:
+                                        status = re.sub("-",".",status)
+                elif not len(fhtml):
+                        status = "ErrHostNoDir"
+
+                f.close()
+                """if host hasn't directory information, no need to save tmp file"""
+                if status != "ErrHostNoDir" and re.match("Err", status):
+                        logpath = d.getVar('LOG_DIR', True)
+                        subprocess.call("cp %s %s/" % (f.name, logpath), shell=True)
+                os.unlink(f.name)
+                return status
+
+        """first check whether a uri is provided"""
+        src_uri = d.getVar('SRC_URI', True)
+        if not src_uri:
+                return
+
+        """initialize log files."""
+        logpath = d.getVar('LOG_DIR', True)
+        bb.utils.mkdirhier(logpath)
+        logfile = os.path.join(logpath, "checkpkg.csv")
+
+        """generate package information from .bb file"""
+        pname = d.getVar('PN', True)
+
+        if pname.find("-native") != -1:
+            pnstripped = pname.split("-native")
+            bb.note("Native Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        if pname.find("-cross") != -1:
+            pnstripped = pname.split("-cross")
+            bb.note("cross Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        if pname.find("-initial") != -1:
+            pnstripped = pname.split("-initial")
+            bb.note("initial Split: %s" % pnstripped)
+            localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True))
+            bb.data.update_data(localdata)
+
+        pdesc = localdata.getVar('DESCRIPTION', True)
+        pgrp = localdata.getVar('SECTION', True)
+        pversion = localdata.getVar('PV', True)
+        plicense = localdata.getVar('LICENSE', True)
+        psection = localdata.getVar('SECTION', True)
+        phome = localdata.getVar('HOMEPAGE', True)
+        prelease = localdata.getVar('PR', True)
+        pdepends = localdata.getVar('DEPENDS', True)
+        pbugtracker = localdata.getVar('BUGTRACKER', True)
+        ppe = localdata.getVar('PE', True)
+        psrcuri = localdata.getVar('SRC_URI', True)
+        maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
+
+        found = 0
+        for uri in src_uri.split():
+                m = re.compile('(?P<type>[^:]*)').match(uri)
+                if not m:
+                        raise MalformedUrl(uri)
+                elif m.group('type') in ('http', 'https', 'ftp', 'cvs', 'svn', 'git'):
+                        found = 1
+                        pproto = m.group('type')
+                        break
+        if not found:
+                pproto = "file"
+        pupver = "N/A"
+        pstatus = "ErrUnknown"
+
+        (type, host, path, user, pswd, parm) = bb.decodeurl(uri)
+        if type in ['http', 'https', 'ftp']:
+                pcurver = d.getVar('PV', True)
+        else:
+                pcurver = d.getVar("SRCREV", True)
+
+        if type in ['http', 'https', 'ftp']:
+                newver = pcurver
+                altpath = path
+                dirver = "-"
+                curname = "-"
+        
+                """
+                match version number amid the path, such as "5.7" in:
+                        http://download.gnome.org/sources/${PN}/5.7/${PN}-${PV}.tar.gz        
+                N.B. how about sth. like "../5.7/5.8/..."? Not find such example so far :-P
+                """
+                m = re.search(r"[^/]*(\d+\.)+\d+([\-_]r\d+)*/", path)
+                if m:
+                        altpath = path.split(m.group())[0]
+                        dirver = m.group().strip("/")
+        
+                        """use new path and remove param. for wget only param is md5sum"""
+                        alturi = bb.encodeurl([type, host, altpath, user, pswd, {}])
+        
+                        newver = check_new_dir(alturi, dirver, d)
+                        altpath = path
+                        if not re.match("Err", newver) and dirver != newver:
+                                altpath = altpath.replace(dirver, newver, True)
+                                
+                """Now try to acquire all remote files in current directory"""
+                if not re.match("Err", newver):
+                        curname = altpath.split("/")[-1]
+        
+                        """get remote name by skipping pacakge name"""
+                        m = re.search(r"/.*/", altpath)
+                        if not m:
+                                altpath = "/"
+                        else:
+                                altpath = m.group()
+        
+                        alturi = bb.encodeurl([type, host, altpath, user, pswd, {}])
+                        newver = check_new_version(alturi, curname, d)
+                        while(newver == "ErrHostNoDir"):
+                                if alturi == "/download":
+                                        break
+                                else:
+                                        alturi = "/".join(alturi.split("/")[0:-2]) + "/download"
+                                        newver = check_new_version(alturi, curname, d)
+                        if not re.match("Err", newver):
+                                pupver = newver
+                                if pupver != pcurver:
+                                        pstatus = "UPDATE"
+                                else:
+                                        pstatus = "MATCH"
+        
+                if re.match("Err", newver):
+                        pstatus = newver + ":" + altpath + ":" + dirver + ":" + curname
+        elif type == 'git':
+                if user:
+                        gituser = user + '@'
+                else:
+                        gituser = ""
+
+                if 'protocol' in parm:
+                        gitproto = parm['protocol']
+                else:
+                        gitproto = "git"
+                gitcmd = "git ls-remote %s://%s%s%s *tag* 2>&1" % (gitproto, gituser, host, path)
+                gitcmd2 = "git ls-remote %s://%s%s%s HEAD 2>&1" % (gitproto, gituser, host, path)
+                tmp = os.popen(gitcmd).read()
+                tmp2 = os.popen(gitcmd2).read()
+                #This is for those repo have tag like: refs/tags/1.2.2
+                if tmp:
+                        tmpline = tmp.split("\n")
+                        verflag = 0
+                        for line in tmpline:
+                                if len(line)==0:
+                                        break;
+                                puptag = line.split("/")[-1]
+                                puptag = re.search("[0-9][0-9|\.|_]+[0-9]", puptag)
+                                if puptag == None:
+                                        continue;
+                                puptag = puptag.group()
+                                puptag = re.sub("_",".",puptag)
+                                plocaltag = pversion.split("+")[0]
+                                if "git" in plocaltag:
+                                        plocaltag = plocaltag.split("-")[0]
+                                result = bb.utils.vercmp(("0", puptag, ""), ("0", plocaltag, ""))
+                                if result > 0:
+                                        verflag = 1
+                                        pstatus = "UPDATE"
+                                        pupver = puptag
+                                elif verflag == 0 :
+                                        pupver = plocaltag
+                                        pstatus = "MATCH"
+                #This is for those no tag repo
+                elif tmp2:
+                        pupver = tmp2.split("\t")[0]
+                        if pupver in pversion:
+                                pstatus = "MATCH"
+                        else:
+                                pstatus = "UPDATE"
+                else:
+                        pstatus = "ErrGitAccess"
+        elif type == 'svn':
+                options = []
+                if user:
+                        options.append("--username %s" % user)
+                if pswd:
+                        options.append("--password %s" % pswd)
+                svnproto = 'svn'
+                if 'proto' in parm:
+                        svnproto = parm['proto']
+                if 'rev' in parm:
+                        pcurver = parm['rev']
+
+                svncmd = "svn info %s %s://%s%s/%s/ 2>&1" % (" ".join(options), svnproto, host, path, parm["module"])
+                print svncmd
+                svninfo = os.popen(svncmd).read()
+                for line in svninfo.split("\n"):
+                        if re.search("^Last Changed Rev:", line):
+                                pupver = line.split(" ")[-1]
+                                if pupver in pversion:
+                                        pstatus = "MATCH"
+                                else:
+                                        pstatus = "UPDATE"
+
+                if re.match("Err", pstatus):
+                        pstatus = "ErrSvnAccess"
+        elif type == 'cvs':
+                pupver = "HEAD"
+                pstatus = "UPDATE"
+        elif type == 'file':
+                """local file is always up-to-date"""
+                pupver = pcurver
+                pstatus = "MATCH"
+        else:
+                pstatus = "ErrUnsupportedProto"
+
+        if re.match("Err", pstatus):
+                pstatus += ":%s%s" % (host, path)
+
+        """Read from manual distro tracking fields as alternative"""
+        pmver = d.getVar("RECIPE_UPSTREAM_VERSION", True)
+        if not pmver:
+                pmver = "N/A"
+                pmstatus = "ErrNoRecipeData"
+        else:
+                if pmver == pcurver:
+                        pmstatus = "MATCH"
+                else:
+                        pmstatus = "UPDATE"
+        
+        psrcuri = psrcuri.split()[0]
+        pdepends = "".join(pdepends.split("\t"))
+        pdesc = "".join(pdesc.split("\t"))
+        lf = bb.utils.lockfile("%s.lock" % logfile)
+        f = open(logfile, "a")
+        f.write("%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\n" % \
+                  (pname,pversion,pupver,plicense,psection, phome,prelease, pdepends,pbugtracker,ppe,pdesc,pstatus,pmver,psrcuri,maintainer))
+        f.close()
+        bb.utils.unlockfile(lf)
 }
 
 addtask checkpkgall after do_checkpkg
 do_checkpkgall[recrdeptask] = "do_checkpkgall do_checkpkg"
 do_checkpkgall[nostamp] = "1"
 do_checkpkgall() {
-	:
+        :
 }
 
 addhandler distro_check_eventhandler
@@ -679,7 +679,7 @@ addtask distro_checkall after do_distro_check
 do_distro_checkall[recrdeptask] = "do_distro_checkall do_distro_check"
 do_distro_checkall[nostamp] = "1"
 do_distro_checkall() {
-	:
+        :
 }
 #
 #Check Missing License Text.
@@ -727,7 +727,7 @@ addtask checklicenseall after do_checklicense
 do_checklicenseall[recrdeptask] = "do_checklicenseall do_checklicense"
 do_checklicenseall[nostamp] = "1"
 do_checklicenseall() {
-	:
+        :
 }
 
 
-- 
1.7.7.6





More information about the Openembedded-core mailing list