[OE-core] [PATCH] bbclass files: Fix typos/grammar in numerous .bbclass comments.

Robert P. J. Day rpjday at crashcourse.ca
Wed Aug 6 11:51:55 UTC 2014


Various non-functional changes to a number of .bbclass files:

 * Spelling
 * Grammar
 * Ridiculously long lines

Signed-off-by: Robert P. J. Day <rpjday at crashcourse.ca>

---

  while perusing numerous .bbclass files to document a few things for
myself, i figured i might as well do some proofreading. all of this is
strictly non-functional changes to comment lines in .bbclass files.

diff --git a/meta/classes/allarch.bbclass b/meta/classes/allarch.bbclass
index c953e7c..0a64588 100644
--- a/meta/classes/allarch.bbclass
+++ b/meta/classes/allarch.bbclass
@@ -1,5 +1,6 @@
 #
-# This class is used for architecture independent recipes/data files (usally scripts)
+# This class is used for architecture-independent recipes/data files, such as
+# configuration files, media files or scripts.
 #

 # Expand STAGING_DIR_HOST since for cross-canadian/native/nativesdk, this will
@@ -15,8 +16,8 @@ python () {
         # No need for virtual/libc or a cross compiler
         d.setVar("INHIBIT_DEFAULT_DEPS","1")

-        # Set these to a common set of values, we shouldn't be using them other that for WORKDIR directory
-        # naming anyway
+        # Set these to a common set of values, we shouldn't be using them
+        # other than for WORKDIR directory naming, anyway.
         d.setVar("TARGET_ARCH", "allarch")
         d.setVar("TARGET_OS", "linux")
         d.setVar("TARGET_CC_ARCH", "none")
@@ -33,7 +34,7 @@ python () {
         # packages.
         d.setVar("LDFLAGS", "")

-        # No need to do shared library processing or debug symbol handling
+        # No need to do shared library processing or debug symbol handling.
         d.setVar("EXCLUDE_FROM_SHLIBS", "1")
         d.setVar("INHIBIT_PACKAGE_DEBUG_SPLIT", "1")
         d.setVar("INHIBIT_PACKAGE_STRIP", "1")
diff --git a/meta/classes/archiver.bbclass b/meta/classes/archiver.bbclass
index efd413b..7f04387 100644
--- a/meta/classes/archiver.bbclass
+++ b/meta/classes/archiver.bbclass
@@ -1,7 +1,7 @@
 # ex:ts=4:sw=4:sts=4:et
 # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*-
 #
-# This bbclass is used for creating archive for:
+# This bbclass is used for creating archives for:
 # 1) original (or unpacked) source: ARCHIVER_MODE[src] = "original"
 # 2) patched source: ARCHIVER_MODE[src] = "patched" (default)
 # 3) configured source: ARCHIVER_MODE[src] = "configured"
@@ -48,7 +48,6 @@ do_ar_original[dirs] = "${ARCHIVER_OUTDIR} ${ARCHIVER_WORKDIR}"

 # This is a convenience for the shell script to use it

-
 python () {
     pn = d.getVar('PN', True)

@@ -71,7 +70,7 @@ python () {
         d.appendVarFlag('do_deploy_archives', 'depends', ' %s:do_ar_patched' % pn)
     elif ar_src == "configured":
         # We can't use "addtask do_ar_configured after do_configure" since it
-        # will cause the deptask of do_populate_sysroot to run not matter what
+        # will cause the deptask of do_populate_sysroot to run no matter what
         # archives we need, so we add the depends here.
         d.appendVarFlag('do_ar_configured', 'depends', ' %s:do_configure' % pn)
         d.appendVarFlag('do_deploy_archives', 'depends', ' %s:do_ar_configured' % pn)
@@ -122,7 +121,7 @@ python () {
             d.setVarFlag('do_unpack_and_patch', 'stamp-base-clean', flag_clean)
 }

-# Take all the sources for a recipe and puts them in WORKDIR/archiver-work/.
+# Take all the sources for a recipe and put them in WORKDIR/archiver-work/.
 # Files in SRC_URI are copied directly, anything that's a directory
 # (e.g. git repositories) is "unpacked" and then put into a tarball.
 python do_ar_original() {
@@ -190,8 +189,7 @@ python do_ar_configured() {
         bb.note('Archiving the configured source...')
         # The libtool-native's do_configure will remove the
         # ${STAGING_DATADIR}/aclocal/libtool.m4, so we can't re-run the
-        # do_configure, we archive the already configured ${S} to
-        # instead of.
+        # do_configure, we archive the already configured ${S} instead.
         if d.getVar('PN', True) != 'libtool-native':
             # Change the WORKDIR to make do_configure run in another dir.
             d.setVar('WORKDIR', d.getVar('ARCHIVER_WORKDIR', True))
@@ -276,9 +274,9 @@ python do_unpack_and_patch() {
     # Change the WORKDIR to make do_unpack do_patch run in another dir.
     d.setVar('WORKDIR', d.getVar('ARCHIVER_WORKDIR', True))

-    # The changed 'WORKDIR' also casued 'B' changed, create dir 'B' for the
-    # possibly requiring of the following tasks (such as some recipes's
-    # do_patch required 'B' existed).
+    # The changed 'WORKDIR' also caused 'B' to change, create dir 'B' for
+    # possibly requiring the following tasks (such as some recipe's
+    # do_patch requiring that 'B' exists).
     bb.utils.mkdirhier(d.getVar('B', True))

     # The kernel source is ready after do_validate_branches
diff --git a/meta/classes/base.bbclass b/meta/classes/base.bbclass
index 8114cf6..3c24727 100644
--- a/meta/classes/base.bbclass
+++ b/meta/classes/base.bbclass
@@ -68,7 +68,7 @@ def base_dep_prepend(d):
     #

     deps = ""
-    # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command.  Whether or  not
+    # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command.  Whether or not
     # we need that built is the responsibility of the patch function / class, not
     # the application.
     if not d.getVar('INHIBIT_DEFAULT_DEPS'):
@@ -81,8 +81,8 @@ BASEDEPENDS = "${@base_dep_prepend(d)}"
 DEPENDS_prepend="${BASEDEPENDS} "

 FILESPATH = "${@base_set_filespath(["${FILE_DIRNAME}/${BP}", "${FILE_DIRNAME}/${BPN}", "${FILE_DIRNAME}/files"], d)}"
-# THISDIR only works properly with imediate expansion as it has to run
-# in the context of the location its used (:=)
+# THISDIR only works properly with immediate expansion as it has to run
+# in the context of the location where it's used (:=)
 THISDIR = "${@os.path.dirname(d.getVar('FILE', True))}"

 def extra_path_elements(d):
@@ -311,8 +311,8 @@ python base_eventhandler() {
         bb.plain('\n%s\n%s\n' % (statusheader, '\n'.join(statuslines)))

     # This code is to silence warnings where the SDK variables overwrite the
-    # target ones and we'd see dulpicate key names overwriting each other
-    # for various PREFERRED_PROVIDERS
+    # target ones and we'd see duplicate key names overwriting each other
+    # for various PREFERRED_PROVIDERs.
     if isinstance(e, bb.event.RecipePreFinalise):
         if e.data.getVar("TARGET_PREFIX", True) == e.data.getVar("SDK_PREFIX", True):
             e.data.delVar("PREFERRED_PROVIDER_virtual/${TARGET_PREFIX}binutils")
@@ -465,7 +465,7 @@ python () {
         else:
             appendVar('EXTRA_OECONF', extraconf)

-    # If PRINC is set, try and increase the PR value by the amount specified
+    # If PRINC is set, try and increase the PR value by the amount specified.
     # The PR server is now the preferred way to handle PR changes based on
     # the checksum of the recipe (including bbappend).  The PRINC is now
     # obsolete.  Return a warning to the user.
@@ -495,7 +495,7 @@ python () {
                  " whitelisted in LICENSE_FLAGS_WHITELIST")

     # If we're building a target package we need to use fakeroot (pseudo)
-    # in order to capture permissions, owners, groups and special files
+    # in order to capture permissions, owners, groups and special files.
     if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('cross', d):
         d.setVarFlag('do_unpack', 'umask', '022')
         d.setVarFlag('do_configure', 'umask', '022')
@@ -587,12 +587,12 @@ python () {
     elif "osc://" in srcuri:
         d.appendVarFlag('do_fetch', 'depends', ' osc-native:do_populate_sysroot')

-    # *.lz4 should depends on lz4-native for unpacking
+    # *.lz4 should depend on lz4-native for unpacking
     # Not endswith because of "*.patch.lz4;patch=1". Need bb.fetch.decodeurl in future
     if '.lz4' in srcuri:
         d.appendVarFlag('do_unpack', 'depends', ' lz4-native:do_populate_sysroot')

-    # *.xz should depends on xz-native for unpacking
+    # *.xz should depend on xz-native for unpacking
     # Not endswith because of "*.patch.xz;patch=1". Need bb.fetch.decodeurl in future
     if '.xz' in srcuri:
         d.appendVarFlag('do_unpack', 'depends', ' xz-native:do_populate_sysroot')
@@ -616,7 +616,7 @@ python () {
         return

     #
-    # We always try to scan SRC_URI for urls with machine overrides
+    # We always try to scan SRC_URI for URLs with machine overrides
     # unless the package sets SRC_URI_OVERRIDES_PACKAGE_ARCH=0
     #
     override = d.getVar('SRC_URI_OVERRIDES_PACKAGE_ARCH', True)
diff --git a/meta/classes/blacklist.bbclass b/meta/classes/blacklist.bbclass
index a0141a8..b988808 100644
--- a/meta/classes/blacklist.bbclass
+++ b/meta/classes/blacklist.bbclass
@@ -1,15 +1,11 @@
-# anonymous support class from originally from angstrom
+# anonymous support class originally from angstrom
 #
-# To use the blacklist, a distribution should include this
-# class in the INHERIT_DISTRO
-#
-# No longer use ANGSTROM_BLACKLIST, instead use a table of
-# recipes in PNBLACKLIST
+# This class is already included by default from defaultsetup.conf.
 #
 # Features:
 #
 # * To add a package to the blacklist, set:
-#   PNBLACKLIST[pn] = "message"
+#   PNBLACKLIST[pn] = "message explaining rejection"
 #

 # Cope with PNBLACKLIST flags for multilib case
diff --git a/meta/classes/boot-directdisk.bbclass b/meta/classes/boot-directdisk.bbclass
index 995d3e7..8c8039b 100644
--- a/meta/classes/boot-directdisk.bbclass
+++ b/meta/classes/boot-directdisk.bbclass
@@ -1,19 +1,19 @@
 # boot-directdisk.bbclass
-# (loosly based off bootimg.bbclass Copyright (C) 2004, Advanced Micro Devices, Inc.)
+# (loosely based off bootimg.bbclass Copyright (C) 2004, Advanced Micro Devices, Inc.)
 #
 # Create an image which can be placed directly onto a harddisk using dd and then
 # booted.
 #
-# This uses syslinux. extlinux would have been nice but required the ext2/3
+# This uses SYSLINUX. extlinux would have been nice but required the ext2/3
 # partition to be mounted. grub requires to run itself as part of the install
 # process.
 #
-# The end result is a 512 boot sector populated with an MBR and partition table
-# followed by an msdos fat16 partition containing syslinux and a linux kernel
-# completed by the ext2/3 rootfs.
+# The end result is a 512-byte boot sector populated with an MBR and
+# partition table followed by an MSDOS fat16 partition containing SYSLINUX
+# and a linux kernel completed by the ext2/3 rootfs.
 #
-# We have to push the msdos parition table size > 16MB so fat 16 is used as parted
-# won't touch fat12 partitions.
+# We have to push the MSDOS partition table size > 16MB so fat16 is used
+# as parted won't touch fat12 partitions.

 # External variables needed

diff --git a/meta/classes/bugzilla.bbclass b/meta/classes/bugzilla.bbclass
index 3fc8956..234d964 100644
--- a/meta/classes/bugzilla.bbclass
+++ b/meta/classes/bugzilla.bbclass
@@ -1,7 +1,7 @@
 #
 # Small event handler to automatically open URLs and file
-# bug reports at a bugzilla of your choiche
-# it uses XML-RPC interface, so you must have it enabled
+# bug reports at a bugzilla of your choice;
+# it uses XML-RPC interface, so you must have it enabled.
 #
 # Before using you must define BUGZILLA_USER, BUGZILLA_PASS credentials,
 # BUGZILLA_XMLRPC - uri of xmlrpc.cgi,
diff --git a/meta/classes/buildhistory.bbclass b/meta/classes/buildhistory.bbclass
index 20382ce..7c6d384 100644
--- a/meta/classes/buildhistory.bbclass
+++ b/meta/classes/buildhistory.bbclass
@@ -24,7 +24,7 @@ sstate_install[vardepsexclude] += "buildhistory_emit_pkghistory"
 SSTATEPOSTINSTFUNCS[vardepvalueexclude] .= "| buildhistory_emit_pkghistory"

 #
-# Write out metadata about this package for comparision when writing future packages
+# Write out metadata about this package for comparison when writing future packages
 #
 python buildhistory_emit_pkghistory() {
     if not d.getVar('BB_CURRENTTASK', True) in ['packagedata', 'packagedata_setscene']:
@@ -431,7 +431,8 @@ buildhistory_get_sdk_installed_target() {

 buildhistory_list_files() {
 	# List the files in the specified directory, but exclude date/time etc.
-	# This awk script is somewhat messy, but handles where the size is not printed for device files under pseudo
+	# This awk script is somewhat messy, but handles where the size is
+	# not printed for device files under pseudo
 	( cd $1 && find . -printf "%M %-10u %-10g %10s %p -> %l\n" | sort -k5 | sed 's/ * -> $//' > $2 )
 }

@@ -490,7 +491,8 @@ ROOTFS_POSTPROCESS_COMMAND =+ " buildhistory_list_installed_image ;\

 IMAGE_POSTPROCESS_COMMAND += " buildhistory_get_imageinfo ; "

-# We want these to be the last run so that we get called after complementary package installation
+# We want these to be the last run so that we get called after
+# complementary package installation.
 POPULATE_SDK_POST_TARGET_COMMAND_append = " buildhistory_list_installed_sdk_target ;\
                                             buildhistory_get_sdk_installed_target ; "
 POPULATE_SDK_POST_HOST_COMMAND_append = " buildhistory_list_installed_sdk_host ;\
@@ -505,7 +507,8 @@ def buildhistory_get_layers(d):
     return layertext

 def buildhistory_get_metadata_revs(d):
-    # We want an easily machine-readable format here, so get_layers_branch_rev isn't quite what we want
+    # We want an easily machine-readable format here, so get_layers_branch_rev
+    # isn't quite what we want
     layers = (d.getVar("BBLAYERS", True) or "").split()
     medadata_revs = ["%-17s = %s:%s" % (os.path.basename(i), \
         base_get_metadata_git_branch(i, None).strip(), \
@@ -555,7 +558,8 @@ def buildhistory_get_cmdline(d):

 buildhistory_commit() {
 	if [ ! -d ${BUILDHISTORY_DIR} ] ; then
-		# Code above that creates this dir never executed, so there can't be anything to commit
+		# Code above that creates this dir never executed, so there
+		# can't be anything to commit
 		return
 	fi

diff --git a/meta/classes/core-image.bbclass b/meta/classes/core-image.bbclass
index 1b36cba..b897c5b 100644
--- a/meta/classes/core-image.bbclass
+++ b/meta/classes/core-image.bbclass
@@ -7,8 +7,8 @@ LIC_FILES_CHKSUM = "file://${COREBASE}/LICENSE;md5=4d92cd373abda3937c2bc47fbc49d

 # IMAGE_FEATURES control content of the core reference images
 #
-# By default we install packagegroup-core-boot and packagegroup-base packages - this gives us
-# working (console only) rootfs.
+# By default we install packagegroup-core-boot and packagegroup-base packages;
+# this gives us working (console only) rootfs.
 #
 # Available IMAGE_FEATURES:
 #
diff --git a/meta/classes/cross-canadian.bbclass b/meta/classes/cross-canadian.bbclass
index 6da43fe..54322b4 100644
--- a/meta/classes/cross-canadian.bbclass
+++ b/meta/classes/cross-canadian.bbclass
@@ -1,5 +1,5 @@
 #
-# NOTE - When using this class the user is repsonsible for ensuring that
+# NOTE - When using this class the user is responsible for ensuring that
 # TRANSLATED_TARGET_ARCH is added into PN. This ensures that if the TARGET_ARCH
 # is changed, another nativesdk xxx-canadian-cross can be installed
 #
@@ -114,7 +114,7 @@ do_populate_sysroot[stamp-extra-info] = ""

 USE_NLS = "${SDKUSE_NLS}"

-# We have to us TARGET_ARCH but we care about the absolute value
+# We have to use TARGET_ARCH but we care about the absolute value
 # and not any particular tune that is enabled.
 TARGET_ARCH[vardepsexclude] = "TUNE_ARCH"

diff --git a/meta/classes/crosssdk.bbclass b/meta/classes/crosssdk.bbclass
index 261a374..27a3e2b 100644
--- a/meta/classes/crosssdk.bbclass
+++ b/meta/classes/crosssdk.bbclass
@@ -29,7 +29,7 @@ baselib = "lib"
 do_populate_sysroot[stamp-extra-info] = ""
 do_packagedata[stamp-extra-info] = ""

-# Need to force this to ensure consitency accross architectures
+# Need to force this to ensure consistency accross architectures
 EXTRA_OECONF_FPU = ""

 USE_NLS = "no"
diff --git a/meta/classes/debian.bbclass b/meta/classes/debian.bbclass
index 1ddb56f..2eca2db 100644
--- a/meta/classes/debian.bbclass
+++ b/meta/classes/debian.bbclass
@@ -1,7 +1,7 @@
-# Debian package renaming only occurs when a package is built
+# Debian package renaming only occurs when a package is built.
 # We therefore have to make sure we build all runtime packages
-# before building the current package to make the packages runtime
-# depends are correct
+# before building the current package to make sure the packages
+# runtime depends are correct.
 #
 # Custom library package names can be defined setting
 # DEBIANNAME_ + pkgname to the desired name.
diff --git a/meta/classes/devshell.bbclass b/meta/classes/devshell.bbclass
index 41164a3..be994dd 100644
--- a/meta/classes/devshell.bbclass
+++ b/meta/classes/devshell.bbclass
@@ -22,8 +22,8 @@ do_devshell[nostamp] = "1"

 # devshell and fakeroot/pseudo need careful handling since only the final
 # command should run under fakeroot emulation, any X connection should
-# be done as the normal user. We therfore carefully construct the envionment
-# manually
+# be done as the normal user. We therefore carefully construct the environment
+# manually.
 python () {
     if d.getVarFlag("do_devshell", "fakeroot"):
        # We need to signal our code that we want fakeroot however we
diff --git a/meta/classes/grub-efi.bbclass b/meta/classes/grub-efi.bbclass
index 47bd35e..189b102 100644
--- a/meta/classes/grub-efi.bbclass
+++ b/meta/classes/grub-efi.bbclass
@@ -13,7 +13,7 @@
 # ${LABELS} - a list of targets for the automatic config
 # ${APPEND} - an override list of append strings for each label
 # ${GRUB_OPTS} - additional options to add to the config, ';' delimited # (optional)
-# ${GRUB_TIMEOUT} - timeout before executing the deault label (optional)
+# ${GRUB_TIMEOUT} - timeout before executing the default label (optional)

 do_bootimg[depends] += "${MLPREFIX}grub-efi:do_deploy"
 do_bootdirectdisk[depends] += "${MLPREFIX}grub-efi:do_deploy"
diff --git a/meta/classes/icecc.bbclass b/meta/classes/icecc.bbclass
index 3ec8c06..db44703 100644
--- a/meta/classes/icecc.bbclass
+++ b/meta/classes/icecc.bbclass
@@ -3,26 +3,30 @@
 # Stages directories with symlinks from gcc/g++ to icecc, for both
 # native and cross compilers. Depending on each configure or compile,
 # the directories are added at the head of the PATH list and ICECC_CXX
-# and ICEC_CC are set.
+# and ICECC_CC are set.
 #
 # For the cross compiler, creates a tar.gz of our toolchain and sets
 # ICECC_VERSION accordingly.
 #
-# The class now handles all 3 different compile 'stages' (i.e native ,cross-kernel and target) creating the
-# necessary environment tar.gz file to be used by the remote machines.
-# It also supports meta-toolchain generation
+# The class now handles all 3 different compile 'stages' (i.e native,
+# cross-kernel and target) creating the necessary environment tar.gz file
+# to be used by the remote machines. It also supports meta-toolchain generation.
 #
-# If ICECC_PATH is not set in local.conf then the class will try to locate it using 'bb.utils.which'
-# but nothing is sure ;)
+# If ICECC_PATH is not set in local.conf then the class will try to locate it
+# using 'bb.utils.which' but nothing is sure ;)
 #
-# If ICECC_ENV_EXEC is set in local.conf, then it should point to the icecc-create-env script provided by the user
-# or the default one provided by icecc-create-env.bb will be used
-# (NOTE that this is a modified version of the script need it and *not the one that comes with icecc*
+# If ICECC_ENV_EXEC is set in local.conf, then it should point to the
+# icecc-create-env script provided by the user or the default one
+# provided by icecc-create-env.bb will be used.
+# (NOTE that this is a modified version of the script and *not the one
+# that comes with icecc*.
 #
-# User can specify if specific packages or packages belonging to class should not use icecc to distribute
-# compile jobs to remote machines, but handled locally, by defining ICECC_USER_CLASS_BL and ICECC_USER_PACKAGE_BL
-# with the appropriate values in local.conf. In addition the user can force to enable icecc for packages
-# which set an empty PARALLEL_MAKE variable by defining ICECC_USER_PACKAGE_WL.
+# User can specify if specific packages or packages belonging to class
+# should not use icecc to distribute compile jobs to remote machines,
+# but handled locally, by defining ICECC_USER_CLASS_BL and ICECC_USER_PACKAGE_BL
+# with the appropriate values in local.conf. In addition, the user can
+# force to enable icecc for packages # which set an empty PARALLEL_MAKE
+# variable by defining ICECC_USER_PACKAGE_WL.
 #
 #########################################################################################
 #Error checking is kept to minimum so double check any parameters you pass to the class
@@ -33,7 +37,7 @@ BB_HASHBASE_WHITELIST += "ICECC_PARALLEL_MAKE ICECC_DISABLED ICECC_USER_PACKAGE_
 ICECC_ENV_EXEC ?= "${STAGING_BINDIR_NATIVE}/icecc-create-env"

 def icecc_dep_prepend(d):
-    # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command.  Whether or  not
+    # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command.  Whether or not
     # we need that built is the responsibility of the patch function / class, not
     # the application.
     if not d.getVar('INHIBIT_DEFAULT_DEPS'):
@@ -66,7 +70,7 @@ def create_path(compilers, bb, d):
     if icc_is_kernel(bb, d):
         staging += "-kernel"

-    #check if the icecc path is set by the user
+    # Check if the icecc path is set by the user
     icecc = get_icecc(d)

     # Create the dir if necessary
diff --git a/meta/classes/insane.bbclass b/meta/classes/insane.bbclass
index 55bfaf2..51b1993 100644
--- a/meta/classes/insane.bbclass
+++ b/meta/classes/insane.bbclass
@@ -8,7 +8,7 @@
 #  -Check the RUNTIME path for the $TMPDIR
 #  -Check if .la files wrongly point to workdir
 #  -Check if .pc files wrongly point to workdir
-#  -Check if packages contains .debug directories or .so files
+#  -Check if packages contain .debug directories or .so files
 #   where they should be in -dev or -dbg
 #  -Check if config.log contains traces to broken autoconf tests
 #  -Ensure that binaries in base_[bindir|sbindir|libdir] do not link
@@ -24,7 +24,7 @@ QADEPENDS_class-native = ""
 QADEPENDS_class-nativesdk = ""
 QA_SANE = "True"

-# Elect whether a given type of error is a warning or error, they may
+# Select whether a given type of error is a warning or error, they may
 # have been set by other files.
 WARN_QA ?= "ldflags useless-rpaths rpaths staticdev libdir xorg-driver-abi \
             textrel already-stripped incompatible-license files-invalid \
@@ -236,8 +236,8 @@ def package_qa_check_useless_rpaths(file, name, d, elf, messages):
         if m:
             rpath = m.group(1)
             if rpath_eq(rpath, libdir) or rpath_eq(rpath, base_libdir):
-                # The dynamic linker searches both these places anyway.  There is no point in
-                # looking there again.
+                # The dynamic linker searches both these places anyway.
+                # There is no point in looking there again.
                 messages["useless-rpaths"] = "%s: %s contains probably-redundant RPATH %s" % (name, package_qa_clean_path(file, d), rpath)

 QAPATHTEST[dev-so] = "package_qa_check_dev"
diff --git a/meta/classes/kernel-grub.bbclass b/meta/classes/kernel-grub.bbclass
index a63f482..b9079e9 100644
--- a/meta/classes/kernel-grub.bbclass
+++ b/meta/classes/kernel-grub.bbclass
@@ -4,14 +4,14 @@
 # you to fall back to the original kernel as well.
 #
 # - In kernel-image's preinstall scriptlet, it backs up original kernel to avoid
-#   probable confliction with the new one.
+#   probable conflict with the new one.
 #
 # - In kernel-image's postinstall scriptlet, it modifies grub's config file to
 #   updates the new kernel as the boot priority.
 #

 pkg_preinst_kernel-image_append () {
-	# Parsing confliction
+	# Parsing conflict
 	[ -f "$D/boot/grub/menu.list" ] && grubcfg="$D/boot/grub/menu.list"
 	[ -f "$D/boot/grub/grub.cfg" ] && grubcfg="$D/boot/grub/grub.cfg"
 	if [ -n "$grubcfg" ]; then
diff --git a/meta/classes/kernel-yocto.bbclass b/meta/classes/kernel-yocto.bbclass
index 6010dc9..8fcc7fe 100644
--- a/meta/classes/kernel-yocto.bbclass
+++ b/meta/classes/kernel-yocto.bbclass
@@ -44,7 +44,7 @@ def find_kernel_feature_dirs(d):

     return feature_dirs

-# find the master/machine source branch. In the same way that the fetcher proceses
+# find the master/machine source branch. In the same way that the fetcher processes
 # git repositories in the SRC_URI we take the first repo found, first branch.
 def get_machine_branch(d, default):
     fetch = bb.fetch2.Fetch([], d)
diff --git a/meta/classes/libc-package.bbclass b/meta/classes/libc-package.bbclass
index c1bc399..2d399b8 100644
--- a/meta/classes/libc-package.bbclass
+++ b/meta/classes/libc-package.bbclass
@@ -1,6 +1,6 @@
 #
-# This class knows how to package up [e]glibc. Its shared since prebuild binary toolchains
-# may need packaging and its pointless to duplicate this code.
+# This class knows how to package up [e]glibc. It's shared since prebuild binary
+# toolchains may need packaging and it's pointless to duplicate this code.
 #
 # Caller should set GLIBC_INTERNAL_USE_BINARY_LOCALE to one of:
 #  "compile" - Use QEMU to generate the binary locale files
diff --git a/meta/classes/license.bbclass b/meta/classes/license.bbclass
index 601f561..1042405 100644
--- a/meta/classes/license.bbclass
+++ b/meta/classes/license.bbclass
@@ -110,7 +110,7 @@ python do_populate_lic() {
     copy_license_files(lic_files_paths, destdir)
 }

-# it would be better to copy them in do_install_append, but find_license_filesa is python
+# it would be better to copy them in do_install_append, but find_license_files is python
 python perform_packagecopy_prepend () {
     enabled = oe.data.typed_value('LICENSE_CREATE_PACKAGE', d)
     if d.getVar('CLASSOVERRIDE', True) == 'class-target' and enabled:
diff --git a/meta/classes/module.bbclass b/meta/classes/module.bbclass
index ad6f7af..d8450ff 100644
--- a/meta/classes/module.bbclass
+++ b/meta/classes/module.bbclass
@@ -26,7 +26,7 @@ module_do_install() {

 EXPORT_FUNCTIONS do_compile do_install

-# add all splitted modules to PN RDEPENDS, PN can be empty now
+# add all split modules to PN RDEPENDS, PN can be empty now
 KERNEL_MODULES_META_PACKAGE = "${PN}"
 FILES_${PN} = ""
 ALLOW_EMPTY_${PN} = "1"
diff --git a/meta/classes/package.bbclass b/meta/classes/package.bbclass
index 6a552d9..77186f8 100644
--- a/meta/classes/package.bbclass
+++ b/meta/classes/package.bbclass
@@ -26,7 +26,7 @@
 #    a list of affected files in FILER{PROVIDES,DEPENDS}FLIST_pkg
 #
 # h) package_do_shlibs - Look at the shared libraries generated and autotmatically add any
-#    depenedencies found. Also stores the package name so anyone else using this library
+#    dependencies found. Also stores the package name so anyone else using this library
 #    knows which package to depend on.
 #
 # i) package_do_pkgconfig - Keep track of which packages need and provide which .pc files
@@ -288,7 +288,7 @@ def splitdebuginfo(file, debugfile, debugsrcdir, sourcefile, d):
     return 0

 def copydebugsources(debugsrcdir, d):
-    # The debug src information written out to sourcefile is further procecessed
+    # The debug src information written out to sourcefile is further processed
     # and copied to the destination here.

     import stat
@@ -814,7 +814,7 @@ python split_and_strip_files () {
                     continue
                 if not s:
                     continue
-                # Check its an excutable
+                # Check it's an excutable
                 if (s[stat.ST_MODE] & stat.S_IXUSR) or (s[stat.ST_MODE] & stat.S_IXGRP) or (s[stat.ST_MODE] & stat.S_IXOTH) \
                         or ((file.startswith(libdir) or file.startswith(baselibdir)) and ".so" in f):
                     # If it's a symlink, and points to an ELF file, we capture the readlink target
@@ -844,7 +844,7 @@ python split_and_strip_files () {
                         elffiles[file] = elf_file

     #
-    # First lets process debug splitting
+    # First, let's process debug splitting
     #
     if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1'):
         hardlinkmap = {}
@@ -1463,7 +1463,7 @@ python package_do_shlibs() {
             rpath = []
             p = sub.Popen([d.expand("${HOST_PREFIX}otool"), '-l', file],stdout=sub.PIPE,stderr=sub.PIPE)
             err, out = p.communicate()
-            # If returned succesfully, process stderr for results
+            # If returned successfully, process stderr for results
             if p.returncode == 0:
                 for l in err.split("\n"):
                     l = l.strip()
@@ -1472,7 +1472,7 @@ python package_do_shlibs() {

         p = sub.Popen([d.expand("${HOST_PREFIX}otool"), '-L', file],stdout=sub.PIPE,stderr=sub.PIPE)
         err, out = p.communicate()
-        # If returned succesfully, process stderr for results
+        # If returned successfully, process stderr for results
         if p.returncode == 0:
             for l in err.split("\n"):
                 l = l.strip()
@@ -1938,9 +1938,9 @@ python do_package () {
     # Optimisations
     ###########################################################################

-    # Contunually rexpanding complex expressions is inefficient, particularly when
-    # we write to the datastore and invalidate the expansion cache. This code
-    # pre-expands some frequently used variables
+    # Continually re-expanding complex expressions is inefficient, particularly
+    # when we write to the datastore and invalidate the expansion cache. This
+    # code pre-expands some frequently used variables.

     def expandVar(x, d):
         d.setVar(x, d.getVar(x, True))
diff --git a/meta/classes/package_rpm.bbclass b/meta/classes/package_rpm.bbclass
index 0a32b3e..b24b91a 100644
--- a/meta/classes/package_rpm.bbclass
+++ b/meta/classes/package_rpm.bbclass
@@ -7,7 +7,7 @@ RPMBUILD="rpmbuild"

 PKGWRITEDIRRPM = "${WORKDIR}/deploy-rpms"

-# Maintaining the perfile dependencies has singificant overhead when writing the
+# Maintaining the per-file dependencies has significant overhead when writing the
 # packages. When set, this value merges them for efficiency.
 MERGEPERFILEDEPS = "1"

@@ -171,7 +171,7 @@ python write_specfile () {
             depends = bb.utils.join_deps(newdeps_dict)
             d.setVar(varname, depends.strip())

-    # We need to change the style the dependency from BB to RPM
+    # We need to change the style of the dependency from BB to RPM.
     # This needs to happen AFTER the mapping_rename_hook
     def print_deps(variable, tag, array, d):
         depends = variable
@@ -635,7 +635,7 @@ python do_package_rpm () {
         return

     # Construct the spec file...
-    # If the spec file already exist, and has not been stored into
+    # If the spec file already exists, and has not been stored into
     # pseudo's files.db, it maybe cause rpmbuild src.rpm fail,
     # so remove it before doing rpmbuild src.rpm.
     srcname    = strip_multilib(d.getVar('PN', True), d)
diff --git a/meta/classes/rm_work.bbclass b/meta/classes/rm_work.bbclass
index f0f6d18..7d656de 100644
--- a/meta/classes/rm_work.bbclass
+++ b/meta/classes/rm_work.bbclass
@@ -1,7 +1,7 @@
 #
 # Removes source after build
 #
-# To use it add that line to conf/local.conf:
+# To use it add this line to conf/local.conf:
 #
 # INHERIT += "rm_work"
 #
@@ -31,7 +31,7 @@ do_rm_work () {
     for dir in *
     do
         # Retain only logs and other files in temp, safely ignore
-        # failures of removing pseudo folers on NFS2/3 server.
+        # failures of removing pseudo folders on NFS2/3 server.
         if [ $dir = 'pseudo' ]; then
             rm -rf $dir 2> /dev/null || true
         elif [ $dir != 'temp' ]; then
@@ -39,7 +39,7 @@ do_rm_work () {
         fi
     done

-    # Need to add pseudo back or subsqeuent work in this workdir
+    # Need to add pseudo back or subsequent work in this workdir
     # might fail since setscene may not rerun to recreate it
     mkdir -p ${WORKDIR}/pseudo/

diff --git a/meta/classes/spdx.bbclass b/meta/classes/spdx.bbclass
index 55ce3af..d6302eb 100644
--- a/meta/classes/spdx.bbclass
+++ b/meta/classes/spdx.bbclass
@@ -1,5 +1,5 @@
 # This class integrates real-time license scanning, generation of SPDX standard
-# output and verifiying license info during the building process.
+# output and verifying license info during the building process.
 # It is a combination of efforts from the OE-Core, SPDX and Fossology projects.
 #
 # For more information on FOSSology:

-- 

========================================================================
Robert P. J. Day                                 Ottawa, Ontario, CANADA
                        http://crashcourse.ca

Twitter:                                       http://twitter.com/rpjday
LinkedIn:                               http://ca.linkedin.com/in/rpjday
========================================================================




More information about the Openembedded-core mailing list