[OE-core] My thoughts on the future of OE?
Richard Purdie
richard.purdie at linuxfoundation.org
Thu May 1 17:02:41 UTC 2014
I was asked what I thought were things that needed discussion at OEDAM.
Sadly I won't be there but I thought it might help to write down my
thoughts in a few areas.
Developer Workflow
------------------
Firstly, I think the big piece we need to address as a project is
"developer workflow" as this is where people are struggling using it.
Unfortunately "developer workflow" means different things to different
people? Which one do I mean then? I actually mean all of them. As some
examples:
-----------------------------------------------------------------------
* A kernel developer wanting to rebuild a kernel
[on/off target, with the ADT/SDK or a recipe]
* A kernel developer wanting to build a kernel module
[on/off target, with the ADT/SDK or a recipe]
* An application developer wanting to build a single App
[on/off target, with the ADT/SDK or a recipe]
* An application developer wanting to (re)build a library, linking an
App to it
[on/off target, with the ADT/SDK or a recipe]
* A user wanting to rebuild an image with a package added
[on and off target - feeds or a build]
* A user wanting to rebuild an image with more advanced changes
The user may want to skip the image creation step and deploy data
straight onto an existing running target using rsync.
Their application/kernel may be in an external directory already under
SCM control.
There are requests for "real time" viewing of the build logs
There are requests for the python devshell to better be able to poke
around the datastore
There are requests for a shell like environment with commands
interacting with a memory resident bitbake.
In a team environment how should patch review work? Which tools would we
recommend to help people? Where does and autobuilder fit into this?
Gerrit? How to handle bugs/regressions/features?
Also, when something fails, how do they get help and fix it? My dream
here is an error reporting server and over time, we can attach "fixes"
to those failures showing people how to potentially fix their problems.
-----------------------------------------------------------------------
So my first ask is that we actually try and write down all these
different cases which is no small task in itself. I've a start of a list
above, we should probably put this into the wiki and have people add
their own use cases (or use cases of those around them in their company
etc.). The trouble is there are some many different variants!
Once we have some idea of the cases, we can start to put together some
kind of plan about how we intend to help the given use cases and to try
and prioritise them. Perhaps we should put some kind of weighting
against them in the wiki and people can increase the numbers of say
their top three desires.
Whilst this looks like an impossible problem, the good news is that I
believe we can solve it and we actually already have a rather powerful
toolbox to help work on it. I have a rough idea of a roadmap that can
move us forward:
a) locked sstate - I know its not in master yet but I believe this is
key to allowing some of the use cases where you want to change some
number of things but keep the rest of the system the same.
b) Rework the ADT so it consists of a build of prebuilt objects and
locked sstate and bitbake is hiding behind the scenes. There would be
some kind of setup command (which would extract the toolchain and core
libs for example) and then they could use it like the existing ADT.
c) We could then document workflows where you could build an extra app
into an image and so on.
d) In parallel with the above, we need to look at things like
externalsrc and see if there are ways we can better integrate it. Maybe
some helper scripts? Can it work with the new ADT/SDK?
Hopefully the above is enough to seed some discussion! :)
Automated XYZ
-------------
We've made great steps forward with the automation we have. I would like
to see us move forward in a few key areas:
a) Allow automated ptest running as part of our qemu image tests. This
is harder since it means:
i) installing ptest packages
ii) collecting up the results in some wya that can be parsed
iii) allow comparison of existing results with previous runs
iv) report regressions
b) Implement automated recipe upgrade detection. We have proof of
concept code for this and there are some logical steps for how to
proceed:
i) I'd like to see it integrated into the fetcher codebase. There
are some proof of concept patches at:
http://git.yoctoproject.org/cgit.cgi/poky-contrib/commit/?h=rpurdie/t222&id=46ba4664a8b1d2d59de5fba92e7e77929a0ee24d
(and the preceding patches)
ii) We will probably need to markup the recipes themselves with
some new data. Some example data:
http://git.yoctoproject.org/cgit.cgi/poky/tree/meta-yocto/conf/distro/include/package_regex.inc
iii) Feed the data from the above code into the package reporting system
The parts are there and have been proven to roughly work. The next
step would be to bring them into the core and start using them
properly. Its not an easy thing to get right but I think the
potential benefits are worth it.
c) Implement code which attempts automated package upgrades and then
reports the success/failure. This depends on a) and b) for best
results.
d) Continue to move forward with automated testing on real hardware
too. The idea is that the above tests in a) should also be usable
here so we can test significantly more of the software stack than
we've ever been able to before.
e) Continue to add test cases for:
i) the images (ptest and others)
ii) oe-selftest
iii) bitbake-selftest
iv) add new tests like tests for the toaster UI
The basic idea here is that existing work done manually isn't going to
scale. We have all the foundations to be able to do something few
projects have ever done before and automate this on a scale not
previously seen. This means it would stop people from having to do
"boring" work and concentrate on the interesting pieces where the
problems are. It also means we should be able to track down regressions
much earlier and deal with them more quickly.
How soon will all of this work? Probably not overnight but I believe its
something to aim for.
Contributors to the Project
---------------------------
This is an open question. How do we attract more developers to work on
the project? This applies both to layers like OE-Core and outside that,
we could all use some help in the maintainership, the development and
the testing. What blocks people from joining us? How can we encourage
more people to participate?
Unfortunately when things are working for people they tend to think that
ok, they'll go and do something else (like work on their product!) but
is there a way we could get some involvement from them? Lots of small
pieces of help could build to a large whole...
Conclusion
----------
I've probably rambled enough. The project is I believe in a strong
position but the above are areas I think we need to work on if we want
to truly fulfil the project's potential.
Cheers,
Richard
More information about the Openembedded-core
mailing list