Welcome to the Yocto Project Documentation
Yocto Project Quick Build
Welcome!
This short document steps you through the process for a typical image build using the Yocto Project. The document also introduces how to configure a build for specific hardware. You will use Yocto Project to build a reference embedded OS called Poky.
Note
The examples in this paper assume you are using a native Linux system running a recent Ubuntu Linux distribution. If the machine you want to use Yocto Project on to build an image (Build Host) is not a native Linux system, you can still perform these steps by using CROss PlatformS (CROPS) and setting up a Poky container. See the Setting Up to Use CROss PlatformS (CROPS) section in the Yocto Project Development Tasks Manual for more information.
You may use Windows Subsystem For Linux v2 to set up a build host using Windows 10.
Note
The Yocto Project is not compatible with WSLv1, it is compatible but not officially supported nor validated with WSLv2, if you still decide to use WSL please upgrade to WSLv2.
See the Setting Up to Use Windows Subsystem For Linux (WSLv2) section in the Yocto Project Development Tasks Manual for more information.
If you want more conceptual or background information on the Yocto Project, see the Yocto Project Overview and Concepts Manual.
Compatible Linux Distribution
Make sure your Build Host meets the following requirements:
50 Gbytes of free disk space
Runs a supported Linux distribution (i.e. recent releases of Fedora, openSUSE, CentOS, Debian, or Ubuntu). For a list of Linux distributions that support the Yocto Project, see the Supported Linux Distributions section in the Yocto Project Reference Manual. For detailed information on preparing your build host, see the Preparing the Build Host section in the Yocto Project Development Tasks Manual.
Git 1.8.3.1 or greater
tar 1.28 or greater
Python 3.6.0 or greater.
gcc 5.0 or greater.
If your build host does not meet any of these three listed version requirements, you can take steps to prepare the system so that you can still use the Yocto Project. See the Required Git, tar, Python and gcc Versions section in the Yocto Project Reference Manual for information.
Build Host Packages
You must install essential host packages on your build host. The following command installs the host packages based on an Ubuntu distribution:
$ sudo apt-get install gawk wget git diffstat unzip texinfo gcc build-essential chrpath socat cpio python3 python3-pip python3-pexpect xz-utils debianutils iputils-ping python3-git python3-jinja2 libegl1-mesa libsdl1.2-dev pylint3 xterm python3-subunit mesa-common-dev
Note
For host package requirements on all supported Linux distributions, see the Required Packages for the Build Host section in the Yocto Project Reference Manual.
Use Git to Clone Poky
Once you complete the setup instructions for your machine, you need to get a copy of the Poky repository on your build host. Use the following commands to clone the Poky repository.
$ git clone git://git.yoctoproject.org/poky
Cloning into 'poky'...
remote: Counting
objects: 432160, done. remote: Compressing objects: 100%
(102056/102056), done. remote: Total 432160 (delta 323116), reused
432037 (delta 323000) Receiving objects: 100% (432160/432160), 153.81 MiB | 8.54 MiB/s, done.
Resolving deltas: 100% (323116/323116), done.
Checking connectivity... done.
Go to Releases wiki page, and choose a release
codename (such as hardknott
), corresponding to either the
latest stable release or a Long Term Support release.
Then move to the poky
directory and take a look at existing branches:
$ cd poky
$ git branch -a
.
.
.
remotes/origin/HEAD -> origin/master
remotes/origin/dunfell
remotes/origin/dunfell-next
.
.
.
remotes/origin/gatesgarth
remotes/origin/gatesgarth-next
.
.
.
remotes/origin/master
remotes/origin/master-next
.
.
.
For this example, check out the hardknott
branch based on the
Hardknott
release:
$ git checkout -t origin/hardknott -b my-hardknott
Branch 'my-hardknott' set up to track remote branch 'hardknott' from 'origin'.
Switched to a new branch 'my-hardknott'
The previous Git checkout command creates a local branch named
my-hardknott
. The files available to you in that branch
exactly match the repository’s files in the hardknott
release branch.
Note that you can regularly type the following command in the same directory to keep your local files in sync with the release branch:
$ git pull
For more options and information about accessing Yocto Project related repositories, see the Locating Yocto Project Source Files section in the Yocto Project Development Tasks Manual.
Building Your Image
Use the following steps to build your image. The build process creates an entire Linux distribution, including the toolchain, from source.
Note
If you are working behind a firewall and your build host is not set up for proxies, you could encounter problems with the build process when fetching source code (e.g. fetcher failures or Git failures).
If you do not know your proxy settings, consult your local network infrastructure resources and get that information. A good starting point could also be to check your web browser settings. Finally, you can find more information on the “Working Behind a Network Proxy” page of the Yocto Project Wiki.
Initialize the Build Environment: From within the
poky
directory, run the oe-init-build-env environment setup script to define Yocto Project’s build environment on your build host.$ cd poky $ source oe-init-build-env You had no conf/local.conf file. This configuration file has therefore been created for you with some default values. You may wish to edit it to, for example, select a different MACHINE (target hardware). See conf/local.conf for more information as common configuration options are commented. You had no conf/bblayers.conf file. This configuration file has therefore been created for you with some default values. To add additional metadata layers into your configuration please add entries to conf/bblayers.conf. The Yocto Project has extensive documentation about OE including a reference manual which can be found at: https://docs.yoctoproject.org For more information about OpenEmbedded see their website: http://www.openembedded.org/ ### Shell environment set up for builds. ### You can now run 'bitbake <target>' Common targets are: core-image-minimal core-image-sato meta-toolchain meta-ide-support You can also run generated QEMU images with a command like 'runqemu qemux86-64'
Among other things, the script creates the Build Directory, which is
build
in this case and is located in the Source Directory. After the script runs, your current working directory is set to the Build Directory. Later, when the build completes, the Build Directory contains all the files created during the build.Examine Your Local Configuration File: When you set up the build environment, a local configuration file named
local.conf
becomes available in aconf
subdirectory of the Build Directory. For this example, the defaults are set to build for aqemux86
target, which is suitable for emulation. The package manager used is set to the RPM package manager.Tip
You can significantly speed up your build and guard against fetcher failures by using mirrors. To use mirrors, add these lines to your local.conf file in the Build directory:
SSTATE_MIRRORS = "\ file://.* http://sstate.yoctoproject.org/dev/PATH;downloadfilename=PATH \n \ file://.* http://sstate.yoctoproject.org/3.2.4/PATH;downloadfilename=PATH \n \ file://.* http://sstate.yoctoproject.org/3.3.1/PATH;downloadfilename=PATH \n \ "
The previous examples showed how to add sstate paths for Yocto Project 3.2.4, 3.3.1, and a development area. For a complete index of sstate locations, see http://sstate.yoctoproject.org/.
Start the Build: Continue with the following command to build an OS image for the target, which is
core-image-sato
in this example:$ bitbake core-image-sato
For information on using the
bitbake
command, see the BitBake section in the Yocto Project Overview and Concepts Manual, or see the “BitBake Command” section in the BitBake User Manual.Simulate Your Image Using QEMU: Once this particular image is built, you can start QEMU, which is a Quick EMUlator that ships with the Yocto Project:
$ runqemu qemux86-64
If you want to learn more about running QEMU, see the Using the Quick EMUlator (QEMU) chapter in the Yocto Project Development Tasks Manual.
Exit QEMU: Exit QEMU by either clicking on the shutdown icon or by typing
Ctrl-C
in the QEMU transcript window from which you evoked QEMU.
Customizing Your Build for Specific Hardware
So far, all you have done is quickly built an image suitable for emulation only. This section shows you how to customize your build for specific hardware by adding a hardware layer into the Yocto Project development environment.
In general, layers are repositories that contain related sets of instructions and configurations that tell the Yocto Project what to do. Isolating related metadata into functionally specific layers facilitates modular development and makes it easier to reuse the layer metadata.
Note
By convention, layer names start with the string “meta-“.
Follow these steps to add a hardware layer:
Find a Layer: Lots of hardware layers exist. The Yocto Project Source Repositories has many hardware layers. This example adds the meta-altera hardware layer.
Clone the Layer: Use Git to make a local copy of the layer on your machine. You can put the copy in the top level of the copy of the Poky repository created earlier:
$ cd poky $ git clone https://github.com/kraj/meta-altera.git Cloning into 'meta-altera'... remote: Counting objects: 25170, done. remote: Compressing objects: 100% (350/350), done. remote: Total 25170 (delta 645), reused 719 (delta 538), pack-reused 24219 Receiving objects: 100% (25170/25170), 41.02 MiB | 1.64 MiB/s, done. Resolving deltas: 100% (13385/13385), done. Checking connectivity... done.
The hardware layer now exists with other layers inside the Poky reference repository on your build host as
meta-altera
and contains all the metadata needed to support hardware from Altera, which is owned by Intel.Note
It is recommended for layers to have a branch per Yocto Project release. Please make sure to checkout the layer branch supporting the Yocto Project release you’re using.
Change the Configuration to Build for a Specific Machine: The MACHINE variable in the
local.conf
file specifies the machine for the build. For this example, set theMACHINE
variable tocyclone5
. These configurations are used: https://github.com/kraj/meta-altera/blob/master/conf/machine/cyclone5.conf.Note
See the “Examine Your Local Configuration File” step earlier for more information on configuring the build.
Add Your Layer to the Layer Configuration File: Before you can use a layer during a build, you must add it to your
bblayers.conf
file, which is found in the Build Directoryconf
directory.Use the
bitbake-layers add-layer
command to add the layer to the configuration file:$ cd poky/build $ bitbake-layers add-layer ../meta-altera NOTE: Starting bitbake server... Parsing recipes: 100% |##################################################################| Time: 0:00:32 Parsing of 918 .bb files complete (0 cached, 918 parsed). 1401 targets, 123 skipped, 0 masked, 0 errors.
You can find more information on adding layers in the Adding a Layer Using the bitbake-layers Script section.
Completing these steps has added the meta-altera
layer to your Yocto
Project development environment and configured it to build for the
cyclone5
machine.
Note
The previous steps are for demonstration purposes only. If you were
to attempt to build an image for the cyclone5
machine, you should
read the Altera README
.
Creating Your Own General Layer
Maybe you have an application or specific set of behaviors you need to
isolate. You can create your own general layer using the
bitbake-layers create-layer
command. The tool automates layer
creation by setting up a subdirectory with a layer.conf
configuration file, a recipes-example
subdirectory that contains an
example.bb
recipe, a licensing file, and a README
.
The following commands run the tool to create a layer named
meta-mylayer
in the poky
directory:
$ cd poky
$ bitbake-layers create-layer meta-mylayer
NOTE: Starting bitbake server...
Add your new layer with 'bitbake-layers add-layer meta-mylayer'
For more information on layers and how to create them, see the Creating a General Layer Using the bitbake-layers Script section in the Yocto Project Development Tasks Manual.
Where To Go Next
Now that you have experienced using the Yocto Project, you might be asking yourself “What now?”. The Yocto Project has many sources of information including the website, wiki pages, and user manuals:
Website: The Yocto Project Website provides background information, the latest builds, breaking news, full development documentation, and access to a rich Yocto Project Development Community into which you can tap.
Developer Screencast: The Getting Started with the Yocto Project - New Developer Screencast Tutorial provides a 30-minute video created for users unfamiliar with the Yocto Project but familiar with Linux build hosts. While this screencast is somewhat dated, the introductory and fundamental concepts are useful for the beginner.
Yocto Project Overview and Concepts Manual: The Yocto Project Overview and Concepts Manual is a great place to start to learn about the Yocto Project. This manual introduces you to the Yocto Project and its development environment. The manual also provides conceptual information for various aspects of the Yocto Project.
Yocto Project Wiki: The Yocto Project Wiki provides additional information on where to go next when ramping up with the Yocto Project, release information, project planning, and QA information.
Yocto Project Mailing Lists: Related mailing lists provide a forum for discussion, patch submission and announcements. Several mailing lists exist and are grouped according to areas of concern. See the Mailing lists section in the Yocto Project Reference Manual for a complete list of Yocto Project mailing lists.
Comprehensive List of Links and Other Documentation: The Links and Related Documentation section in the Yocto Project Reference Manual provides a comprehensive list of all related links and other user documentation.
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
What I wish I’d known about Yocto Project
Note
Before reading further, make sure you’ve taken a look at the Software Overview page which presents the definitions for many of the terms referenced here. Also, know that some of the information here won’t make sense now, but as you start developing, it is the information you’ll want to keep close at hand. These are best known methods for working with Yocto Project and they are updated regularly.
Using the Yocto Project is fairly easy, until something goes wrong. Without an understanding of how the build process works, you’ll find yourself trying to troubleshoot “a black box”. Here are a few items that new users wished they had known before embarking on their first build with Yocto Project. Feel free to contact us with other suggestions.
Use Git, not the tarball download: If you use git the software will be automatically updated with bug updates because of how git works. If you download the tarball instead, you will need to be responsible for your own updates.
Get to know the layer index: All layers can be found in the layer index. Layers which have applied for Yocto Project Compatible status (structure continuity assurance and testing) can be found in the Yocto Project Compatible index. Generally check the Compatible layer index first, and if you don’t find the necessary layer check the general layer index. The layer index is an original artifact from the Open Embedded Project. As such, that index doesn’t have the curating and testing that the Yocto Project provides on Yocto Project Compatible layer list, but the latter has fewer entries. Know that when you start searching in the layer index that not all layers have the same level of maturity, validation, or usability. Nor do searches prioritize displayed results. There is no easy way to help you through the process of choosing the best layer to suit your needs. Consequently, it is often trial and error, checking the mailing lists, or working with other developers through collaboration rooms that can help you make good choices.
Use existing BSP layers from silicon vendors when possible: Intel, TI, NXP and others have information on what BSP layers to use with their silicon. These layers have names such as “meta-intel” or “meta-ti”. Try not to build layers from scratch. If you do have custom silicon, use one of these layers as a guide or template and familiarize yourself with the Yocto Project Board Support Package Developer’s Guide.
Do not put everything into one layer: Use different layers to logically separate information in your build. As an example, you could have a BSP layer, a GUI layer, a distro configuration, middleware, or an application (e.g. “meta-filesystems”, “meta-python”, “meta-intel”, and so forth). Putting your entire build into one layer limits and complicates future customization and reuse. Isolating information into layers, on the other hand, helps keep simplify future customizations and reuse.
Never modify the POKY layer. Never. Ever. When you update to the next release, you’ll lose all of your work. ALL OF IT.
Don’t be fooled by documentation searching results: Yocto Project documentation is always being updated. Unfortunately, when you use Google to search for Yocto Project concepts or terms, Google consistently searches and retrieves older versions of Yocto Project manuals. For example, searching for a particular topic using Google could result in a “hit” on a Yocto Project manual that is several releases old. To be sure that you are using the most current Yocto Project documentation, use the drop-down menu at the top of any of its page.
Many developers look through the All-in-one ‘Mega’ Manual for a concept or term by doing a search through the whole page. This manual is a concatenation of the core set of Yocto Project manual. Thus, a simple string search using Ctrl-F in this manual produces all the “hits” for a desired term or concept. Once you find the area in which you are interested, you can display the actual manual, if desired. It is also possible to use the search bar in the menu or in the left navigation pane.
Understand the basic concepts of how the build system works: the workflow: Understanding the Yocto Project workflow is important as it can help you both pinpoint where trouble is occurring and how the build is breaking. The workflow breaks down into the following steps:
Fetch – get the source code
Extract – unpack the sources
Patch – apply patches for bug fixes and new capability
Configure – set up your environment specifications
Build – compile and link
Install – copy files to target directories
Package – bundle files for installation
During “fetch”, there may be an inability to find code. During “extract”, there is likely an invalid zip or something similar. In other words, the function of a particular part of the workflow gives you an idea of what might be going wrong.
Know that you can generate a dependency graph and learn how to do it: A dependency graph shows dependencies between recipes, tasks, and targets. You can use the “-g” option with BitBake to generate this graph. When you start a build and the build breaks, you could see packages you have no clue about or have any idea why the build system has included them. The dependency graph can clarify that confusion. You can learn more about dependency graphs and how to generate them in the Generating Dependency Graphs section in the BitBake User Manual.
Here’s how you decode “magic” folder names in tmp/work: The build system fetches, unpacks, preprocesses, and builds. If something goes wrong, the build system reports to you directly the path to a folder where the temporary (build/tmp) files and packages reside resulting from the build. For a detailed example of this process, see the example. Unfortunately this example is on an earlier release of Yocto Project.
When you perform a build, you can use the “-u” BitBake command-line option to specify a user interface viewer into the dependency graph (e.g. knotty, ncurses, or taskexp) that helps you understand the build dependencies better.
You can build more than just images: You can build and run a specific task for a specific package (including devshell) or even a single recipe. When developers first start using the Yocto Project, the instructions found in the Yocto Project Quick Build show how to create an image and then run or flash that image. However, you can actually build just a single recipe. Thus, if some dependency or recipe isn’t working, you can just say “bitbake foo” where “foo” is the name for a specific recipe. As you become more advanced using the Yocto Project, and if builds are failing, it can be useful to make sure the fetch itself works as desired. Here are some valuable links: Using a Development Shell for information on how to build and run a specific task using devshell. Also, the SDK manual shows how to build out a specific recipe.
An ambiguous definition: Package vs Recipe: A recipe contains instructions the build system uses to create packages. Recipes and Packages are the difference between the front end and the result of the build process.
As mentioned, the build system takes the recipe and creates packages from the recipe’s instructions. The resulting packages are related to the one thing the recipe is building but are different parts (packages) of the build (i.e. the main package, the doc package, the debug symbols package, the separate utilities package, and so forth). The build system splits out the packages so that you don’t need to install the packages you don’t want or need, which is advantageous because you are building for small devices when developing for embedded and IoT.
You will want to learn about and know what’s packaged in rootfs.
Create your own image recipe: There are a number of ways to create your own image recipe. We suggest you create your own image recipe as opposed to appending an existing recipe. It is trivial and easy to write an image recipe. Again, do not try appending to an existing image recipe. Create your own and do it right from the start.
Finally, here is a list of the basic skills you will need as a systems developer. You must be able to:
deal with corporate proxies
add a package to an image
understand the difference between a recipe and package
build a package by itself and why that’s useful
find out what packages are created by a recipe
find out what files are in a package
find out what files are in an image
add an ssh server to an image (enable transferring of files to target)
know the anatomy of a recipe
know how to create and use layers
find recipes (with the OpenEmbedded Layer index)
understand difference between machine and distro settings
find and use the right BSP (machine) for your hardware
find examples of distro features and know where to set them
understanding the task pipeline and executing individual tasks
understand devtool and how it simplifies your workflow
improve build speeds with shared downloads and shared state cache
generate and understand a dependency graph
generate and understand bitbake environment
build an Extensible SDK for applications development
Depending on what you primary interests are with the Yocto Project, you could consider any of the following reading:
Look Through the Yocto Project Development Tasks Manual: This manual contains procedural information grouped to help you get set up, work with layers, customize images, write new recipes, work with libraries, and use QEMU. The information is task-based and spans the breadth of the Yocto Project. See the Yocto Project Development Tasks Manual.
Look Through the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual: This manual describes how to use both the standard SDK and the extensible SDK, which are used primarily for application development. The Using the Extensible SDK also provides example workflows that use devtool. See the section Using devtool in Your SDK Workflow for more information.
Learn About Kernel Development: If you want to see how to work with the kernel and understand Yocto Linux kernels, see the Yocto Project Linux Kernel Development Manual. This manual provides information on how to patch the kernel, modify kernel recipes, and configure the kernel.
Learn About Board Support Packages (BSPs): If you want to learn about BSPs, see the Yocto Project Board Support Package Developer’s Guide. This manual also provides an example BSP creation workflow. See the Board Support Packages (BSP) - Developer’s Guide section.
Learn About Toaster: Toaster is a web interface to the Yocto Project’s OpenEmbedded build system. If you are interested in using this type of interface to create images, see the Toaster User Manual.
Have Available the Yocto Project Reference Manual: Unlike the rest of the Yocto Project manual set, this manual is comprised of material suited for reference rather than procedures. You can get build details, a closer look at how the pieces of the Yocto Project development environment work together, information on various technical details, guidance on migrating to a newer Yocto Project release, reference material on the directory structure, classes, and tasks. The Yocto Project Reference Manual also contains a fairly comprehensive glossary of variables used within the Yocto Project.
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Transitioning to a custom environment for systems development
Note
So you’ve finished the Yocto Project Quick Build and glanced over the document What I wish I’d known about Yocto Project, the latter contains important information learned from other users. You’re well prepared. But now, as you are starting your own project, it isn’t exactly straightforward what to do. And, the documentation is daunting. We’ve put together a few hints to get you started.
Make a list of the processor, target board, technologies, and capabilities that will be part of your project. You will be finding layers with recipes and other metadata that support these things, and adding them to your configuration. (See #3)
Set up your board support. Even if you’re using custom hardware, it might be easier to start with an existing target board that uses the same processor or at least the same architecture as your custom hardware. Knowing the board already has a functioning Board Support Package (BSP) within the project makes it easier for you to get comfortable with project concepts.
Find and acquire the best BSP for your target. Use the Yocto Project curated layer index or even the OpenEmbedded layer index to find and acquire the best BSP for your target board. The Yocto Project layer index BSPs are regularly validated. The best place to get your first BSP is from your silicon manufacturer or board vendor – they can point you to their most qualified efforts. In general, for Intel silicon use meta-intel, for Texas Instruments use meta-ti, and so forth. Choose a BSP that has been tested with the same Yocto Project release that you’ve downloaded. Be aware that some BSPs may not be immediately supported on the very latest release, but they will be eventually.
You might want to start with the build specification that Poky provides (which is reference embedded distribution) and then add your newly chosen layers to that. Here is the information about adding layers.
Based on the layers you’ve chosen, make needed changes in your configuration. For instance, you’ve chosen a machine type and added in the corresponding BSP layer. You’ll then need to change the value of the
MACHINE
variable in your configuration file (build/local.conf) to point to that same machine type. There could be other layer-specific settings you need to change as well. Each layer has aREADME
document that you can look at for this type of usage information.Add a new layer for any custom recipes and metadata you create. Use the
bitbake-layers create-layer
tool for Yocto Project 2.4+ releases. If you are using a Yocto Project release earlier than 2.4, use theyocto-layer create
tool. Thebitbake-layers
tool also provides a number of other useful layer-related commands. See Creating a General Layer Using the bitbake-layers Script section.Create your own layer for the BSP you’re going to use. It is not common that you would need to create an entire BSP from scratch unless you have a really special device. Even if you are using an existing BSP, create your own layer for the BSP. For example, given a 64-bit x86-based machine, copy the conf/intel-corei7-64 definition and give the machine a relevant name (think board name, not product name). Make sure the layer configuration is dependent on the meta-intel layer (or at least, meta-intel remains in your bblayers.conf). Now you can put your custom BSP settings into your layer and you can re-use it for different applications.
Write your own recipe to build additional software support that isn’t already available in the form of a recipe. Creating your own recipe is especially important for custom application software that you want to run on your device. Writing new recipes is a process of refinement. Start by getting each step of the build process working beginning with fetching all the way through packaging. Next, run the software on your target and refine further as needed. See Writing a New Recipe in the Yocto Project Development Tasks Manual for more information.
Now you’re ready to create an image recipe. There are a number of ways to do this. However, it is strongly recommended that you have your own image recipe - don’t try appending to existing image recipes. Recipes for images are trivial to create and you usually want to fully customize their contents.
Build your image and refine it. Add what’s missing and fix anything that’s broken using your knowledge of the workflow to identify where issues might be occurring.
Consider creating your own distribution. When you get to a certain level of customization, consider creating your own distribution rather than using the default reference distribution.
Distribution settings define the packaging back-end (e.g. rpm or other) as well as the package feed and possibly the update solution. You would create your own distribution in a new layer inheriting from Poky but overriding what needs to change for your distribution. If you find yourself adding a lot of configuration to your local.conf file aside from paths and other typical local settings, it’s time to consider creating your own distribution.
You can add product specifications that can customize the distribution if needed in other layers. You can also add other functionality specific to the product. But to update the distribution, not individual products, you update the distribution feature through that layer.
Congratulations! You’re well on your way. Welcome to the Yocto Project community.
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Yocto Project Overview and Concepts Manual
1 The Yocto Project Overview and Concepts Manual
1.1 Welcome
Welcome to the Yocto Project Overview and Concepts Manual! This manual introduces the Yocto Project by providing concepts, software overviews, best-known-methods (BKMs), and any other high-level introductory information suitable for a new Yocto Project user.
The following list describes what you can get from this manual:
Introducing the Yocto Project: This chapter provides an introduction to the Yocto Project. You will learn about features and challenges of the Yocto Project, the layer model, components and tools, development methods, the Poky reference distribution, the OpenEmbedded build system workflow, and some basic Yocto terms.
The Yocto Project Development Environment: This chapter helps you get started understanding the Yocto Project development environment. You will learn about open source, development hosts, Yocto Project source repositories, workflows using Git and the Yocto Project, a Git primer, and information about licensing.
Yocto Project Concepts : This chapter presents various concepts regarding the Yocto Project. You can find conceptual information about components, development, cross-toolchains, and so forth.
This manual does not give you the following:
Step-by-step Instructions for Development Tasks: Instructional procedures reside in other manuals within the Yocto Project documentation set. For example, the Yocto Project Development Tasks Manual provides examples on how to perform various development tasks. As another example, the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual contains detailed instructions on how to install an SDK, which is used to develop applications for target hardware.
Reference Material: This type of material resides in an appropriate reference manual. For example, system variables are documented in the Yocto Project Reference Manual. As another example, the Yocto Project Board Support Package Developer’s Guide contains reference information on BSPs.
Detailed Public Information Not Specific to the Yocto Project: For example, exhaustive information on how to use the Source Control Manager Git is better covered with Internet searches and official Git Documentation than through the Yocto Project documentation.
1.2 Other Information
Because this manual presents information for many different topics, supplemental information is recommended for full comprehension. For additional introductory information on the Yocto Project, see the Yocto Project Website. If you want to build an image with no knowledge of Yocto Project as a way of quickly testing it out, see the Yocto Project Quick Build document. For a comprehensive list of links and other documentation, see the “Links and Related Documentation” section in the Yocto Project Reference Manual.
2 Introducing the Yocto Project
2.1 What is the Yocto Project?
The Yocto Project is an open source collaboration project that helps developers create custom Linux-based systems that are designed for embedded products regardless of the product’s hardware architecture. Yocto Project provides a flexible toolset and a development environment that allows embedded device developers across the world to collaborate through shared technologies, software stacks, configurations, and best practices used to create these tailored Linux images.
Thousands of developers worldwide have discovered that Yocto Project provides advantages in both systems and applications development, archival and management benefits, and customizations used for speed, footprint, and memory utilization. The project is a standard when it comes to delivering embedded software stacks. The project allows software customizations and build interchange for multiple hardware platforms as well as software stacks that can be maintained and scaled.
For further introductory information on the Yocto Project, you might be interested in this article by Drew Moseley and in this short introductory video.
The remainder of this section overviews advantages and challenges tied to the Yocto Project.
2.1.1 Features
The following list describes features and advantages of the Yocto Project:
Widely Adopted Across the Industry: Many semiconductor, operating system, software, and service vendors adopt and support the Yocto Project in their products and services. For a look at the Yocto Project community and the companies involved with the Yocto Project, see the “COMMUNITY” and “ECOSYSTEM” tabs on the Yocto Project home page.
Architecture Agnostic: Yocto Project supports Intel, ARM, MIPS, AMD, PPC and other architectures. Most ODMs, OSVs, and chip vendors create and supply BSPs that support their hardware. If you have custom silicon, you can create a BSP that supports that architecture.
Aside from broad architecture support, the Yocto Project fully supports a wide range of devices emulated by the Quick EMUlator (QEMU).
Images and Code Transfer Easily: Yocto Project output can easily move between architectures without moving to new development environments. Additionally, if you have used the Yocto Project to create an image or application and you find yourself not able to support it, commercial Linux vendors such as Wind River, Mentor Graphics, Timesys, and ENEA could take it and provide ongoing support. These vendors have offerings that are built using the Yocto Project.
Flexibility: Corporations use the Yocto Project many different ways. One example is to create an internal Linux distribution as a code base the corporation can use across multiple product groups. Through customization and layering, a project group can leverage the base Linux distribution to create a distribution that works for their product needs.
Ideal for Constrained Embedded and IoT devices: Unlike a full Linux distribution, you can use the Yocto Project to create exactly what you need for embedded devices. You only add the feature support or packages that you absolutely need for the device. For devices that have display hardware, you can use available system components such as X11, Wayland, GTK+, Qt, Clutter, and SDL (among others) to create a rich user experience. For devices that do not have a display or where you want to use alternative UI frameworks, you can choose to not build these components.
Comprehensive Toolchain Capabilities: Toolchains for supported architectures satisfy most use cases. However, if your hardware supports features that are not part of a standard toolchain, you can easily customize that toolchain through specification of platform-specific tuning parameters. And, should you need to use a third-party toolchain, mechanisms built into the Yocto Project allow for that.
Mechanism Rules Over Policy: Focusing on mechanism rather than policy ensures that you are free to set policies based on the needs of your design instead of adopting decisions enforced by some system software provider.
Uses a Layer Model: The Yocto Project layer infrastructure groups related functionality into separate bundles. You can incrementally add these grouped functionalities to your project as needed. Using layers to isolate and group functionality reduces project complexity and redundancy, allows you to easily extend the system, make customizations, and keep functionality organized.
Supports Partial Builds: You can build and rebuild individual packages as needed. Yocto Project accomplishes this through its Shared State Cache (sstate) scheme. Being able to build and debug components individually eases project development.
Releases According to a Strict Schedule: Major releases occur on a six-month cycle predictably in October and April. The most recent two releases support point releases to address common vulnerabilities and exposures. This predictability is crucial for projects based on the Yocto Project and allows development teams to plan activities.
Rich Ecosystem of Individuals and Organizations: For open source projects, the value of community is very important. Support forums, expertise, and active developers who continue to push the Yocto Project forward are readily available.
Binary Reproducibility: The Yocto Project allows you to be very specific about dependencies and achieves very high percentages of binary reproducibility (e.g. 99.8% for
core-image-minimal
). When distributions are not specific about which packages are pulled in and in what order to support dependencies, other build systems can arbitrarily include packages.License Manifest: The Yocto Project provides a license manifest for review by people who need to track the use of open source licenses (e.g. legal teams).
2.1.2 Challenges
The following list presents challenges you might encounter when developing using the Yocto Project:
Steep Learning Curve: The Yocto Project has a steep learning curve and has many different ways to accomplish similar tasks. It can be difficult to choose how to proceed when varying methods exist by which to accomplish a given task.
Understanding What Changes You Need to Make For Your Design Requires Some Research: Beyond the simple tutorial stage, understanding what changes need to be made for your particular design can require a significant amount of research and investigation. For information that helps you transition from trying out the Yocto Project to using it for your project, see the “What I wish I’d known about Yocto Project” and “Transitioning to a custom environment for systems development” documents on the Yocto Project website.
Project Workflow Could Be Confusing: The Yocto Project workflow <overview-manual/development-environment:the yocto project development environment> could be confusing if you are used to traditional desktop and server software development. In a desktop development environment, mechanisms exist to easily pull and install new packages, which are typically pre-compiled binaries from servers accessible over the Internet. Using the Yocto Project, you must modify your configuration and rebuild to add additional packages.
Working in a Cross-Build Environment Can Feel Unfamiliar: When developing code to run on a target, compilation, execution, and testing done on the actual target can be faster than running a BitBake build on a development host and then deploying binaries to the target for test. While the Yocto Project does support development tools on the target, the additional step of integrating your changes back into the Yocto Project build environment would be required. Yocto Project supports an intermediate approach that involves making changes on the development system within the BitBake environment and then deploying only the updated packages to the target.
The Yocto Project OpenEmbedded Build System produces packages in standard formats (i.e. RPM, DEB, IPK, and TAR). You can deploy these packages into the running system on the target by using utilities on the target such as
rpm
oripk
.Initial Build Times Can be Significant: Long initial build times are unfortunately unavoidable due to the large number of packages initially built from scratch for a fully functioning Linux system. Once that initial build is completed, however, the shared-state (sstate) cache mechanism Yocto Project uses keeps the system from rebuilding packages that have not been “touched” since the last build. The sstate mechanism significantly reduces times for successive builds.
2.2 The Yocto Project Layer Model
The Yocto Project’s “Layer Model” is a development model for embedded and IoT Linux creation that distinguishes the Yocto Project from other simple build systems. The Layer Model simultaneously supports collaboration and customization. Layers are repositories that contain related sets of instructions that tell the OpenEmbedded Build System what to do. You can collaborate, share, and reuse layers.
Layers can contain changes to previous instructions or settings at any time. This powerful override capability is what allows you to customize previously supplied collaborative or community layers to suit your product requirements.
You use different layers to logically separate information in your build. As an example, you could have BSP, GUI, distro configuration, middleware, or application layers. Putting your entire build into one layer limits and complicates future customization and reuse. Isolating information into layers, on the other hand, helps simplify future customizations and reuse. You might find it tempting to keep everything in one layer when working on a single project. However, the more modular your Metadata, the easier it is to cope with future changes.
Note
Use Board Support Package (BSP) layers from silicon vendors when possible.
Familiarize yourself with the Yocto Project curated layer index or the OpenEmbedded layer index. The latter contains more layers but they are less universally validated.
Layers support the inclusion of technologies, hardware components, and software components. The Yocto Project Compatible designation provides a minimum level of standardization that contributes to a strong ecosystem. “YP Compatible” is applied to appropriate products and software components such as BSPs, other OE-compatible layers, and related open-source projects, allowing the producer to use Yocto Project badges and branding assets.
To illustrate how layers are used to keep things modular, consider
machine customizations. These types of customizations typically reside
in a special layer, rather than a general layer, called a BSP Layer.
Furthermore, the machine customizations should be isolated from recipes
and Metadata that support a new GUI environment, for example. This
situation gives you a couple of layers: one for the machine
configurations, and one for the GUI environment. It is important to
understand, however, that the BSP layer can still make machine-specific
additions to recipes within the GUI environment layer without polluting
the GUI layer itself with those machine-specific changes. You can
accomplish this through a recipe that is a BitBake append
(.bbappend
) file, which is described later in this section.
Note
For general information on BSP layer structure, see the Yocto Project Board Support Package Developer’s Guide .
The Source Directory
contains both general layers and BSP layers right out of the box. You
can easily identify layers that ship with a Yocto Project release in the
Source Directory by their names. Layers typically have names that begin
with the string meta-
.
Note
It is not a requirement that a layer name begin with the prefix
meta-
, but it is a commonly accepted standard in the Yocto Project
community.
For example, if you were to examine the tree view
of the poky
repository, you will see several layers: meta
,
meta-skeleton
, meta-selftest
, meta-poky
, and
meta-yocto-bsp
. Each of these repositories represents a distinct
layer.
For procedures on how to create layers, see the “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual.
2.3 Components and Tools
The Yocto Project employs a collection of components and tools used by the project itself, by project developers, and by those using the Yocto Project. These components and tools are open source projects and metadata that are separate from the reference distribution (Poky) and the OpenEmbedded Build System. Most of the components and tools are downloaded separately.
This section provides brief overviews of the components and tools associated with the Yocto Project.
2.3.1 Development Tools
The following list consists of tools that help you develop images and applications using the Yocto Project:
CROPS: CROPS is an open source, cross-platform development framework that leverages Docker Containers. CROPS provides an easily managed, extensible environment that allows you to build binaries for a variety of architectures on Windows, Linux and Mac OS X hosts.
devtool: This command-line tool is available as part of the extensible SDK (eSDK) and is its cornerstone. You can use
devtool
to help build, test, and package software within the eSDK. You can use the tool to optionally integrate what you build into an image built by the OpenEmbedded build system.The
devtool
command employs a number of sub-commands that allow you to add, modify, and upgrade recipes. As with the OpenEmbedded build system, “recipes” represent software packages withindevtool
. When you usedevtool add
, a recipe is automatically created. When you usedevtool modify
, the specified existing recipe is used in order to determine where to get the source code and how to patch it. In both cases, an environment is set up so that when you build the recipe a source tree that is under your control is used in order to allow you to make changes to the source as desired. By default, both new recipes and the source go into a “workspace” directory under the eSDK. Thedevtool upgrade
command updates an existing recipe so that you can build it for an updated set of source files.You can read about the
devtool
workflow in the Yocto Project Application Development and Extensible Software Development Kit (eSDK) Manual in the “Using devtool in Your SDK Workflow” section.Extensible Software Development Kit (eSDK): The eSDK provides a cross-development toolchain and libraries tailored to the contents of a specific image. The eSDK makes it easy to add new applications and libraries to an image, modify the source for an existing component, test changes on the target hardware, and integrate into the rest of the OpenEmbedded build system. The eSDK gives you a toolchain experience supplemented with the powerful set of
devtool
commands tailored for the Yocto Project environment.For information on the eSDK, see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) Manual.
Toaster: Toaster is a web interface to the Yocto Project OpenEmbedded build system. Toaster allows you to configure, run, and view information about builds. For information on Toaster, see the Toaster User Manual.
2.3.2 Production Tools
The following list consists of tools that help production related activities using the Yocto Project:
Auto Upgrade Helper: This utility when used in conjunction with the OpenEmbedded Build System (BitBake and OE-Core) automatically generates upgrades for recipes that are based on new versions of the recipes published upstream. See Using the Auto Upgrade Helper (AUH) for how to set it up.
Recipe Reporting System: The Recipe Reporting System tracks recipe versions available for Yocto Project. The main purpose of the system is to help you manage the recipes you maintain and to offer a dynamic overview of the project. The Recipe Reporting System is built on top of the OpenEmbedded Layer Index, which is a website that indexes OpenEmbedded-Core layers.
Patchwork: Patchwork is a fork of a project originally started by OzLabs. The project is a web-based tracking system designed to streamline the process of bringing contributions into a project. The Yocto Project uses Patchwork as an organizational tool to handle patches, which number in the thousands for every release.
AutoBuilder: AutoBuilder is a project that automates build tests and quality assurance (QA). By using the public AutoBuilder, anyone can determine the status of the current “master” branch of Poky.
Note
AutoBuilder is based on buildbot.
A goal of the Yocto Project is to lead the open source industry with a project that automates testing and QA procedures. In doing so, the project encourages a development community that publishes QA and test plans, publicly demonstrates QA and test plans, and encourages development of tools that automate and test and QA procedures for the benefit of the development community.
You can learn more about the AutoBuilder used by the Yocto Project Autobuilder here.
Cross-Prelink: Prelinking is the process of pre-computing the load addresses and link tables generated by the dynamic linker as compared to doing this at runtime. Doing this ahead of time results in performance improvements when the application is launched and reduced memory usage for libraries shared by many applications.
Historically, cross-prelink is a variant of prelink, which was conceived by Jakub Jelínek a number of years ago. Both prelink and cross-prelink are maintained in the same repository albeit on separate branches. By providing an emulated runtime dynamic linker (i.e.
glibc
-derivedld.so
emulation), the cross-prelink project extends the prelink software’s ability to prelink a sysroot environment. Additionally, the cross-prelink software enables the ability to work in sysroot style environments.The dynamic linker determines standard load address calculations based on a variety of factors such as mapping addresses, library usage, and library function conflicts. The prelink tool uses this information, from the dynamic linker, to determine unique load addresses for executable and linkable format (ELF) binaries that are shared libraries and dynamically linked. The prelink tool modifies these ELF binaries with the pre-computed information. The result is faster loading and often lower memory consumption because more of the library code can be re-used from shared Copy-On-Write (COW) pages.
The original upstream prelink project only supports running prelink on the end target device due to the reliance on the target device’s dynamic linker. This restriction causes issues when developing a cross-compiled system. The cross-prelink adds a synthesized dynamic loader that runs on the host, thus permitting cross-prelinking without ever having to run on a read-write target filesystem.
Pseudo: Pseudo is the Yocto Project implementation of fakeroot, which is used to run commands in an environment that seemingly has root privileges.
During a build, it can be necessary to perform operations that require system administrator privileges. For example, file ownership or permissions might need to be defined. Pseudo is a tool that you can either use directly or through the environment variable
LD_PRELOAD
. Either method allows these operations to succeed as if system administrator privileges exist even when they do not.Thanks to Pseudo, the Yocto Project never needs root privileges to build images for your target system.
You can read more about Pseudo in the “Fakeroot and Pseudo” section.
2.3.3 Open-Embedded Build System Components
The following list consists of components associated with the OpenEmbedded Build System:
BitBake: BitBake is a core component of the Yocto Project and is used by the OpenEmbedded build system to build images. While BitBake is key to the build system, BitBake is maintained separately from the Yocto Project.
BitBake is a generic task execution engine that allows shell and Python tasks to be run efficiently and in parallel while working within complex inter-task dependency constraints. In short, BitBake is a build engine that works through recipes written in a specific format in order to perform sets of tasks.
You can learn more about BitBake in the BitBake User Manual.
OpenEmbedded-Core: OpenEmbedded-Core (OE-Core) is a common layer of metadata (i.e. recipes, classes, and associated files) used by OpenEmbedded-derived systems, which includes the Yocto Project. The Yocto Project and the OpenEmbedded Project both maintain the OpenEmbedded-Core. You can find the OE-Core metadata in the Yocto Project Source Repositories.
Historically, the Yocto Project integrated the OE-Core metadata throughout the Yocto Project source repository reference system (Poky). After Yocto Project Version 1.0, the Yocto Project and OpenEmbedded agreed to work together and share a common core set of metadata (OE-Core), which contained much of the functionality previously found in Poky. This collaboration achieved a long-standing OpenEmbedded objective for having a more tightly controlled and quality-assured core. The results also fit well with the Yocto Project objective of achieving a smaller number of fully featured tools as compared to many different ones.
Sharing a core set of metadata results in Poky as an integration layer on top of OE-Core. You can see that in this figure. The Yocto Project combines various components such as BitBake, OE-Core, script “glue”, and documentation for its build system.
2.3.4 Reference Distribution (Poky)
Poky is the Yocto Project reference distribution. It contains the OpenEmbedded Build System (BitBake and OE-Core) as well as a set of metadata to get you started building your own distribution. See the figure in “What is the Yocto Project?” section for an illustration that shows Poky and its relationship with other parts of the Yocto Project.
To use the Yocto Project tools and components, you can download
(clone
) Poky and use it to bootstrap your own distribution.
Note
Poky does not contain binary files. It is a working example of how to build your own custom Linux distribution from source.
You can read more about Poky in the “Reference Embedded Distribution (Poky)” section.
2.3.5 Packages for Finished Targets
The following lists components associated with packages for finished targets:
Matchbox: Matchbox is an Open Source, base environment for the X Window System running on non-desktop, embedded platforms such as handhelds, set-top boxes, kiosks, and anything else for which screen space, input mechanisms, or system resources are limited.
Matchbox consists of a number of interchangeable and optional applications that you can tailor to a specific, non-desktop platform to enhance usability in constrained environments.
You can find the Matchbox source in the Yocto Project Source Repositories.
Opkg: Open PacKaGe management (opkg) is a lightweight package management system based on the itsy package (ipkg) management system. Opkg is written in C and resembles Advanced Package Tool (APT) and Debian Package (dpkg) in operation.
Opkg is intended for use on embedded Linux devices and is used in this capacity in the OpenEmbedded and OpenWrt projects, as well as the Yocto Project.
Note
As best it can, opkg maintains backwards compatibility with ipkg and conforms to a subset of Debian’s policy manual regarding control files.
You can find the opkg source in the Yocto Project Source Repositories.
2.3.6 Archived Components
The Build Appliance is a virtual machine image that enables you to build and boot a custom embedded Linux image with the Yocto Project using a non-Linux development system.
Historically, the Build Appliance was the second of three methods by which you could use the Yocto Project on a system that was not native to Linux.
Hob: Hob, which is now deprecated and is no longer available since the 2.1 release of the Yocto Project provided a rudimentary, GUI-based interface to the Yocto Project. Toaster has fully replaced Hob.
Build Appliance: Post Hob, the Build Appliance became available. It was never recommended that you use the Build Appliance as a day-to-day production development environment with the Yocto Project. Build Appliance was useful as a way to try out development in the Yocto Project environment.
CROPS: The final and best solution available now for developing using the Yocto Project on a system not native to Linux is with CROPS.
2.4 Development Methods
The Yocto Project development environment usually involves a Build Host and target hardware. You use the Build Host to build images and develop applications, while you use the target hardware to execute deployed software.
This section provides an introduction to the choices or development methods you have when setting up your Build Host. Depending on your particular workflow preference and the type of operating system your Build Host runs, several choices exist that allow you to use the Yocto Project.
Note
For additional detail about the Yocto Project development environment, see the “The Yocto Project Development Environment” chapter.
Native Linux Host: By far the best option for a Build Host. A system running Linux as its native operating system allows you to develop software by directly using the BitBake tool. You can accomplish all aspects of development from a regular shell in a supported Linux distribution.
For information on how to set up a Build Host on a system running Linux as its native operating system, see the “Setting Up a Native Linux Host” section in the Yocto Project Development Tasks Manual.
CROss PlatformS (CROPS): Typically, you use CROPS, which leverages Docker Containers, to set up a Build Host that is not running Linux (e.g. Microsoft Windows or macOS).
Note
You can, however, use CROPS on a Linux-based system.
CROPS is an open source, cross-platform development framework that provides an easily managed, extensible environment for building binaries targeted for a variety of architectures on Windows, macOS, or Linux hosts. Once the Build Host is set up using CROPS, you can prepare a shell environment to mimic that of a shell being used on a system natively running Linux.
For information on how to set up a Build Host with CROPS, see the “Setting Up to Use CROss PlatformS (CROPS)” section in the Yocto Project Development Tasks Manual.
Windows Subsystem For Linux (WSLv2): You may use Windows Subsystem For Linux v2 to set up a Build Host using Windows 10.
Note
The Yocto Project is not compatible with WSLv1, it is compatible but not officially supported nor validated with WSLv2, if you still decide to use WSL please upgrade to WSLv2.
The Windows Subsystem For Linux allows Windows 10 to run a real Linux kernel inside of a lightweight virtual machine (VM).
For information on how to set up a Build Host with WSLv2, see the “Setting Up to Use Windows Subsystem For Linux (WSLv2)” section in the Yocto Project Development Tasks Manual.
Toaster: Regardless of what your Build Host is running, you can use Toaster to develop software using the Yocto Project. Toaster is a web interface to the Yocto Project’s OpenEmbedded Build System. The interface allows you to configure and run your builds. Information about builds is collected and stored in a database. You can use Toaster to configure and start builds on multiple remote build servers.
For information about and how to use Toaster, see the Toaster User Manual.
2.5 Reference Embedded Distribution (Poky)
“Poky”, which is pronounced Pock-ee, is the name of the Yocto Project’s reference distribution or Reference OS Kit. Poky contains the OpenEmbedded Build System (BitBake and OpenEmbedded-Core (OE-Core)) as well as a set of Metadata to get you started building your own distro. In other words, Poky is a base specification of the functionality needed for a typical embedded system as well as the components from the Yocto Project that allow you to build a distribution into a usable binary image.
Poky is a combined repository of BitBake, OpenEmbedded-Core (which is
found in meta
), meta-poky
, meta-yocto-bsp
, and documentation
provided all together and known to work well together. You can view
these items that make up the Poky repository in the
Source Repositories.
Note
If you are interested in all the contents of the poky Git repository, see the “Top-Level Core Components” section in the Yocto Project Reference Manual.
The following figure illustrates what generally comprises Poky:
BitBake is a task executor and scheduler that is the heart of the OpenEmbedded build system.
meta-poky
, which is Poky-specific metadata.meta-yocto-bsp
, which are Yocto Project-specific Board Support Packages (BSPs).OpenEmbedded-Core (OE-Core) metadata, which includes shared configurations, global variable definitions, shared classes, packaging, and recipes. Classes define the encapsulation and inheritance of build logic. Recipes are the logical units of software and images to be built.
Documentation, which contains the Yocto Project source files used to make the set of user manuals.
Note
While Poky is a “complete” distribution specification and is tested and put through QA, you cannot use it as a product “out of the box” in its current form.
To use the Yocto Project tools, you can use Git to clone (download) the Poky repository then use your local copy of the reference distribution to bootstrap your own distribution.
Note
Poky does not contain binary files. It is a working example of how to build your own custom Linux distribution from source.
Poky has a regular, well established, six-month release cycle under its own version. Major releases occur at the same time major releases (point releases) occur for the Yocto Project, which are typically in the Spring and Fall. For more information on the Yocto Project release schedule and cadence, see the “Yocto Project Releases and the Stable Release Process” chapter in the Yocto Project Reference Manual.
Much has been said about Poky being a “default configuration”. A default configuration provides a starting image footprint. You can use Poky out of the box to create an image ranging from a shell-accessible minimal image all the way up to a Linux Standard Base-compliant image that uses a GNOME Mobile and Embedded (GMAE) based reference user interface called Sato.
One of the most powerful properties of Poky is that every aspect of a build is controlled by the metadata. You can use metadata to augment these base image types by adding metadata layers <overview-manual/yp-intro:the yocto project layer model> that extend functionality. These layers can provide, for example, an additional software stack for an image type, add a board support package (BSP) for additional hardware, or even create a new image type.
Metadata is loosely grouped into configuration files or package recipes.
A recipe is a collection of non-executable metadata used by BitBake to
set variables or define additional build-time tasks. A recipe contains
fields such as the recipe description, the recipe version, the license
of the package and the upstream source repository. A recipe might also
indicate that the build process uses autotools, make, distutils or any
other build process, in which case the basic functionality can be
defined by the classes it inherits from the OE-Core layer’s class
definitions in ./meta/classes
. Within a recipe you can also define
additional tasks as well as task prerequisites. Recipe syntax through
BitBake also supports both _prepend
and _append
operators as a
method of extending task functionality. These operators inject code into
the beginning or end of a task. For information on these BitBake
operators, see the
“Appending and Prepending (Override Style Syntax)”
section in the BitBake User’s Manual.
2.6 The OpenEmbedded Build System Workflow
The OpenEmbedded Build System uses a “workflow” to accomplish image and SDK generation. The following figure overviews that workflow:
Following is a brief summary of the “workflow”:
Developers specify architecture, policies, patches and configuration details.
The build system fetches and downloads the source code from the specified location. The build system supports standard methods such as tarballs or source code repositories systems such as Git.
Once source code is downloaded, the build system extracts the sources into a local work area where patches are applied and common steps for configuring and compiling the software are run.
The build system then installs the software into a temporary staging area where the binary package format you select (DEB, RPM, or IPK) is used to roll up the software.
Different QA and sanity checks run throughout entire build process.
After the binaries are created, the build system generates a binary package feed that is used to create the final root file image.
The build system generates the file system image and a customized Extensible SDK (eSDK) for application development in parallel.
For a very detailed look at this workflow, see the “OpenEmbedded Build System Concepts” section.
2.7 Some Basic Terms
It helps to understand some basic fundamental terms when learning the Yocto Project. Although a list of terms exists in the “Yocto Project Terms” section of the Yocto Project Reference Manual, this section provides the definitions of some terms helpful for getting started:
Configuration Files: Files that hold global definitions of variables, user-defined variables, and hardware configuration information. These files tell the OpenEmbedded Build System what to build and what to put into the image to support a particular platform.
Extensible Software Development Kit (eSDK): A custom SDK for application developers. This eSDK allows developers to incorporate their library and programming changes back into the image to make their code available to other application developers. For information on the eSDK, see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
Layer: A collection of related recipes. Layers allow you to consolidate related metadata to customize your build. Layers also isolate information used when building for multiple architectures. Layers are hierarchical in their ability to override previous specifications. You can include any number of available layers from the Yocto Project and customize the build by adding your own layers after them. You can search the Layer Index for layers used within Yocto Project.
For more detailed information on layers, see the “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual. For a discussion specifically on BSP Layers, see the “BSP Layers” section in the Yocto Project Board Support Packages (BSP) Developer’s Guide.
Metadata: A key element of the Yocto Project is the Metadata that is used to construct a Linux distribution and is contained in the files that the OpenEmbedded build system parses when building an image. In general, Metadata includes recipes, configuration files, and other information that refers to the build instructions themselves, as well as the data used to control what things get built and the effects of the build. Metadata also includes commands and data used to indicate what versions of software are used, from where they are obtained, and changes or additions to the software itself (patches or auxiliary files) that are used to fix bugs or customize the software for use in a particular situation. OpenEmbedded-Core is an important set of validated metadata.
OpenEmbedded Build System: The terms “BitBake” and “build system” are sometimes used for the OpenEmbedded Build System.
BitBake is a task scheduler and execution engine that parses instructions (i.e. recipes) and configuration data. After a parsing phase, BitBake creates a dependency tree to order the compilation, schedules the compilation of the included code, and finally executes the building of the specified custom Linux image (distribution). BitBake is similar to the
make
tool.During a build process, the build system tracks dependencies and performs a native or cross-compilation of each package. As a first step in a cross-build setup, the framework attempts to create a cross-compiler toolchain (i.e. Extensible SDK) suited for the target platform.
OpenEmbedded-Core (OE-Core): OE-Core is metadata comprised of foundation recipes, classes, and associated files that are meant to be common among many different OpenEmbedded-derived systems, including the Yocto Project. OE-Core is a curated subset of an original repository developed by the OpenEmbedded community that has been pared down into a smaller, core set of continuously validated recipes. The result is a tightly controlled and quality-assured core set of recipes.
You can see the Metadata in the
meta
directory of the Yocto Project Source Repositories.Packages: In the context of the Yocto Project, this term refers to a recipe’s packaged output produced by BitBake (i.e. a “baked recipe”). A package is generally the compiled binaries produced from the recipe’s sources. You “bake” something by running it through BitBake.
It is worth noting that the term “package” can, in general, have subtle meanings. For example, the packages referred to in the “Required Packages for the Build Host” section in the Yocto Project Reference Manual are compiled binaries that, when installed, add functionality to your host Linux distribution.
Another point worth noting is that historically within the Yocto Project, recipes were referred to as packages - thus, the existence of several BitBake variables that are seemingly mis-named, (e.g. PR, PV, and PE).
Poky: Poky is a reference embedded distribution and a reference test configuration. Poky provides the following:
A base-level functional distro used to illustrate how to customize a distribution.
A means by which to test the Yocto Project components (i.e. Poky is used to validate the Yocto Project).
A vehicle through which you can download the Yocto Project.
Poky is not a product level distro. Rather, it is a good starting point for customization.
Note
Poky is an integration layer on top of OE-Core.
Recipe: The most common form of metadata. A recipe contains a list of settings and tasks (i.e. instructions) for building packages that are then used to build the binary image. A recipe describes where you get source code and which patches to apply. Recipes describe dependencies for libraries or for other recipes as well as configuration and compilation options. Related recipes are consolidated into a layer.
3 The Yocto Project Development Environment
This chapter takes a look at the Yocto Project development environment. The chapter provides Yocto Project Development environment concepts that help you understand how work is accomplished in an open source environment, which is very different as compared to work accomplished in a closed, proprietary environment.
Specifically, this chapter addresses open source philosophy, source repositories, workflows, Git, and licensing.
3.1 Open Source Philosophy
Open source philosophy is characterized by software development directed by peer production and collaboration through an active community of developers. Contrast this to the more standard centralized development models used by commercial software companies where a finite set of developers produces a product for sale using a defined set of procedures that ultimately result in an end product whose architecture and source material are closed to the public.
Open source projects conceptually have differing concurrent agendas, approaches, and production. These facets of the development process can come from anyone in the public (community) who has a stake in the software project. The open source environment contains new copyright, licensing, domain, and consumer issues that differ from the more traditional development environment. In an open source environment, the end product, source material, and documentation are all available to the public at no cost.
A benchmark example of an open source project is the Linux kernel, which was initially conceived and created by Finnish computer science student Linus Torvalds in 1991. Conversely, a good example of a non-open source project is the Windows family of operating systems developed by Microsoft Corporation.
Wikipedia has a good historical description of the Open Source Philosophy here. You can also find helpful information on how to participate in the Linux Community here.
3.2 The Development Host
A development host or Build Host is key to using the Yocto Project. Because the goal of the Yocto Project is to develop images or applications that run on embedded hardware, development of those images and applications generally takes place on a system not intended to run the software - the development host.
You need to set up a development host in order to use it with the Yocto Project. Most find that it is best to have a native Linux machine function as the development host. However, it is possible to use a system that does not run Linux as its operating system as your development host. When you have a Mac or Windows-based system, you can set it up as the development host by using CROPS, which leverages Docker Containers. Once you take the steps to set up a CROPS machine, you effectively have access to a shell environment that is similar to what you see when using a Linux-based development host. For the steps needed to set up a system using CROPS, see the “Setting Up to Use CROss PlatformS (CROPS)” section in the Yocto Project Development Tasks Manual.
If your development host is going to be a system that runs a Linux distribution, steps still exist that you must take to prepare the system for use with the Yocto Project. You need to be sure that the Linux distribution on the system is one that supports the Yocto Project. You also need to be sure that the correct set of host packages are installed that allow development using the Yocto Project. For the steps needed to set up a development host that runs Linux, see the “Setting Up a Native Linux Host” section in the Yocto Project Development Tasks Manual.
Once your development host is set up to use the Yocto Project, several methods exist for you to do work in the Yocto Project environment:
Command Lines, BitBake, and Shells: Traditional development in the Yocto Project involves using the OpenEmbedded Build System, which uses BitBake, in a command-line environment from a shell on your development host. You can accomplish this from a host that is a native Linux machine or from a host that has been set up with CROPS. Either way, you create, modify, and build images and applications all within a shell-based environment using components and tools available through your Linux distribution and the Yocto Project.
For a general flow of the build procedures, see the “Building a Simple Image” section in the Yocto Project Development Tasks Manual.
Board Support Package (BSP) Development: Development of BSPs involves using the Yocto Project to create and test layers that allow easy development of images and applications targeted for specific hardware. To development BSPs, you need to take some additional steps beyond what was described in setting up a development host.
The Yocto Project Board Support Package Developer’s Guide provides BSP-related development information. For specifics on development host preparation, see the “Preparing Your Build Host to Work With BSP Layers” section in the Yocto Project Board Support Package (BSP) Developer’s Guide.
Kernel Development: If you are going to be developing kernels using the Yocto Project you likely will be using
devtool
. A workflow usingdevtool
makes kernel development quicker by reducing iteration cycle times.The Yocto Project Linux Kernel Development Manual provides kernel-related development information. For specifics on development host preparation, see the “Preparing the Build Host to Work on the Kernel” section in the Yocto Project Linux Kernel Development Manual.
Using Toaster: The other Yocto Project development method that involves an interface that effectively puts the Yocto Project into the background is Toaster. Toaster provides an interface to the OpenEmbedded build system. The interface enables you to configure and run your builds. Information about builds is collected and stored in a database. You can use Toaster to configure and start builds on multiple remote build servers.
For steps that show you how to set up your development host to use Toaster and on how to use Toaster in general, see the Toaster User Manual.
3.3 Yocto Project Source Repositories
The Yocto Project team maintains complete source repositories for all Yocto Project files at https://git.yoctoproject.org/cgit/cgit.cgi/. This web-based source code browser is organized into categories by function such as IDE Plugins, Matchbox, Poky, Yocto Linux Kernel, and so forth. From the interface, you can click on any particular item in the “Name” column and see the URL at the bottom of the page that you need to clone a Git repository for that particular item. Having a local Git repository of the Source Directory, which is usually named “poky”, allows you to make changes, contribute to the history, and ultimately enhance the Yocto Project’s tools, Board Support Packages, and so forth.
For any supported release of Yocto Project, you can also go to the
Yocto Project Website and select the “DOWNLOADS”
item from the “SOFTWARE” menu and get a released tarball of the poky
repository, any supported BSP tarball, or Yocto Project tools. Unpacking
these tarballs gives you a snapshot of the released files.
Note
The recommended method for setting up the Yocto Project Source Directory and the files for supported BSPs (e.g.,
meta-intel
) is to use Git to create a local copy of the upstream repositories.Be sure to always work in matching branches for both the selected BSP repository and the Source Directory (i.e.
poky
) repository. For example, if you have checked out the “master” branch ofpoky
and you are going to usemeta-intel
, be sure to checkout the “master” branch ofmeta-intel
.
In summary, here is where you can get the project files needed for development:
Source Repositories: This area contains IDE Plugins, Matchbox, Poky, Poky Support, Tools, Yocto Linux Kernel, and Yocto Metadata Layers. You can create local copies of Git repositories for each of these areas.
For steps on how to view and access these upstream Git repositories, see the “Accessing Source Repositories” Section in the Yocto Project Development Tasks Manual.
Index of /releases: This is an index of releases such as Poky, Pseudo, installers for cross-development toolchains, miscellaneous support and all released versions of Yocto Project in the form of images or tarballs. Downloading and extracting these files does not produce a local copy of the Git repository but rather a snapshot of a particular release or image.
For steps on how to view and access these files, see the “Accessing Index of Releases” section in the Yocto Project Development Tasks Manual.
“DOWNLOADS” page for the Yocto Project Website :
The Yocto Project website includes a “DOWNLOADS” page accessible through the “SOFTWARE” menu that allows you to download any Yocto Project release, tool, and Board Support Package (BSP) in tarball form. The tarballs are similar to those found in the Index of /releases: area.
For steps on how to use the “DOWNLOADS” page, see the “Using the Downloads Page” section in the Yocto Project Development Tasks Manual.
3.4 Git Workflows and the Yocto Project
Developing using the Yocto Project likely requires the use of Git. Git is a free, open source distributed version control system used as part of many collaborative design environments. This section provides workflow concepts using the Yocto Project and Git. In particular, the information covers basic practices that describe roles and actions in a collaborative development environment.
Note
If you are familiar with this type of development environment, you might not want to read this section.
The Yocto Project files are maintained using Git in “branches” whose Git histories track every change and whose structures provide branches for all diverging functionality. Although there is no need to use Git, many open source projects do so.
For the Yocto Project, a key individual called the “maintainer” is responsible for the integrity of the “master” branch of a given Git repository. The “master” branch is the “upstream” repository from which final or most recent builds of a project occur. The maintainer is responsible for accepting changes from other developers and for organizing the underlying branch structure to reflect release strategies and so forth.
Note
For information on finding out who is responsible for (maintains) a particular area of code in the Yocto Project, see the “Submitting a Change to the Yocto Project” section of the Yocto Project Development Tasks Manual.
The Yocto Project poky
Git repository also has an upstream
contribution Git repository named poky-contrib
. You can see all the
branches in this repository using the web interface of the
Source Repositories organized within the “Poky Support”
area. These branches hold changes (commits) to the project that have
been submitted or committed by the Yocto Project development team and by
community members who contribute to the project. The maintainer
determines if the changes are qualified to be moved from the “contrib”
branches into the “master” branch of the Git repository.
Developers (including contributing community members) create and maintain cloned repositories of upstream branches. The cloned repositories are local to their development platforms and are used to develop changes. When a developer is satisfied with a particular feature or change, they “push” the change to the appropriate “contrib” repository.
Developers are responsible for keeping their local repository up-to-date with whatever upstream branch they are working against. They are also responsible for straightening out any conflicts that might arise within files that are being worked on simultaneously by more than one person. All this work is done locally on the development host before anything is pushed to a “contrib” area and examined at the maintainer’s level.
A somewhat formal method exists by which developers commit changes and push them into the “contrib” area and subsequently request that the maintainer include them into an upstream branch. This process is called “submitting a patch” or “submitting a change.” For information on submitting patches and changes, see the “Submitting a Change to the Yocto Project” section in the Yocto Project Development Tasks Manual.
In summary, a single point of entry exists for changes into a “master” or development branch of the Git repository, which is controlled by the project’s maintainer. And, a set of developers exist who independently develop, test, and submit changes to “contrib” areas for the maintainer to examine. The maintainer then chooses which changes are going to become a permanent part of the project.
While each development environment is unique, there are some best practices or methods that help development run smoothly. The following list describes some of these practices. For more information about Git workflows, see the workflow topics in the Git Community Book.
Make Small Changes: It is best to keep the changes you commit small as compared to bundling many disparate changes into a single commit. This practice not only keeps things manageable but also allows the maintainer to more easily include or refuse changes.
Make Complete Changes: It is also good practice to leave the repository in a state that allows you to still successfully build your project. In other words, do not commit half of a feature, then add the other half as a separate, later commit. Each commit should take you from one buildable project state to another buildable state.
Use Branches Liberally: It is very easy to create, use, and delete local branches in your working Git repository on the development host. You can name these branches anything you like. It is helpful to give them names associated with the particular feature or change on which you are working. Once you are done with a feature or change and have merged it into your local master branch, simply discard the temporary branch.
Merge Changes: The
git merge
command allows you to take the changes from one branch and fold them into another branch. This process is especially helpful when more than a single developer might be working on different parts of the same feature. Merging changes also automatically identifies any collisions or “conflicts” that might happen as a result of the same lines of code being altered by two different developers.Manage Branches: Because branches are easy to use, you should use a system where branches indicate varying levels of code readiness. For example, you can have a “work” branch to develop in, a “test” branch where the code or change is tested, a “stage” branch where changes are ready to be committed, and so forth. As your project develops, you can merge code across the branches to reflect ever-increasing stable states of the development.
Use Push and Pull: The push-pull workflow is based on the concept of developers “pushing” local commits to a remote repository, which is usually a contribution repository. This workflow is also based on developers “pulling” known states of the project down into their local development repositories. The workflow easily allows you to pull changes submitted by other developers from the upstream repository into your work area ensuring that you have the most recent software on which to develop. The Yocto Project has two scripts named
create-pull-request
andsend-pull-request
that ship with the release to facilitate this workflow. You can find these scripts in thescripts
folder of the Source Directory. For information on how to use these scripts, see the “Using Scripts to Push a Change Upstream and Request a Pull” section in the Yocto Project Development Tasks Manual.Patch Workflow: This workflow allows you to notify the maintainer through an email that you have a change (or patch) you would like considered for the “master” branch of the Git repository. To send this type of change, you format the patch and then send the email using the Git commands
git format-patch
andgit send-email
. For information on how to use these scripts, see the “Submitting a Change to the Yocto Project” section in the Yocto Project Development Tasks Manual.
3.5 Git
The Yocto Project makes extensive use of Git, which is a free, open source distributed version control system. Git supports distributed development, non-linear development, and can handle large projects. It is best that you have some fundamental understanding of how Git tracks projects and how to work with Git if you are going to use the Yocto Project for development. This section provides a quick overview of how Git works and provides you with a summary of some essential Git commands.
Note
For more information on Git, see https://git-scm.com/documentation.
If you need to download Git, it is recommended that you add Git to your system through your distribution’s “software store” (e.g. for Ubuntu, use the Ubuntu Software feature). For the Git download page, see https://git-scm.com/download.
For information beyond the introductory nature in this section, see the “Locating Yocto Project Source Files” section in the Yocto Project Development Tasks Manual.
3.5.2 Basic Commands
Git has an extensive set of commands that lets you manage changes and perform collaboration over the life of a project. Conveniently though, you can manage with a small set of basic operations and workflows once you understand the basic philosophy behind Git. You do not have to be an expert in Git to be functional. A good place to look for instruction on a minimal set of Git commands is here.
The following list of Git commands briefly describes some basic Git operations as a way to get started. As with any set of commands, this list (in most cases) simply shows the base command and omits the many arguments it supports. See the Git documentation for complete descriptions and strategies on how to use these commands:
git init: Initializes an empty Git repository. You cannot use Git commands unless you have a
.git
repository.git clone: Creates a local clone of a Git repository that is on equal footing with a fellow developer’s Git repository or an upstream repository.
git add: Locally stages updated file contents to the index that Git uses to track changes. You must stage all files that have changed before you can commit them.
git commit: Creates a local “commit” that documents the changes you made. Only changes that have been staged can be committed. Commits are used for historical purposes, for determining if a maintainer of a project will allow the change, and for ultimately pushing the change from your local Git repository into the project’s upstream repository.
git status: Reports any modified files that possibly need to be staged and gives you a status of where you stand regarding local commits as compared to the upstream repository.
git checkout branch-name: Changes your local working branch and in this form assumes the local branch already exists. This command is analogous to “cd”.
git checkout -b working-branch upstream-branch: Creates and checks out a working branch on your local machine. The local branch tracks the upstream branch. You can use your local branch to isolate your work. It is a good idea to use local branches when adding specific features or changes. Using isolated branches facilitates easy removal of changes if they do not work out.
git branch: Displays the existing local branches associated with your local repository. The branch that you have currently checked out is noted with an asterisk character.
git branch -D branch-name: Deletes an existing local branch. You need to be in a local branch other than the one you are deleting in order to delete branch-name.
git pull –rebase: Retrieves information from an upstream Git repository and places it in your local Git repository. You use this command to make sure you are synchronized with the repository from which you are basing changes (.e.g. the “master” branch). The “–rebase” option ensures that any local commits you have in your branch are preserved at the top of your local branch.
git push repo-name local-branch:upstream-branch: Sends all your committed local changes to the upstream Git repository that your local repository is tracking (e.g. a contribution repository). The maintainer of the project draws from these repositories to merge changes (commits) into the appropriate branch of project’s upstream repository.
git merge: Combines or adds changes from one local branch of your repository with another branch. When you create a local Git repository, the default branch is named “master”. A typical workflow is to create a temporary branch that is based off “master” that you would use for isolated work. You would make your changes in that isolated branch, stage and commit them locally, switch to the “master” branch, and then use the
git merge
command to apply the changes from your isolated branch into the currently checked out branch (e.g. “master”). After the merge is complete and if you are done with working in that isolated branch, you can safely delete the isolated branch.git cherry-pick commits: Choose and apply specific commits from one branch into another branch. There are times when you might not be able to merge all the changes in one branch with another but need to pick out certain ones.
gitk: Provides a GUI view of the branches and changes in your local Git repository. This command is a good way to graphically see where things have diverged in your local repository.
Note
You need to install the gitk package on your development system to use this command.
git log: Reports a history of your commits to the repository. This report lists all commits regardless of whether you have pushed them upstream or not.
git diff: Displays line-by-line differences between a local working file and the same file as understood by Git. This command is useful to see what you have changed in any given file.
3.6 Licensing
Because open source projects are open to the public, they have different licensing structures in place. License evolution for both Open Source and Free Software has an interesting history. If you are interested in this history, you can find basic information here:
In general, the Yocto Project is broadly licensed under the Massachusetts Institute of Technology (MIT) License. MIT licensing permits the reuse of software within proprietary software as long as the license is distributed with that software. MIT is also compatible with the GNU General Public License (GPL). Patches to the Yocto Project follow the upstream licensing scheme. You can find information on the MIT license here. You can find information on the GNU GPL here.
When you build an image using the Yocto Project, the build process uses
a known list of licenses to ensure compliance. You can find this list in
the Source Directory at
meta/files/common-licenses
. Once the build completes, the list of
all licenses found and used during that build are kept in the
Build Directory at
tmp/deploy/licenses
.
If a module requires a license that is not in the base list, the build process generates a warning during the build. These tools make it easier for a developer to be certain of the licenses with which their shipped products must comply. However, even with these tools it is still up to the developer to resolve potential licensing issues.
The base list of licenses used by the build process is a combination of the Software Package Data Exchange (SPDX) list and the Open Source Initiative (OSI) projects. SPDX Group is a working group of the Linux Foundation that maintains a specification for a standard format for communicating the components, licenses, and copyrights associated with a software package. OSI is a corporation dedicated to the Open Source Definition and the effort for reviewing and approving licenses that conform to the Open Source Definition (OSD).
You can find a list of the combined SPDX and OSI licenses that the Yocto
Project uses in the meta/files/common-licenses
directory in your
Source Directory.
For information that can help you maintain compliance with various open source licensing during the lifecycle of a product created using the Yocto Project, see the “Maintaining Open Source License Compliance During Your Product’s Lifecycle” section in the Yocto Project Development Tasks Manual.
4 Yocto Project Concepts
This chapter provides explanations for Yocto Project concepts that go beyond the surface of “how-to” information and reference (or look-up) material. Concepts such as components, the OpenEmbedded Build System workflow, cross-development toolchains, shared state cache, and so forth are explained.
4.1 Yocto Project Components
The BitBake task executor together with various types of configuration files form the OpenEmbedded-Core (OE-Core). This section overviews these components by describing their use and how they interact.
BitBake handles the parsing and execution of the data files. The data itself is of various types:
Recipes: Provides details about particular pieces of software.
Class Data: Abstracts common build information (e.g. how to build a Linux kernel).
Configuration Data: Defines machine-specific settings, policy decisions, and so forth. Configuration data acts as the glue to bind everything together.
BitBake knows how to combine multiple data sources together and refers to each data source as a layer. For information on layers, see the “Understanding and Creating Layers” section of the Yocto Project Development Tasks Manual.
Following are some brief details on these core components. For additional information on how these components interact during a build, see the “OpenEmbedded Build System Concepts” section.
4.1.1 BitBake
BitBake is the tool at the heart of the OpenEmbedded Build System and is responsible for parsing the Metadata, generating a list of tasks from it, and then executing those tasks.
This section briefly introduces BitBake. If you want more information on BitBake, see the BitBake User Manual.
To see a list of the options BitBake supports, use either of the following commands:
$ bitbake -h
$ bitbake --help
The most common usage for BitBake is bitbake recipename
, where
recipename
is the name of the recipe you want to build (referred
to as the “target”). The target often equates to the first part of a
recipe’s filename (e.g. “foo” for a recipe named foo_1.3.0-r0.bb
).
So, to process the matchbox-desktop_1.2.3.bb
recipe file, you might
type the following:
$ bitbake matchbox-desktop
Several different
versions of matchbox-desktop
might exist. BitBake chooses the one
selected by the distribution configuration. You can get more details
about how BitBake chooses between different target versions and
providers in the
“Preferences” section
of the BitBake User Manual.
BitBake also tries to execute any dependent tasks first. So for example,
before building matchbox-desktop
, BitBake would build a cross
compiler and glibc
if they had not already been built.
A useful BitBake option to consider is the -k
or --continue
option. This option instructs BitBake to try and continue processing the
job as long as possible even after encountering an error. When an error
occurs, the target that failed and those that depend on it cannot be
remade. However, when you use this option other dependencies can still
be processed.
4.1.2 Recipes
Files that have the .bb
suffix are “recipes” files. In general, a
recipe contains information about a single piece of software. This
information includes the location from which to download the unaltered
source, any source patches to be applied to that source (if needed),
which special configuration options to apply, how to compile the source
files, and how to package the compiled output.
The term “package” is sometimes used to refer to recipes. However, since
the word “package” is used for the packaged output from the OpenEmbedded
build system (i.e. .ipk
or .deb
files), this document avoids
using the term “package” when referring to recipes.
4.1.3 Classes
Class files (.bbclass
) contain information that is useful to share
between recipes files. An example is the
autotools class,
which contains common settings for any application that Autotools uses.
The “Classes” chapter in the
Yocto Project Reference Manual provides details about classes and how to
use them.
4.1.4 Configurations
The configuration files (.conf
) define various configuration
variables that govern the OpenEmbedded build process. These files fall
into several areas that define machine configuration options,
distribution configuration options, compiler tuning options, general
common configuration options, and user configuration options in
conf/local.conf
, which is found in the Build Directory.
4.2 Layers
Layers are repositories that contain related metadata (i.e. sets of instructions) that tell the OpenEmbedded build system how to build a target. The Yocto Project Layer Model facilitates collaboration, sharing, customization, and reuse within the Yocto Project development environment. Layers logically separate information for your project. For example, you can use a layer to hold all the configurations for a particular piece of hardware. Isolating hardware-specific configurations allows you to share other metadata by using a different layer where that metadata might be common across several pieces of hardware.
Many layers exist that work in the Yocto Project development environment. The Yocto Project Curated Layer Index and OpenEmbedded Layer Index both contain layers from which you can use or leverage.
By convention, layers in the Yocto Project follow a specific form.
Conforming to a known structure allows BitBake to make assumptions
during builds on where to find types of metadata. You can find
procedures and learn about tools (i.e. bitbake-layers
) for creating
layers suitable for the Yocto Project in the
“Understanding and Creating Layers”
section of the Yocto Project Development Tasks Manual.
4.3 OpenEmbedded Build System Concepts
This section takes a more detailed look inside the build process used by the OpenEmbedded Build System, which is the build system specific to the Yocto Project. At the heart of the build system is BitBake, the task executor.
The following diagram represents the high-level workflow of a build. The remainder of this section expands on the fundamental input, output, process, and metadata logical blocks that make up the workflow.
In general, the build’s workflow consists of several functional areas:
User Configuration: metadata you can use to control the build process.
Metadata Layers: Various layers that provide software, machine, and distro metadata.
Source Files: Upstream releases, local projects, and SCMs.
Build System: Processes under the control of BitBake. This block expands on how BitBake fetches source, applies patches, completes compilation, analyzes output for package generation, creates and tests packages, generates images, and generates cross-development tools.
Package Feeds: Directories containing output packages (RPM, DEB or IPK), which are subsequently used in the construction of an image or Software Development Kit (SDK), produced by the build system. These feeds can also be copied and shared using a web server or other means to facilitate extending or updating existing images on devices at runtime if runtime package management is enabled.
Images: Images produced by the workflow.
Application Development SDK: Cross-development tools that are produced along with an image or separately with BitBake.
4.3.1 User Configuration
User configuration helps define the build. Through user configuration, you can tell BitBake the target architecture for which you are building the image, where to store downloaded source, and other build properties.
The following figure shows an expanded representation of the “User Configuration” box of the general workflow figure:
BitBake needs some basic configuration files in order to complete a
build. These files are *.conf
files. The minimally necessary ones
reside as example files in the build/conf
directory of the
Source Directory. For simplicity,
this section refers to the Source Directory as the “Poky Directory.”
When you clone the Poky Git repository
or you download and unpack a Yocto Project release, you can set up the
Source Directory to be named anything you want. For this discussion, the
cloned repository uses the default name poky
.
Note
The Poky repository is primarily an aggregation of existing repositories. It is not a canonical upstream source.
The meta-poky
layer inside Poky contains a conf
directory that
has example configuration files. These example files are used as a basis
for creating actual configuration files when you source
oe-init-build-env, which is the
build environment script.
Sourcing the build environment script creates a
Build Directory if one does not
already exist. BitBake uses the Build Directory for all its work during
builds. The Build Directory has a conf
directory that contains
default versions of your local.conf
and bblayers.conf
configuration files. These default configuration files are created only
if versions do not already exist in the Build Directory at the time you
source the build environment setup script.
Because the Poky repository is fundamentally an aggregation of existing repositories, some users might be familiar with running the oe-init-build-env script in the context of separate OpenEmbedded-Core (OE-Core) and BitBake repositories rather than a single Poky repository. This discussion assumes the script is executed from within a cloned or unpacked version of Poky.
Depending on where the script is sourced, different sub-scripts are
called to set up the Build Directory (Yocto or OpenEmbedded).
Specifically, the script scripts/oe-setup-builddir
inside the poky
directory sets up the Build Directory and seeds the directory (if
necessary) with configuration files appropriate for the Yocto Project
development environment.
Note
The
scripts/oe-setup-builddir
script uses the
$TEMPLATECONF
variable to determine which sample configuration files to locate.
The local.conf
file provides many basic variables that define a
build environment. Here is a list of a few. To see the default
configurations in a local.conf
file created by the build environment
script, see the
local.conf.sample
in the meta-poky
layer:
Target Machine Selection: Controlled by the MACHINE variable.
Download Directory: Controlled by the DL_DIR variable.
Shared State Directory: Controlled by the SSTATE_DIR variable.
Build Output: Controlled by the TMPDIR variable.
Distribution Policy: Controlled by the DISTRO variable.
Packaging Format: Controlled by the PACKAGE_CLASSES variable.
SDK Target Architecture: Controlled by the SDKMACHINE variable.
Extra Image Packages: Controlled by the EXTRA_IMAGE_FEATURES variable.
Note
Configurations set in the conf/local.conf file can also be set in the conf/site.conf and conf/auto.conf configuration files.
The bblayers.conf
file tells BitBake what layers you want considered
during the build. By default, the layers listed in this file include
layers minimally needed by the build system. However, you must manually
add any custom layers you have created. You can find more information on
working with the bblayers.conf
file in the
“Enabling Your Layer”
section in the Yocto Project Development Tasks Manual.
The files site.conf
and auto.conf
are not created by the
environment initialization script. If you want the site.conf
file,
you need to create that yourself. The auto.conf
file is typically
created by an autobuilder:
site.conf: You can use the
conf/site.conf
configuration file to configure multiple build directories. For example, suppose you had several build environments and they shared some common features. You can set these default build properties here. A good example is perhaps the packaging format to use through the PACKAGE_CLASSES variable.One useful scenario for using the
conf/site.conf
file is to extend your BBPATH variable to include the path to aconf/site.conf
. Then, when BitBake looks for Metadata usingBBPATH
, it finds theconf/site.conf
file and applies your common configurations found in the file. To override configurations in a particular build directory, alter the similar configurations within that build directory’sconf/local.conf
file.auto.conf: The file is usually created and written to by an autobuilder. The settings put into the file are typically the same as you would find in the
conf/local.conf
or theconf/site.conf
files.
You can edit all configuration files to further define any particular build environment. This process is represented by the “User Configuration Edits” box in the figure.
When you launch your build with the bitbake target
command, BitBake
sorts out the configurations to ultimately define your build
environment. It is important to understand that the
OpenEmbedded Build System reads the
configuration files in a specific order: site.conf
, auto.conf
,
and local.conf
. And, the build system applies the normal assignment
statement rules as described in the
“Syntax and Operators” chapter
of the BitBake User Manual. Because the files are parsed in a specific
order, variable assignments for the same variable could be affected. For
example, if the auto.conf
file and the local.conf
set variable1
to different values, because the build system parses local.conf
after auto.conf
, variable1 is assigned the value from the
local.conf
file.
4.3.2 Metadata, Machine Configuration, and Policy Configuration
The previous section described the user configurations that define BitBake’s global behavior. This section takes a closer look at the layers the build system uses to further control the build. These layers provide Metadata for the software, machine, and policies.
In general, three types of layer input exists. You can see them below the “User Configuration” box in the general workflow figure <overview-manual/concepts:openembedded build system concepts>:
Metadata (.bb + Patches): Software layers containing user-supplied recipe files, patches, and append files. A good example of a software layer might be the meta-qt5 layer from the OpenEmbedded Layer Index. This layer is for version 5.0 of the popular Qt cross-platform application development framework for desktop, embedded and mobile.
Machine BSP Configuration: Board Support Package (BSP) layers (i.e. “BSP Layer” in the following figure) providing machine-specific configurations. This type of information is specific to a particular target architecture. A good example of a BSP layer from the Reference Distribution (Poky) is the meta-yocto-bsp layer.
Policy Configuration: Distribution Layers (i.e. “Distro Layer” in the following figure) providing top-level or general policies for the images or SDKs being built for a particular distribution. For example, in the Poky Reference Distribution the distro layer is the meta-poky layer. Within the distro layer is a
conf/distro
directory that contains distro configuration files (e.g. poky.conf that contain many policy configurations for the Poky distribution.
The following figure shows an expanded representation of these three layers from the general workflow figure:
In general, all layers have a similar structure. They all contain a
licensing file (e.g. COPYING.MIT
) if the layer is to be distributed,
a README
file as good practice and especially if the layer is to be
distributed, a configuration directory, and recipe directories. You can
learn about the general structure for layers used with the Yocto Project
in the
“Creating Your Own Layer”
section in the
Yocto Project Development Tasks Manual. For a general discussion on
layers and the many layers from which you can draw, see the
“Layers” and
“The Yocto Project Layer Model” sections both
earlier in this manual.
If you explored the previous links, you discovered some areas where many layers that work with the Yocto Project exist. The Source Repositories also shows layers categorized under “Yocto Metadata Layers.”
Note
Layers exist in the Yocto Project Source Repositories that cannot be found in the OpenEmbedded Layer Index. These layers are either deprecated or experimental in nature.
BitBake uses the conf/bblayers.conf
file, which is part of the user
configuration, to find what layers it should be using as part of the
build.
4.3.2.1 Distro Layer
The distribution layer provides policy configurations for your
distribution. Best practices dictate that you isolate these types of
configurations into their own layer. Settings you provide in
conf/distro/distro.conf
override similar settings that BitBake finds
in your conf/local.conf
file in the Build Directory.
The following list provides some explanation and references for what you typically find in the distribution layer:
classes: Class files (
.bbclass
) hold common functionality that can be shared among recipes in the distribution. When your recipes inherit a class, they take on the settings and functions for that class. You can read more about class files in the “Classes” chapter of the Yocto Reference Manual.conf: This area holds configuration files for the layer (
conf/layer.conf
), the distribution (conf/distro/distro.conf
), and any distribution-wide include files.recipes-:* Recipes and append files that affect common functionality across the distribution. This area could include recipes and append files to add distribution-specific configuration, initialization scripts, custom image recipes, and so forth. Examples of
recipes-*
directories arerecipes-core
andrecipes-extra
. Hierarchy and contents within arecipes-*
directory can vary. Generally, these directories contain recipe files (*.bb
), recipe append files (*.bbappend
), directories that are distro-specific for configuration files, and so forth.
4.3.2.2 BSP Layer
The BSP Layer provides machine configurations that target specific hardware. Everything in this layer is specific to the machine for which you are building the image or the SDK. A common structure or form is defined for BSP layers. You can learn more about this structure in the Yocto Project Board Support Package Developer’s Guide.
Note
In order for a BSP layer to be considered compliant with the Yocto Project, it must meet some structural requirements.
The BSP Layer’s configuration directory contains configuration files for
the machine (conf/machine/machine.conf
) and, of course, the layer
(conf/layer.conf
).
The remainder of the layer is dedicated to specific recipes by function:
recipes-bsp
, recipes-core
, recipes-graphics
,
recipes-kernel
, and so forth. Metadata can exist for multiple
formfactors, graphics support systems, and so forth.
Note
While the figure shows several recipes-* directories, not all these directories appear in all BSP layers.
4.3.2.3 Software Layer
The software layer provides the Metadata for additional software packages used during the build. This layer does not include Metadata that is specific to the distribution or the machine, which are found in their respective layers.
This layer contains any recipes, append files, and patches, that your project needs.
4.3.3 Sources
In order for the OpenEmbedded build system to create an image or any target, it must be able to access source files. The general workflow figure represents source files using the “Upstream Project Releases”, “Local Projects”, and “SCMs (optional)” boxes. The figure represents mirrors, which also play a role in locating source files, with the “Source Materials” box.
The method by which source files are ultimately organized is a function of the project. For example, for released software, projects tend to use tarballs or other archived files that can capture the state of a release guaranteeing that it is statically represented. On the other hand, for a project that is more dynamic or experimental in nature, a project might keep source files in a repository controlled by a Source Control Manager (SCM) such as Git. Pulling source from a repository allows you to control the point in the repository (the revision) from which you want to build software. Finally, a combination of the two might exist, which would give the consumer a choice when deciding where to get source files.
BitBake uses the SRC_URI
variable to point to source files regardless of their location. Each
recipe must have a SRC_URI
variable that points to the source.
Another area that plays a significant role in where source files come
from is pointed to by the
DL_DIR variable. This area is
a cache that can hold previously downloaded source. You can also
instruct the OpenEmbedded build system to create tarballs from Git
repositories, which is not the default behavior, and store them in the
DL_DIR
by using the
BB_GENERATE_MIRROR_TARBALLS
variable.
Judicious use of a DL_DIR
directory can save the build system a trip
across the Internet when looking for files. A good method for using a
download directory is to have DL_DIR
point to an area outside of
your Build Directory. Doing so allows you to safely delete the Build
Directory if needed without fear of removing any downloaded source file.
The remainder of this section provides a deeper look into the source files and the mirrors. Here is a more detailed look at the source file area of the general workflow figure:
4.3.3.1 Upstream Project Releases
Upstream project releases exist anywhere in the form of an archived file (e.g. tarball or zip file). These files correspond to individual recipes. For example, the figure uses specific releases each for BusyBox, Qt, and Dbus. An archive file can be for any released product that can be built using a recipe.
4.3.3.2 Local Projects
Local projects are custom bits of software the user provides. These bits reside somewhere local to a project - perhaps a directory into which the user checks in items (e.g. a local directory containing a development source tree used by the group).
The canonical method through which to include a local project is to use
the externalsrc
class to include that local project. You use either the local.conf
or a recipe’s append file to override or set the recipe to point to the
local directory on your disk to pull in the whole source tree.
4.3.3.3 Source Control Managers (Optional)
Another place from which the build system can get source files is with fetchers employing various Source Control Managers (SCMs) such as Git or Subversion. In such cases, a repository is cloned or checked out. The do_fetch task inside BitBake uses the SRC_URI variable and the argument’s prefix to determine the correct fetcher module.
Note
For information on how to have the OpenEmbedded build system generate tarballs for Git repositories and place them in the DL_DIR directory, see the BB_GENERATE_MIRROR_TARBALLS variable in the Yocto Project Reference Manual.
When fetching a repository, BitBake uses the SRCREV variable to determine the specific revision from which to build.
4.3.3.4 Source Mirror(s)
Two kinds of mirrors exist: pre-mirrors and regular mirrors. The PREMIRRORS and MIRRORS variables point to these, respectively. BitBake checks pre-mirrors before looking upstream for any source files. Pre-mirrors are appropriate when you have a shared directory that is not a directory defined by the DL_DIR variable. A Pre-mirror typically points to a shared directory that is local to your organization.
Regular mirrors can be any site across the Internet that is used as an alternative location for source code should the primary site not be functioning for some reason or another.
4.3.4 Package Feeds
When the OpenEmbedded build system generates an image or an SDK, it gets the packages from a package feed area located in the Build Directory. The general workflow figure shows this package feeds area in the upper-right corner.
This section looks a little closer into the package feeds area used by the build system. Here is a more detailed look at the area:
Package feeds are an intermediary step in the build process. The OpenEmbedded build system provides classes to generate different package types, and you specify which classes to enable through the PACKAGE_CLASSES variable. Before placing the packages into package feeds, the build process validates them with generated output quality assurance checks through the insane class.
The package feed area resides in the Build Directory. The directory the build system uses to temporarily store packages is determined by a combination of variables and the particular package manager in use. See the “Package Feeds” box in the illustration and note the information to the right of that area. In particular, the following defines where package files are kept:
DEPLOY_DIR: Defined as
tmp/deploy
in the Build Directory.DEPLOY_DIR_*
: Depending on the package manager used, the package type sub-folder. Given RPM, IPK, or DEB packaging and tarball creation, the DEPLOY_DIR_RPM, DEPLOY_DIR_IPK, DEPLOY_DIR_DEB, or DEPLOY_DIR_TAR, variables are used, respectively.PACKAGE_ARCH: Defines architecture-specific sub-folders. For example, packages could exist for the i586 or qemux86 architectures.
BitBake uses the
do_package_write_*
tasks to generate packages and place them into the package holding area
(e.g. do_package_write_ipk
for IPK packages). See the
“do_package_write_deb”,
“do_package_write_ipk”,
“do_package_write_rpm”,
and
“do_package_write_tar”
sections in the Yocto Project Reference Manual for additional
information. As an example, consider a scenario where an IPK packaging
manager is being used and package architecture support for both i586 and
qemux86 exist. Packages for the i586 architecture are placed in
build/tmp/deploy/ipk/i586
, while packages for the qemux86
architecture are placed in build/tmp/deploy/ipk/qemux86
.
4.3.5 BitBake Tool
The OpenEmbedded build system uses BitBake to produce images and Software Development Kits (SDKs). You can see from the general workflow figure, the BitBake area consists of several functional areas. This section takes a closer look at each of those areas.
Note
Separate documentation exists for the BitBake tool. See the BitBake User Manual for reference material on BitBake.
4.3.5.1 Source Fetching
The first stages of building a recipe are to fetch and unpack the source code:
The do_fetch and do_unpack tasks fetch the source files and unpack them into the Build Directory.
Note
For every local file (e.g. file:// ) that is part of a recipe’s SRC_URI statement, the OpenEmbedded build system takes a checksum of the file for the recipe and inserts the checksum into the signature for the do_fetch task. If any local file has been modified, the do_fetch task and all tasks that depend on it are re-executed.
By default, everything is accomplished in the Build Directory, which has a defined structure. For additional general information on the Build Directory, see the “build/” section in the Yocto Project Reference Manual.
Each recipe has an area in the Build Directory where the unpacked source code resides. The S variable points to this area for a recipe’s unpacked source code. The name of that directory for any given recipe is defined from several different variables. The preceding figure and the following list describe the Build Directory’s hierarchy:
TMPDIR: The base directory where the OpenEmbedded build system performs all its work during the build. The default base directory is the
tmp
directory.PACKAGE_ARCH: The architecture of the built package or packages. Depending on the eventual destination of the package or packages (i.e. machine architecture, Build Host, SDK, or specific machine),
PACKAGE_ARCH
varies. See the variable’s description for details.TARGET_OS: The operating system of the target device. A typical value would be “linux” (e.g. “qemux86-poky-linux”).
PN: The name of the recipe used to build the package. This variable can have multiple meanings. However, when used in the context of input files,
PN
represents the name of the recipe.WORKDIR: The location where the OpenEmbedded build system builds a recipe (i.e. does the work to create the package).
S: Contains the unpacked source files for a given recipe.
Note
In the previous figure, notice that two sample hierarchies exist: one based on package architecture (i.e. PACKAGE_ARCH ) and one based on a machine (i.e. MACHINE ). The underlying structures are identical. The differentiator being what the OpenEmbedded build system is using as a build target (e.g. general architecture, a build host, an SDK, or a specific machine).
4.3.5.2 Patching
Once source code is fetched and unpacked, BitBake locates patch files and applies them to the source files:
The do_patch task uses a recipe’s SRC_URI statements and the FILESPATH variable to locate applicable patch files.
Default processing for patch files assumes the files have either
*.patch
or *.diff
file types. You can use SRC_URI
parameters
to change the way the build system recognizes patch files. See the
do_patch task for more
information.
BitBake finds and applies multiple patches for a single recipe in the
order in which it locates the patches. The FILESPATH
variable
defines the default set of directories that the build system uses to
search for patch files. Once found, patches are applied to the recipe’s
source files, which are located in the
S directory.
For more information on how the source directories are created, see the “Source Fetching” section. For more information on how to create patches and how the build system processes patches, see the “Patching Code” section in the Yocto Project Development Tasks Manual. You can also see the “Use devtool modify to Modify the Source of an Existing Component” section in the Yocto Project Application Development and the Extensible Software Development Kit (SDK) manual and the “Using Traditional Kernel Development to Patch the Kernel” section in the Yocto Project Linux Kernel Development Manual.
4.3.5.3 Configuration, Compilation, and Staging
After source code is patched, BitBake executes tasks that configure and compile the source code. Once compilation occurs, the files are copied to a holding area (staged) in preparation for packaging:
This step in the build process consists of the following tasks:
do_prepare_recipe_sysroot: This task sets up the two sysroots in
${
WORKDIR}
(i.e.recipe-sysroot
andrecipe-sysroot-native
) so that during the packaging phase the sysroots can contain the contents of the do_populate_sysroot tasks of the recipes on which the recipe containing the tasks depends. A sysroot exists for both the target and for the native binaries, which run on the host system.do_configure: This task configures the source by enabling and disabling any build-time and configuration options for the software being built. Configurations can come from the recipe itself as well as from an inherited class. Additionally, the software itself might configure itself depending on the target for which it is being built.
The configurations handled by the do_configure task are specific to configurations for the source code being built by the recipe.
If you are using the autotools class, you can add additional configuration options by using the EXTRA_OECONF or PACKAGECONFIG_CONFARGS variables. For information on how this variable works within that class, see the autotools class here.
do_compile: Once a configuration task has been satisfied, BitBake compiles the source using the do_compile task. Compilation occurs in the directory pointed to by the B variable. Realize that the
B
directory is, by default, the same as the S directory.do_install: After compilation completes, BitBake executes the do_install task. This task copies files from the
B
directory and places them in a holding area pointed to by the D variable. Packaging occurs later using files from this holding directory.
4.3.5.4 Package Splitting
After source code is configured, compiled, and staged, the build system analyzes the results and splits the output into packages:
The do_package and do_packagedata tasks combine to analyze the files found in the D directory and split them into subsets based on available packages and files. Analysis involves the following as well as other items: splitting out debugging symbols, looking at shared library dependencies between packages, and looking at package relationships.
The do_packagedata
task creates package metadata based on the
analysis such that the build system can generate the final packages. The
do_populate_sysroot
task stages (copies) a subset of the files installed by the
do_install task into
the appropriate sysroot. Working, staged, and intermediate results of
the analysis and package splitting process use several areas:
PKGD: The destination directory (i.e.
package
) for packages before they are split into individual packages.PKGDESTWORK: A temporary work area (i.e.
pkgdata
) used by thedo_package
task to save package metadata.PKGDEST: The parent directory (i.e.
packages-split
) for packages after they have been split.PKGDATA_DIR: A shared, global-state directory that holds packaging metadata generated during the packaging process. The packaging process copies metadata from
PKGDESTWORK
to thePKGDATA_DIR
area where it becomes globally available.STAGING_DIR_HOST: The path for the sysroot for the system on which a component is built to run (i.e.
recipe-sysroot
).STAGING_DIR_NATIVE: The path for the sysroot used when building components for the build host (i.e.
recipe-sysroot-native
).STAGING_DIR_TARGET: The path for the sysroot used when a component that is built to execute on a system and it generates code for yet another machine (e.g. cross-canadian recipes).
The FILES variable defines the files that go into each package in PACKAGES. If you want details on how this is accomplished, you can look at package.bbclass.
Depending on the type of packages being created (RPM, DEB, or IPK), the
do_package_write_*
task creates the actual packages and places them in the Package Feed
area, which is ${TMPDIR}/deploy
. You can see the
“Package Feeds” section for more detail on
that part of the build process.
Note
Support for creating feeds directly from the deploy/* directories does not exist. Creating such feeds usually requires some kind of feed maintenance mechanism that would upload the new packages into an official package feed (e.g. the Ångström distribution). This functionality is highly distribution-specific and thus is not provided out of the box.
4.3.5.5 Image Generation
Once packages are split and stored in the Package Feeds area, the build system uses BitBake to generate the root filesystem image:
The image generation process consists of several stages and depends on several tasks and variables. The do_rootfs task creates the root filesystem (file and directory structure) for an image. This task uses several key variables to help create the list of packages to actually install:
IMAGE_INSTALL: Lists out the base set of packages from which to install from the Package Feeds area.
PACKAGE_EXCLUDE: Specifies packages that should not be installed into the image.
IMAGE_FEATURES: Specifies features to include in the image. Most of these features map to additional packages for installation.
PACKAGE_CLASSES: Specifies the package backend (e.g. RPM, DEB, or IPK) to use and consequently helps determine where to locate packages within the Package Feeds area.
IMAGE_LINGUAS: Determines the language(s) for which additional language support packages are installed.
PACKAGE_INSTALL: The final list of packages passed to the package manager for installation into the image.
With IMAGE_ROOTFS
pointing to the location of the filesystem under construction and the
PACKAGE_INSTALL
variable providing the final list of packages to
install, the root file system is created.
Package installation is under control of the package manager (e.g. dnf/rpm, opkg, or apt/dpkg) regardless of whether or not package management is enabled for the target. At the end of the process, if package management is not enabled for the target, the package manager’s data files are deleted from the root filesystem. As part of the final stage of package installation, post installation scripts that are part of the packages are run. Any scripts that fail to run on the build host are run on the target when the target system is first booted. If you are using a read-only root filesystem, all the post installation scripts must succeed on the build host during the package installation phase since the root filesystem on the target is read-only.
The final stages of the do_rootfs
task handle post processing. Post
processing includes creation of a manifest file and optimizations.
The manifest file (.manifest
) resides in the same directory as the
root filesystem image. This file lists out, line-by-line, the installed
packages. The manifest file is useful for the
testimage class,
for example, to determine whether or not to run specific tests. See the
IMAGE_MANIFEST
variable for additional information.
Optimizing processes that are run across the image include mklibs
,
prelink
, and any other post-processing commands as defined by the
ROOTFS_POSTPROCESS_COMMAND
variable. The mklibs
process optimizes the size of the libraries,
while the prelink
process optimizes the dynamic linking of shared
libraries to reduce start up time of executables.
After the root filesystem is built, processing begins on the image through the do_image task. The build system runs any pre-processing commands as defined by the IMAGE_PREPROCESS_COMMAND variable. This variable specifies a list of functions to call before the build system creates the final image output files.
The build system dynamically creates do_image_*
tasks as needed,
based on the image types specified in the
IMAGE_FSTYPES variable.
The process turns everything into an image file or a set of image files
and can compress the root filesystem image to reduce the overall size of
the image. The formats used for the root filesystem depend on the
IMAGE_FSTYPES
variable. Compression depends on whether the formats
support compression.
As an example, a dynamically created task when creating a particular image type would take the following form:
do_image_type
So, if the type
as specified by the IMAGE_FSTYPES
were ext4
, the dynamically
generated task would be as follows:
do_image_ext4
The final task involved in image creation is the do_image_complete task. This task completes the image by applying any image post processing as defined through the IMAGE_POSTPROCESS_COMMAND variable. The variable specifies a list of functions to call once the build system has created the final image output files.
Note
The entire image generation process is run under Pseudo. Running under Pseudo ensures that the files in the root filesystem have correct ownership.
4.3.5.6 SDK Generation
The OpenEmbedded build system uses BitBake to generate the Software Development Kit (SDK) installer scripts for both the standard SDK and the extensible SDK (eSDK):
Note
For more information on the cross-development toolchain generation, see the “Cross-Development Toolchain Generation” section. For information on advantages gained when building a cross-development toolchain using the do_populate_sdk task, see the “Building an SDK Installer” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
Like image generation, the SDK script process consists of several stages and depends on many variables. The do_populate_sdk and do_populate_sdk_ext tasks use these key variables to help create the list of packages to actually install. For information on the variables listed in the figure, see the “Application Development SDK” section.
The do_populate_sdk
task helps create the standard SDK and handles
two parts: a target part and a host part. The target part is the part
built for the target hardware and includes libraries and headers. The
host part is the part of the SDK that runs on the
SDKMACHINE.
The do_populate_sdk_ext
task helps create the extensible SDK and
handles host and target parts differently than its counter part does for
the standard SDK. For the extensible SDK, the task encapsulates the
build system, which includes everything needed (host and target) for the
SDK.
Regardless of the type of SDK being constructed, the tasks perform some
cleanup after which a cross-development environment setup script and any
needed configuration files are created. The final output is the
Cross-development toolchain installation script (.sh
file), which
includes the environment setup script.
4.3.5.7 Stamp Files and the Rerunning of Tasks
For each task that completes successfully, BitBake writes a stamp file into the STAMPS_DIR directory. The beginning of the stamp file’s filename is determined by the STAMP variable, and the end of the name consists of the task’s name and current input checksum.
Note
This naming scheme assumes that BB_SIGNATURE_HANDLER is “OEBasicHash”, which is almost always the case in current OpenEmbedded.
To determine if a task needs to be rerun, BitBake checks if a stamp file with a matching input checksum exists for the task. If such a stamp file exists, the task’s output is assumed to exist and still be valid. If the file does not exist, the task is rerun.
Note
The stamp mechanism is more general than the shared state (sstate) cache mechanism described in the “Setscene Tasks and Shared State” section. BitBake avoids rerunning any task that has a valid stamp file, not just tasks that can be accelerated through the sstate cache.
However, you should realize that stamp files only serve as a marker that some work has been done and that these files do not record task output. The actual task output would usually be somewhere in TMPDIR (e.g. in some recipe’s WORKDIR.) What the sstate cache mechanism adds is a way to cache task output that can then be shared between build machines.
Since STAMPS_DIR
is usually a subdirectory of TMPDIR
, removing
TMPDIR
will also remove STAMPS_DIR
, which means tasks will
properly be rerun to repopulate TMPDIR
.
If you want some task to always be considered “out of date”, you can mark it with the nostamp varflag. If some other task depends on such a task, then that task will also always be considered out of date, which might not be what you want.
For details on how to view information about a task’s signature, see the “Viewing Task Variable Dependencies” section in the Yocto Project Development Tasks Manual.
4.3.6 Images
The images produced by the build system are compressed forms of the root filesystem and are ready to boot on a target device. You can see from the general workflow figure that BitBake output, in part, consists of images. This section takes a closer look at this output:
Note
For a list of example images that the Yocto Project provides, see the “Images” chapter in the Yocto Project Reference Manual.
The build process writes images out to the Build Directory
inside the
tmp/deploy/images/machine/
folder as shown in the figure. This
folder contains any files expected to be loaded on the target device.
The DEPLOY_DIR variable
points to the deploy
directory, while the
DEPLOY_DIR_IMAGE
variable points to the appropriate directory containing images for the
current configuration.
kernel-image: A kernel binary file. The KERNEL_IMAGETYPE variable determines the naming scheme for the kernel image file. Depending on this variable, the file could begin with a variety of naming strings. The
deploy/images/
machine directory can contain multiple image files for the machine.root-filesystem-image: Root filesystems for the target device (e.g.
*.ext3
or*.bz2
files). The IMAGE_FSTYPES variable determines the root filesystem image type. Thedeploy/images/
machine directory can contain multiple root filesystems for the machine.kernel-modules: Tarballs that contain all the modules built for the kernel. Kernel module tarballs exist for legacy purposes and can be suppressed by setting the MODULE_TARBALL_DEPLOY variable to “0”. The
deploy/images/
machine directory can contain multiple kernel module tarballs for the machine.bootloaders: If applicable to the target machine, bootloaders supporting the image. The
deploy/images/
machine directory can contain multiple bootloaders for the machine.symlinks: The
deploy/images/
machine folder contains a symbolic link that points to the most recently built file for each machine. These links might be useful for external scripts that need to obtain the latest version of each file.
4.3.7 Application Development SDK
In the general workflow figure, the
output labeled “Application Development SDK” represents an SDK. The SDK
generation process differs depending on whether you build an extensible
SDK (e.g. bitbake -c populate_sdk_ext
imagename) or a standard SDK
(e.g. bitbake -c populate_sdk
imagename). This section takes a
closer look at this output:
The specific form of this output is a set of files that includes a
self-extracting SDK installer (*.sh
), host and target manifest
files, and files used for SDK testing. When the SDK installer file is
run, it installs the SDK. The SDK consists of a cross-development
toolchain, a set of libraries and headers, and an SDK environment setup
script. Running this installer essentially sets up your
cross-development environment. You can think of the cross-toolchain as
the “host” part because it runs on the SDK machine. You can think of the
libraries and headers as the “target” part because they are built for
the target hardware. The environment setup script is added so that you
can initialize the environment before using the tools.
Note
The Yocto Project supports several methods by which you can set up this cross-development environment. These methods include downloading pre-built SDK installers or building and installing your own SDK installer.
For background information on cross-development toolchains in the Yocto Project development environment, see the “Cross-Development Toolchain Generation” section.
For information on setting up a cross-development environment, see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
All the output files for an SDK are written to the deploy/sdk
folder
inside the Build Directory as
shown in the previous figure. Depending on the type of SDK, several
variables exist that help configure these files. The following list
shows the variables associated with an extensible SDK:
DEPLOY_DIR: Points to the
deploy
directory.SDK_EXT_TYPE: Controls whether or not shared state artifacts are copied into the extensible SDK. By default, all required shared state artifacts are copied into the SDK.
SDK_INCLUDE_PKGDATA: Specifies whether or not packagedata is included in the extensible SDK for all recipes in the “world” target.
SDK_INCLUDE_TOOLCHAIN: Specifies whether or not the toolchain is included when building the extensible SDK.
SDK_LOCAL_CONF_WHITELIST: A list of variables allowed through from the build system configuration into the extensible SDK configuration.
SDK_LOCAL_CONF_BLACKLIST: A list of variables not allowed through from the build system configuration into the extensible SDK configuration.
SDK_INHERIT_BLACKLIST: A list of classes to remove from the INHERIT value globally within the extensible SDK configuration.
This next list, shows the variables associated with a standard SDK:
DEPLOY_DIR: Points to the
deploy
directory.SDKMACHINE: Specifies the architecture of the machine on which the cross-development tools are run to create packages for the target hardware.
SDKIMAGE_FEATURES: Lists the features to include in the “target” part of the SDK.
TOOLCHAIN_HOST_TASK: Lists packages that make up the host part of the SDK (i.e. the part that runs on the
SDKMACHINE
). When you usebitbake -c populate_sdk imagename
to create the SDK, a set of default packages apply. This variable allows you to add more packages.TOOLCHAIN_TARGET_TASK: Lists packages that make up the target part of the SDK (i.e. the part built for the target hardware).
SDKPATH: Defines the default SDK installation path offered by the installation script.
SDK_HOST_MANIFEST: Lists all the installed packages that make up the host part of the SDK. This variable also plays a minor role for extensible SDK development as well. However, it is mainly used for the standard SDK.
SDK_TARGET_MANIFEST: Lists all the installed packages that make up the target part of the SDK. This variable also plays a minor role for extensible SDK development as well. However, it is mainly used for the standard SDK.
4.4 Cross-Development Toolchain Generation
The Yocto Project does most of the work for you when it comes to creating The Cross-Development Toolchain. This section provides some technical background on how cross-development toolchains are created and used. For more information on toolchains, you can also see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
In the Yocto Project development environment, cross-development toolchains are used to build images and applications that run on the target hardware. With just a few commands, the OpenEmbedded build system creates these necessary toolchains for you.
The following figure shows a high-level build environment regarding toolchain construction and use.
Most of the work occurs on the Build Host. This is the machine used to
build images and generally work within the the Yocto Project
environment. When you run
BitBake to create an image, the
OpenEmbedded build system uses the host gcc
compiler to bootstrap a
cross-compiler named gcc-cross
. The gcc-cross
compiler is what
BitBake uses to compile source files when creating the target image. You
can think of gcc-cross
simply as an automatically generated
cross-compiler that is used internally within BitBake only.
Note
The extensible SDK does not use gcc-cross-canadian since this SDK ships a copy of the OpenEmbedded build system and the sysroot within it contains gcc-cross .
The chain of events that occurs when the standard toolchain is bootstrapped:
binutils-cross -> linux-libc-headers -> gcc-cross -> libgcc-initial -> glibc -> libgcc -> gcc-runtime
gcc
: The compiler, GNU Compiler Collection (GCC).binutils-cross
: The binary utilities needed in order to run thegcc-cross
phase of the bootstrap operation and build the headers for the C library.linux-libc-headers
: Headers needed for the cross-compiler and C library build.libgcc-initial
: An initial version of the gcc support library needed to bootstrapglibc
.libgcc
: The final version of the gcc support library which can only be built once there is a C library to link against.glibc
: The GNU C Library.gcc-cross
: The final stage of the bootstrap process for the cross-compiler. This stage results in the actual cross-compiler that BitBake uses when it builds an image for a targeted device.This tool is a “native” tool (i.e. it is designed to run on the build host).
gcc-runtime
: Runtime libraries resulting from the toolchain bootstrapping process. This tool produces a binary that consists of the runtime libraries need for the targeted device.
You can use the OpenEmbedded build system to build an installer for the
relocatable SDK used to develop applications. When you run the
installer, it installs the toolchain, which contains the development
tools (e.g., gcc-cross-canadian
, binutils-cross-canadian
, and
other nativesdk-*
tools), which are tools native to the SDK (i.e.
native to SDK_ARCH), you
need to cross-compile and test your software. The figure shows the
commands you use to easily build out this toolchain. This
cross-development toolchain is built to execute on the
SDKMACHINE, which might or
might not be the same machine as the Build Host.
Note
If your target architecture is supported by the Yocto Project, you can take advantage of pre-built images that ship with the Yocto Project and already contain cross-development toolchain installers.
Here is the bootstrap process for the relocatable toolchain:
gcc -> binutils-crosssdk -> gcc-crosssdk-initial -> linux-libc-headers -> glibc-initial -> nativesdk-glibc -> gcc-crosssdk -> gcc-cross-canadian
gcc
: The build host’s GNU Compiler Collection (GCC).binutils-crosssdk
: The bare minimum binary utilities needed in order to run thegcc-crosssdk-initial
phase of the bootstrap operation.gcc-crosssdk-initial
: An early stage of the bootstrap process for creating the cross-compiler. This stage builds enough of thegcc-crosssdk
and supporting pieces so that the final stage of the bootstrap process can produce the finished cross-compiler. This tool is a “native” binary that runs on the build host.linux-libc-headers
: Headers needed for the cross-compiler.glibc-initial
: An initial version of the Embedded GLIBC needed to bootstrapnativesdk-glibc
.nativesdk-glibc
: The Embedded GLIBC needed to bootstrap thegcc-crosssdk
.gcc-crosssdk
: The final stage of the bootstrap process for the relocatable cross-compiler. Thegcc-crosssdk
is a transitory compiler and never leaves the build host. Its purpose is to help in the bootstrap process to create the eventualgcc-cross-canadian
compiler, which is relocatable. This tool is also a “native” package (i.e. it is designed to run on the build host).gcc-cross-canadian
: The final relocatable cross-compiler. When run on the SDKMACHINE, this tool produces executable code that runs on the target device. Only one cross-canadian compiler is produced per architecture since they can be targeted at different processor optimizations using configurations passed to the compiler through the compile commands. This circumvents the need for multiple compilers and thus reduces the size of the toolchains.
Note
For information on advantages gained when building a cross-development toolchain installer, see the “Building an SDK Installer” appendix in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
4.6 Automatically Added Runtime Dependencies
The OpenEmbedded build system automatically adds common types of runtime
dependencies between packages, which means that you do not need to
explicitly declare the packages using
RDEPENDS. Three automatic
mechanisms exist (shlibdeps
, pcdeps
, and depchains
) that
handle shared libraries, package configuration (pkg-config) modules, and
-dev
and -dbg
packages, respectively. For other types of runtime
dependencies, you must manually declare the dependencies.
shlibdeps
: During the do_package task of each recipe, all shared libraries installed by the recipe are located. For each shared library, the package that contains the shared library is registered as providing the shared library. More specifically, the package is registered as providing the soname of the library. The resulting shared-library-to-package mapping is saved globally in PKGDATA_DIR by the do_packagedata task.Simultaneously, all executables and shared libraries installed by the recipe are inspected to see what shared libraries they link against. For each shared library dependency that is found,
PKGDATA_DIR
is queried to see if some package (likely from a different recipe) contains the shared library. If such a package is found, a runtime dependency is added from the package that depends on the shared library to the package that contains the library.The automatically added runtime dependency also includes a version restriction. This version restriction specifies that at least the current version of the package that provides the shared library must be used, as if “package (>= version)” had been added to
RDEPENDS
. This forces an upgrade of the package containing the shared library when installing the package that depends on the library, if needed.If you want to avoid a package being registered as providing a particular shared library (e.g. because the library is for internal use only), then add the library to PRIVATE_LIBS inside the package’s recipe.
pcdeps
: During thedo_package
task of each recipe, all pkg-config modules (*.pc
files) installed by the recipe are located. For each module, the package that contains the module is registered as providing the module. The resulting module-to-package mapping is saved globally inPKGDATA_DIR
by thedo_packagedata
task.Simultaneously, all pkg-config modules installed by the recipe are inspected to see what other pkg-config modules they depend on. A module is seen as depending on another module if it contains a “Requires:” line that specifies the other module. For each module dependency,
PKGDATA_DIR
is queried to see if some package contains the module. If such a package is found, a runtime dependency is added from the package that depends on the module to the package that contains the module.Note
The pcdeps mechanism most often infers dependencies between -dev packages.
depchains
: If a packagefoo
depends on a packagebar
, thenfoo-dev
andfoo-dbg
are also made to depend onbar-dev
andbar-dbg
, respectively. Taking the-dev
packages as an example, thebar-dev
package might provide headers and shared library symlinks needed byfoo-dev
, which shows the need for a dependency between the packages.The dependencies added by
depchains
are in the form of RRECOMMENDS.Note
By default,
foo-dev
also has anRDEPENDS
-style dependency onfoo
, because the default value ofRDEPENDS_${PN}-dev
(set in bitbake.conf) includes “${PN}”.To ensure that the dependency chain is never broken,
-dev
and-dbg
packages are always generated by default, even if the packages turn out to be empty. See the ALLOW_EMPTY variable for more information.
The do_package
task depends on the do_packagedata
task of each
recipe in DEPENDS through use
of a [
deptask]
declaration, which guarantees that the required
shared-library/module-to-package mapping information will be available
when needed as long as DEPENDS
has been correctly set.
4.7 Fakeroot and Pseudo
Some tasks are easier to implement when allowed to perform certain
operations that are normally reserved for the root user (e.g.
do_install,
do_package_write*,
do_rootfs, and
do_image*). For example,
the do_install
task benefits from being able to set the UID and GID
of installed files to arbitrary values.
One approach to allowing tasks to perform root-only operations would be to require BitBake to run as root. However, this method is cumbersome and has security issues. The approach that is actually used is to run tasks that benefit from root privileges in a “fake” root environment. Within this environment, the task and its child processes believe that they are running as the root user, and see an internally consistent view of the filesystem. As long as generating the final output (e.g. a package or an image) does not require root privileges, the fact that some earlier steps ran in a fake root environment does not cause problems.
The capability to run tasks in a fake root environment is known as “fakeroot”, which is derived from the BitBake keyword/variable flag that requests a fake root environment for a task.
In the OpenEmbedded Build System, the program that implements
fakeroot is known as Pseudo. Pseudo
overrides system calls by using the environment variable LD_PRELOAD
,
which results in the illusion of running as root. To keep track of
“fake” file ownership and permissions resulting from operations that
require root permissions, Pseudo uses an SQLite 3 database. This
database is stored in
${
WORKDIR}/pseudo/files.db
for individual recipes. Storing the database in a file as opposed to in
memory gives persistence between tasks and builds, which is not
accomplished using fakeroot.
Note
If you add your own task that manipulates the same files or
directories as a fakeroot task, then that task also needs to run
under fakeroot. Otherwise, the task cannot run root-only operations,
and cannot see the fake file ownership and permissions set by the
other task. You need to also add a dependency on
virtual/fakeroot-native:do_populate_sysroot
, giving the following:
fakeroot do_mytask () {
...
}
do_mytask[depends] += "virtual/fakeroot-native:do_populate_sysroot"
For more information, see the FAKEROOT* variables in the BitBake User Manual. You can also reference the “Why Not Fakeroot?” article for background information on Fakeroot and Pseudo.
5 Manual Revision History
Revision |
Date |
Note |
---|---|---|
2.5 |
May 2018 |
The initial document released with the Yocto Project 2.5 Release |
2.6 |
November 2018 |
Released with the Yocto Project 2.6 Release. |
2.7 |
May 2019 |
Released with the Yocto Project 2.7 Release. |
3.0 |
October 2019 |
Released with the Yocto Project 3.0 Release. |
3.1 |
April 2020 |
Released with the Yocto Project 3.1 Release. |
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Yocto Project Reference Manual
1 System Requirements
Welcome to the Yocto Project Reference Manual! This manual provides reference information for the current release of the Yocto Project, and is most effectively used after you have an understanding of the basics of the Yocto Project. The manual is neither meant to be read as a starting point to the Yocto Project, nor read from start to finish. Rather, use this manual to find variable definitions, class descriptions, and so forth as needed during the course of using the Yocto Project.
For introductory information on the Yocto Project, see the Yocto Project Website and the “The Yocto Project Development Environment” chapter in the Yocto Project Overview and Concepts Manual.
If you want to use the Yocto Project to quickly build an image without having to understand concepts, work through the Yocto Project Quick Build document. You can find “how-to” information in the Yocto Project Development Tasks Manual. You can find Yocto Project overview and conceptual information in the Yocto Project Overview and Concepts Manual.
Note
For more information about the Yocto Project Documentation set, see the Links and Related Documentation section.
1.1 Supported Linux Distributions
Currently, the Yocto Project is supported on the following distributions:
Ubuntu 16.04 (LTS)
Ubuntu 18.04 (LTS)
Ubuntu 20.04
Fedora 30
Fedora 31
Fedora 32
CentOS 7.x
CentOS 8.x
Debian GNU/Linux 8.x (Jessie)
Debian GNU/Linux 9.x (Stretch)
Debian GNU/Linux 10.x (Buster)
openSUSE Leap 15.1
Note
While the Yocto Project Team attempts to ensure all Yocto Project releases are one hundred percent compatible with each officially supported Linux distribution, instances might exist where you encounter a problem while using the Yocto Project on a specific distribution.
Yocto Project releases are tested against the stable Linux distributions in the above list. The Yocto Project should work on other distributions but validation is not performed against them.
In particular, the Yocto Project does not support and currently has no plans to support rolling-releases or development distributions due to their constantly changing nature. We welcome patches and bug reports, but keep in mind that our priority is on the supported platforms listed below.
You may use Windows Subsystem For Linux v2 to set up a build host using Windows 10, but validation is not performed against build hosts using WSLv2.
The Yocto Project is not compatible with WSLv1, it is compatible but not officially supported nor validated with WSLv2, if you still decide to use WSL please upgrade to WSLv2.
If you encounter problems, please go to Yocto Project Bugzilla and submit a bug. We are interested in hearing about your experience. For information on how to submit a bug, see the Yocto Project Bugzilla wiki page and the “Submitting a Defect Against the Yocto Project” section in the Yocto Project Development Tasks Manual.
1.2 Required Packages for the Build Host
The list of packages you need on the host development system can be large when covering all build scenarios using the Yocto Project. This section describes required packages according to Linux distribution and function.
1.2.1 Ubuntu and Debian
The following list shows the required packages by function given a supported Ubuntu or Debian Linux distribution:
Note
If your build system has the
oss4-dev
package installed, you might experience QEMU build failures due to the package installing its own custom/usr/include/linux/soundcard.h
on the Debian system. If you run into this situation, either of the following solutions exist:$ sudo apt-get build-dep qemu $ sudo apt-get remove oss4-dev
For Debian-8,
python3-git
andpylint3
are no longer available viaapt-get
.$ sudo pip3 install GitPython pylint==1.9.5
Essentials: Packages needed to build an image on a headless system:
$ sudo apt-get install gawk wget git diffstat unzip texinfo gcc build-essential chrpath socat cpio python3 python3-pip python3-pexpect xz-utils debianutils iputils-ping python3-git python3-jinja2 libegl1-mesa libsdl1.2-dev pylint3 xterm python3-subunit mesa-common-dev
Documentation: Packages needed if you are going to build out the Yocto Project documentation manuals:
$ sudo apt-get install make python3-pip $ sudo pip3 install sphinx sphinx_rtd_theme pyyaml
Note
It is currently not possible to build out documentation from Debian 8 (Jessie) because of outdated
pip3
andpython3
.python3-sphinx
is too outdated.
1.2.2 Fedora Packages
The following list shows the required packages by function given a supported Fedora Linux distribution:
Essentials: Packages needed to build an image for a headless system:
$ sudo dnf install gawk make wget tar bzip2 gzip python3 unzip perl patch diffutils diffstat git cpp gcc gcc-c++ glibc-devel texinfo chrpath ccache perl-Data-Dumper perl-Text-ParseWords perl-Thread-Queue perl-bignum socat python3-pexpect findutils which file cpio python python3-pip xz python3-GitPython python3-jinja2 SDL-devel xterm rpcgen mesa-libGL-devel
Documentation: Packages needed if you are going to build out the Yocto Project documentation manuals:
$ sudo dnf install make python3-pip which $ sudo pip3 install sphinx sphinx_rtd_theme pyyaml
1.2.3 openSUSE Packages
The following list shows the required packages by function given a supported openSUSE Linux distribution:
Essentials: Packages needed to build an image for a headless system:
$ sudo zypper install python gcc gcc-c++ git chrpath make wget python-xml diffstat makeinfo python-curses patch socat python3 python3-curses tar python3-pip python3-pexpect xz which python3-Jinja2 Mesa-libEGL1 libSDL-devel xterm rpcgen Mesa-dri-devel $ sudo pip3 install GitPython
Documentation: Packages needed if you are going to build out the Yocto Project documentation manuals:
$ sudo zypper install make python3-pip which $ sudo pip3 install sphinx sphinx_rtd_theme pyyaml
1.2.4 CentOS-7 Packages
The following list shows the required packages by function given a supported CentOS-7 Linux distribution:
Essentials: Packages needed to build an image for a headless system:
$ sudo yum install -y epel-release $ sudo yum makecache $ sudo yum install gawk make wget tar bzip2 gzip python3 unzip perl patch diffutils diffstat git cpp gcc gcc-c++ glibc-devel texinfo chrpath socat perl-Data-Dumper perl-Text-ParseWords perl-Thread-Queue python36-pip xz which SDL-devel xterm mesa-libGL-devel $ sudo pip3 install GitPython jinja2
Note
Extra Packages for Enterprise Linux (i.e.
epel-release
) is a collection of packages from Fedora built on RHEL/CentOS for easy installation of packages not included in enterprise Linux by default. You need to install these packages separately.The
makecache
command consumes additional Metadata fromepel-release
.
Documentation: Packages needed if you are going to build out the Yocto Project documentation manuals:
$ sudo yum install make python3-pip which $ sudo pip3 install sphinx sphinx_rtd_theme pyyaml
1.2.5 CentOS-8 Packages
The following list shows the required packages by function given a supported CentOS-8 Linux distribution:
Essentials: Packages needed to build an image for a headless system:
$ sudo dnf install -y epel-release $ sudo dnf config-manager --set-enabled PowerTools $ sudo dnf makecache $ sudo dnf install gawk make wget tar bzip2 gzip python3 unzip perl patch diffutils diffstat git cpp gcc gcc-c++ glibc-devel texinfo chrpath ccache socat perl-Data-Dumper perl-Text-ParseWords perl-Thread-Queue python3-pip python3-GitPython python3-jinja2 python3-pexpect xz which SDL-devel xterm rpcgen mesa-libGL-devel
Note
Extra Packages for Enterprise Linux (i.e.
epel-release
) is a collection of packages from Fedora built on RHEL/CentOS for easy installation of packages not included in enterprise Linux by default. You need to install these packages separately.The
PowerTools
repo provides additional packages such asrpcgen
andtexinfo
.The
makecache
command consumes additional Metadata fromepel-release
.
Documentation: Packages needed if you are going to build out the Yocto Project documentation manuals:
$ sudo dnf install make python3-pip which $ sudo pip3 install sphinx sphinx_rtd_theme pyyaml
1.3 Required Git, tar, Python and gcc Versions
In order to use the build system, your host development system must meet the following version requirements for Git, tar, and Python:
Git 1.8.3.1 or greater
tar 1.28 or greater
Python 3.6.0 or greater
If your host development system does not meet all these requirements,
you can resolve this by installing a buildtools
tarball that
contains these tools. You can get the tarball one of two ways: download
a pre-built tarball or use BitBake to build the tarball.
In addition, your host development system must meet the following version requirement for gcc:
gcc 5.0 or greater
If your host development system does not meet this requirement, you can
resolve this by installing a buildtools-extended
tarball that
contains additional tools, the equivalent of the Debian/Ubuntu build-essential
package.
In the sections that follow, three different methods will be described for
installing the buildtools
or buildtools-extended
toolset.
1.3.1 Installing a Pre-Built buildtools
Tarball with install-buildtools
script
The install-buildtools
script is the easiest of the three methods by
which you can get these tools. It downloads a pre-built buildtools
installer and automatically installs the tools for you:
Execute the
install-buildtools
script. Here is an example:$ cd poky $ scripts/install-buildtools --without-extended-buildtools \ --base-url https://downloads.yoctoproject.org/releases/yocto \ --release yocto-3.3.1 \ --installer-version 3.3.1
During execution, the buildtools tarball will be downloaded, the checksum of the download will be verified, the installer will be run for you, and some basic checks will be run to make sure the installation is functional.
To avoid the need of
sudo
privileges, theinstall-buildtools
script will by default tell the installer to install in:/path/to/poky/buildtools
If your host development system needs the additional tools provided in the
buildtools-extended
tarball, you can instead execute theinstall-buildtools
script with the default parameters:$ cd poky $ scripts/install-buildtools
Source the tools environment setup script by using a command like the following:
$ source /path/to/poky/buildtools/environment-setup-x86_64-pokysdk-linux
Of course, you need to supply your installation directory and be sure to use the right file (i.e. i586 or x86_64).
After you have sourced the setup script, the tools are added to
PATH
and any other environment variables required to run the tools are initialized. The results are working versions versions of Git, tar, Python andchrpath
. And in the case of thebuildtools-extended
tarball, additional working versions of tools includinggcc
,make
and the other tools included inpackagegroup-core-buildessential
.
1.3.2 Downloading a Pre-Built buildtools
Tarball
If you would prefer not to use the install-buildtools
script, you can instead
download and run a pre-built buildtools installer yourself with the following
steps:
Locate and download the
*.sh
at https://downloads.yoctoproject.org/releases/yocto/yocto-3.3.1/buildtools/Execute the installation script. Here is an example for the traditional installer:
$ sh ~/Downloads/x86_64-buildtools-nativesdk-standalone-3.3.1.sh
Here is an example for the extended installer:
$ sh ~/Downloads/x86_64-buildtools-extended-nativesdk-standalone-3.3.1.sh
During execution, a prompt appears that allows you to choose the installation directory. For example, you could choose the following:
/home/your-username/buildtools
Source the tools environment setup script by using a command like the following:
$ source /home/your_username/buildtools/environment-setup-i586-poky-linux
Of course, you need to supply your installation directory and be sure to use the right file (i.e. i585 or x86-64).
After you have sourced the setup script, the tools are added to
PATH
and any other environment variables required to run the tools are initialized. The results are working versions versions of Git, tar, Python andchrpath
. And in the case of thebuildtools-extended
tarball, additional working versions of tools includinggcc
,make
and the other tools included inpackagegroup-core-buildessential
.
1.3.3 Building Your Own buildtools
Tarball
Building and running your own buildtools installer applies only when you
have a build host that can already run BitBake. In this case, you use
that machine to build the .sh
file and then take steps to transfer
and run it on a machine that does not meet the minimal Git, tar, and
Python (or gcc) requirements.
Here are the steps to take to build and run your own buildtools installer:
On the machine that is able to run BitBake, be sure you have set up your build environment with the setup script (oe-init-build-env).
Run the BitBake command to build the tarball:
$ bitbake buildtools-tarball
or run the BitBake command to build the extended tarball:
$ bitbake buildtools-extended-tarball
Note
The SDKMACHINE variable in your
local.conf
file determines whether you build tools for a 32-bit or 64-bit system.Once the build completes, you can find the
.sh
file that installs the tools in thetmp/deploy/sdk
subdirectory of the Build Directory. The installer file has the string “buildtools” (or “buildtools-extended”) in the name.Transfer the
.sh
file from the build host to the machine that does not meet the Git, tar, or Python (or gcc) requirements.On the machine that does not meet the requirements, run the
.sh
file to install the tools. Here is an example for the traditional installer:$ sh ~/Downloads/x86_64-buildtools-nativesdk-standalone-3.3.1.sh
Here is an example for the extended installer:
$ sh ~/Downloads/x86_64-buildtools-extended-nativesdk-standalone-3.3.1.sh
During execution, a prompt appears that allows you to choose the installation directory. For example, you could choose the following:
/home/your_username/buildtools
Source the tools environment setup script by using a command like the following:
$ source /home/your_username/buildtools/environment-setup-x86_64-poky-linux
Of course, you need to supply your installation directory and be sure to use the right file (i.e. i586 or x86_64).
After you have sourced the setup script, the tools are added to
PATH
and any other environment variables required to run the tools are initialized. The results are working versions versions of Git, tar, Python andchrpath
. And in the case of thebuildtools-extended
tarball, additional working versions of tools includinggcc
,make
and the other tools included inpackagegroup-core-buildessential
.
2 Yocto Project Terms
Following is a list of terms and definitions users new to the Yocto Project development environment might find helpful. While some of these terms are universal, the list includes them just in case:
- Append Files
Files that append build information to a recipe file. Append files are known as BitBake append files and
.bbappend
files. The OpenEmbedded build system expects every append file to have a corresponding recipe (.bb
) file. Furthermore, the append file and corresponding recipe file must use the same root filename. The filenames can differ only in the file type suffix used (e.g.formfactor_0.0.bb
andformfactor_0.0.bbappend
).Information in append files extends or overrides the information in the similarly-named recipe file. For an example of an append file in use, see the “Using .bbappend Files in Your Layer” section in the Yocto Project Development Tasks Manual.
When you name an append file, you can use the “
%
” wildcard character to allow for matching recipe names. For example, suppose you have an append file named as follows:busybox_1.21.%.bbappend
That append file would match any
busybox_1.21.
x.bb
version of the recipe. So, the append file would match any of the following recipe names:busybox_1.21.1.bb busybox_1.21.2.bb busybox_1.21.3.bb busybox_1.21.10.bb busybox_1.21.25.bb
Note
The use of the “%” character is limited in that it only works directly in front of the .bbappend portion of the append file’s name. You cannot use the wildcard character in any other location of the name.
- BitBake
The task executor and scheduler used by the OpenEmbedded build system to build images. For more information on BitBake, see the BitBake User Manual.
- Board Support Package (BSP)
A group of drivers, definitions, and other components that provide support for a specific hardware configuration. For more information on BSPs, see the Yocto Project Board Support Package Developer’s Guide.
- Build Directory
This term refers to the area used by the OpenEmbedded build system for builds. The area is created when you
source
the setup environment script that is found in the Source Directory (i.e. oe-init-build-env). The TOPDIR variable points to the Build Directory.You have a lot of flexibility when creating the Build Directory. Following are some examples that show how to create the directory. The examples assume your Source Directory is named
poky
:Create the Build Directory inside your Source Directory and let the name of the Build Directory default to
build
:$ cd poky $ source oe-init-build-env
Create the Build Directory inside your home directory and specifically name it
test-builds
:$ source poky/oe-init-build-env test-builds
Provide a directory path and specifically name the Build Directory. Any intermediate folders in the pathname must exist. This next example creates a Build Directory named
YP-25.0.1
within the existing directorymybuilds
:$ source poky/oe-init-build-env mybuilds/YP-25.0.1
Note
By default, the Build Directory contains TMPDIR, which is a temporary directory the build system uses for its work.
TMPDIR
cannot be under NFS. Thus, by default, the Build Directory cannot be under NFS. However, if you need the Build Directory to be under NFS, you can set this up by settingTMPDIR
in yourlocal.conf
file to use a local drive. Doing so effectively separatesTMPDIR
from TOPDIR, which is the Build Directory.- Build Host
The system used to build images in a Yocto Project Development environment. The build system is sometimes referred to as the development host.
- Classes
Files that provide for logic encapsulation and inheritance so that commonly used patterns can be defined once and then easily used in multiple recipes. For reference information on the Yocto Project classes, see the “Classes” chapter. Class files end with the
.bbclass
filename extension.- Configuration File
Files that hold global definitions of variables, user-defined variables, and hardware configuration information. These files tell the OpenEmbedded build system what to build and what to put into the image to support a particular platform.
Configuration files end with a
.conf
filename extension. Theconf/local.conf
configuration file in the Build Directory contains user-defined variables that affect every build. Themeta-poky/conf/distro/poky.conf
configuration file defines Yocto “distro” configuration variables used only when building with this policy. Machine configuration files, which are located throughout the Source Directory, define variables for specific hardware and are only used when building for that target (e.g. themachine/beaglebone.conf
configuration file defines variables for the Texas Instruments ARM Cortex-A8 development board).- Container Layer
Layers that hold other layers. An example of a container layer is OpenEmbedded’s meta-openembedded layer. The
meta-openembedded
layer contains manymeta-*
layers.- Cross-Development Toolchain
In general, a cross-development toolchain is a collection of software development tools and utilities that run on one architecture and allow you to develop software for a different, or targeted, architecture. These toolchains contain cross-compilers, linkers, and debuggers that are specific to the target architecture.
The Yocto Project supports two different cross-development toolchains:
A toolchain only used by and within BitBake when building an image for a target architecture.
A relocatable toolchain used outside of BitBake by developers when developing applications that will run on a targeted device.
Creation of these toolchains is simple and automated. For information on toolchain concepts as they apply to the Yocto Project, see the “Cross-Development Toolchain Generation” section in the Yocto Project Overview and Concepts Manual. You can also find more information on using the relocatable toolchain in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
- Extensible Software Development Kit (eSDK)
A custom SDK for application developers. This eSDK allows developers to incorporate their library and programming changes back into the image to make their code available to other application developers.
For information on the eSDK, see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
- Image
An image is an artifact of the BitBake build process given a collection of recipes and related Metadata. Images are the binary output that run on specific hardware or QEMU and are used for specific use-cases. For a list of the supported image types that the Yocto Project provides, see the “Images” chapter.
- Layer
A collection of related recipes. Layers allow you to consolidate related metadata to customize your build. Layers also isolate information used when building for multiple architectures. Layers are hierarchical in their ability to override previous specifications. You can include any number of available layers from the Yocto Project and customize the build by adding your layers after them. You can search the Layer Index for layers used within Yocto Project.
For introductory information on layers, see the “The Yocto Project Layer Model” section in the Yocto Project Overview and Concepts Manual. For more detailed information on layers, see the “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual. For a discussion specifically on BSP Layers, see the “BSP Layers” section in the Yocto Project Board Support Packages (BSP) Developer’s Guide.
- Metadata
A key element of the Yocto Project is the Metadata that is used to construct a Linux distribution and is contained in the files that the OpenEmbedded Build System parses when building an image. In general, Metadata includes recipes, configuration files, and other information that refers to the build instructions themselves, as well as the data used to control what things get built and the effects of the build. Metadata also includes commands and data used to indicate what versions of software are used, from where they are obtained, and changes or additions to the software itself (patches or auxiliary files) that are used to fix bugs or customize the software for use in a particular situation. OpenEmbedded-Core is an important set of validated metadata.
In the context of the kernel (“kernel Metadata”), the term refers to the kernel config fragments and features contained in the yocto-kernel-cache Git repository.
- OpenEmbedded-Core (OE-Core)
OE-Core is metadata comprised of foundational recipes, classes, and associated files that are meant to be common among many different OpenEmbedded-derived systems, including the Yocto Project. OE-Core is a curated subset of an original repository developed by the OpenEmbedded community that has been pared down into a smaller, core set of continuously validated recipes. The result is a tightly controlled and an quality-assured core set of recipes.
You can see the Metadata in the
meta
directory of the Yocto Project Source Repositories.- OpenEmbedded Build System
The build system specific to the Yocto Project. The OpenEmbedded build system is based on another project known as “Poky”, which uses BitBake as the task executor. Throughout the Yocto Project documentation set, the OpenEmbedded build system is sometimes referred to simply as “the build system”. If other build systems, such as a host or target build system are referenced, the documentation clearly states the difference.
Note
For some historical information about Poky, see the Poky term.
- Package
In the context of the Yocto Project, this term refers to a recipe’s packaged output produced by BitBake (i.e. a “baked recipe”). A package is generally the compiled binaries produced from the recipe’s sources. You “bake” something by running it through BitBake.
It is worth noting that the term “package” can, in general, have subtle meanings. For example, the packages referred to in the “Required Packages for the Build Host” section are compiled binaries that, when installed, add functionality to your Linux distribution.
Another point worth noting is that historically within the Yocto Project, recipes were referred to as packages - thus, the existence of several BitBake variables that are seemingly mis-named, (e.g. PR, PV, and PE).
- Package Groups
Arbitrary groups of software Recipes. You use package groups to hold recipes that, when built, usually accomplish a single task. For example, a package group could contain the recipes for a company’s proprietary or value-add software. Or, the package group could contain the recipes that enable graphics. A package group is really just another recipe. Because package group files are recipes, they end with the
.bb
filename extension.- Poky
Poky, which is pronounced Pock-ee, is a reference embedded distribution and a reference test configuration. Poky provides the following:
A base-level functional distro used to illustrate how to customize a distribution.
A means by which to test the Yocto Project components (i.e. Poky is used to validate the Yocto Project).
A vehicle through which you can download the Yocto Project.
Poky is not a product level distro. Rather, it is a good starting point for customization.
Note
Poky began as an open-source project initially developed by OpenedHand. OpenedHand developed Poky from the existing OpenEmbedded build system to create a commercially supportable build system for embedded Linux. After Intel Corporation acquired OpenedHand, the poky project became the basis for the Yocto Project’s build system.
- Recipe
A set of instructions for building packages. A recipe describes where you get source code, which patches to apply, how to configure the source, how to compile it and so on. Recipes also describe dependencies for libraries or for other recipes. Recipes represent the logical unit of execution, the software to build, the images to build, and use the
.bb
file extension.- Reference Kit
A working example of a system, which includes a BSP as well as a build host and other components, that can work on specific hardware.
- Source Directory
This term refers to the directory structure created as a result of creating a local copy of the
poky
Git repositorygit://git.yoctoproject.org/poky
or expanding a releasedpoky
tarball.Note
Creating a local copy of the poky Git repository is the recommended method for setting up your Source Directory.
Sometimes you might hear the term “poky directory” used to refer to this directory structure.
Note
The OpenEmbedded build system does not support file or directory names that contain spaces. Be sure that the Source Directory you use does not contain these types of names.
The Source Directory contains BitBake, Documentation, Metadata and other files that all support the Yocto Project. Consequently, you must have the Source Directory in place on your development system in order to do any development using the Yocto Project.
When you create a local copy of the Git repository, you can name the repository anything you like. Throughout much of the documentation, “poky” is used as the name of the top-level folder of the local copy of the poky Git repository. So, for example, cloning the
poky
Git repository results in a local Git repository whose top-level folder is also named “poky”.While it is not recommended that you use tarball expansion to set up the Source Directory, if you do, the top-level directory name of the Source Directory is derived from the Yocto Project release tarball. For example, downloading and unpacking https://downloads.yoctoproject.org/releases/yocto/yocto-3.3.1/poky-hardknott-25.0.1.tar.bz2 results in a Source Directory whose root folder is named
poky-hardknott-25.0.1
.It is important to understand the differences between the Source Directory created by unpacking a released tarball as compared to cloning
git://git.yoctoproject.org/poky
. When you unpack a tarball, you have an exact copy of the files based on the time of release - a fixed release point. Any changes you make to your local files in the Source Directory are on top of the release and will remain local only. On the other hand, when you clone thepoky
Git repository, you have an active development repository with access to the upstream repository’s branches and tags. In this case, any local changes you make to the local Source Directory can be later applied to active development branches of the upstreampoky
Git repository.For more information on concepts related to Git repositories, branches, and tags, see the “Repositories, Tags, and Branches” section in the Yocto Project Overview and Concepts Manual.
- Task
A unit of execution for BitBake (e.g. do_compile, do_fetch, do_patch, and so forth).
- Toaster
A web interface to the Yocto Project’s OpenEmbedded Build System. The interface enables you to configure and run your builds. Information about builds is collected and stored in a database. For information on Toaster, see the Toaster User Manual.
- Upstream
A reference to source code or repositories that are not local to the development system but located in a master area that is controlled by the maintainer of the source code. For example, in order for a developer to work on a particular piece of code, they need to first get a copy of it from an “upstream” source.
3 Yocto Project Releases and the Stable Release Process
The Yocto Project release process is predictable and consists of both major and minor (point) releases. This brief chapter provides information on how releases are named, their life cycle, and their stability.
3.1 Major and Minor Release Cadence
The Yocto Project delivers major releases (e.g. 3.3.1) using a six month cadence roughly timed each April and October of the year. Following are examples of some major YP releases with their codenames also shown. See the “Major Release Codenames” section for information on codenames used with major releases.
2.2 (Morty)
2.1 (Krogoth)
2.0 (Jethro)
While the cadence is never perfect, this timescale facilitates regular releases that have strong QA cycles while not overwhelming users with too many new releases. The cadence is predictable and avoids many major holidays in various geographies.
The Yocto project delivers minor (point) releases on an unscheduled basis and are usually driven by the accumulation of enough significant fixes or enhancements to the associated major release. Following are some example past point releases:
2.1.1
2.1.2
2.2.1
The point release indicates a point in the major release branch where a full QA cycle and release process validates the content of the new branch.
Note
Realize that there can be patches merged onto the stable release branches as and when they become available.
3.2 Major Release Codenames
Each major release receives a codename that identifies the release in the Yocto Project Source Repositories. The concept is that branches of Metadata with the same codename are likely to be compatible and thus work together.
Note
Codenames are associated with major releases because a Yocto Project release number (e.g. 3.3.1) could conflict with a given layer or company versioning scheme. Codenames are unique, interesting, and easily identifiable.
Releases are given a nominal release version as well but the codename is used in repositories for this reason. You can find information on Yocto Project releases and codenames at https://wiki.yoctoproject.org/wiki/Releases.
3.3 Stable Release Process
Once released, the release enters the stable release process at which time a person is assigned as the maintainer for that stable release. This maintainer monitors activity for the release by investigating and handling nominated patches and backport activity. Only fixes and enhancements that have first been applied on the “master” branch (i.e. the current, in-development branch) are considered for backporting to a stable release.
Note
The current Yocto Project policy regarding backporting is to consider bug fixes and security fixes only. Policy dictates that features are not backported to a stable release. This policy means generic recipe version upgrades are unlikely to be accepted for backporting. The exception to this policy occurs when a strong reason exists such as the fix happens to also be the preferred upstream approach.
Stable release branches have strong maintenance for about a year after their initial release. Should significant issues be found for any release regardless of its age, fixes could be backported to older releases. For issues that are not backported given an older release, Community LTS trees and branches exist where community members share patches for older releases. However, these types of patches do not go through the same release process as do point releases. You can find more information about stable branch maintenance at https://wiki.yoctoproject.org/wiki/Stable_branch_maintenance.
3.4 Testing and Quality Assurance
Part of the Yocto Project development and release process is quality assurance through the execution of test strategies. Test strategies provide the Yocto Project team a way to ensure a release is validated. Additionally, because the test strategies are visible to you as a developer, you can validate your projects. This section overviews the available test infrastructure used in the Yocto Project. For information on how to run available tests on your projects, see the “Performing Automated Runtime Testing” section in the Yocto Project Development Tasks Manual.
The QA/testing infrastructure is woven into the project to the point where core developers take some of it for granted. The infrastructure consists of the following pieces:
bitbake-selftest
: A standalone command that runs unit tests on key pieces of BitBake and its fetchers.sanity.bbclass: This automatically included class checks the build environment for missing tools (e.g.
gcc
) or common misconfigurations such as MACHINE set incorrectly.insane.bbclass: This class checks the generated output from builds for sanity. For example, if building for an ARM target, did the build produce ARM binaries. If, for example, the build produced PPC binaries then there is a problem.
testimage.bbclass: This class performs runtime testing of images after they are built. The tests are usually used with QEMU to boot the images and check the combined runtime result boot operation and functions. However, the test can also use the IP address of a machine to test.
ptest: Runs tests against packages produced during the build for a given piece of software. The test allows the packages to be run within a target image.
oe-selftest
: Tests combination BitBake invocations. These tests operate outside the OpenEmbedded build system itself. Theoe-selftest
can run all tests by default or can run selected tests or test suites.Note
Running
oe-selftest
requires host packages beyond the “Essential” grouping. See the Required Packages for the Build Host section for more information.
Originally, much of this testing was done manually. However, significant effort has been made to automate the tests so that more people can use them and the Yocto Project development team can run them faster and more efficiently.
The Yocto Project’s main Autobuilder (https://autobuilder.yoctoproject.org)
publicly tests each Yocto Project release’s code in the
OpenEmbedded-Core (OE-Core), Poky, and BitBake repositories. The testing
occurs for both the current state of the “master” branch and also for
submitted patches. Testing for submitted patches usually occurs in the
“ross/mut” branch in the poky-contrib
repository (i.e. the
master-under-test branch) or in the “master-next” branch in the poky
repository.
Note
You can find all these branches in the Yocto Project Source Repositories .
Testing within these public branches ensures in a publicly visible way that all of the main supposed architectures and recipes in OE-Core successfully build and behave properly.
Various features such as multilib
, sub architectures (e.g. x32
,
poky-tiny
, musl
, no-x11
and and so forth),
bitbake-selftest
, and oe-selftest
are tested as part of the QA
process of a release. Complete testing and validation for a release
takes the Autobuilder workers several hours.
Note
The Autobuilder workers are non-homogeneous, which means regular testing across a variety of Linux distributions occurs. The Autobuilder is limited to only testing QEMU-based setups and not real hardware.
Finally, in addition to the Autobuilder’s tests, the Yocto Project QA team also performs testing on a variety of platforms, which includes actual hardware, to ensure expected results.
4 Migrating to a Newer Yocto Project Release
This chapter provides information you can use to migrate work to a newer Yocto Project release. You can find the same information in the release notes for a given release.
4.1 General Migration Considerations
Some considerations are not tied to a specific Yocto Project release. This section presents information you should consider when migrating to any new Yocto Project release.
Dealing with Customized Recipes:
Issues could arise if you take older recipes that contain customizations and simply copy them forward expecting them to work after you migrate to new Yocto Project metadata. For example, suppose you have a recipe in your layer that is a customized version of a core recipe copied from the earlier release, rather than through the use of an append file. When you migrate to a newer version of Yocto Project, the metadata (e.g. perhaps an include file used by the recipe) could have changed in a way that would break the build. Say, for example, a function is removed from an include file and the customized recipe tries to call that function.
You could “forward-port” all your customizations in your recipe so that everything works for the new release. However, this is not the optimal solution as you would have to repeat this process with each new release if changes occur that give rise to problems.
The better solution (where practical) is to use append files (
*.bbappend
) to capture any customizations you want to make to a recipe. Doing so, isolates your changes from the main recipe making them much more manageable. However, sometimes it is not practical to use an append file. A good example of this is when introducing a newer or older version of a recipe in another layer.Updating Append Files:
Since append files generally only contain your customizations, they often do not need to be adjusted for new releases. However, if the
.bbappend
file is specific to a particular version of the recipe (i.e. its name does not use the % wildcard) and the version of the recipe to which it is appending has changed, then you will at a minimum need to rename the append file to match the name of the recipe file. A mismatch between an append file and its corresponding recipe file (.bb
) will trigger an error during parsing.Depending on the type of customization the append file applies, other incompatibilities might occur when you upgrade. For example, if your append file applies a patch and the recipe to which it is appending is updated to a newer version, the patch might no longer apply. If this is the case and assuming the patch is still needed, you must modify the patch file so that it does apply.
4.2 Moving to the Yocto Project 1.3 Release (danny)
This section provides migration information for moving to the Yocto Project 1.3 Release (codename “danny”) from the prior release.
4.2.1 Local Configuration
Differences include changes for
SSTATE_MIRRORS and bblayers.conf
.
4.2.1.1 SSTATE_MIRRORS
The shared state cache (sstate-cache), as pointed to by SSTATE_DIR, by default now has two-character subdirectories to prevent issues arising from too many files in the same directory. Also, native sstate-cache packages, which are built to run on the host system, will go into a subdirectory named using the distro ID string. If you copy the newly structured sstate-cache to a mirror location (either local or remote) and then point to it in SSTATE_MIRRORS, you need to append “PATH” to the end of the mirror URL so that the path used by BitBake before the mirror substitution is appended to the path used to access the mirror. Here is an example:
SSTATE_MIRRORS = "file://.* http://someserver.tld/share/sstate/PATH"
4.2.1.2 bblayers.conf
The meta-yocto
layer consists of two parts that correspond to the
Poky reference distribution and the reference hardware Board Support
Packages (BSPs), respectively: meta-yocto
and meta-yocto-bsp
.
When running BitBake for the first time after upgrading, your
conf/bblayers.conf
file will be updated to handle this change and
you will be asked to re-run or restart for the changes to take effect.
4.2.2 Recipes
Differences include changes for the following:
4.2.2.1 Python Function Whitespace
All Python functions must now use four spaces for indentation.
Previously, an inconsistent mix of spaces and tabs existed, which made
extending these functions using _append
or _prepend
complicated
given that Python treats whitespace as syntactically significant. If you
are defining or extending any Python functions (e.g.
populate_packages
, do_unpack
, do_patch
and so forth) in
custom recipes or classes, you need to ensure you are using consistent
four-space indentation.
4.2.2.2 proto= in SRC_URI
Any use of proto=
in SRC_URI needs to be
changed to protocol=
. In particular, this applies to the following
URIs:
svn://
bzr://
hg://
osc://
Other URIs were already using protocol=
. This change improves
consistency.
4.2.2.3 nativesdk
The suffix nativesdk
is now implemented as a prefix, which
simplifies a lot of the packaging code for nativesdk
recipes. All
custom nativesdk
recipes, which are relocatable packages that are
native to SDK_ARCH, and any references need to
be updated to use nativesdk-*
instead of *-nativesdk
.
4.2.2.4 Task Recipes
“Task” recipes are now known as “Package groups” and have been renamed
from task-*.bb
to packagegroup-*.bb
. Existing references to the
previous task-*
names should work in most cases as there is an
automatic upgrade path for most packages. However, you should update
references in your own recipes and configurations as they could be
removed in future releases. You should also rename any custom task-*
recipes to packagegroup-*
, and change them to inherit
packagegroup
instead of task
, as well as taking the opportunity
to remove anything now handled by packagegroup.bbclass
, such as
providing -dev
and -dbg
packages, setting
LIC_FILES_CHKSUM, and so forth. See the
“packagegroup.bbclass” section for
further details.
4.2.2.5 IMAGE_FEATURES
Image recipes that previously included apps-console-core
in
IMAGE_FEATURES should now include splash
instead to enable the boot-up splash screen. Retaining
apps-console-core
will still include the splash screen but generates a
warning. The apps-x11-core
and apps-x11-games
IMAGE_FEATURES
features have been removed.
4.2.2.6 Removed Recipes
The following recipes have been removed. For most of them, it is unlikely that you would have any references to them in your own Metadata. However, you should check your metadata against this list to be sure:
libx11-trim
: Replaced bylibx11
, which has a negligible size difference with modern Xorg.xserver-xorg-lite
: Usexserver-xorg
, which has a negligible size difference when DRI and GLX modules are not installed.xserver-kdrive
: Effectively unmaintained for many years.mesa-xlib
: No longer serves any purpose.galago
: Replaced by telepathy.gail
: Functionality was integrated into GTK+ 2.13.eggdbus
: No longer needed.gcc-*-intermediate
: The build has been restructured to avoid the need for this step.libgsmd
: Unmaintained for many years. Functionality now provided byofono
instead.contacts, dates, tasks, eds-tools: Largely unmaintained PIM application suite. It has been moved to
meta-gnome
inmeta-openembedded
.
In addition to the previously listed changes, the meta-demoapps
directory has also been removed because the recipes in it were not being
maintained and many had become obsolete or broken. Additionally, these
recipes were not parsed in the default configuration. Many of these
recipes are already provided in an updated and maintained form within
the OpenEmbedded community layers such as meta-oe
and
meta-gnome
. For the remainder, you can now find them in the
meta-extras
repository, which is in the
Source Repositories at
https://git.yoctoproject.org/cgit/cgit.cgi/meta-extras/.
4.2.3 Linux Kernel Naming
The naming scheme for kernel output binaries has been changed to now include PE as part of the filename:
KERNEL_IMAGE_BASE_NAME ?= "${KERNEL_IMAGETYPE}-${PE}-${PV}-${PR}-${MACHINE}-${DATETIME}"
Because the PE
variable is not set by default, these binary files
could result with names that include two dash characters. Here is an
example:
bzImage--3.10.9+git0+cd502a8814_7144bcc4b8-r0-qemux86-64-20130830085431.bin
4.3 Moving to the Yocto Project 1.4 Release (dylan)
This section provides migration information for moving to the Yocto Project 1.4 Release (codename “dylan”) from the prior release.
4.3.1 BitBake
Differences include the following:
Comment Continuation: If a comment ends with a line continuation (\) character, then the next line must also be a comment. Any instance where this is not the case, now triggers a warning. You must either remove the continuation character, or be sure the next line is a comment.
Package Name Overrides: The runtime package specific variables RDEPENDS, RRECOMMENDS, RSUGGESTS, RPROVIDES, RCONFLICTS, RREPLACES, FILES, ALLOW_EMPTY, and the pre, post, install, and uninstall script functions
pkg_preinst
,pkg_postinst
,pkg_prerm
, andpkg_postrm
should always have a package name override. For example, useRDEPENDS_${PN}
for the main package instead ofRDEPENDS
. BitBake uses more strict checks when it parses recipes.
4.3.2 Build Behavior
Differences include the following:
Shared State Code: The shared state code has been optimized to avoid running unnecessary tasks. For example, the following no longer populates the target sysroot since that is not necessary:
$ bitbake -c rootfs some-image
Instead, the system just needs to extract the output package contents, re-create the packages, and construct the root filesystem. This change is unlikely to cause any problems unless you have missing declared dependencies.
Scanning Directory Names: When scanning for files in SRC_URI, the build system now uses FILESOVERRIDES instead of OVERRIDES for the directory names. In general, the values previously in
OVERRIDES
are now inFILESOVERRIDES
as well. However, if you relied upon an additional value you previously added toOVERRIDES
, you might now need to add it toFILESOVERRIDES
unless you are already adding it through the MACHINEOVERRIDES or DISTROOVERRIDES variables, as appropriate. For more related changes, see the “Variables” section.
4.3.3 Proxies and Fetching Source
A new oe-git-proxy
script has been added to replace previous methods
of handling proxies and fetching source from Git. See the
meta-yocto/conf/site.conf.sample
file for information on how to use
this script.
4.3.4 Custom Interfaces File (netbase change)
If you have created your own custom etc/network/interfaces
file by
creating an append file for the netbase
recipe, you now need to
create an append file for the init-ifupdown
recipe instead, which
you can find in the Source Directory at
meta/recipes-core/init-ifupdown
. For information on how to use
append files, see the
“Using .bbappend Files in Your Layer”
section in the Yocto Project Development Tasks Manual.
4.3.5 Remote Debugging
Support for remote debugging with the Eclipse IDE is now separated into
an image feature (eclipse-debug
) that corresponds to the
packagegroup-core-eclipse-debug
package group. Previously, the
debugging feature was included through the tools-debug
image
feature, which corresponds to the packagegroup-core-tools-debug
package group.
4.3.6 Variables
The following variables have changed:
SANITY_TESTED_DISTROS
: This variable now uses a distribution ID, which is composed of the host distributor ID followed by the release. Previously, SANITY_TESTED_DISTROS was composed of the description field. For example, “Ubuntu 12.10” becomes “Ubuntu-12.10”. You do not need to worry about this change if you are not specifically setting this variable, or if you are specifically setting it to “”.SRC_URI
: The${
PN}
,${
PF}
,${
P}
, andFILE_DIRNAME
directories have been dropped from the default value of the FILESPATH variable, which is used as the search path for finding files referred to in SRC_URI. If you have a recipe that relied upon these directories, which would be unusual, then you will need to add the appropriate paths within the recipe or, alternatively, rearrange the files. The most common locations are still covered by${
BP}
,${
BPN}
, and “files”, which all remain in the default value of FILESPATH.
4.3.7 Target Package Management with RPM
If runtime package management is enabled and the RPM backend is selected, Smart is now installed for package download, dependency resolution, and upgrades instead of Zypper. For more information on how to use Smart, run the following command on the target:
smart --help
4.3.8 Recipes Moved
The following recipes were moved from their previous locations because they are no longer used by anything in the OpenEmbedded-Core:
clutter-box2d
: Now resides in themeta-oe
layer.evolution-data-server
: Now resides in themeta-gnome
layer.gthumb
: Now resides in themeta-gnome
layer.gtkhtml2
: Now resides in themeta-oe
layer.gupnp
: Now resides in themeta-multimedia
layer.gypsy
: Now resides in themeta-oe
layer.libcanberra
: Now resides in themeta-gnome
layer.libgdata
: Now resides in themeta-gnome
layer.libmusicbrainz
: Now resides in themeta-multimedia
layer.metacity
: Now resides in themeta-gnome
layer.polkit
: Now resides in themeta-oe
layer.zeroconf
: Now resides in themeta-networking
layer.
4.3.9 Removals and Renames
The following list shows what has been removed or renamed:
evieext
: Removed because it has been removed fromxserver
since 2008.Gtk+ DirectFB: Removed support because upstream Gtk+ no longer supports it as of version 2.18.
libxfontcache / xfontcacheproto
: Removed because they were removed from the Xorg server in 2008.libxp / libxprintapputil / libxprintutil / printproto
: Removed because the XPrint server was removed from Xorg in 2008.libxtrap / xtrapproto
: Removed because their functionality was broken upstream.linux-yocto 3.0 kernel: Removed with linux-yocto 3.8 kernel being added. The linux-yocto 3.2 and linux-yocto 3.4 kernels remain as part of the release.
lsbsetup
: Removed with functionality now provided bylsbtest
.matchbox-stroke
: Removed because it was never more than a proof-of-concept.matchbox-wm-2 / matchbox-theme-sato-2
: Removed because they are not maintained. However,matchbox-wm
andmatchbox-theme-sato
are still provided.mesa-dri
: Renamed tomesa
.mesa-xlib
: Removed because it was no longer useful.mutter
: Removed because nothing ever uses it and the recipe is very old.orinoco-conf
: Removed because it has become obsolete.update-modules
: Removed because it is no longer used. The kernel modulepostinstall
andpostrm
scripts can now do the same task without the use of this script.web
: Removed because it is not maintained. Superseded byweb-webkit
.xf86bigfontproto
: Removed because upstream it has been disabled by default since 2007. Nothing usesxf86bigfontproto
.xf86rushproto
: Removed because its dependency inxserver
was spurious and it was removed in 2005.zypper / libzypp / sat-solver
: Removed and been functionally replaced with Smart (python-smartpm
) when RPM packaging is used and package management is enabled on the target.
4.4 Moving to the Yocto Project 1.5 Release (dora)
This section provides migration information for moving to the Yocto Project 1.5 Release (codename “dora”) from the prior release.
4.4.1 Host Dependency Changes
The OpenEmbedded build system now has some additional requirements on the host system:
Python 2.7.3+
Tar 1.24+
Git 1.7.8+
Patched version of Make if you are using 3.82. Most distributions that provide Make 3.82 use the patched version.
If the Linux distribution you are using on your build host does not provide packages for these, you can install and use the Buildtools tarball, which provides an SDK-like environment containing them.
For more information on this requirement, see the “Required Git, tar, Python and gcc Versions” section.
4.4.2 atom-pc
Board Support Package (BSP)
The atom-pc
hardware reference BSP has been replaced by a
genericx86
BSP. This BSP is not necessarily guaranteed to work on
all x86 hardware, but it will run on a wider range of systems than the
atom-pc
did.
Note
Additionally, a genericx86-64
BSP has been added for 64-bit Atom
systems.
4.4.3 BitBake
The following changes have been made that relate to BitBake:
BitBake now supports a
_remove
operator. The addition of this operator means you will have to rename any items in recipe space (functions, variables) whose names currently contain_remove_
or end with_remove
to avoid unexpected behavior.BitBake’s global method pool has been removed. This method is not particularly useful and led to clashes between recipes containing functions that had the same name.
The “none” server backend has been removed. The “process” server backend has been serving well as the default for a long time now.
The
bitbake-runtask
script has been removed.${
P}
and${
PF}
are no longer added to PROVIDES by default inbitbake.conf
. These version-specificPROVIDES
items were seldom used. Attempting to use them could result in two versions being built simultaneously rather than just one version due to the way BitBake resolves dependencies.
4.4.4 QA Warnings
The following changes have been made to the package QA checks:
If you have customized ERROR_QA or WARN_QA values in your configuration, check that they contain all of the issues that you wish to be reported. Previous Yocto Project versions contained a bug that meant that any item not mentioned in
ERROR_QA
orWARN_QA
would be treated as a warning. Consequently, several important items were not already in the default value ofWARN_QA
. All of the possible QA checks are now documented in the “insane.bbclass” section.An additional QA check has been added to check if
/usr/share/info/dir
is being installed. Your recipe should delete this file within do_install if “make install” is installing it.If you are using the
buildhistory
class, the check for the package version going backwards is now controlled using a standard QA check. Thus, if you have customized yourERROR_QA
orWARN_QA
values and still wish to have this check performed, you should add “version-going-backwards” to your value for one or the other variables depending on how you wish it to be handled. See the documented QA checks in the “insane.bbclass” section.
4.4.5 Directory Layout Changes
The following directory changes exist:
Output SDK installer files are now named to include the image name and tuning architecture through the SDK_NAME variable.
Images and related files are now installed into a directory that is specific to the machine, instead of a parent directory containing output files for multiple machines. The DEPLOY_DIR_IMAGE variable continues to point to the directory containing images for the current MACHINE and should be used anywhere there is a need to refer to this directory. The
runqemu
script now uses this variable to find images and kernel binaries and will use BitBake to determine the directory. Alternatively, you can set theDEPLOY_DIR_IMAGE
variable in the external environment.When buildhistory is enabled, its output is now written under the Build Directory rather than TMPDIR. Doing so makes it easier to delete
TMPDIR
and preserve the build history. Additionally, data for produced SDKs is now split by IMAGE_NAME.The
pkgdata
directory produced as part of the packaging process has been collapsed into a single machine-specific directory. This directory is located undersysroots
and uses a machine-specific name (i.e.tmp/sysroots/machine/pkgdata
).
4.4.6 Shortened Git SRCREV
Values
BitBake will now shorten revisions from Git repositories from the normal 40 characters down to 10 characters within SRCPV for improved usability in path and file names. This change should be safe within contexts where these revisions are used because the chances of spatially close collisions is very low. Distant collisions are not a major issue in the way the values are used.
4.4.7 IMAGE_FEATURES
The following changes have been made that relate to IMAGE_FEATURES:
The value of
IMAGE_FEATURES
is now validated to ensure invalid feature items are not added. Some users mistakenly add package names to this variable instead of using IMAGE_INSTALL in order to have the package added to the image, which does not work. This change is intended to catch those kinds of situations. ValidIMAGE_FEATURES
are drawn fromPACKAGE_GROUP
definitions, COMPLEMENTARY_GLOB and a new “validitems” varflag onIMAGE_FEATURES
. The “validitems” varflag change allows additional features to be added if they are not provided using the previous two mechanisms.The previously deprecated “apps-console-core”
IMAGE_FEATURES
item is no longer supported. Add “splash” toIMAGE_FEATURES
if you wish to have the splash screen enabled, since this is all that apps-console-core was doing.
4.4.8 /run
The /run
directory from the Filesystem Hierarchy Standard 3.0 has
been introduced. You can find some of the implications for this change
here.
The change also means that recipes that install files to /var/run
must be changed. You can find a guide on how to make these changes
here.
4.4.9 Removal of Package Manager Database Within Image Recipes
The image core-image-minimal
no longer adds
remove_packaging_data_files
to
ROOTFS_POSTPROCESS_COMMAND.
This addition is now handled automatically when “package-management” is
not in IMAGE_FEATURES. If you have custom
image recipes that make this addition, you should remove the lines, as
they are not needed and might interfere with correct operation of
postinstall scripts.
4.4.10 Images Now Rebuild Only on Changes Instead of Every Time
The do_rootfs and other related image construction tasks are no longer marked as “nostamp”. Consequently, they will only be re-executed when their inputs have changed. Previous versions of the OpenEmbedded build system always rebuilt the image when requested rather when necessary.
4.4.11 Task Recipes
The previously deprecated task.bbclass
has now been dropped. For
recipes that previously inherited from this class, you should rename
them from task-*
to packagegroup-*
and inherit packagegroup
instead.
For more information, see the “packagegroup.bbclass” section.
4.4.12 BusyBox
By default, we now split BusyBox into two binaries: one that is suid
root for those components that need it, and another for the rest of the
components. Splitting BusyBox allows for optimization that eliminates
the tinylogin
recipe as recommended by upstream. You can disable
this split by setting
BUSYBOX_SPLIT_SUID to “0”.
4.4.13 Automated Image Testing
A new automated image testing framework has been added through the
testimage.bbclass class. This
framework replaces the older imagetest-qemu
framework.
You can learn more about performing automated image tests in the “Performing Automated Runtime Testing” section in the Yocto Project Development Tasks Manual.
4.4.14 Build History
Following are changes to Build History:
Installed package sizes:
installed-package-sizes.txt
for an image now records the size of the files installed by each package instead of the size of each compressed package archive file.The dependency graphs (
depends*.dot
) now use the actual package names instead of replacing dashes, dots and plus signs with underscores.The
buildhistory-diff
andbuildhistory-collect-srcrevs
utilities have improved command-line handling. Use the--help
option for each utility for more information on the new syntax.
For more information on Build History, see the “Maintaining Build Output Quality” section in the Yocto Project Development Tasks Manual.
4.4.15 udev
Following are changes to udev
:
udev
no longer brings inudev-extraconf
automatically through RRECOMMENDS, since this was originally intended to be optional. If you need the extra rules, then addudev-extraconf
to your image.udev
no longer brings inpciutils-ids
orusbutils-ids
throughRRECOMMENDS
. These are not needed byudev
itself and removing them saves around 350KB.
4.4.16 Removed and Renamed Recipes
The
linux-yocto
3.2 kernel has been removed.libtool-nativesdk
has been renamed tonativesdk-libtool
.tinylogin
has been removed. It has been replaced by a suid portion of Busybox. See the “BusyBox” section for more information.external-python-tarball
has been renamed tobuildtools-tarball
.web-webkit
has been removed. It has been functionally replaced bymidori
.imake
has been removed. It is no longer needed by any other recipe.transfig-native
has been removed. It is no longer needed by any other recipe.anjuta-remote-run
has been removed. Anjuta IDE integration has not been officially supported for several releases.
4.4.17 Other Changes
Following is a list of short entries describing other changes:
run-postinsts
: Make this generic.base-files
: Remove the unnecessarymedia/
xxx directories.alsa-state
: Provide an emptyasound.conf
by default.classes/image
: Ensure BAD_RECOMMENDATIONS supports pre-renamed package names.classes/rootfs_rpm
: ImplementBAD_RECOMMENDATIONS
for RPM.systemd
: Removesystemd_unitdir
ifsystemd
is not in DISTRO_FEATURES.systemd
: Removeinit.d
dir ifsystemd
unit file is present andsysvinit
is not a distro feature.libpam
: Deny all services for theOTHER
entries.image.bbclass
: Moveruntime_mapping_rename
to avoid conflict withmultilib
. See YOCTO #4993 in Bugzilla for more information.linux-dtb
: Use kernel build system to generate thedtb
files.kern-tools
: Switch from guilt to newkgit-s2q
tool.
4.5 Moving to the Yocto Project 1.6 Release (daisy)
This section provides migration information for moving to the Yocto Project 1.6 Release (codename “daisy”) from the prior release.
4.5.1 archiver
Class
The archiver class has been rewritten and its configuration has been simplified. For more details on the source archiver, see the “Maintaining Open Source License Compliance During Your Product’s Lifecycle” section in the Yocto Project Development Tasks Manual.
4.5.2 Packaging Changes
The following packaging changes have been made:
The
binutils
recipe no longer produces abinutils-symlinks
package.update-alternatives
is now used to handle the preferredbinutils
variant on the target instead.The tc (traffic control) utilities have been split out of the main
iproute2
package and put into theiproute2-tc
package.The
gtk-engines
schemas have been moved to a dedicatedgtk-engines-schemas
package.The
armv7a
with thumb package architecture suffix has changed. The suffix for these packages with the thumb optimization enabled is “t2” as it should be. Use of this suffix was not the case in the 1.5 release. Architecture names will change within package feeds as a result.
4.5.3 BitBake
The following changes have been made to BitBake.
4.5.3.1 Matching Branch Requirement for Git Fetching
When fetching source from a Git repository using SRC_URI, BitBake will now validate the SRCREV value against the branch. You can specify the branch using the following form:
SRC_URI = "git://server.name/repository;branch=branchname"
If you do not specify a branch, BitBake looks in the default “master” branch.
Alternatively, if you need to bypass this check (e.g. if you are
fetching a revision corresponding to a tag that is not on any branch),
you can add “;nobranch=1” to the end of the URL within SRC_URI
.
4.5.3.2 Python Definition substitutions
BitBake had some previously deprecated Python definitions within its
bb
module removed. You should use their sub-module counterparts
instead:
bb.MalformedUrl
: Usebb.fetch.MalformedUrl
.bb.encodeurl
: Usebb.fetch.encodeurl
.bb.decodeurl
: Usebb.fetch.decodeurl
bb.mkdirhier
: Usebb.utils.mkdirhier
.bb.movefile
: Usebb.utils.movefile
.bb.copyfile
: Usebb.utils.copyfile
.bb.which
: Usebb.utils.which
.bb.vercmp_string
: Usebb.utils.vercmp_string
.bb.vercmp
: Usebb.utils.vercmp
.
4.5.3.3 SVK Fetcher
The SVK fetcher has been removed from BitBake.
4.5.3.4 Console Output Error Redirection
The BitBake console UI will now output errors to stderr
instead of
stdout
. Consequently, if you are piping or redirecting the output of
bitbake
to somewhere else, and you wish to retain the errors, you
will need to add 2>&1
(or something similar) to the end of your
bitbake
command line.
4.5.3.5 task-
taskname Overrides
task-
taskname overrides have been adjusted so that tasks whose
names contain underscores have the underscores replaced by hyphens for
the override so that they now function properly. For example, the task
override for do_populate_sdk is
task-populate-sdk
.
4.5.4 Changes to Variables
The following variables have changed. For information on the OpenEmbedded build system variables, see the “Variables Glossary” Chapter.
4.5.4.1 TMPDIR
TMPDIR can no longer be on an NFS mount. NFS does
not offer full POSIX locking and inode consistency and can cause
unexpected issues if used to store TMPDIR
.
The check for this occurs on startup. If TMPDIR
is detected on an
NFS mount, an error occurs.
4.5.4.2 PRINC
The PRINC
variable has been deprecated and triggers a warning if
detected during a build. For PR increments on changes,
use the PR service instead. You can find out more about this service in
the “Working With a PR Service”
section in the Yocto Project Development Tasks Manual.
4.5.4.3 IMAGE_TYPES
The “sum.jffs2” option for IMAGE_TYPES has been replaced by the “jffs2.sum” option, which fits the processing order.
4.5.4.4 COPY_LIC_MANIFEST
The COPY_LIC_MANIFEST variable must now be set to “1” rather than any value in order to enable it.
4.5.4.5 COPY_LIC_DIRS
The COPY_LIC_DIRS variable must now be set to “1” rather than any value in order to enable it.
4.5.4.6 PACKAGE_GROUP
The PACKAGE_GROUP
variable has been renamed to
FEATURE_PACKAGES to more accurately
reflect its purpose. You can still use PACKAGE_GROUP
but the
OpenEmbedded build system produces a warning message when it encounters
the variable.
4.5.4.7 Preprocess and Post Process Command Variable Behavior
The following variables now expect a semicolon separated list of functions to call and not arbitrary shell commands:
For migration purposes, you can simply wrap shell commands in a shell function and then call the function. Here is an example:
my_postprocess_function() {
echo "hello" > ${IMAGE_ROOTFS}/hello.txt
}
ROOTFS_POSTPROCESS_COMMAND += "my_postprocess_function; "
4.5.5 Package Test (ptest)
Package Tests (ptest) are built but not installed by default. For
information on using Package Tests, see the
“Testing Packages With ptest”
section in the Yocto Project Development Tasks Manual. For information on the
ptest
class, see the “ptest.bbclass”
section.
4.5.6 Build Changes
Separate build and source directories have been enabled by default for
selected recipes where it is known to work (a whitelist) and for all
recipes that inherit the cmake class. In
future releases the autotools class
will enable a separate build directory by default as well. Recipes
building Autotools-based software that fails to build with a separate
build directory should be changed to inherit from the
autotools-brokensep class instead of
the autotools
or autotools_stage
classes.
4.5.7 qemu-native
qemu-native
now builds without SDL-based graphical output support by
default. The following additional lines are needed in your
local.conf
to enable it:
PACKAGECONFIG_pn-qemu-native = "sdl"
ASSUME_PROVIDED += "libsdl-native"
Note
The default local.conf
contains these statements. Consequently, if you
are building a headless system and using a default local.conf
file, you will need comment these two lines out.
4.5.8 core-image-basic
core-image-basic
has been renamed to core-image-full-cmdline
.
In addition to core-image-basic
being renamed,
packagegroup-core-basic
has been renamed to
packagegroup-core-full-cmdline
to match.
4.5.9 Licensing
The top-level LICENSE
file has been changed to better describe the
license of the various components of OpenEmbedded-Core (OE-Core). However,
the licensing itself remains unchanged.
Normally, this change would not cause any side-effects. However, some
recipes point to this file within
LIC_FILES_CHKSUM (as
${COREBASE}/LICENSE
) and thus the accompanying checksum must be
changed from 3f40d7994397109285ec7b81fdeb3b58 to
4d92cd373abda3937c2bc47fbc49d690. A better alternative is to have
LIC_FILES_CHKSUM
point to a file describing the license that is
distributed with the source that the recipe is building, if possible,
rather than pointing to ${COREBASE}/LICENSE
.
4.5.10 CFLAGS
Options
The “-fpermissive” option has been removed from the default
CFLAGS value. You need to take action on
individual recipes that fail when building with this option. You need to
either patch the recipes to fix the issues reported by the compiler, or
you need to add “-fpermissive” to CFLAGS
in the recipes.
4.5.11 Custom Image Output Types
Custom image output types, as selected using IMAGE_FSTYPES, must declare their dependencies on other image types (if any) using a new IMAGE_TYPEDEP variable.
4.5.12 Tasks
The do_package_write
task has been removed. The task is no longer
needed.
4.5.13 update-alternative
Provider
The default update-alternatives
provider has been changed from
opkg
to opkg-utils
. This change resolves some troublesome
circular dependencies. The runtime package has also been renamed from
update-alternatives-cworth
to update-alternatives-opkg
.
4.5.14 virtclass
Overrides
The virtclass
overrides are now deprecated. Use the equivalent class
overrides instead (e.g. virtclass-native
becomes class-native
.)
4.5.15 Removed and Renamed Recipes
The following recipes have been removed:
packagegroup-toolset-native
- This recipe is largely unused.linux-yocto-3.8
- Support for the Linux yocto 3.8 kernel has been dropped. Support for the 3.10 and 3.14 kernels have been added with thelinux-yocto-3.10
andlinux-yocto-3.14
recipes.ocf-linux
- This recipe has been functionally replaced usingcryptodev-linux
.genext2fs
-genext2fs
is no longer used by the build system and is unmaintained upstream.js
- This provided an ancient version of Mozilla’s javascript engine that is no longer needed.zaurusd
- The recipe has been moved to themeta-handheld
layer.eglibc 2.17
- Replaced by theeglibc 2.19
recipe.gcc 4.7.2
- Replaced by the now stablegcc 4.8.2
.external-sourcery-toolchain
- this recipe is now maintained in themeta-sourcery
layer.linux-libc-headers-yocto 3.4+git
- Now using version 3.10 of thelinux-libc-headers
by default.meta-toolchain-gmae
- This recipe is obsolete.packagegroup-core-sdk-gmae
- This recipe is obsolete.packagegroup-core-standalone-gmae-sdk-target
- This recipe is obsolete.
4.5.16 Removed Classes
The following classes have become obsolete and have been removed:
module_strip
pkg_metainfo
pkg_distribute
image-empty
4.5.17 Reference Board Support Packages (BSPs)
The following reference BSPs changes occurred:
The BeagleBoard (
beagleboard
) ARM reference hardware has been replaced by the BeagleBone (beaglebone
) hardware.The RouterStation Pro (
routerstationpro
) MIPS reference hardware has been replaced by the EdgeRouter Lite (edgerouter
) hardware.
The previous reference BSPs for the beagleboard
and
routerstationpro
machines are still available in a new
meta-yocto-bsp-old
layer in the
Source Repositories at
https://git.yoctoproject.org/cgit/cgit.cgi/meta-yocto-bsp-old/.
4.6 Moving to the Yocto Project 1.7 Release (dizzy)
This section provides migration information for moving to the Yocto Project 1.7 Release (codename “dizzy”) from the prior release.
4.6.1 Changes to Setting QEMU PACKAGECONFIG
Options in local.conf
The QEMU recipe now uses a number of
PACKAGECONFIG options to enable various
optional features. The method used to set defaults for these options
means that existing local.conf
files will need to be modified to
append to PACKAGECONFIG
for qemu-native
and nativesdk-qemu
instead of setting it. In other words, to enable graphical output for
QEMU, you should now have these lines in local.conf
:
PACKAGECONFIG_append_pn-qemu-native = " sdl"
PACKAGECONFIG_append_pn-nativesdk-qemu = " sdl"
4.6.2 Minimum Git version
The minimum Git
version required on the
build host is now 1.7.8 because the --list
option is now required by
BitBake’s Git fetcher. As always, if your host distribution does not
provide a version of Git that meets this requirement, you can use the
buildtools-tarball
that does. See the
“Required Git, tar, Python and gcc Versions”
section for more information.
4.6.3 Autotools Class Changes
The following autotools class changes occurred:
A separate build directory is now used by default: The
autotools
class has been changed to use a directory for building (B), which is separate from the source directory (S). This is commonly referred to asB != S
, or an out-of-tree build.If the software being built is already capable of building in a directory separate from the source, you do not need to do anything. However, if the software is not capable of being built in this manner, you will need to either patch the software so that it can build separately, or you will need to change the recipe to inherit the autotools-brokensep class instead of the
autotools
orautotools_stage
classes.The
--foreign
option is no longer passed toautomake
when runningautoconf
: This option tellsautomake
that a particular software package does not follow the GNU standards and therefore should not be expected to distribute certain files such asChangeLog
,AUTHORS
, and so forth. Because the majority of upstream software packages already tellautomake
to enable foreign mode themselves, the option is mostly superfluous. However, some recipes will need patches for this change. You can easily make the change by patchingconfigure.ac
so that it passes “foreign” toAM_INIT_AUTOMAKE()
. See this commit for an example showing how to make the patch.
4.6.4 Binary Configuration Scripts Disabled
Some of the core recipes that package binary configuration scripts now
disable the scripts due to the scripts previously requiring error-prone
path substitution. Software that links against these libraries using
these scripts should use the much more robust pkg-config
instead.
The list of recipes changed in this version (and their configuration
scripts) is as follows:
directfb (directfb-config)
freetype (freetype-config)
gpgme (gpgme-config)
libassuan (libassuan-config)
libcroco (croco-6.0-config)
libgcrypt (libgcrypt-config)
libgpg-error (gpg-error-config)
libksba (ksba-config)
libpcap (pcap-config)
libpcre (pcre-config)
libpng (libpng-config, libpng16-config)
libsdl (sdl-config)
libusb-compat (libusb-config)
libxml2 (xml2-config)
libxslt (xslt-config)
ncurses (ncurses-config)
neon (neon-config)
npth (npth-config)
pth (pth-config)
taglib (taglib-config)
Additionally, support for pkg-config
has been added to some recipes in the
previous list in the rare cases where the upstream software package does
not already provide it.
4.6.5 eglibc 2.19
Replaced with glibc 2.20
Because eglibc
and glibc
were already fairly close, this
replacement should not require any significant changes to other software
that links to eglibc
. However, there were a number of minor changes
in glibc 2.20
upstream that could require patching some software
(e.g. the removal of the _BSD_SOURCE
feature test macro).
glibc 2.20
requires version 2.6.32 or greater of the Linux kernel.
Thus, older kernels will no longer be usable in conjunction with it.
For full details on the changes in glibc 2.20
, see the upstream
release notes
here.
4.6.6 Kernel Module Autoloading
The module_autoload_* variable is now
deprecated and a new
KERNEL_MODULE_AUTOLOAD variable
should be used instead. Also, module_conf_*
must now be used in conjunction with a new
KERNEL_MODULE_PROBECONF variable.
The new variables no longer require you to specify the module name as
part of the variable name. This change not only simplifies usage but
also allows the values of these variables to be appropriately
incorporated into task signatures and thus trigger the appropriate tasks
to re-execute when changed. You should replace any references to
module_autoload_*
with KERNEL_MODULE_AUTOLOAD
, and add any
modules for which module_conf_*
is specified to
KERNEL_MODULE_PROBECONF
.
4.6.7 QA Check Changes
The following changes have occurred to the QA check process:
Additional QA checks
file-rdeps
andbuild-deps
have been added in order to verify that file dependencies are satisfied (e.g. package contains a script requiring/bin/bash
) and build-time dependencies are declared, respectively. For more information, please see the “QA Error and Warning Messages” chapter.Package QA checks are now performed during a new do_package_qa task rather than being part of the do_package task. This allows more parallel execution. This change is unlikely to be an issue except for highly customized recipes that disable packaging tasks themselves by marking them as
noexec
. For those packages, you will need to disable thedo_package_qa
task as well.Files being overwritten during the do_populate_sysroot task now trigger an error instead of a warning. Recipes should not be overwriting files written to the sysroot by other recipes. If you have these types of recipes, you need to alter them so that they do not overwrite these files.
You might now receive this error after changes in configuration or metadata resulting in orphaned files being left in the sysroot. If you do receive this error, the way to resolve the issue is to delete your TMPDIR or to move it out of the way and then re-start the build. Anything that has been fully built up to that point and does not need rebuilding will be restored from the shared state cache and the rest of the build will be able to proceed as normal.
4.6.8 Removed Recipes
The following recipes have been removed:
x-load
: This recipe has been superseded by U-Boot SPL for all Cortex-based TI SoCs. For legacy boards, themeta-ti
layer, which contains a maintained recipe, should be used instead.ubootchart
: This recipe is obsolete. Abootchart2
recipe has been added to functionally replace it.linux-yocto 3.4
: Support for the linux-yocto 3.4 kernel has been dropped. Support for the 3.10 and 3.14 kernels remains, while support for version 3.17 has been added.eglibc
has been removed in favor ofglibc
. See the “eglibc 2.19 Replaced with glibc 2.20” section for more information.
4.6.9 Miscellaneous Changes
The following miscellaneous change occurred:
The build history feature now writes
build-id.txt
instead ofbuild-id
. Additionally,build-id.txt
now contains the full build header as printed by BitBake upon starting the build. You should manually remove old “build-id” files from your existing build history repositories to avoid confusion. For information on the build history feature, see the “Maintaining Build Output Quality” section in the Yocto Project Development Tasks Manual.
4.7 Moving to the Yocto Project 1.8 Release (fido)
This section provides migration information for moving to the Yocto Project 1.8 Release (codename “fido”) from the prior release.
4.7.1 Removed Recipes
The following recipes have been removed:
owl-video
: Functionality replaced bygst-player
.gaku
: Functionality replaced bygst-player
.gnome-desktop
: This recipe is now available inmeta-gnome
and is no longer needed.gsettings-desktop-schemas
: This recipe is now available inmeta-gnome
and is no longer needed.python-argparse
: Theargparse
module is already provided in the default Python distribution in a package namedpython-argparse
. Consequently, the separatepython-argparse
recipe is no longer needed.telepathy-python, libtelepathy, telepathy-glib, telepathy-idle, telepathy-mission-control
: All these recipes have moved tometa-oe
and are consequently no longer needed by any recipes in OpenEmbedded-Core.linux-yocto_3.10
andlinux-yocto_3.17
: Support for the linux-yocto 3.10 and 3.17 kernels has been dropped. Support for the 3.14 kernel remains, while support for 3.19 kernel has been added.poky-feed-config-opkg
: This recipe has become obsolete and is no longer needed. Usedistro-feed-config
frommeta-oe
instead.libav 0.8.x
:libav 9.x
is now used.sed-native
: No longer needed. A working version ofsed
is expected to be provided by the host distribution.
4.7.2 BlueZ 4.x / 5.x Selection
Proper built-in support for selecting BlueZ 5.x in preference to the
default of 4.x now exists. To use BlueZ 5.x, simply add “bluez5” to your
DISTRO_FEATURES value. If you had
previously added append files (*.bbappend
) to make this selection,
you can now remove them.
Additionally, a bluetooth
class has been added to make selection of
the appropriate bluetooth support within a recipe a little easier. If
you wish to make use of this class in a recipe, add something such as
the following:
inherit bluetooth
PACKAGECONFIG ??= "${@bb.utils.contains('DISTRO_FEATURES', 'bluetooth', '${BLUEZ}', '', d)}"
PACKAGECONFIG[bluez4] = "--enable-bluetooth,--disable-bluetooth,bluez4"
PACKAGECONFIG[bluez5] = "--enable-bluez5,--disable-bluez5,bluez5"
4.7.3 Kernel Build Changes
The kernel build process was changed to place the source in a common
shared work area and to place build artifacts separately in the source
code tree. In theory, migration paths have been provided for most common
usages in kernel recipes but this might not work in all cases. In
particular, users need to ensure that ${S}
(source files) and
${B}
(build artifacts) are used correctly in functions such as
do_configure and
do_install. For kernel recipes that do not
inherit from kernel-yocto
or include linux-yocto.inc
, you might
wish to refer to the linux.inc
file in the meta-oe
layer for the
kinds of changes you need to make. For reference, here is the
commit
where the linux.inc
file in meta-oe
was updated.
Recipes that rely on the kernel source code and do not inherit the
module classes might need to add explicit dependencies on the
do_shared_workdir
kernel task, for example:
do_configure[depends] += "virtual/kernel:do_shared_workdir"
4.7.4 SSL 3.0 is Now Disabled in OpenSSL
SSL 3.0 is now disabled when building OpenSSL. Disabling SSL 3.0 avoids
any lingering instances of the POODLE vulnerability. If you feel you
must re-enable SSL 3.0, then you can add an append file (*.bbappend
)
for the openssl
recipe to remove “-no-ssl3” from
EXTRA_OECONF.
4.7.5 Default Sysroot Poisoning
gcc's
default sysroot and include directories are now “poisoned”. In
other words, the sysroot and include directories are being redirected to
a non-existent location in order to catch when host directories are
being used due to the correct options not being passed. This poisoning
applies both to the cross-compiler used within the build and to the
cross-compiler produced in the SDK.
If this change causes something in the build to fail, it almost certainly means the various compiler flags and commands are not being passed correctly to the underlying piece of software. In such cases, you need to take corrective steps.
4.7.6 Rebuild Improvements
Changes have been made to the base, autotools, and cmake classes to clean out generated files when the do_configure task needs to be re-executed.
One of the improvements is to attempt to run “make clean” during the
do_configure
task if a Makefile
exists. Some software packages
do not provide a working clean target within their make files. If you
have such recipes, you need to set
CLEANBROKEN to “1” within the recipe, for example:
CLEANBROKEN = "1"
4.7.7 QA Check and Validation Changes
The following QA Check and Validation Changes have occurred:
Usage of
PRINC
previously triggered a warning. It now triggers an error. You should remove any remaining usage ofPRINC
in any recipe or append file.An additional QA check has been added to detect usage of
${D}
in FILES values where D values should not be used at all. The same check ensures that$D
is used inpkg_preinst/pkg_postinst/pkg_prerm/pkg_postrm
functions instead of${D}
.S now needs to be set to a valid value within a recipe. If
S
is not set in the recipe, the directory is not automatically created. IfS
does not point to a directory that exists at the time the do_unpack task finishes, a warning will be shown.LICENSE is now validated for correct formatting of multiple licenses. If the format is invalid (e.g. multiple licenses are specified with no operators to specify how the multiple licenses interact), then a warning will be shown.
4.7.8 Miscellaneous Changes
The following miscellaneous changes have occurred:
The
send-error-report
script now expects a “-s” option to be specified before the server address. This assumes a server address is being specified.The
oe-pkgdata-util
script now expects a “-p” option to be specified before thepkgdata
directory, which is now optional. If thepkgdata
directory is not specified, the script will run BitBake to query PKGDATA_DIR from the build environment.
4.8 Moving to the Yocto Project 2.0 Release (jethro)
This section provides migration information for moving to the Yocto Project 2.0 Release (codename “jethro”) from the prior release.
4.8.1 GCC 5
The default compiler is now GCC 5.2. This change has required fixes for compilation errors in a number of other recipes.
One important example is a fix for when the Linux kernel freezes at boot
time on ARM when built with GCC 5. If you are using your own kernel
recipe or source tree and building for ARM, you will likely need to
apply this
patch.
The standard linux-yocto
kernel source tree already has a workaround
for the same issue.
For further details, see https://gcc.gnu.org/gcc-5/changes.html and the porting guide at https://gcc.gnu.org/gcc-5/porting_to.html.
Alternatively, you can switch back to GCC 4.9 or 4.8 by setting
GCCVERSION
in your configuration, as follows:
GCCVERSION = "4.9%"
4.8.2 Gstreamer 0.10 Removed
Gstreamer 0.10 has been removed in favor of Gstreamer 1.x. As part of
the change, recipes for Gstreamer 0.10 and related software are now
located in meta-multimedia
. This change results in Qt4 having Phonon
and Gstreamer support in QtWebkit disabled by default.
4.8.3 Removed Recipes
The following recipes have been moved or removed:
bluez4
: The recipe is obsolete and has been moved due tobluez5
becoming fully integrated. Thebluez4
recipe now resides inmeta-oe
.gamin
: The recipe is obsolete and has been removed.gnome-icon-theme
: The recipe’s functionally has been replaced byadwaita-icon-theme
.Gstreamer 0.10 Recipes: Recipes for Gstreamer 0.10 have been removed in favor of the recipes for Gstreamer 1.x.
insserv
: The recipe is obsolete and has been removed.libunique
: The recipe is no longer used and has been moved tometa-oe
.midori
: The recipe’s functionally has been replaced byepiphany
.python-gst
: The recipe is obsolete and has been removed since it only contains bindings for Gstreamer 0.10.qt-mobility
: The recipe is obsolete and has been removed since it requiresGstreamer 0.10
, which has been replaced.subversion
: All 1.6.x versions of this recipe have been removed.webkit-gtk
: The older 1.8.3 version of this recipe has been removed in favor ofwebkitgtk
.
4.8.4 BitBake datastore improvements
The method by which BitBake’s datastore handles overrides has changed.
Overrides are now applied dynamically and bb.data.update_data()
is
now a no-op. Thus, bb.data.update_data()
is no longer required in
order to apply the correct overrides. In practice, this change is
unlikely to require any changes to Metadata. However, these minor
changes in behavior exist:
All potential overrides are now visible in the variable history as seen when you run the following:
$ bitbake -e
d.delVar('VARNAME')
andd.setVar('VARNAME', None)
result in the variable and all of its overrides being cleared out. Before the change, only the non-overridden values were cleared.
4.8.5 Shell Message Function Changes
The shell versions of the BitBake message functions (i.e. bbdebug
,
bbnote
, bbwarn
, bbplain
, bberror
, and bbfatal
) are
now connected through to their BitBake equivalents bb.debug()
,
bb.note()
, bb.warn()
, bb.plain()
, bb.error()
, and
bb.fatal()
, respectively. Thus, those message functions that you
would expect to be printed by the BitBake UI are now actually printed.
In practice, this change means two things:
If you now see messages on the console that you did not previously see as a result of this change, you might need to clean up the calls to
bbwarn
,bberror
, and so forth. Or, you might want to simply remove the calls.The
bbfatal
message function now suppresses the full error log in the UI, which means any calls tobbfatal
where you still wish to see the full error log should be replaced bydie
orbbfatal_log
.
4.8.6 Extra Development/Debug Package Cleanup
The following recipes have had extra dev/dbg
packages removed:
acl
apmd
aspell
attr
augeas
bzip2
cogl
curl
elfutils
gcc-target
libgcc
libtool
libxmu
opkg
pciutils
rpm
sysfsutils
tiff
xz
All of the above recipes now conform to the standard packaging scheme
where a single -dev
, -dbg
, and -staticdev
package exists per
recipe.
4.8.7 Recipe Maintenance Tracking Data Moved to OE-Core
Maintenance tracking data for recipes that was previously part of
meta-yocto
has been moved to OpenEmbedded-Core (OE-Core). The change
includes package_regex.inc
and distro_alias.inc
, which are
typically enabled when using the distrodata
class. Additionally, the
contents of upstream_tracking.inc
has now been split out to the
relevant recipes.
4.8.8 Automatic Stale Sysroot File Cleanup
Stale files from recipes that no longer exist in the current
configuration are now automatically removed from sysroot as well as
removed from any other place managed by shared state. This automatic
cleanup means that the build system now properly handles situations such
as renaming the build system side of recipes, removal of layers from
bblayers.conf
, and DISTRO_FEATURES
changes.
Additionally, work directories for old versions of recipes are now pruned. If you wish to disable pruning old work directories, you can set the following variable in your configuration:
SSTATE_PRUNE_OBSOLETEWORKDIR = "0"
4.8.9 linux-yocto
Kernel Metadata Repository Now Split from Source
The linux-yocto
tree has up to now been a combined set of kernel
changes and configuration (meta) data carried in a single tree. While
this format is effective at keeping kernel configuration and source
modifications synchronized, it is not always obvious to developers how
to manipulate the Metadata as compared to the source.
Metadata processing has now been removed from the
kernel-yocto class and the external
Metadata repository yocto-kernel-cache
, which has always been used
to seed the linux-yocto
“meta” branch. This separate linux-yocto
cache repository is now the primary location for this data. Due to this
change, linux-yocto
is no longer able to process combined trees.
Thus, if you need to have your own combined kernel repository, you must
do the split there as well and update your recipes accordingly. See the
meta/recipes-kernel/linux/linux-yocto_4.1.bb
recipe for an example.
4.8.10 Additional QA checks
The following QA checks have been added:
Added a “host-user-contaminated” check for ownership issues for packaged files outside of
/home
. The check looks for files that are incorrectly owned by the user that ran BitBake instead of owned by a valid user in the target system.Added an “invalid-chars” check for invalid (non-UTF8) characters in recipe metadata variable values (i.e. DESCRIPTION, SUMMARY, LICENSE, and SECTION). Some package managers do not support these characters.
Added an “invalid-packageconfig” check for any options specified in PACKAGECONFIG that do not match any
PACKAGECONFIG
option defined for the recipe.
4.8.11 Miscellaneous Changes
These additional changes exist:
gtk-update-icon-cache
has been renamed togtk-icon-utils
.The
tools-profile
IMAGE_FEATURES item as well as its corresponding packagegroup andpackagegroup-core-tools-profile
no longer bring inoprofile
. Bringing inoprofile
was originally added to aid compilation on resource-constrained targets. However, this aid has not been widely used and is not likely to be used going forward due to the more powerful target platforms and the existence of better cross-compilation tools.The IMAGE_FSTYPES variable’s default value now specifies
ext4
instead ofext3
.All support for the
PRINC
variable has been removed.The
packagegroup-core-full-cmdline
packagegroup no longer brings inlighttpd
due to the fact that bringing inlighttpd
is not really in line with the packagegroup’s purpose, which is to add full versions of command-line tools that by default are provided bybusybox
.
4.9 Moving to the Yocto Project 2.1 Release (krogoth)
This section provides migration information for moving to the Yocto Project 2.1 Release (codename “krogoth”) from the prior release.
4.9.1 Variable Expansion in Python Functions
Variable expressions, such as ${VARNAME}
no longer expand
automatically within Python functions. Suppressing expansion was done to
allow Python functions to construct shell scripts or other code for
situations in which you do not want such expressions expanded. For any
existing code that relies on these expansions, you need to change the
expansions to expand the value of individual variables through
d.getVar()
. To alternatively expand more complex expressions, use
d.expand()
.
4.9.2 Overrides Must Now be Lower-Case
The convention for overrides has always been for them to be lower-case
characters. This practice is now a requirement as BitBake’s datastore
now assumes lower-case characters in order to give a slight performance
boost during parsing. In practical terms, this requirement means that
anything that ends up in OVERRIDES must now
appear in lower-case characters (e.g. values for MACHINE
,
TARGET_ARCH
, DISTRO
, and also recipe names if
_pn-
recipename overrides are to be effective).
4.9.3 Expand Parameter to getVar()
and getVarFlag()
is Now Mandatory
The expand parameter to getVar()
and getVarFlag()
previously
defaulted to False if not specified. Now, however, no default exists so
one must be specified. You must change any getVar()
calls that do
not specify the final expand parameter to calls that do specify the
parameter. You can run the following sed
command at the base of a
layer to make this change:
sed -e 's:\(\.getVar([^,()]*\)):\1, False):g' -i `grep -ril getVar *`
sed -e 's:\(\.getVarFlag([^,()]*,[^,()]*\)):\1, False):g' -i `grep -ril getVarFlag *`
Note
The reason for this change is that it prepares the way for changing the default to True in a future Yocto Project release. This future change is a much more sensible default than False. However, the change needs to be made gradually as a sudden change of the default would potentially cause side-effects that would be difficult to detect.
4.9.4 Makefile Environment Changes
EXTRA_OEMAKE now defaults to “” instead of
“-e MAKEFLAGS=”. Setting EXTRA_OEMAKE
to “-e MAKEFLAGS=” by default
was a historical accident that has required many classes (e.g.
autotools
, module
) and recipes to override this default in order
to work with sensible build systems. When upgrading to the release, you
must edit any recipe that relies upon this old default by either setting
EXTRA_OEMAKE
back to “-e MAKEFLAGS=” or by explicitly setting any
required variable value overrides using EXTRA_OEMAKE
, which is
typically only needed when a Makefile sets a default value for a
variable that is inappropriate for cross-compilation using the “=”
operator rather than the “?=” operator.
4.9.5 libexecdir
Reverted to ${prefix}/libexec
The use of ${libdir}/${BPN}
as libexecdir
is different as
compared to all other mainstream distributions, which either uses
${prefix}/libexec
or ${libdir}
. The use is also contrary to the
GNU Coding Standards (i.e.
https://www.gnu.org/prep/standards/html_node/Directory-Variables.html)
that suggest ${prefix}/libexec
and also notes that any
package-specific nesting should be done by the package itself. Finally,
having libexecdir
change between recipes makes it very difficult for
different recipes to invoke binaries that have been installed into
libexecdir
. The Filesystem Hierarchy Standard (i.e.
https://refspecs.linuxfoundation.org/FHS_3.0/fhs/ch04s07.html) now
recognizes the use of ${prefix}/libexec/
, giving distributions the
choice between ${prefix}/lib
or ${prefix}/libexec
without
breaking FHS.
4.9.6 ac_cv_sizeof_off_t
is No Longer Cached in Site Files
For recipes inheriting the autotools
class, ac_cv_sizeof_off_t
is no longer cached in the site files for
autoconf
. The reason for this change is because the
ac_cv_sizeof_off_t
value is not necessarily static per architecture
as was previously assumed. Rather, the value changes based on whether
large file support is enabled. For most software that uses autoconf
,
this change should not be a problem. However, if you have a recipe that
bypasses the standard do_configure task
from the autotools
class and the software the recipe is building
uses a very old version of autoconf
, the recipe might be incapable
of determining the correct size of off_t
during do_configure
.
The best course of action is to patch the software as necessary to allow
the default implementation from the autotools
class to work such
that autoreconf
succeeds and produces a working configure script,
and to remove the overridden do_configure
task such that the default
implementation does get used.
4.9.7 Image Generation is Now Split Out from Filesystem Generation
Previously, for image recipes the do_rootfs task assembled the filesystem and then from that filesystem generated images. With this Yocto Project release, image generation is split into separate do_image tasks for clarity both in operation and in the code.
For most cases, this change does not present any problems. However, if
you have made customizations that directly modify the do_rootfs
task
or that mention do_rootfs
, you might need to update those changes.
In particular, if you had added any tasks after do_rootfs
, you
should make edits so that those tasks are after the
do_image_complete task rather than
after do_rootfs
so that your added tasks run at the correct
time.
A minor part of this restructuring is that the post-processing definitions and functions have been moved from the image class to the rootfs-postcommands class. Functionally, however, they remain unchanged.
4.9.8 Removed Recipes
The following recipes have been removed in the 2.1 release:
gcc
version 4.8: Versions 4.9 and 5.3 remain.qt4
: All support for Qt 4.x has been moved out to a separatemeta-qt4
layer because Qt 4 is no longer supported upstream.x11vnc
: Moved to themeta-oe
layer.linux-yocto-3.14
: No longer supported.linux-yocto-3.19
: No longer supported.libjpeg
: Replaced by thelibjpeg-turbo
recipe.pth
: Became obsolete.liboil
: Recipe is no longer needed and has been moved to themeta-multimedia
layer.gtk-theme-torturer
: Recipe is no longer needed and has been moved to themeta-gnome
layer.gnome-mime-data
: Recipe is no longer needed and has been moved to themeta-gnome
layer.udev
: Replaced by theeudev
recipe for compatibility when usingsysvinit
with newer kernels.python-pygtk
: Recipe became obsolete.adt-installer
: Recipe became obsolete. See the “ADT Removed” section for more information.
4.9.9 Class Changes
The following classes have changed:
autotools_stage
: Removed because the autotools class now provides its functionality. Recipes that inherited fromautotools_stage
should now inherit fromautotools
instead.boot-directdisk
: Merged into theimage-vm
class. Theboot-directdisk
class was rarely directly used. Consequently, this change should not cause any issues.bootimg
: Merged into the image-live class. Thebootimg
class was rarely directly used. Consequently, this change should not cause any issues.packageinfo
: Removed due to its limited use by the Hob UI, which has itself been removed.
4.9.10 Build System User Interface Changes
The following changes have been made to the build system user interface:
Hob GTK+-based UI: Removed because it is unmaintained and based on the outdated GTK+ 2 library. The Toaster web-based UI is much more capable and is actively maintained. See the “Using the Toaster Web Interface” section in the Toaster User Manual for more information on this interface.
“puccho” BitBake UI: Removed because is unmaintained and no longer useful.
4.9.11 ADT Removed
The Application Development Toolkit (ADT) has been removed because its functionality almost completely overlapped with the standard SDK and the extensible SDK. For information on these SDKs and how to build and use them, see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
Note
The Yocto Project Eclipse IDE Plug-in is still supported and is not affected by this change.
4.9.12 Poky Reference Distribution Changes
The following changes have been made for the Poky distribution:
The
meta-yocto
layer has been renamed tometa-poky
to better match its purpose, which is to provide the Poky reference distribution. Themeta-yocto-bsp
layer retains its original name since it provides reference machines for the Yocto Project and it is otherwise unrelated to Poky. References tometa-yocto
in yourconf/bblayers.conf
should automatically be updated, so you should not need to change anything unless you are relying on this naming elsewhere.The uninative class is now enabled by default in Poky. This class attempts to isolate the build system from the host distribution’s C library and makes re-use of native shared state artifacts across different host distributions practical. With this class enabled, a tarball containing a pre-built C library is downloaded at the start of the build.
The
uninative
class is enabled through themeta/conf/distro/include/yocto-uninative.inc
file, which for those not using the Poky distribution, can include to easily enable the same functionality.Alternatively, if you wish to build your own
uninative
tarball, you can do so by building theuninative-tarball
recipe, making it available to your build machines (e.g. over HTTP/HTTPS) and setting a similar configuration as the one set byyocto-uninative.inc
.Static library generation, for most cases, is now disabled by default in the Poky distribution. Disabling this generation saves some build time as well as the size used for build output artifacts.
Disabling this library generation is accomplished through a
meta/conf/distro/include/no-static-libs.inc
, which for those not using the Poky distribution can easily include to enable the same functionality.Any recipe that needs to opt-out of having the “–disable-static” option specified on the configure command line either because it is not a supported option for the configure script or because static libraries are needed should set the following variable:
DISABLE_STATIC = ""
The separate
poky-tiny
distribution now uses the musl C library instead of a heavily pared downglibc
. Using musl results in a smaller distribution and facilitates much greater maintainability because musl is designed to have a small footprint.If you have used
poky-tiny
and have customized theglibc
configuration you will need to redo those customizations with musl when upgrading to the new release.
4.9.13 Packaging Changes
The following changes have been made to packaging:
The
runuser
andmountpoint
binaries, which were previously in the mainutil-linux
package, have been split out into theutil-linux-runuser
andutil-linux-mountpoint
packages, respectively.The
python-elementtree
package has been merged into thepython-xml
package.
4.9.14 Tuning File Changes
The following changes have been made to the tuning files:
The “no-thumb-interwork” tuning feature has been dropped from the ARM tune include files. Because interworking is required for ARM EABI, attempting to disable it through a tuning feature no longer makes sense.
Note
Support for ARM OABI was deprecated in gcc 4.7.
The
tune-cortexm*.inc
andtune-cortexr4.inc
files have been removed because they are poorly tested. Until the OpenEmbedded build system officially gains support for CPUs without an MMU, these tuning files would probably be better maintained in a separate layer if needed.
4.9.15 Supporting GObject Introspection
This release supports generation of GLib Introspective Repository (GIR) files through GObject introspection, which is the standard mechanism for accessing GObject-based software from runtime environments. You can enable, disable, and test the generation of this data. See the “Enabling GObject Introspection Support” section in the Yocto Project Development Tasks Manual for more information.
4.9.16 Miscellaneous Changes
These additional changes exist:
The minimum Git version has been increased to 1.8.3.1. If your host distribution does not provide a sufficiently recent version, you can install the buildtools, which will provide it. See the Required Git, tar, Python and gcc Versions section for more information on the buildtools tarball.
The buggy and incomplete support for the RPM version 4 package manager has been removed. The well-tested and maintained support for RPM version 5 remains.
Previously, the following list of packages were removed if package-management was not in IMAGE_FEATURES, regardless of any dependencies:
update-rc.d base-passwd shadow update-alternatives run-postinsts
With the Yocto Project 2.1 release, these packages are only removed if “read-only-rootfs” is in
IMAGE_FEATURES
, since they might still be needed for a read-write image even in the absence of a package manager (e.g. if users need to be added, modified, or removed at runtime).The devtool modify command now defaults to extracting the source since that is most commonly expected. The “-x” or “–extract” options are now no-ops. If you wish to provide your own existing source tree, you will now need to specify either the “-n” or “–no-extract” options when running
devtool modify
.If the formfactor for a machine is either not supplied or does not specify whether a keyboard is attached, then the default is to assume a keyboard is attached rather than assume no keyboard. This change primarily affects the Sato UI.
The
.debug
directory packaging is now automatic. If your recipe builds software that installs binaries into directories other than the standard ones, you no longer need to take care of settingFILES_${PN}-dbg
to pick up the resulting.debug
directories as these directories are automatically found and added.Inaccurate disk and CPU percentage data has been dropped from
buildstats
output. This data has been replaced withgetrusage()
data and corrected IO statistics. You will probably need to update any custom code that reads thebuildstats
data.The
meta/conf/distro/include/package_regex.inc
is now deprecated. The contents of this file have been moved to individual recipes.Note
Because this file will likely be removed in a future Yocto Project release, it is suggested that you remove any references to the file that might be in your configuration.
The
v86d/uvesafb
has been removed from thegenericx86
andgenericx86-64
reference machines, which are provided by themeta-yocto-bsp
layer. Most modern x86 boards do not rely on this file and it only adds kernel error messages during startup. If you do still need to supportuvesafb
, you can simply addv86d
to your image.Build sysroot paths are now removed from debug symbol files. Removing these paths means that remote GDB using an unstripped build system sysroot will no longer work (although this was never documented to work). The supported method to accomplish something similar is to set
IMAGE_GEN_DEBUGFS
to “1”, which will generate a companion debug image containing unstripped binaries and associated debug sources alongside the image.
4.10 Moving to the Yocto Project 2.2 Release (morty)
This section provides migration information for moving to the Yocto Project 2.2 Release (codename “morty”) from the prior release.
4.10.1 Minimum Kernel Version
The minimum kernel version for the target system and for SDK is now
3.2.0, due to the upgrade to glibc 2.24
. Specifically, for
AArch64-based targets the version is 3.14. For Nios II-based targets,
the minimum kernel version is 3.19.
Note
For x86 and x86_64, you can reset OLDEST_KERNEL to anything down to 2.6.32 if desired.
4.10.2 Staging Directories in Sysroot Has Been Simplified
The way directories are staged in sysroot has been simplified and introduces the new SYSROOT_DIRS, SYSROOT_DIRS_NATIVE, and SYSROOT_DIRS_BLACKLIST. See the v2 patch series on the OE-Core Mailing List for additional information.
4.10.3 Removal of Old Images and Other Files in tmp/deploy
Now Enabled
Removal of old images and other files in tmp/deploy/
is now enabled
by default due to a new staging method used for those files. As a result
of this change, the RM_OLD_IMAGE
variable is now redundant.
4.10.4 Python Changes
The following changes for Python occurred:
4.10.4.1 BitBake Now Requires Python 3.4+
BitBake requires Python 3.4 or greater.
4.10.4.2 UTF-8 Locale Required on Build Host
A UTF-8 locale is required on the build host due to Python 3. Since C.UTF-8 is not a standard, the default is en_US.UTF-8.
4.10.4.3 Metadata Must Now Use Python 3 Syntax
The metadata is now required to use Python 3 syntax. For help preparing metadata, see any of the many Python 3 porting guides available. Alternatively, you can reference the conversion commits for Bitbake and you can use OpenEmbedded-Core (OE-Core) as a guide for changes. Following are particular areas of interest:
subprocess command-line pipes needing locale decoding
the syntax for octal values changed
the
iter*()
functions changed nameiterators now return views, not lists
changed names for Python modules
4.10.4.4 Target Python Recipes Switched to Python 3
Most target Python recipes have now been switched to Python 3. Unfortunately, systems using RPM as a package manager and providing online package-manager support through SMART still require Python 2.
Note
Python 2 and recipes that use it can still be built for the target as with previous versions.
4.10.4.5 buildtools-tarball
Includes Python 3
buildtools-tarball
now includes Python 3.
4.10.5 uClibc Replaced by musl
uClibc has been removed in favor of musl. Musl has matured, is better maintained, and is compatible with a wider range of applications as compared to uClibc.
4.10.6 ${B}
No Longer Default Working Directory for Tasks
${
B}
is no longer the default working
directory for tasks. Consequently, any custom tasks you define now need
to either have the
[
dirs]
flag
set, or the task needs to change into the appropriate working directory
manually (e.g using cd
for a shell task).
Note
The preferred method is to use the [dirs] flag.
4.10.7 runqemu
Ported to Python
runqemu
has been ported to Python and has changed behavior in some
cases. Previous usage patterns continue to be supported.
The new runqemu
is a Python script. Machine knowledge is no longer
hardcoded into runqemu
. You can choose to use the qemuboot
configuration file to define the BSP’s own arguments and to make it
bootable with runqemu
. If you use a configuration file, use the
following form:
image-name-machine.qemuboot.conf
The configuration file
enables fine-grained tuning of options passed to QEMU without the
runqemu
script hard-coding any knowledge about different machines.
Using a configuration file is particularly convenient when trying to use
QEMU with machines other than the qemu*
machines in
OpenEmbedded-Core (OE-Core). The qemuboot.conf
file is generated by the
qemuboot
class when the root filesystem is being build (i.e. build
rootfs). QEMU boot arguments can be set in BSP’s configuration file and
the qemuboot
class will save them to qemuboot.conf
.
If you want to use runqemu
without a configuration file, use the
following command form:
$ runqemu machine rootfs kernel [options]
Supported machines are as follows:
qemuarm
qemuarm64
qemux86
qemux86-64
qemuppc
qemumips
qemumips64
qemumipsel
qemumips64el
Consider the
following example, which uses the qemux86-64
machine, provides a
root filesystem, provides an image, and uses the nographic
option:
$ runqemu qemux86-64 tmp/deploy/images/qemux86-64/core-image-minimal-qemux86-64.ext4 tmp/deploy/images/qemux86-64/bzImage nographic
Following is a list of variables that can be set in configuration files
such as bsp.conf
to enable the BSP to be booted by runqemu
:
Note
“QB” means “QEMU Boot”.
QB_SYSTEM_NAME: QEMU name (e.g. "qemu-system-i386")
QB_OPT_APPEND: Options to append to QEMU (e.g. "-show-cursor")
QB_DEFAULT_KERNEL: Default kernel to boot (e.g. "bzImage")
QB_DEFAULT_FSTYPE: Default FSTYPE to boot (e.g. "ext4")
QB_MEM: Memory (e.g. "-m 512")
QB_MACHINE: QEMU machine (e.g. "-machine virt")
QB_CPU: QEMU cpu (e.g. "-cpu qemu32")
QB_CPU_KVM: Similar to QB_CPU except used for kvm support (e.g. "-cpu kvm64")
QB_KERNEL_CMDLINE_APPEND: Options to append to the kernel's -append
option (e.g. "console=ttyS0 console=tty")
QB_DTB: QEMU dtb name
QB_AUDIO_DRV: QEMU audio driver (e.g. "alsa", set it when support audio)
QB_AUDIO_OPT: QEMU audio option (e.g. "-soundhw ac97,es1370"), which is used
when QB_AUDIO_DRV is set.
QB_KERNEL_ROOT: Kernel's root (e.g. /dev/vda)
QB_TAP_OPT: Network option for 'tap' mode (e.g.
"-netdev tap,id=net0,ifname=@TAP@,script=no,downscript=no -device virtio-net-device,netdev=net0").
runqemu will replace "@TAP@" with the one that is used, such as tap0, tap1 ...
QB_SLIRP_OPT: Network option for SLIRP mode (e.g. "-netdev user,id=net0 -device virtio-net-device,netdev=net0")
QB_ROOTFS_OPT: Used as rootfs (e.g.
"-drive id=disk0,file=@ROOTFS@,if=none,format=raw -device virtio-blk-device,drive=disk0").
runqemu will replace "@ROOTFS@" with the one which is used, such as
core-image-minimal-qemuarm64.ext4.
QB_SERIAL_OPT: Serial port (e.g. "-serial mon:stdio")
QB_TCPSERIAL_OPT: tcp serial port option (e.g.
" -device virtio-serial-device -chardev socket,id=virtcon,port=@PORT@,host=127.0.0.1 -device virtconsole,chardev=virtcon"
runqemu will replace "@PORT@" with the port number which is used.
To use runqemu
, set IMAGE_CLASSES as
follows and run runqemu
:
Note
For command-line syntax, use runqemu help
.
IMAGE_CLASSES += "qemuboot"
4.10.8 Default Linker Hash Style Changed
The default linker hash style for gcc-cross
is now “sysv” in order
to catch recipes that are building software without using the
OpenEmbedded LDFLAGS. This change could result in
seeing some “No GNU_HASH in the elf binary” QA issues when building such
recipes. You need to fix these recipes so that they use the expected
LDFLAGS
. Depending on how the software is built, the build system
used by the software (e.g. a Makefile) might need to be patched.
However, sometimes making this fix is as simple as adding the following
to the recipe:
TARGET_CC_ARCH += "${LDFLAGS}"
4.10.9 KERNEL_IMAGE_BASE_NAME
no Longer Uses KERNEL_IMAGETYPE
The KERNEL_IMAGE_BASE_NAME
variable no longer uses the
KERNEL_IMAGETYPE variable to create the
image’s base name. Because the OpenEmbedded build system can now build
multiple kernel image types, this part of the kernel image base name as
been removed leaving only the following:
KERNEL_IMAGE_BASE_NAME ?= "${PKGE}-${PKGV}-${PKGR}-${MACHINE}-${DATETIME}"
If you have recipes or
classes that use KERNEL_IMAGE_BASE_NAME
directly, you might need to
update the references to ensure they continue to work.
4.10.10 BitBake Changes
The following changes took place for BitBake:
The “goggle” UI and standalone image-writer tool have been removed as they both require GTK+ 2.0 and were not being maintained.
The Perforce fetcher now supports SRCREV for specifying the source revision to use, be it
${
AUTOREV}
, changelist number, p4date, or label, in preference to separate SRC_URI parameters to specify these. This change is more in-line with how the other fetchers work for source control systems. Recipes that fetch from Perforce will need to be updated to useSRCREV
in place of specifying the source revision withinSRC_URI
.Some of BitBake’s internal code structures for accessing the recipe cache needed to be changed to support the new multi-configuration functionality. These changes will affect external tools that use BitBake’s tinfoil module. For information on these changes, see the changes made to the scripts supplied with OpenEmbedded-Core: 1 and 2.
The task management code has been rewritten to avoid using ID indirection in order to improve performance. This change is unlikely to cause any problems for most users. However, the setscene verification function as pointed to by
BB_SETSCENE_VERIFY_FUNCTION
needed to change signature. Consequently, a new variable namedBB_SETSCENE_VERIFY_FUNCTION2
has been added allowing multiple versions of BitBake to work with suitably written metadata, which includes OpenEmbedded-Core and Poky. Anyone with custom BitBake task scheduler code might also need to update the code to handle the new structure.
4.10.11 Swabber has Been Removed
Swabber, a tool that was intended to detect host contamination in the build process, has been removed, as it has been unmaintained and unused for some time and was never particularly effective. The OpenEmbedded build system has since incorporated a number of mechanisms including enhanced QA checks that mean that there is less of a need for such a tool.
4.10.12 Removed Recipes
The following recipes have been removed:
augeas
: No longer needed and has been moved tometa-oe
.directfb
: Unmaintained and has been moved tometa-oe
.gcc
: Removed 4.9 version. Versions 5.4 and 6.2 are still present.gnome-doc-utils
: No longer needed.gtk-doc-stub
: Replaced bygtk-doc
.gtk-engines
: No longer needed and has been moved tometa-gnome
.gtk-sato-engine
: Became obsolete.libglade
: No longer needed and has been moved tometa-oe
.libmad
: Unmaintained and functionally replaced bylibmpg123
.libmad
has been moved tometa-oe
.libowl
: Became obsolete.libxsettings-client
: No longer needed.oh-puzzles
: Functionally replaced bypuzzles
.oprofileui
: Became obsolete. OProfile has been largely supplanted by perf.packagegroup-core-directfb.bb
: Removed.core-image-directfb.bb
: Removed.pointercal
: No longer needed and has been moved tometa-oe
.python-imaging
: No longer needed and moved tometa-python
python-pyrex
: No longer needed and moved tometa-python
.sato-icon-theme
: Became obsolete.swabber-native
: Swabber has been removed. See the entry on Swabber.tslib
: No longer needed and has been moved tometa-oe
.uclibc
: Removed in favor of musl.xtscal
: No longer needed and moved tometa-oe
4.10.13 Removed Classes
The following classes have been removed:
distutils-native-base
: No longer needed.distutils3-native-base
: No longer needed.sdl
: Only set DEPENDS and SECTION, which are better set within the recipe instead.sip
: Mostly unused.swabber
: See the entry on Swabber.
4.10.14 Minor Packaging Changes
The following minor packaging changes have occurred:
grub
: Splitgrub-editenv
into its own package.systemd
: Split container and vm related units into a new package, systemd-container.util-linux
: Movedprlimit
to a separateutil-linux-prlimit
package.
4.10.15 Miscellaneous Changes
The following miscellaneous changes have occurred:
package_regex.inc
: Removed because the definitionspackage_regex.inc
previously contained have been moved to their respective recipes.Both
devtool add
andrecipetool create
now use a fixed SRCREV by default when fetching from a Git repository. You can override this in either case to use${
AUTOREV}
instead by using the-a
orDASHDASHautorev
command-line optiondistcc
: GTK+ UI is now disabled by default.packagegroup-core-tools-testapps
: Removed Piglit.image.bbclass
: Renamed COMPRESS(ION) to CONVERSION. This change means thatCOMPRESSIONTYPES
,COMPRESS_DEPENDS
andCOMPRESS_CMD
are deprecated in favor ofCONVERSIONTYPES
,CONVERSION_DEPENDS
andCONVERSION_CMD
. TheCOMPRESS*
variable names will still work in the 2.2 release but metadata that does not need to be backwards-compatible should be changed to use the new names as theCOMPRESS*
ones will be removed in a future release.gtk-doc
: A full version ofgtk-doc
is now made available. However, some old software might not be capable of using the current version ofgtk-doc
to build documentation. You need to change recipes that build such software so that they explicitly disable building documentation withgtk-doc
.
4.11 Moving to the Yocto Project 2.3 Release (pyro)
This section provides migration information for moving to the Yocto Project 2.3 Release (codename “pyro”) from the prior release.
4.11.1 Recipe-specific Sysroots
The OpenEmbedded build system now uses one sysroot per recipe to resolve long-standing issues with configuration script auto-detection of undeclared dependencies. Consequently, you might find that some of your previously written custom recipes are missing declared dependencies, particularly those dependencies that are incidentally built earlier in a typical build process and thus are already likely to be present in the shared sysroot in previous releases.
Consider the following:
Declare Build-Time Dependencies: Because of this new feature, you must explicitly declare all build-time dependencies for your recipe. If you do not declare these dependencies, they are not populated into the sysroot for the recipe.
Specify Pre-Installation and Post-Installation Native Tool Dependencies: You must specifically specify any special native tool dependencies of
pkg_preinst
andpkg_postinst
scripts by using the PACKAGE_WRITE_DEPS variable. Specifying these dependencies ensures that these tools are available if these scripts need to be run on the build host during the do_rootfs task.As an example, see the
dbus
recipe. You will see that this recipe has apkg_postinst
that callssystemctl
if “systemd” is in DISTRO_FEATURES. In the example,systemd-systemctl-native
is added toPACKAGE_WRITE_DEPS
, which is also conditional on “systemd” being inDISTRO_FEATURES
.Examine Recipes that Use
SSTATEPOSTINSTFUNCS
: You need to examine any recipe that usesSSTATEPOSTINSTFUNCS
and determine steps to take.Functions added to
SSTATEPOSTINSTFUNCS
are still called as they were in previous Yocto Project releases. However, since a separate sysroot is now being populated for every recipe and if existing functions being called throughSSTATEPOSTINSTFUNCS
are doing relocation, then you will need to change these to use a post-installation script that is installed by a function added to SYSROOT_PREPROCESS_FUNCS.For an example, see the
pixbufcache
class inmeta/classes/
in the Yocto Project Source Repositories.Note
The SSTATEPOSTINSTFUNCS variable itself is now deprecated in favor of the do_populate_sysroot[postfuncs] task. Consequently, if you do still have any function or functions that need to be called after the sysroot component is created for a recipe, then you would be well advised to take steps to use a post installation script as described previously. Taking these steps prepares your code for when SSTATEPOSTINSTFUNCS is removed in a future Yocto Project release.
Specify the Sysroot when Using Certain External Scripts: Because the shared sysroot is now gone, the scripts
oe-find-native-sysroot
andoe-run-native
have been changed such that you need to specify which recipe’s STAGING_DIR_NATIVE is used.
Note
You can find more information on how recipe-specific sysroots work in the “staging.bbclass” section.
4.11.2 PATH
Variable
Within the environment used to run build tasks, the environment variable
PATH
is now sanitized such that the normal native binary paths
(/bin
, /sbin
, /usr/bin
and so forth) are removed and a
directory containing symbolic links linking only to the binaries from
the host mentioned in the HOSTTOOLS and
HOSTTOOLS_NONFATAL variables is added
to PATH
.
Consequently, any native binaries provided by the host that you need to call needs to be in one of these two variables at the configuration level.
Alternatively, you can add a native recipe (i.e. -native
) that
provides the binary to the recipe’s DEPENDS
value.
Note
PATH
is not sanitized in the same way within devshell
.
If it were, you would have difficulty running host tools for
development and debugging within the shell.
4.11.3 Changes to Scripts
The following changes to scripts took place:
oe-find-native-sysroot
: The usage for theoe-find-native-sysroot
script has changed to the following:$ . oe-find-native-sysroot recipe
You must now supply a recipe for recipe as part of the command. Prior to the Yocto Project 2.3 release, it was not necessary to provide the script with the command.
oe-run-native
: The usage for theoe-run-native
script has changed to the following:$ oe-run-native native_recipe tool
You must supply the name of the native recipe and the tool you want to run as part of the command. Prior to the Yocto Project 2.3 release, it was not necessary to provide the native recipe with the command.
cleanup-workdir
: Thecleanup-workdir
script has been removed because the script was found to be deleting files it should not have, which lead to broken build trees. Rather than trying to delete portions of TMPDIR and getting it wrong, it is recommended that you deleteTMPDIR
and have it restored from shared state (sstate) on subsequent builds.wipe-sysroot
: Thewipe-sysroot
script has been removed as it is no longer needed with recipe-specific sysroots.
4.11.4 Changes to Functions
The previously deprecated bb.data.getVar()
, bb.data.setVar()
,
and related functions have been removed in favor of d.getVar()
,
d.setVar()
, and so forth.
You need to fix any references to these old functions.
4.11.5 BitBake Changes
The following changes took place for BitBake:
BitBake’s Graphical Dependency Explorer UI Replaced: BitBake’s graphical dependency explorer UI
depexp
was replaced bytaskexp
(“Task Explorer”), which provides a graphical way of exploring thetask-depends.dot
file. The data presented by Task Explorer is much more accurate than the data that was presented bydepexp
. Being able to visualize the data is an often requested feature as standard*.dot
file viewers cannot usual cope with the size of thetask-depends.dot
file.BitBake “-g” Output Changes: The
package-depends.dot
andpn-depends.dot
files as previously generated using thebitbake -g
command have been removed. Arecipe-depends.dot
file is now generated as a collapsed version oftask-depends.dot
instead.The reason for this change is because
package-depends.dot
andpn-depends.dot
largely date back to a time before task-based execution and do not take into account task-level dependencies between recipes, which could be misleading.Mirror Variable Splitting Changes: Mirror variables including MIRRORS, PREMIRRORS, and SSTATE_MIRRORS can now separate values entirely with spaces. Consequently, you no longer need “\n”. BitBake looks for pairs of values, which simplifies usage. There should be no change required to existing mirror variable values themselves.
The Subversion (SVN) Fetcher Uses an “ssh” Parameter and Not an “rsh” Parameter: The SVN fetcher now takes an “ssh” parameter instead of an “rsh” parameter. This new optional parameter is used when the “protocol” parameter is set to “svn+ssh”. You can only use the new parameter to specify the
ssh
program used by SVN. The SVN fetcher passes the new parameter through theSVN_SSH
environment variable during the do_fetch task.See the “Subversion (SVN) Fetcher (svn://)” section in the BitBake User Manual for additional information.
BB_SETSCENE_VERIFY_FUNCTION
andBB_SETSCENE_VERIFY_FUNCTION2
Removed: Because the mechanism they were part of is no longer necessary with recipe-specific sysroots, theBB_SETSCENE_VERIFY_FUNCTION
andBB_SETSCENE_VERIFY_FUNCTION2
variables have been removed.
4.11.6 Absolute Symbolic Links
Absolute symbolic links (symlinks) within staged files are no longer
permitted and now trigger an error. Any explicit creation of symlinks
can use the lnr
script, which is a replacement for ln -r
.
If the build scripts in the software that the recipe is building are
creating a number of absolute symlinks that need to be corrected, you
can inherit relative_symlinks
within the recipe to turn those
absolute symlinks into relative symlinks.
4.11.7 GPLv2 Versions of GPLv3 Recipes Moved
Older GPLv2 versions of GPLv3 recipes have moved to a separate
meta-gplv2
layer.
If you use INCOMPATIBLE_LICENSE to
exclude GPLv3 or set PREFERRED_VERSION
to substitute a GPLv2 version of a GPLv3 recipe, then you must add the
meta-gplv2
layer to your configuration.
Note
You can find meta-gplv2
layer in the OpenEmbedded layer index at
https://layers.openembedded.org/layerindex/branch/master/layer/meta-gplv2.
These relocated GPLv2 recipes do not receive the same level of maintenance as other core recipes. The recipes do not get security fixes and upstream no longer maintains them. In fact, the upstream community is actively hostile towards people that use the old versions of the recipes. Moving these recipes into a separate layer both makes the different needs of the recipes clearer and clearly identifies the number of these recipes.
Note
The long-term solution might be to move to BSD-licensed replacements of the GPLv3 components for those that need to exclude GPLv3-licensed components from the target system. This solution will be investigated for future Yocto Project releases.
4.11.8 Package Management Changes
The following package management changes took place:
Smart package manager is replaced by DNF package manager. Smart has become unmaintained upstream, is not ported to Python 3.x. Consequently, Smart needed to be replaced. DNF is the only feasible candidate.
The change in functionality is that the on-target runtime package management from remote package feeds is now done with a different tool that has a different set of command-line options. If you have scripts that call the tool directly, or use its API, they need to be fixed.
For more information, see the DNF Documentation.
Rpm 5.x is replaced with Rpm 4.x. This is done for two major reasons:
DNF is API-incompatible with Rpm 5.x and porting it and maintaining the port is non-trivial.
Rpm 5.x itself has limited maintenance upstream, and the Yocto Project is one of the very few remaining users.
Berkeley DB 6.x is removed and Berkeley DB 5.x becomes the default:
Version 6.x of Berkeley DB has largely been rejected by the open source community due to its AGPLv3 license. As a result, most mainstream open source projects that require DB are still developed and tested with DB 5.x.
In OE-core, the only thing that was requiring DB 6.x was Rpm 5.x. Thus, no reason exists to continue carrying DB 6.x in OE-core.
createrepo
is replaced withcreaterepo_c
.createrepo_c
is the current incarnation of the tool that generates remote repository metadata. It is written in C as compared tocreaterepo
, which is written in Python.createrepo_c
is faster and is maintained.Architecture-independent RPM packages are “noarch” instead of “all”.
This change was made because too many places in DNF/RPM4 stack already make that assumption. Only the filenames and the architecture tag has changed. Nothing else has changed in OE-core system, particularly in the allarch.bbclass class.
Signing of remote package feeds using
PACKAGE_FEED_SIGN
is not currently supported. This issue will be fully addressed in a future Yocto Project release. See defect 11209 for more information on a solution to package feed signing with RPM in the Yocto Project 2.3 release.OPKG now uses the libsolv backend for resolving package dependencies by default. This is vastly superior to OPKG’s internal ad-hoc solver that was previously used. This change does have a small impact on disk (around 500 KB) and memory footprint.
Note
For further details on this change, see the commit message.
4.11.9 Removed Recipes
The following recipes have been removed:
linux-yocto 4.8
: Version 4.8 has been removed. Versions 4.1 (LTSI), 4.4 (LTS), 4.9 (LTS/LTSI) and 4.10 are now present.python-smartpm
: Functionally replaced bydnf
.createrepo
: Replaced by thecreaterepo-c
recipe.rpmresolve
: No longer needed with the move to RPM 4 as RPM itself is used instead.gstreamer
: Removed the GStreamer Git version recipes as they have been stale.1.10.
x recipes are still present.alsa-conf-base
: Merged intoalsa-conf
sincelibasound
depended on both. Essentially, no way existed to install only one of these.tremor
: Moved tometa-multimedia
. Fixed-integer Vorbis decoding is not needed by current hardware. Thus, GStreamer’s ivorbis plugin has been disabled by default eliminating the need for thetremor
recipe in OpenEmbedded-Core (OE-Core).gummiboot
: Replaced bysystemd-boot
.
4.11.10 Wic Changes
The following changes have been made to Wic:
Note
For more information on Wic, see the “Creating Partitioned Images Using Wic” section in the Yocto Project Development Tasks Manual.
Default Output Directory Changed: Wic’s default output directory is now the current directory by default instead of the unusual
/var/tmp/wic
.The “-o” and “–outdir” options remain unchanged and are used to specify your preferred output directory if you do not want to use the default directory.
fsimage Plug-in Removed: The Wic fsimage plugin has been removed as it duplicates functionality of the rawcopy plugin.
4.11.11 QA Changes
The following QA checks have changed:
unsafe-references-in-binaries
: Theunsafe-references-in-binaries
QA check, which was disabled by default, has now been removed. This check was intended to detect binaries in/bin
that link to libraries in/usr/lib
and have the case where the user has/usr
on a separate filesystem to/
.The removed QA check was buggy. Additionally,
/usr
residing on a separate partition from/
is now a rare configuration. Consequently,unsafe-references-in-binaries
was removed.file-rdeps
: Thefile-rdeps
QA check is now an error by default instead of a warning. Because it is an error instead of a warning, you need to address missing runtime dependencies.For additional information, see the insane class and the “Errors and Warnings” section.
4.11.12 Miscellaneous Changes
The following miscellaneous changes have occurred:
In this release, a number of recipes have been changed to ignore the
largefile
DISTRO_FEATURES item, enabling large file support unconditionally. This feature has always been enabled by default. Disabling the feature has not been widely tested.Note
Future releases of the Yocto Project will remove entirely the ability to disable the largefile feature, which would make it unconditionally enabled everywhere.
If the DISTRO_VERSION value contains the value of the DATE variable, which is the default between Poky releases, the
DATE
value is explicitly excluded from/etc/issue
and/etc/issue.net
, which is displayed at the login prompt, in order to avoid conflicts with Multilib enabled. Regardless, theDATE
value is inaccurate if thebase-files
recipe is restored from shared state (sstate) rather than rebuilt.If you need the build date recorded in
/etc/issue*
or anywhere else in your image, a better method is to define a post-processing function to do it and have the function called from ROOTFS_POSTPROCESS_COMMAND. Doing so ensures the value is always up-to-date with the created image.Dropbear’s
init
script now disables DSA host keys by default. This change is in line with the systemd service file, which supports RSA keys only, and with recent versions of OpenSSH, which deprecates DSA host keys.The buildhistory class now correctly uses tabs as separators between all columns in
installed-package-sizes.txt
in order to aid import into other tools.The
USE_LDCONFIG
variable has been replaced with the “ldconfig”DISTRO_FEATURES
feature. Distributions that previously set:USE_LDCONFIG = "0"
should now instead use the following:
DISTRO_FEATURES_BACKFILL_CONSIDERED_append = " ldconfig"
The default value of COPYLEFT_LICENSE_INCLUDE now includes all versions of AGPL licenses in addition to GPL and LGPL.
Note
The default list is not intended to be guaranteed as a complete safe list. You should seek legal advice based on what you are distributing if you are unsure.
Kernel module packages are now suffixed with the kernel version in order to allow module packages from multiple kernel versions to co-exist on a target system. If you wish to return to the previous naming scheme that does not include the version suffix, use the following:
KERNEL_MODULE_PACKAGE_SUFFIX = ""
Removal of
libtool
*.la
files is now enabled by default. The*.la
files are not actually needed on Linux and relocating them is an unnecessary burden.If you need to preserve these
.la
files (e.g. in a custom distribution), you must change INHERIT_DISTRO such that “remove-libtool” is not included in the value.Extensible SDKs built for GCC 5+ now refuse to install on a distribution where the host GCC version is 4.8 or 4.9. This change resulted from the fact that the installation is known to fail due to the way the
uninative
shared state (sstate) package is built. See the uninative class for additional information.All native and nativesdk recipes now use a separate
DISTRO_FEATURES
value instead of sharing the value used by recipes for the target, in order to avoid unnecessary rebuilds.The
DISTRO_FEATURES
fornative
recipes is DISTRO_FEATURES_NATIVE added to an intersection ofDISTRO_FEATURES
and DISTRO_FEATURES_FILTER_NATIVE.For nativesdk recipes, the corresponding variables are DISTRO_FEATURES_NATIVESDK and DISTRO_FEATURES_FILTER_NATIVESDK.
The
FILESDIR
variable, which was previously deprecated and rarely used, has now been removed. You should change any recipes that setFILESDIR
to set FILESPATH instead.The
MULTIMACH_HOST_SYS
variable has been removed as it is no longer needed with recipe-specific sysroots.
4.12 Moving to the Yocto Project 2.4 Release (rocko)
This section provides migration information for moving to the Yocto Project 2.4 Release (codename “rocko”) from the prior release.
4.12.1 Memory Resident Mode
A persistent mode is now available in BitBake’s default operation,
replacing its previous “memory resident mode” (i.e.
oe-init-build-env-memres
). Now you only need to set
BB_SERVER_TIMEOUT to a timeout (in
seconds) and BitBake’s server stays resident for that amount of time
between invocations. The oe-init-build-env-memres
script has been
removed since a separate environment setup script is no longer needed.
4.12.2 Packaging Changes
This section provides information about packaging changes that have occurred:
python3
Changes:The main “python3” package now brings in all of the standard Python 3 distribution rather than a subset. This behavior matches what is expected based on traditional Linux distributions. If you wish to install a subset of Python 3, specify
python-core
plus one or more of the individual packages that are still produced.python3
: Thebz2.py
,lzma.py
, and_compression.py
scripts have been moved from thepython3-misc
package to thepython3-compression
package.
binutils
: Thelibbfd
library is now packaged in a separate “libbfd” package. This packaging saves space when certain tools (e.g.perf
) are installed. In such cases, the tools only needlibbfd
rather than all the packages inbinutils
.util-linux
Changes:The
su
program is now packaged in a separate “util-linux-su” package, which is only built when “pam” is listed in the DISTRO_FEATURES variable.util-linux
should not be installed unless it is needed becausesu
is normally provided through the shadow file format. The mainutil-linux
package has runtime dependencies (i.e. RDEPENDS) on theutil-linux-su
package when “pam” is inDISTRO_FEATURES
.The
switch_root
program is now packaged in a separate “util-linux-switch-root” package for small initramfs images that do not need the wholeutil-linux
package or the busybox binary, which are both much larger thanswitch_root
. The mainutil-linux
package has a recommended runtime dependency (i.e. RRECOMMENDS) on theutil-linux-switch-root
package.The
ionice
program is now packaged in a separate “util-linux-ionice” package. The mainutil-linux
package has a recommended runtime dependency (i.e.RRECOMMENDS
) on theutil-linux-ionice
package.
initscripts
: Thesushell
program is now packaged in a separate “initscripts-sushell” package. This packaging change allows systems to pullsushell
in whenselinux
is enabled. The change also eliminates needing to pull in the entireinitscripts
package. The maininitscripts
package has a runtime dependency (i.e.RDEPENDS
) on thesushell
package when “selinux” is inDISTRO_FEATURES
.glib-2.0
: Theglib-2.0
package now has a recommended runtime dependency (i.e.RRECOMMENDS
) on theshared-mime-info
package, since large portions of GIO are not useful without the MIME database. You can remove the dependency by using the BAD_RECOMMENDATIONS variable ifshared-mime-info
is too large and is not required.Go Standard Runtime: The Go standard runtime has been split out from the main
go
recipe into a separatego-runtime
recipe.
4.12.3 Removed Recipes
The following recipes have been removed:
acpitests
: This recipe is not maintained.autogen-native
: No longer required by Grub, oe-core, or meta-oe.bdwgc
: Nothing in OpenEmbedded-Core requires this recipe. It has moved to meta-oe.byacc
: This recipe was only needed by rpm 5.x and has moved to meta-oe.gcc (5.4)
: The 5.4 series dropped the recipe in favor of 6.3 / 7.2.gnome-common
: Deprecated upstream and no longer needed.go-bootstrap-native
: Go 1.9 does its own bootstrapping so this recipe has been removed.guile
: This recipe was only needed byautogen-native
andremake
. The recipe is no longer needed by either of these programs.libclass-isa-perl
: This recipe was previously needed for LSB 4, no longer needed.libdumpvalue-perl
: This recipe was previously needed for LSB 4, no longer needed.libenv-perl
: This recipe was previously needed for LSB 4, no longer needed.libfile-checktree-perl
: This recipe was previously needed for LSB 4, no longer needed.libi18n-collate-perl
: This recipe was previously needed for LSB 4, no longer needed.libiconv
: This recipe was only needed foruclibc
, which was removed in the previous release.glibc
andmusl
have their own implementations.meta-mingw
still needslibiconv
, so it has been moved tometa-mingw
.libpng12
: This recipe was previously needed for LSB. The currentlibpng
is 1.6.x.libpod-plainer-perl
: This recipe was previously needed for LSB 4, no longer needed.linux-yocto (4.1)
: This recipe was removed in favor of 4.4, 4.9, 4.10 and 4.12.mailx
: This recipe was previously only needed for LSB compatibility, and upstream is defunct.mesa (git version only)
: The git version recipe was stale with respect to the release version.ofono (git version only)
: The git version recipe was stale with respect to the release version.portmap
: This recipe is obsolete and is superseded byrpcbind
.python3-pygpgme
: This recipe is old and unmaintained. It was previously required bydnf
, which has switched to officialgpgme
Python bindings.python-async
: This recipe has been removed in favor of the Python 3 version.python-gitdb
: This recipe has been removed in favor of the Python 3 version.python-git
: This recipe was removed in favor of the Python 3 version.python-mako
: This recipe was removed in favor of the Python 3 version.python-pexpect
: This recipe was removed in favor of the Python 3 version.python-ptyprocess
: This recipe was removed in favor of Python the 3 version.python-pycurl
: Nothing is using this recipe in OpenEmbedded-Core (i.e.meta-oe
).python-six
: This recipe was removed in favor of the Python 3 version.python-smmap
: This recipe was removed in favor of the Python 3 version.remake
: Usingremake
as the provider ofvirtual/make
is broken. Consequently, this recipe is not needed in OpenEmbedded-Core.
4.12.4 Kernel Device Tree Move
Kernel Device Tree support is now easier to enable in a kernel recipe.
The Device Tree code has moved to a
kernel-devicetree class.
Functionality is automatically enabled for any recipe that inherits the
kernel class and sets the
KERNEL_DEVICETREE variable. The
previous mechanism for doing this,
meta/recipes-kernel/linux/linux-dtb.inc
, is still available to avoid
breakage, but triggers a deprecation warning. Future releases of the
Yocto Project will remove meta/recipes-kernel/linux/linux-dtb.inc
.
It is advisable to remove any require
statements that request
meta/recipes-kernel/linux/linux-dtb.inc
from any custom kernel
recipes you might have. This will avoid breakage in post 2.4 releases.
4.12.5 Package QA Changes
The following package QA changes took place:
The “unsafe-references-in-scripts” QA check has been removed.
If you refer to
${COREBASE}/LICENSE
within LIC_FILES_CHKSUM you receive a warning because this file is a description of the license for OE-Core. Use${COMMON_LICENSE_DIR}/MIT
if your recipe is MIT-licensed and you cannot use the preferred method of referring to a file within the source tree.
4.12.6 README
File Changes
The following are changes to README
files:
The main Poky
README
file has been moved to themeta-poky
layer and has been renamedREADME.poky
. A symlink has been created so that references to the old location work.The
README.hardware
file has been moved tometa-yocto-bsp
. A symlink has been created so that references to the old location work.A
README.qemu
file has been created with coverage of theqemu*
machines.
4.12.7 Miscellaneous Changes
The following are additional changes:
The
ROOTFS_PKGMANAGE_BOOTSTRAP
variable and any references to it have been removed. You should remove this variable from any custom recipes.The
meta-yocto
directory has been removed.Note
In the Yocto Project 2.1 release meta-yocto was renamed to meta-poky and the meta-yocto subdirectory remained to avoid breaking existing configurations.
The
maintainers.inc
file, which tracks maintainers by listing a primary person responsible for each recipe in OE-Core, has been moved frommeta-poky
to OE-Core (i.e. frommeta-poky/conf/distro/include
tometa/conf/distro/include
).The buildhistory class now makes a single commit per build rather than one commit per subdirectory in the repository. This behavior assumes the commits are enabled with BUILDHISTORY_COMMIT = “1”, which is typical. Previously, the
buildhistory
class made one commit per subdirectory in the repository in order to make it easier to see the changes for a particular subdirectory. To view a particular change, specify that subdirectory as the last parameter on thegit show
orgit diff
commands.The
x86-base.inc
file, which is included by all x86-based machine configurations, now sets IMAGE_FSTYPES using?=
to “live” rather than appending with+=
. This change makes the default easier to override.BitBake fires multiple “BuildStarted” events when multiconfig is enabled (one per configuration). For more information, see the “Events” section in the BitBake User Manual.
By default, the
security_flags.inc
file sets a GCCPIE variable with an option to enable Position Independent Executables (PIE) withingcc
. Enabling PIE in the GNU C Compiler (GCC), makes Return Oriented Programming (ROP) attacks much more difficult to execute.OE-Core now provides a
bitbake-layers
plugin that implements a “create-layer” subcommand. The implementation of this subcommand has resulted in theyocto-layer
script being deprecated and will likely be removed in the next Yocto Project release.The
vmdk
,vdi
, andqcow2
image file types are now used in conjunction with the “wic” image type throughCONVERSION_CMD
. Consequently, the equivalent image types are nowwic.vmdk
,wic.vdi
, andwic.qcow2
, respectively.do_image_<type>[depends]
has replacedIMAGE_DEPENDS_<type>
. If you have your own classes that implement custom image types, then you need to update them.OpenSSL 1.1 has been introduced. However, the default is still 1.0.x through the PREFERRED_VERSION variable. This preference is set is due to the remaining compatibility issues with other software. The PROVIDES variable in the openssl 1.0 recipe now includes “openssl10” as a marker that can be used in DEPENDS within recipes that build software that still depend on OpenSSL 1.0.
To ensure consistent behavior, BitBake’s “-r” and “-R” options (i.e. prefile and postfile), which are used to read or post-read additional configuration files from the command line, now only affect the current BitBake command. Before these BitBake changes, these options would “stick” for future executions.
4.13 Moving to the Yocto Project 2.5 Release (sumo)
This section provides migration information for moving to the Yocto Project 2.5 Release (codename “sumo”) from the prior release.
4.13.1 Packaging Changes
This section provides information about packaging changes that have occurred:
bind-libs
: The libraries packaged by the bind recipe are in a separatebind-libs
package.libfm-gtk
: Thelibfm
GTK+ bindings are split into a separatelibfm-gtk
package.flex-libfl
: The flex recipe splits out libfl into a separateflex-libfl
package to avoid too many dependencies being pulled in where only the library is needed.grub-efi
: Thegrub-efi
configuration is split into a separategrub-bootconf
recipe. However, the dependency relationship fromgrub-efi
is through a virtual/grub-bootconf provider making it possible to have your own recipe provide the dependency. Alternatively, you can use a BitBake append file to bring the configuration back into thegrub-efi
recipe.armv7a Legacy Package Feed Support: Legacy support is removed for transitioning from
armv7a
toarmv7a-vfp-neon
in package feeds, which was previously enabled by settingPKGARCHCOMPAT_ARMV7A
. This transition occurred in 2011 and active package feeds should by now be updated to the new naming.
4.13.2 Removed Recipes
The following recipes have been removed:
gcc
: The version 6.4 recipes are replaced by 7.x.gst-player
: Renamed togst-examples
as per upstream.hostap-utils
: This software package is obsolete.latencytop
: This recipe is no longer maintained upstream. The last release was in 2009.libpfm4
: The only file that requires this recipe isoprofile
, which has been removed.linux-yocto
: The version 4.4, 4.9, and 4.10 recipes have been removed. Versions 4.12, 4.14, and 4.15 remain.man
: This recipe has been replaced by modernman-db
mkelfimage
: This tool has been removed in the upstream coreboot project, and is no longer needed with the removal of the ELF image type.nativesdk-postinst-intercept
: This recipe is not maintained.neon
: This software package is no longer maintained upstream and is no longer needed by anything in OpenEmbedded-Core.oprofile
: The functionality of this recipe is replaced byperf
and keeping compatibility on an ongoing basis withmusl
is difficult.pax
: This software package is obsolete.stat
: This software package is not maintained upstream.coreutils
provides a modern stat binary.zisofs-tools-native
: This recipe is no longer needed because the compressed ISO image feature has been removed.
4.13.3 Scripts and Tools Changes
The following are changes to scripts and tools:
yocto-bsp
,yocto-kernel
, andyocto-layer
: Theyocto-bsp
,yocto-kernel
, andyocto-layer
scripts previously shipped with poky but not in OpenEmbedded-Core have been removed. These scripts are not maintained and are outdated. In many cases, they are also limited in scope. Thebitbake-layers create-layer
command is a direct replacement foryocto-layer
. See the documentation to create a BSP or kernel recipe in the “BSP Kernel Recipe Example” section.devtool finish
:devtool finish
now exits with an error if there are uncommitted changes or a rebase/am in progress in the recipe’s source repository. If this error occurs, there might be uncommitted changes that will not be included in updates to the patches applied by the recipe. A -f/–force option is provided for situations that the uncommitted changes are inconsequential and you want to proceed regardless.scripts/oe-setup-rpmrepo
script: The functionality ofscripts/oe-setup-rpmrepo
is replaced bybitbake package-index
.scripts/test-dependencies.sh
script: The script is largely made obsolete by the recipe-specific sysroots functionality introduced in the previous release.
4.13.4 BitBake Changes
The following are BitBake changes:
The
--runall
option has changed. There are two different behaviors people might want:Behavior A: For a given target (or set of targets) look through the task graph and run task X only if it is present and will be built.
Behavior B: For a given target (or set of targets) look through the task graph and run task X if any recipe in the taskgraph has such a target, even if it is not in the original task graph.
The
--runall
option now performs “Behavior B”. Previously--runall
behaved like “Behavior A”. A--runonly
option has been added to retain the ability to perform “Behavior A”.Several explicit “run this task for all recipes in the dependency tree” tasks have been removed (e.g.
fetchall
,checkuriall
, and the*all
tasks provided by thedistrodata
andarchiver
classes). There is a BitBake option to complete this for any arbitrary task. For example:bitbake <target> -c fetchall
should now be replaced with:
bitbake <target> --runall=fetch
4.13.5 Python and Python 3 Changes
The following are auto-packaging changes to Python and Python 3:
The script-managed python-*-manifest.inc
files that were previously
used to generate Python and Python 3 packages have been replaced with a
JSON-based file that is easier to read and maintain. A new task is
available for maintainers of the Python recipes to update the JSON file
when upgrading to new Python versions. You can now edit the file
directly instead of having to edit a script and run it to update the
file.
One particular change to note is that the Python recipes no longer have
build-time provides for their packages. This assumes python-foo
is
one of the packages provided by the Python recipe. You can no longer run
bitbake python-foo
or have a
DEPENDS on python-foo
,
but doing either of the following causes the package to work as
expected:
IMAGE_INSTALL_append = " python-foo"
or
RDEPENDS_${PN} = "python-foo"
The earlier build-time provides behavior was a quirk of the way the Python manifest file was created. For more information on this change please see this commit.
4.13.6 Miscellaneous Changes
The following are additional changes:
The
kernel
class supports building packages for multiple kernels. If your kernel recipe or.bbappend
file mentions packaging at all, you should replace references to the kernel in package names with${KERNEL_PACKAGE_NAME}
. For example, if you disable automatic installation of the kernel image usingRDEPENDS_kernel-base = ""
you can avoid warnings usingRDEPENDS_${KERNEL_PACKAGE_NAME}-base = ""
instead.The
buildhistory
class commits changes to the repository by default so you no longer need to setBUILDHISTORY_COMMIT = "1"
. If you want to disable commits you need to setBUILDHISTORY_COMMIT = "0"
in your configuration.The
beaglebone
reference machine has been renamed tobeaglebone-yocto
. Thebeaglebone-yocto
BSP is a reference implementation using only mainline components available in OpenEmbedded-Core andmeta-yocto-bsp
, whereas Texas Instruments maintains a full-featured BSP in themeta-ti
layer. This rename avoids the previous name clash that existed between the two BSPs.The
update-alternatives
class no longer works with SysVinit
scripts because this usage has been problematic. Also, thesysklogd
recipe no longer usesupdate-alternatives
because it is incompatible with other implementations.By default, the cmake class uses
ninja
instead ofmake
for building. This improves build performance. If a recipe is broken withninja
, then the recipe can setOECMAKE_GENERATOR = "Unix Makefiles"
to change back tomake
.The previously deprecated
base_*
functions have been removed in favor of their replacements inmeta/lib/oe
andbitbake/lib/bb
. These are typically used from recipes and classes. Any references to the old functions must be updated. The following table shows the removed functions and their replacements:Removed
Replacement
base_path_join()
oe.path.join()
base_path_relative()
oe.path.relative()
base_path_out()
oe.path.format_display()
base_read_file()
oe.utils.read_file()
base_ifelse()
oe.utils.ifelse()
base_conditional()
oe.utils.conditional()
base_less_or_equal()
oe.utils.less_or_equal()
base_version_less_or_equal()
oe.utils.version_less_or_equal()
base_contains()
bb.utils.contains()
base_both_contain()
oe.utils.both_contain()
base_prune_suffix()
oe.utils.prune_suffix()
oe_filter()
oe.utils.str_filter()
oe_filter_out()
oe.utils.str_filter_out() (or use the _remove operator)
Using
exit 1
to explicitly defer a postinstall script until first boot is now deprecated since it is not an obvious mechanism and can mask actual errors. If you want to explicitly defer a postinstall to first boot on the target rather than atrootfs
creation time, usepkg_postinst_ontarget()
or callpostinst_intercept delay_to_first_boot
frompkg_postinst()
. Any failure of apkg_postinst()
script (includingexit 1
) will trigger a warning duringdo_rootfs
.For more information, see the “Post-Installation Scripts” section in the Yocto Project Development Tasks Manual.
The
elf
image type has been removed. This image type was removed because themkelfimage
tool that was required to create it is no longer provided by coreboot upstream and required updating every timebinutils
updated.Support for .iso image compression (previously enabled through
COMPRESSISO = "1"
) has been removed. The userspace tools (zisofs-tools
) are unmaintained andsquashfs
provides better performance and compression. In order to build a live image with squashfs+lz4 compression enabled you should now setLIVE_ROOTFS_TYPE = "squashfs-lz4"
and ensure thatlive
is inIMAGE_FSTYPES
.Recipes with an unconditional dependency on
libpam
are only buildable withpam
inDISTRO_FEATURES
. If the dependency is truly optional then it is recommended that the dependency be conditional uponpam
being inDISTRO_FEATURES
.For EFI-based machines, the bootloader (
grub-efi
by default) is installed into the image at /boot. Wic can be used to split the bootloader into separate boot and rootfs partitions if necessary.Patches whose context does not match exactly (i.e. where patch reports “fuzz” when applying) will generate a warning. For an example of this see this commit.
Layers are expected to set
LAYERSERIES_COMPAT_layername
to match the version(s) of OpenEmbedded-Core they are compatible with. This is specified as codenames using spaces to separate multiple values (e.g. “rocko sumo”). If a layer does not setLAYERSERIES_COMPAT_layername
, a warning will is shown. If a layer sets a value that does not include the current version (“sumo” for the 2.5 release), then an error will be produced.The
TZ
environment variable is set to “UTC” within the build environment in order to fix reproducibility problems in some recipes.
4.14 Moving to the Yocto Project 2.6 Release (thud)
This section provides migration information for moving to the Yocto Project 2.6 Release (codename “thud”) from the prior release.
4.14.1 GCC 8.2 is Now Used by Default
The GNU Compiler Collection version 8.2 is now used by default for compilation. For more information on what has changed in the GCC 8.x release, see https://gcc.gnu.org/gcc-8/changes.html.
If you still need to compile with version 7.x, GCC 7.3 is also provided. You can select this version by setting the and can be selected by setting the GCCVERSION variable to “7.%” in your configuration.
4.14.2 Removed Recipes
The following recipes have been removed:
beecrypt: No longer needed since moving to RPM 4.
bigreqsproto: Replaced by
xorgproto
.calibrateproto: Removed in favor of
xinput
.compositeproto: Replaced by
xorgproto
.damageproto: Replaced by
xorgproto
.dmxproto: Replaced by
xorgproto
.dri2proto: Replaced by
xorgproto
.dri3proto: Replaced by
xorgproto
.eee-acpi-scripts: Became obsolete.
fixesproto: Replaced by
xorgproto
.fontsproto: Replaced by
xorgproto
.fstests: Became obsolete.
gccmakedep: No longer used.
glproto: Replaced by
xorgproto
.gnome-desktop3: No longer needed. This recipe has moved to
meta-oe
.icon-naming-utils: No longer used since the Sato theme was removed in 2016.
inputproto: Replaced by
xorgproto
.kbproto: Replaced by
xorgproto
.libusb-compat: Became obsolete.
libuser: Became obsolete.
libnfsidmap: No longer an external requirement since
nfs-utils
2.2.1.libnfsidmap
is now integrated.libxcalibrate: No longer needed with
xinput
mktemp: Became obsolete. The
mktemp
command is provided by bothbusybox
andcoreutils
.ossp-uuid: Is not being maintained and has mostly been replaced by
uuid.h
inutil-linux
.pax-utils: No longer needed. Previous QA tests that did use this recipe are now done at build time.
pcmciautils: Became obsolete.
pixz: No longer needed.
xz
now supports multi-threaded compression.presentproto: Replaced by
xorgproto
.randrproto: Replaced by
xorgproto
.recordproto: Replaced by
xorgproto
.renderproto: Replaced by
xorgproto
.resourceproto: Replaced by
xorgproto
.scrnsaverproto: Replaced by
xorgproto
.trace-cmd: Became obsolete.
perf
replaced this recipe’s functionally.videoproto: Replaced by
xorgproto
.wireless-tools: Became obsolete. Superseded by
iw
.xcmiscproto: Replaced by
xorgproto
.xextproto: Replaced by
xorgproto
.xf86dgaproto: Replaced by
xorgproto
.xf86driproto: Replaced by
xorgproto
.xf86miscproto: Replaced by
xorgproto
.xf86-video-omapfb: Became obsolete. Use kernel modesetting driver instead.
xf86-video-omap: Became obsolete. Use kernel modesetting driver instead.
xf86vidmodeproto: Replaced by
xorgproto
.xineramaproto: Replaced by
xorgproto
.xproto: Replaced by
xorgproto
.yasm: No longer needed since previous usages are now satisfied by
nasm
.
4.14.3 Packaging Changes
The following packaging changes have been made:
cmake:
cmake.m4
andtoolchain
files have been moved to the main package.iptables: The
iptables
modules have been split into separate packages.alsa-lib:
libasound
is now in the mainalsa-lib
package instead oflibasound
.glibc:
libnss-db
is now in its own package along with a/var/db/makedbs.sh
script to update databases.python and python3: The main package has been removed from the recipe. You must install specific packages or
python-modules
/python3-modules
for everything.systemtap: Moved
systemtap-exporter
into its own package.
4.14.4 XOrg Protocol dependencies
The *proto
upstream repositories have been combined into one
“xorgproto” repository. Thus, the corresponding recipes have also been
combined into a single xorgproto
recipe. Any recipes that depend
upon the older *proto
recipes need to be changed to depend on the
newer xorgproto
recipe instead.
For names of recipes removed because of this repository change, see the Removed Recipes section.
4.14.5 distutils
and distutils3
Now Prevent Fetching Dependencies During the do_configure
Task
Previously, it was possible for Python recipes that inherited the
distutils
and
distutils3 classes to fetch code
during the do_configure task to satisfy
dependencies mentioned in setup.py
if those dependencies were not
provided in the sysroot (i.e. recipes providing the dependencies were
missing from DEPENDS).
Note
This change affects classes beyond just the two mentioned (i.e.
distutils
and distutils3
). Any recipe that inherits distutils*
classes are affected. For example, the setuptools
and setuptools3
recipes are affected since they inherit the distutils*
classes.
Fetching these types of dependencies that are not provided in the
sysroot negatively affects the ability to reproduce builds. This type of
fetching is now explicitly disabled. Consequently, any missing
dependencies in Python recipes that use these classes now result in an
error during the do_configure
task.
4.14.6 linux-yocto
Configuration Audit Issues Now Correctly Reported
Due to a bug, the kernel configuration audit functionality was not
writing out any resulting warnings during the build. This issue is now
corrected. You might notice these warnings now if you have a custom
kernel configuration with a linux-yocto
style kernel recipe.
4.14.7 Image/Kernel Artifact Naming Changes
The following changes have been made:
Name variables (e.g. IMAGE_NAME) use a new
IMAGE_VERSION_SUFFIX
variable instead of DATETIME. UsingIMAGE_VERSION_SUFFIX
allows easier and more direct changes.The
IMAGE_VERSION_SUFFIX
variable is set in thebitbake.conf
configuration file as follows:IMAGE_VERSION_SUFFIX = "-${DATETIME}"
Several variables have changed names for consistency:
Old Variable Name New Variable Name ======================================================== KERNEL_IMAGE_BASE_NAME KERNEL_IMAGE_NAME KERNEL_IMAGE_SYMLINK_NAME KERNEL_IMAGE_LINK_NAME MODULE_TARBALL_BASE_NAME MODULE_TARBALL_NAME MODULE_TARBALL_SYMLINK_NAME MODULE_TARBALL_LINK_NAME INITRAMFS_BASE_NAME INITRAMFS_NAME
The
MODULE_IMAGE_BASE_NAME
variable has been removed. The module tarball name is now controlled directly with the MODULE_TARBALL_NAME variable.The KERNEL_DTB_NAME and KERNEL_DTB_LINK_NAME variables have been introduced to control kernel Device Tree Binary (DTB) artifact names instead of mangling
KERNEL_IMAGE_*
variables.The KERNEL_FIT_NAME and KERNEL_FIT_LINK_NAME variables have been introduced to specify the name of flattened image tree (FIT) kernel images similar to other deployed artifacts.
The MODULE_TARBALL_NAME and MODULE_TARBALL_LINK_NAME variable values no longer include the “module-” prefix or “.tgz” suffix. These parts are now hardcoded so that the values are consistent with other artifact naming variables.
Added the INITRAMFS_LINK_NAME variable so that the symlink can be controlled similarly to other artifact types.
INITRAMFS_NAME now uses “${PKGE}-${PKGV}-${PKGR}-${MACHINE}${IMAGE_VERSION_SUFFIX}” instead of “${PV}-${PR}-${MACHINE}-${DATETIME}”, which makes it consistent with other variables.
4.14.8 SERIAL_CONSOLE
Deprecated
The SERIAL_CONSOLE variable has been
functionally replaced by the
SERIAL_CONSOLES variable for some time.
With the Yocto Project 2.6 release, SERIAL_CONSOLE
has been
officially deprecated.
SERIAL_CONSOLE
will continue to work as before for the 2.6 release.
However, for the sake of future compatibility, it is recommended that
you replace all instances of SERIAL_CONSOLE
with
SERIAL_CONSOLES
.
Note
The only difference in usage is that SERIAL_CONSOLES
expects entries to be separated using semicolons as compared to
SERIAL_CONSOLE
, which expects spaces.
4.14.9 Configure Script Reports Unknown Options as Errors
If the configure script reports an unknown option, this now triggers a QA error instead of a warning. Any recipes that previously got away with specifying such unknown options now need to be fixed.
4.14.10 Override Changes
The following changes have occurred:
The
virtclass-native
andvirtclass-nativesdk
Overrides Have Been Removed: Thevirtclass-native
andvirtclass-nativesdk
overrides have been deprecated since 2012 in favor ofclass-native
andclass-nativesdk
, respectively. Bothvirtclass-native
andvirtclass-nativesdk
are now dropped.Note
The
virtclass-multilib-
overrides for multilib are still valid.The
forcevariable
Override Now Has a Higher Priority Thanlibc
Overrides: Theforcevariable
override is documented to be the highest priority override. However, due to a long-standing quirk of how OVERRIDES is set, thelibc
overrides (e.g.libc-glibc
,libc-musl
, and so forth) erroneously had a higher priority. This issue is now corrected.It is likely this change will not cause any problems. However, it is possible with some unusual configurations that you might see a change in behavior if you were relying on the previous behavior. Be sure to check how you use
forcevariable
andlibc-*
overrides in your custom layers and configuration files to ensure they make sense.The
build-${BUILD_OS}
Override Has Been Removed: Thebuild-${BUILD_OS}
, which is typicallybuild-linux
, override has been removed because building on a host operating system other than a recent version of Linux is neither supported nor recommended. Dropping the override avoids giving the impression that other host operating systems might be supported.The “_remove” operator now preserves whitespace. Consequently, when specifying list items to remove, be aware that leading and trailing whitespace resulting from the removal is retained.
See the “Removal (Override Style Syntax)” section in the BitBake User Manual for a detailed example.
4.14.11 systemd
Configuration is Now Split Into systemd-conf
The configuration for the systemd
recipe has been moved into a
system-conf
recipe. Moving this configuration to a separate recipe
avoids the systemd
recipe from becoming machine-specific for cases
where machine-specific configurations need to be applied (e.g. for
qemu*
machines).
Currently, the new recipe packages the following files:
${sysconfdir}/machine-id
${sysconfdir}/systemd/coredump.conf
${sysconfdir}/systemd/journald.conf
${sysconfdir}/systemd/logind.conf
${sysconfdir}/systemd/system.conf
${sysconfdir}/systemd/user.conf
If you previously used bbappend files to append the systemd
recipe to
change any of the listed files, you must do so for the systemd-conf
recipe instead.
4.14.12 Automatic Testing Changes
This section provides information about automatic testing changes:
TEST_IMAGE
Variable Removed: Prior to this release, you set theTEST_IMAGE
variable to “1” to enable automatic testing for successfully built images. TheTEST_IMAGE
variable no longer exists and has been replaced by the TESTIMAGE_AUTO variable.Inheriting the
testimage
andtestsdk
Classes: Best practices now dictate that you use the IMAGE_CLASSES variable rather than the INHERIT variable when you inherit the testimage and testsdk classes used for automatic testing.
4.14.13 OpenSSL Changes
OpenSSL has been upgraded from 1.0 to 1.1. By default, this upgrade could cause problems for recipes that have both versions in their dependency chains. The problem is that both versions cannot be installed together at build time.
Note
It is possible to have both versions of the library at runtime.
4.14.14 BitBake Changes
The server logfile bitbake-cookerdaemon.log
is now always placed in
the Build Directory instead of the current
directory.
4.14.15 Security Changes
The Poky distribution now uses security compiler flags by default. Inclusion of these flags could cause new failures due to stricter checking for various potential security issues in code.
4.14.16 Post Installation Changes
You must explicitly mark post installs to defer to the target. If you
want to explicitly defer a postinstall to first boot on the target
rather than at rootfs creation time, use pkg_postinst_ontarget()
or
call postinst_intercept delay_to_first_boot
from pkg_postinst()
.
Any failure of a pkg_postinst()
script (including exit 1) triggers
an error during the do_rootfs task.
For more information on post-installation behavior, see the “Post-Installation Scripts” section in the Yocto Project Development Tasks Manual.
4.14.17 Python 3 Profile-Guided Optimization
The python3
recipe now enables profile-guided optimization. Using
this optimization requires a little extra build time in exchange for
improved performance on the target at runtime. Additionally, the
optimization is only enabled if the current
MACHINE has support for user-mode emulation in
QEMU (i.e. “qemu-usermode” is in
MACHINE_FEATURES, which it is by
default).
If you wish to disable Python profile-guided optimization regardless of
the value of MACHINE_FEATURES
, then ensure that
PACKAGECONFIG for the python3
recipe
does not contain “pgo”. You could accomplish the latter using the
following at the configuration level:
PACKAGECONFIG_remove_pn-python3 = "pgo"
Alternatively, you can set PACKAGECONFIG
using an append file
for the python3
recipe.
4.14.18 Miscellaneous Changes
The following miscellaneous changes occurred:
Default to using the Thumb-2 instruction set for armv7a and above. If you have any custom recipes that build software that needs to be built with the ARM instruction set, change the recipe to set the instruction set as follows:
ARM_INSTRUCTION_SET = "arm"
run-postinsts
no longer uses/etc/*-postinsts
fordpkg/opkg
in favor of built-in postinst support. RPM behavior remains unchanged.The
NOISO
andNOHDD
variables are no longer used. You now control building*.iso
and*.hddimg
image types directly by using the IMAGE_FSTYPES variable.The
scripts/contrib/mkefidisk.sh
has been removed in favor of Wic.kernel-modules
has been removed from RRECOMMENDS forqemumips
andqemumips64
machines. Removal also impacts thex86-base.inc
file.Note
genericx86
andgenericx86-64
retainkernel-modules
as part of theRRECOMMENDS
variable setting.The
LGPLv2_WHITELIST_GPL-3.0
variable has been removed. If you are setting this variable in your configuration, set or append it to theWHITELIST_GPL-3.0
variable instead.${ASNEEDED}
is now included in the TARGET_LDFLAGS variable directly. The remaining definitions frommeta/conf/distro/include/as-needed.inc
have been moved to corresponding recipes.Support for DSA host keys has been dropped from the OpenSSH recipes. If you are still using DSA keys, you must switch over to a more secure algorithm as recommended by OpenSSH upstream.
The
dhcp
recipe now uses thedhcpd6.conf
configuration file indhcpd6.service
for IPv6 DHCP rather than re-usingdhcpd.conf
, which is now reserved for IPv4.
4.15 Moving to the Yocto Project 2.7 Release (warrior)
This section provides migration information for moving to the Yocto Project 2.7 Release (codename “warrior”) from the prior release.
4.15.1 BitBake Changes
The following changes have been made to BitBake:
BitBake now checks anonymous Python functions and pure Python functions (e.g.
def funcname:
) in the metadata for tab indentation. If found, BitBake produces a warning.Bitbake now checks BBFILE_COLLECTIONS for duplicate entries and triggers an error if any are found.
4.15.2 Eclipse Support Removed
Support for the Eclipse IDE has been removed. Support continues for those releases prior to 2.7 that did include support. The 2.7 release does not include the Eclipse Yocto plugin.
4.15.3 qemu-native
Splits the System and User-Mode Parts
The system and user-mode parts of qemu-native
are now split.
qemu-native
provides the user-mode components and
qemu-system-native
provides the system components. If you have
recipes that depend on QEMU’s system emulation functionality at build
time, they should now depend upon qemu-system-native
instead of
qemu-native
.
4.15.4 The upstream-tracking.inc
File Has Been Removed
The previously deprecated upstream-tracking.inc
file is now removed.
Any UPSTREAM_TRACKING*
variables are now set in the corresponding
recipes instead.
Remove any references you have to the upstream-tracking.inc
file in
your configuration.
4.15.5 The DISTRO_FEATURES_LIBC
Variable Has Been Removed
The DISTRO_FEATURES_LIBC
variable is no longer used. The ability to
configure glibc using kconfig has been removed for quite some time
making the libc-*
features set no longer effective.
Remove any references you have to DISTRO_FEATURES_LIBC
in your own
layers.
4.15.6 License Value Corrections
The following corrections have been made to the LICENSE values set by recipes:
socat: Corrected
LICENSE
to be “GPLv2” rather than “GPLv2+”.libgfortran: Set license to “GPL-3.0-with-GCC-exception”.
elfutils: Removed “Elfutils-Exception” and set to “GPLv2” for shared libraries
4.15.7 Packaging Changes
This section provides information about packaging changes.
bind
: Thensupdate
binary has been moved to thebind-utils
package.Debug split: The default debug split has been changed to create separate source packages (i.e. package_name
-dbg
and package_name-src
). If you are currently usingdbg-pkgs
in IMAGE_FEATURES to bring in debug symbols and you still need the sources, you must now also addsrc-pkgs
toIMAGE_FEATURES
. Source packages remain in the target portion of the SDK by default, unless you have set your own value for SDKIMAGE_FEATURES that does not includesrc-pkgs
.Mount all using
util-linux
:/etc/default/mountall
has moved into the -mount sub-package.Splitting binaries using
util-linux
:util-linux
now splits each binary into its own package for fine-grained control. The mainutil-linux
package pulls in the individual binary packages using the RRECOMMENDS and RDEPENDS variables. As a result, existing images should not see any changes assuming NO_RECOMMENDATIONS is not set.netbase/base-files
:/etc/hosts
has moved fromnetbase
tobase-files
.tzdata
: The main package has been converted to an empty meta package that pulls in alltzdata
packages by default.lrzsz
: This package has been removed frompackagegroup-self-hosted
andpackagegroup-core-tools-testapps
. The X/Y/ZModem support is less likely to be needed on modern systems. If you are relying on these packagegroups to include thelrzsz
package in your image, you now need to explicitly add the package.
4.15.8 Removed Recipes
The following recipes have been removed:
gcc: Drop version 7.3 recipes. Version 8.3 now remains.
linux-yocto: Drop versions 4.14 and 4.18 recipes. Versions 4.19 and 5.0 remain.
go: Drop version 1.9 recipes. Versions 1.11 and 1.12 remain.
xvideo-tests: Became obsolete.
libart-lgpl: Became obsolete.
gtk-icon-utils-native: These tools are now provided by gtk+3-native
gcc-cross-initial: No longer needed. gcc-cross/gcc-crosssdk is now used instead.
gcc-crosssdk-initial: No longer needed. gcc-cross/gcc-crosssdk is now used instead.
glibc-initial: Removed because the benefits of having it for site_config are currently outweighed by the cost of building the recipe.
4.15.9 Removed Classes
The following classes have been removed:
distutils-tools: This class was never used.
bugzilla.bbclass: Became obsolete.
distrodata: This functionally has been replaced by a more modern tinfoil-based implementation.
4.15.10 Miscellaneous Changes
The following miscellaneous changes occurred:
The
distro
subdirectory of the Poky repository has been removed from the top-levelscripts
directory.Perl now builds for the target using perl-cross for better maintainability and improved build performance. This change should not present any problems unless you have heavily customized your Perl recipe.
arm-tunes
: Removed the “-march” option if mcpu is already added.update-alternatives
: Convert file renames to PACKAGE_PREPROCESS_FUNCSbase/pixbufcache
: Obsoletesstatecompletions
code has been removed.inetutils
: This recipe has rsh disabled.
4.16 Moving to the Yocto Project 3.0 Release (zeus)
This section provides migration information for moving to the Yocto Project 3.0 Release (codename “zeus”) from the prior release.
4.16.1 Init System Selection
Changing the init system manager previously required setting a number of
different variables. You can now change the manager by setting the
INIT_MANAGER
variable and the corresponding include files (i.e.
conf/distro/include/init-manager-*.conf
). Include files are provided
for four values: “none”, “sysvinit”, “systemd”, and “mdev-busybox”. The
default value, “none”, for INIT_MANAGER
should allow your current
settings to continue working. However, it is advisable to explicitly set
INIT_MANAGER
.
4.16.2 LSB Support Removed
Linux Standard Base (LSB) as a standard is not current, and is not well suited for embedded applications. Support can be continued in a separate layer if needed. However, presently LSB support has been removed from the core.
As a result of this change, the poky-lsb
derivative distribution
configuration that was also used for testing alternative configurations
has been replaced with a poky-altcfg
distribution that has LSB parts
removed.
4.16.3 Removed Recipes
The following recipes have been removed.
core-image-lsb-dev
: Part of removed LSB support.core-image-lsb
: Part of removed LSB support.core-image-lsb-sdk
: Part of removed LSB support.cve-check-tool
: Functionally replaced by thecve-update-db
recipe andcve-check
class.eglinfo
: No longer maintained.eglinfo
frommesa-demos
is an adequate and maintained alternative.gcc-8.3
: Version 8.3 removed. Replaced by 9.2.gnome-themes-standard
: Only needed by gtk+ 2.x, which has been removed.gtk+
: GTK+ 2 is obsolete and has been replaced by gtk+3.irda-utils
: Has become obsolete. IrDA support has been removed from the Linux kernel in version 4.17 and later.libnewt-python
:libnewt
Python support merged into mainlibnewt
recipe.libsdl
: Replaced by newerlibsdl2
.libx11-diet
: Became obsolete.libxx86dga
: Removed obsolete client library.libxx86misc
: Removed. Library is redundant.linux-yocto
: Version 5.0 removed, which is now redundant (5.2 / 4.19 present).lsbinitscripts
: Part of removed LSB support.lsb
: Part of removed LSB support.lsbtest
: Part of removed LSB support.openssl10
: Replaced by neweropenssl
version 1.1.packagegroup-core-lsb
: Part of removed LSB support.python-nose
: Removed the Python 2.x version of the recipe.python-numpy
: Removed the Python 2.x version of the recipe.python-scons
: Removed the Python 2.x version of the recipe.source-highlight
: No longer needed.stress
: Replaced bystress-ng
.vulkan
: Split intovulkan-loader
,vulkan-headers
, andvulkan-tools
.weston-conf
: Functionality moved toweston-init
.
4.16.4 Packaging Changes
The following packaging changes have occurred.
The Epiphany browser has been dropped from
packagegroup-self-hosted
as it has not been needed insidebuild-appliance-image
for quite some time and was causing resource problems.libcap-ng
Python support has been moved to a separatelibcap-ng-python
recipe to streamline the build process when the Python bindings are not needed.libdrm
now packages the fileamdgpu.ids
into a separatelibdrm-amdgpu
package.python3
: Therunpy
module is now in thepython3-core
package as it is required to support the common “python3 -m” command usage.distcc
now provides separatedistcc-client
anddistcc-server
packages as typically one or the other are needed, rather than both.python*-setuptools
recipes now separately package thepkg_resources
module in apython-pkg-resources
/python3-pkg-resources
package as the module is useful independent of the rest of the setuptools package. The mainpython-setuptools
/python3-setuptools
package depends on this new package so you should only need to update dependencies unless you want to take advantage of the increased granularity.
4.16.5 CVE Checking
cve-check-tool
has been functionally replaced by a new
cve-update-db
recipe and functionality built into the cve-check
class. The result uses NVD JSON data feeds rather than the deprecated
XML feeds that cve-check-tool
was using, supports CVSSv3 scoring,
and makes other improvements.
Additionally, the CVE_CHECK_CVE_WHITELIST
variable has been replaced
by CVE_CHECK_WHITELIST
.
4.16.6 Bitbake Changes
The following BitBake changes have occurred.
addtask
statements now properly validate dependent tasks. Previously, an invalid task was silently ignored. With this change, the invalid task generates a warning.Other invalid
addtask
anddeltask
usages now trigger these warnings: “multiple target tasks arguments with addtask / deltask”, and “multiple before/after clauses”.The “multiconfig” prefix is now shortened to “mc”. “multiconfig” will continue to work, however it may be removed in a future release.
The
bitbake -g
command no longer generates arecipe-depends.dot
file as the contents (i.e. a reprocessed version oftask-depends.dot
) were confusing.The
bb.build.FuncFailed
exception, previously raised bybb.build.exec_func()
when certain other exceptions have occurred, has been removed. The real underlying exceptions will be raised instead. If you have calls tobb.build.exec_func()
in custom classes ortinfoil-using
scripts, any references tobb.build.FuncFailed
should be cleaned up.Additionally, the
bb.build.exec_func()
no longer accepts the “pythonexception” parameter. The function now always raises exceptions. Remove this argument in any calls tobb.build.exec_func()
in custom classes or scripts.The BB_SETSCENE_VERIFY_FUNCTION2 is no longer used. In the unlikely event that you have any references to it, they should be removed.
The
RunQueueExecuteScenequeue
andRunQueueExecuteTasks
events have been removed since setscene tasks are now executed as part of the normal runqueue. Any event handling code in custom classes or scripts that handles these two events need to be updated.The arguments passed to functions used with BB_HASHCHECK_FUNCTION have changed. If you are using your own custom hash check function, see https://git.yoctoproject.org/cgit/cgit.cgi/poky/commit/?id=40a5e193c4ba45c928fccd899415ea56b5417725 for details.
Task specifications in
BB_TASKDEPDATA
and class implementations used in signature generator classes now use “<fn>:<task>” everywhere rather than the “.” delimiter that was being used in some places. This change makes it consistent with all areas in the code. Custom signature generator classes and code that readsBB_TASKDEPDATA
need to be updated to use ‘:’ as a separator rather than ‘.’.
4.16.7 Sanity Checks
The following sanity check changes occurred.
SRC_URI is now checked for usage of two problematic items:
“${PN}” prefix/suffix use - Warnings always appear if ${PN} is used. You must fix the issue regardless of whether multiconfig or anything else that would cause prefixing/suffixing to happen.
Github archive tarballs - these are not guaranteed to be stable. Consequently, it is likely that the tarballs will be refreshed and thus the SRC_URI checksums will fail to apply. It is recommended that you fetch either an official release tarball or a specific revision from the actual Git repository instead.
Either one of these items now trigger a warning by default. If you wish to disable this check, remove
src-uri-bad
from WARN_QA.The
file-rdeps
runtime dependency check no longer expands RDEPENDS recursively as there is no mechanism to ensure they can be fully computed, and thus races sometimes result in errors either showing up or not. Thus, you might now see errors for missing runtime dependencies that were previously satisfied recursively. Here is an example: package A contains a shell script starting with#!/bin/bash
but has no dependency on bash. However, package A depends on package B, which does depend on bash. You need to add the missing dependency or dependencies to resolve the warning.Setting
DEPENDS_${PN}
anywhere (i.e. typically in a recipe) now triggers an error. The error is triggered because DEPENDS is not a package-specific variable unlike RDEPENDS. You should setDEPENDS
instead.systemd currently does not work well with the musl C library because only upstream officially supports linking the library with glibc. Thus, a warning is shown when building systemd in conjunction with musl.
4.16.8 Miscellaneous Changes
The following miscellaneous changes have occurred.
The
gnome
class has been removed because it now does very little. You should update recipes that previously inherited this class to do the following: inherit gnomebase gtk-icon-cache gconf mimeThe
meta/recipes-kernel/linux/linux-dtb.inc
file has been removed. This file was previously deprecated in favor of setting KERNEL_DEVICETREE in any kernel recipe and only produced a warning. Remove anyinclude
orrequire
statements pointing to this file.TARGET_CFLAGS, TARGET_CPPFLAGS, TARGET_CXXFLAGS, and TARGET_LDFLAGS are no longer exported to the external environment. This change did not require any changes to core recipes, which is a good indicator that no changes will be required. However, if for some reason the software being built by one of your recipes is expecting these variables to be set, then building the recipe will fail. In such cases, you must either export the variable or variables in the recipe or change the scripts so that exporting is not necessary.
You must change the host distro identifier used in NATIVELSBSTRING to use all lowercase characters even if it does not contain a version number. This change is necessary only if you are not using
uninative
and SANITY_TESTED_DISTROS.In the
base-files
recipe, writing the hostname into/etc/hosts
and/etc/hostname
is now done within the main do_install function rather than in thedo_install_basefilesissue
function. The reason for the change is becausedo_install_basefilesissue
is more easily overridden without having to duplicate the hostname functionality. If you have done the latter (e.g. in abase-files
bbappend), then you should remove it from your customizeddo_install_basefilesissue
function.The
wic --expand
command now uses commas to separate “key:value” pairs rather than hyphens.Note
The wic command-line help is not updated.
You must update any scripts or commands where you use
wic --expand
with multiple “key:value” pairs.UEFI image variable settings have been moved from various places to a central
conf/image-uefi.conf
. This change should not influence any existing configuration as themeta/conf/image-uefi.conf
in the core metadata sets defaults that can be overridden in the same manner as before.conf/distro/include/world-broken.inc
has been removed. For cases where certain recipes need to be disabled when using the musl C library, these recipes now haveCOMPATIBLE_HOST_libc-musl
set with a comment that explains why.
4.17 Moving to the Yocto Project 3.1 Release (dunfell)
This section provides migration information for moving to the Yocto Project 3.1 Release (codename “dunfell”) from the prior release.
4.17.1 Minimum system requirements
The following versions / requirements of build host components have been updated:
gcc 5.0
python 3.5
tar 1.28
rpcgen
is now required on the host (part of thelibc-dev-bin
package on Ubuntu, Debian and related distributions, and theglibc
package on RPM-based distributions).
Additionally, the makeinfo
and pod2man
tools are no longer
required on the host.
4.17.2 mpc8315e-rdb machine removed
The MPC8315E-RDB machine is old/obsolete and unobtainable, thus given
the maintenance burden the mpc8315e-rdb
machine configuration that
supported it has been removed in this release. The removal does leave a
gap in official PowerPC reference hardware support; this may change in
future if a suitable machine with accompanying support resources is
found.
4.17.3 Python 2 removed
Due to the expiration of upstream support in January 2020, support for Python 2 has now been removed; it is recommended that you use Python 3 instead. If absolutely needed there is a meta-python2 community layer containing Python 2, related classes and various Python 2-based modules, however it should not be considered as supported.
4.17.4 Reproducible builds now enabled by default
In order to avoid unnecessary differences in output files (aiding binary
reproducibility), the Poky distribution configuration
(DISTRO = "poky"
) now inherits the reproducible_build
class by
default.
4.17.5 Impact of ptest feature is now more significant
The Poky distribution configuration (DISTRO = "poky"
) enables ptests
by default to enable runtime testing of various components. In this
release, a dependency needed to be added that has resulted in a
significant increase in the number of components that will be built just
when building a simple image such as core-image-minimal. If you do not
need runtime tests enabled for core components, then it is recommended
that you remove “ptest” from
DISTRO_FEATURES to save a significant
amount of build time e.g. by adding the following in your configuration:
DISTRO_FEATURES_remove = "ptest"
4.17.6 Removed recipes
The following recipes have been removed:
chkconfig
: obsoleteconsole-tools
: obsoleteenchant
: replaced byenchant2
foomatic-filters
: obsoletelibidn
: no longer needed, moved to meta-oelibmodulemd
: replaced bylibmodulemd-v1
linux-yocto
: drop 4.19, 5.2 version recipes (5.4 now provided)nspr
: no longer needed, moved to meta-oenss
: no longer needed, moved to meta-oepython
: Python 2 removed (Python 3 preferred)python-setuptools
: Python 2 version removed (python3-setuptools preferred)sysprof
: no longer needed, moved to meta-oetexi2html
: obsoleteu-boot-fw-utils
: functionally replaced bylibubootenv
4.17.7 features_check class replaces distro_features_check
The distro_features_check
class has had its functionality expanded,
now supporting ANY_OF_MACHINE_FEATURES
,
REQUIRED_MACHINE_FEATURES
, CONFLICT_MACHINE_FEATURES
,
ANY_OF_COMBINED_FEATURES
, REQUIRED_COMBINED_FEATURES
,
CONFLICT_COMBINED_FEATURES
. As a result the class has now been
renamed to features_check
; the distro_features_check
class still
exists but generates a warning and redirects to the new class. In
preparation for a future removal of the old class it is recommended that
you update recipes currently inheriting distro_features_check
to
inherit features_check
instead.
4.17.8 Removed classes
The following classes have been removed:
distutils-base
: moved to meta-python2distutils
: moved to meta-python2libc-common
: merged into the glibc recipe as nothing else used it.python-dir
: moved to meta-python2pythonnative
: moved to meta-python2setuptools
: moved to meta-python2tinderclient
: dropped as it was obsolete.
4.17.9 SRC_URI checksum behaviour
Previously, recipes by tradition included both SHA256 and MD5 checksums for remotely fetched files in SRC_URI, even though only one is actually mandated. However, the MD5 checksum does not add much given its inherent weakness; thus when a checksum fails only the SHA256 sum will now be printed. The md5sum will still be verified if it is specified.
4.17.10 npm fetcher changes
The npm fetcher has been completely reworked in this release. The npm
fetcher now only fetches the package source itself and no longer the
dependencies; there is now also an npmsw fetcher which explicitly
fetches the shrinkwrap file and the dependencies. This removes the
slightly awkward NPM_LOCKDOWN
and NPM_SHRINKWRAP
variables which
pointed to local files; the lockdown file is no longer needed at all.
Additionally, the package name in npm://
entries in
SRC_URI is now specified using a package
parameter instead of the earlier name
which overlapped with the
generic name
parameter. All recipes using the npm fetcher will need
to be changed as a result.
An example of the new scheme:
SRC_URI = "npm://registry.npmjs.org;package=array-flatten;version=1.1.1 \
npmsw://${THISDIR}/npm-shrinkwrap.json"
Another example where the sources are fetched from git rather than an npm repository:
SRC_URI = "git://github.com/foo/bar.git;protocol=https \
npmsw://${THISDIR}/npm-shrinkwrap.json"
devtool and recipetool have also been updated to match with the npm
fetcher changes. Other than producing working and more complete recipes
for npm sources, there is also a minor change to the command line for
devtool: the --fetch-dev
option has been renamed to --npm-dev
as
it is npm-specific.
4.17.11 Packaging changes
intltool
has been removed frompackagegroup-core-sdk
as it is rarely needed to build modern software - gettext can do most of the things it used to be needed for.intltool
has also been removed frompackagegroup-core-self-hosted
as it is not needed to for standard builds.git:
git-am
,git-difftool
,git-submodule
, andgit-request-pull
are no longer perl-based, so are now installed with the maingit
package instead of withingit-perltools
.The
ldconfig
binary built as part of glibc has now been moved to its ownldconfig
package (note noglibc-
prefix). This package is in the RRECOMMENDS of the mainglibc
package ifldconfig
is present in DISTRO_FEATURES.libevent
now splits each shared library into its own package (as Debian does). Since these are shared libraries and will be pulled in through the normal shared library dependency handling, there should be no impact to existing configurations other than less unnecessary libraries being installed in some cases.linux-firmware now has a new package for
bcm4366c
and includes available NVRAM config files into thebcm43340
,bcm43362
,bcm43430
andbcm4356-pcie
packages.harfbuzz
now splits the newlibharfbuzz-subset.so
library into its own package to reduce the main package size in cases wherelibharfbuzz-subset.so
is not needed.
4.17.12 Additional warnings
Warnings will now be shown at do_package_qa
time in the following
circumstances:
A recipe installs
.desktop
files containingMimeType
keys but does not inherit the newmime-xdg
classA recipe installs
.xml
files into${datadir}/mime/packages
but does not inherit themime
class
4.17.13 wic
image type now used instead of live
by default for x86
conf/machine/include/x86-base.inc
(inherited by most x86 machine
configurations) now specifies wic
instead of live
by default in
IMAGE_FSTYPES. The live
image type will
likely be removed in a future release so it is recommended that you use
wic
instead.
4.17.14 Miscellaneous changes
The undocumented
SRC_DISTRIBUTE_LICENSES
variable has now been removed in favour of a newAVAILABLE_LICENSES
variable which is dynamically set based upon license files found in${COMMON_LICENSE_DIR}
and${LICENSE_PATH}
.The tune definition for big-endian microblaze machines is now
microblaze
instead ofmicroblazeeb
.newlib
no longer has built-in syscalls.libgloss
should then provide the syscalls,crt0.o
and other functions that are no longer part ofnewlib
itself. If you are usingTCLIBC = "newlib"
this now means that you must link applications with bothnewlib
andlibgloss
, whereas beforenewlib
would run in many configurations by itself.
4.18 Moving to the Yocto Project 3.2 Release (gatesgarth)
This section provides migration information for moving to the Yocto Project 3.2 Release (codename “gatesgarth”) from the prior release.
4.18.1 Minimum system requirements
gcc
version 6.0 is now required at minimum on the build host. For older
host distributions where this is not available, you can use the
buildtools-extended-tarball
(easily installable using
scripts/install-buildtools
).
4.18.2 Removed recipes
The following recipes have been removed:
bjam-native
: replaced byboost-build-native
avahi-ui
: folded into the mainavahi
recipe - the GTK UI can be disabled using PACKAGECONFIG foravahi
.build-compare
: no longer needed with the removal of thepackagefeed-stability
classdhcp
: obsolete, functionally replaced bydhcpcd
andkea
libmodulemd-v1
: replaced bylibmodulemd
packagegroup-core-device-devel
: obsolete
4.18.3 Removed classes
The following classes (.bbclass files) have been removed:
spdx
: obsolete - the Yocto Project is a strong supporter of SPDX, but this class was old code using a dated approach and had the potential to be misleading. Themeta-sdpxscanner
layer is a much more modern and active approach to handling this and is recommended as a replacement.packagefeed-stability
: this class had become obsolete with the advent of hash equivalence and reproducible builds.
4.18.4 pseudo path filtering and mismatch behaviour
pseudo now operates on a filtered subset of files. This is a significant change
to the way pseudo operates within OpenEmbedded - by default, pseudo monitors and
logs (adds to its database) any file created or modified whilst in a fakeroot
environment. However, there are large numbers of files that we simply don’t care
about the permissions of whilst in that fakeroot
context, for example ${S}, ${B}, ${T},
${SSTATE_DIR}, the central sstate control directories, and others.
As of this release, new functionality in pseudo is enabled to ignore these directory trees (controlled using a new PSEUDO_IGNORE_PATHS variable) resulting in a cleaner database with less chance of “stray” mismatches if files are modified outside pseudo context. It also should reduce some overhead from pseudo as the interprocess round trip to the server is avoided.
There is a possible complication where some existing recipe may break, for
example, a recipe was found to be writing to ${B}/install
for
make install
in do_install
and since ${B}
is listed as not to be tracked,
there were errors trying to chown root
for files in this location. Another
example was the tcl
recipe where the source directory S
is set to a
subdirectory of the source tree but files were written out to the directory
structure above that subdirectory. For these types of cases in your own recipes,
extend PSEUDO_IGNORE_PATHS
to cover additional paths that pseudo should not
be monitoring.
In addition, pseudo’s behaviour on mismatches has now been changed - rather
than doing what turns out to be a rather dangerous “fixup” if it sees a file
with a different path but the same inode as another file it has previously seen,
pseudo will throw an abort()
and direct you to a wiki page
that explains how to deal with this.
4.18.5 MLPREFIX
now required for multilib when runtime dependencies conditionally added
In order to solve some previously intractable problems with runtime dependencies and multilib, a change was made that now requires the MLPREFIX value to be explicitly prepended to package names being added as dependencies (e.g. in RDEPENDS and RRECOMMENDS values) where the dependency is conditionally added.
If you have anonymous python or in-line python conditionally adding
dependencies in your custom recipes, and you intend for those recipes to
work with multilib, then you will need to ensure that ${MLPREFIX}
is prefixed on the package names in the dependencies, for example
(from the glibc
recipe):
RRECOMMENDS_${PN} = "${@bb.utils.contains('DISTRO_FEATURES', 'ldconfig', '${MLPREFIX}ldconfig', '', d)}"
This also applies when conditionally adding packages to PACKAGES where
those packages have dependencies, for example (from the alsa-plugins
recipe):
PACKAGES += "${@bb.utils.contains('PACKAGECONFIG', 'pulseaudio', 'alsa-plugins-pulseaudio-conf', '', d)}"
...
RDEPENDS_${PN}-pulseaudio-conf += "\
${MLPREFIX}libasound-module-conf-pulse \
${MLPREFIX}libasound-module-ctl-pulse \
${MLPREFIX}libasound-module-pcm-pulse \
"
4.18.6 packagegroup-core-device-devel no longer included in images built for qemu* machines
packagegroup-core-device-devel
was previously added automatically to
images built for qemu*
machines, however the purpose of the group and what
it should contain is no longer clear, and in general, adding userspace
development items to images is best done at the image/class level; thus this
packagegroup was removed.
This packagegroup previously pulled in the following:
distcc-config
nfs-export-root
bash
binutils-symlinks
If you still need any of these in your image built for a qemu*
machine
then you will add them explicitly to IMAGE_INSTALL or another
appropriate place in the dependency chain for your image (if you have not
already done so).
4.18.7 DHCP server/client replaced
The dhcp
software package has become unmaintained and thus has been
functionally replaced by dhcpcd
(client) and kea
(server). You will
need to replace references to the recipe/package names as appropriate - most
commonly, at the package level dhcp-client
should be replaced by
dhcpcd
and dhcp-server
should be replaced by kea
. If you have any
custom configuration files for these they will need to be adapted - refer to
the upstream documentation for dhcpcd
and kea
for further details.
4.18.8 Packaging changes
python3
: theurllib
python package has now moved into the core package, as it is used more commonly than just netclient (e.g. email, xml, mimetypes, pydoc). In addition, thepathlib
module is now also part of the core package.iptables
:iptables-apply
andip6tables-apply
have been split out to their own package to avoid a bash dependency in the mainiptables
package
4.18.9 Package QA check changes
Previously, the following package QA checks triggered warnings, however they can be indicators of genuine underlying problems and are therefore now treated as errors:
In addition, the following new checks were added and default to triggering an error:
shebang-size: Check for shebang (#!) lines longer than 128 characters, which can give an error at runtime depending on the operating system.
unhandled-features-check: Check if any of the variables supported by the features_check class is set while not inheriting the class itself.
missing-update-alternatives: Check if the recipe sets the ALTERNATIVE variable for any of its packages, and does not inherit the update-alternatives class.
A trailing slash or duplicated slashes in the value of S or B will now trigger a warning so that they can be removed and path comparisons can be more reliable - remove any instances of these in your recipes if the warning is displayed.
4.18.10 Globbing no longer supported in file://
entries in SRC_URI
Globbing (*
and ?
wildcards) in file://
URLs within SRC_URI
did not properly support file checksums, thus changes to the source files
would not always change the do_fetch task checksum, and consequently would
not ensure that the changed files would be incorporated in subsequent builds.
Unfortunately it is not practical to make globbing work generically here, so
the decision was taken to remove support for globs in file://
URLs.
If you have any usage of these in your recipes, then you will now need to
either add each of the files that you expect to match explicitly, or
alternatively if you still need files to be pulled in dynamically, put the
files into a subdirectory and reference that instead.
4.18.11 deploy class now cleans DEPLOYDIR
before do_deploy
do_deploy
as implemented in the deploy class now cleans up ${DEPLOYDIR} before running, just as do_install
cleans up ${D} before running. This reduces the risk of DEPLOYDIR
being accidentally contaminated by files from previous runs, possibly even with different config, in case of incremental builds.
Most recipes and classes that inherit the deploy
class or interact with do_deploy
are unlikely to be affected by this unless they add prefuncs
to do_deploy
which also put files into ${DEPLOYDIR}
- these should be refactored to use do_deploy_prepend
instead.
4.18.12 Custom SDK / SDK-style recipes need to include nativesdk-sdk-provides-dummy
All nativesdk
packages require /bin/sh
due to their postinstall scriptlets, thus this package has to be dummy-provided within the SDK and nativesdk-sdk-provides-dummy
now does this. If you have a custom SDK recipe (or your own SDK-style recipe similar to e.g. buildtools-tarball
), you will need to ensure nativesdk-sdk-provides-dummy
or an equivalent is included in TOOLCHAIN_HOST_TASK.
4.18.13 ld.so.conf
now moved back to main glibc
package
There are cases where one doesn’t want ldconfig
on target (e.g. for
read-only root filesystems, it’s rather pointless), yet one still
needs /etc/ld.so.conf
to be present at image build time:
When some recipe installs libraries to a non-standard location, and
therefore installs in a file in /etc/ld.so.conf.d/foo.conf
, we
need /etc/ld.so.conf
containing:
include /etc/ld.so.conf.d/*.conf
in order to get those other locations picked up.
Thus /etc/ld.so.conf
is now in the main glibc
package so that
there’s always an ld.so.conf
present when the build-time ldconfig
runs towards the end of image construction.
The ld.so.conf
and ld.so.conf.d/*.conf
files do not take up
significant space (at least not compared to the ~700kB ldconfig
binary), and they
might be needed in case ldconfig
is installable, so they are left
in place after the image is built. Technically it would be possible to
remove them if desired, though it would not be trivial if you still
wanted the build-time ldconfig to function (ROOTFS_POSTPROCESS_COMMAND
will not work as ldconfig
is run after the functions referred to
by that variable).
4.18.14 Host DRI drivers now used for GL support within runqemu
runqemu
now uses the mesa-native libraries everywhere virgl is used
(i.e. when gl
, gl-es
or egl-headless
options are specified),
but instructs them to load DRI drivers from the host. Unfortunately this
may not work well with proprietary graphics drivers such as those from
Nvidia; if you are using such drivers then you may need to switch to an
alternative (such as Nouveau in the case of Nvidia hardware) or avoid
using the GL options.
4.18.15 initramfs images now use a blank suffix
The reference initramfs images (core-image-minimal-initramfs
,
core-image-tiny-initramfs
and core-image-testmaster-initramfs
) now
set an empty string for IMAGE_NAME_SUFFIX, which otherwise defaults
to ".rootfs"
. These images aren’t root filesystems and thus the rootfs
label didn’t make sense. If you are looking for the output files generated
by these image recipes directly then you will need to adapt to the new
naming without the .rootfs
part.
4.18.16 Image artifact name variables now centralised in image-artifact-names class
The defaults for the following image artifact name variables have been moved
from bitbake.conf to a new image-artifact-names
class:
Image-related classes now inherit this class, and typically these variables
are only referenced within image recipes so those will be unaffected by this
change. However if you have references to these variables in either a recipe
that is not an image or a class that is enabled globally, then those will
now need to be changed to inherit image-artifact-names
.
4.18.17 Miscellaneous changes
Support for the long-deprecated
PACKAGE_GROUP
variable has now been removed - replace any remaining instances with FEATURE_PACKAGES.The
FILESPATHPKG
variable, having been previously deprecated, has now been removed. Replace any remaining references with appropriate use of FILESEXTRAPATHS.Erroneous use of
inherit +=
(instead ofINHERIT +=
) in a configuration file now triggers an error instead of silently being ignored.ptest support has been removed from the
kbd
recipe, as upstream has moved to autotest which is difficult to work with in a cross-compilation environment.oe.utils.is_machine_specific()
andoe.utils.machine_paths()
have been removed as their utility was questionable. In the unlikely event that you have references to these in your own code, then the code will need to be reworked.The
i2ctransfer
module is now disabled by default when buildingbusybox
in order to be consistent with disabling the other i2c tools there. If you do wish the i2ctransfer module to be built in BusyBox then addCONFIG_I2CTRANSFER=y
to your custom BusyBox configuration.In the
Upstream-Status
header convention for patches,Accepted
has been replaced withBackport
as these almost always mean the same thing i.e. the patch is already upstream and may need to be removed in a future recipe upgrade. If you are adding these headers to your own patches then useBackport
to indicate that the patch has been sent upstream.The
tune-supersparc.inc
tune file has been removed as it does not appear to be widely used and no longer works.
4.19 Moving to the Yocto Project 3.3 Release (hardknott)
This section provides migration information for moving to the Yocto Project 3.3 Release (codename “hardknott”) from the prior release.
4.19.1 Minimum system requirements
You will now need at least Python 3.6 installed on your build host. Most recent
distributions provide this, but should you be building on a distribution that
does not have it, you can use the buildtools-tarball
(easily installable
using scripts/install-buildtools
) - see
Required Git, tar, Python and gcc Versions
for details.
4.19.2 Removed recipes
The following recipes have been removed:
go-dep
: obsolete with the advent of go modulesgst-validate
: replaced bygst-devtools
linux-yocto
: removed 5.8 version recipes (5.4 / 5.10 still provided)vulkan-demos
: replaced byvulkan-samples
4.19.3 Single version common license file naming
Some license files in meta/files/common-licenses
have been renamed to match
current SPDX naming conventions:
AGPL-3.0 -> AGPL-3.0-only
GPL-1.0 -> GPL-1.0-only
GPL-2.0 -> GPL-2.0-only
GPL-3.0 -> GPL-3.0-only
LGPL-2.0 -> LGPL-2.0-only
LGPL-2.1 -> LGPL-2.1-only
LGPL-3.0 -> LGPL-3.0-only
Additionally, corresponding “-or-later” suffixed files have been added e.g.
GPL-2.0-or-later
.
It is not required that you change LICENSE values as there are mappings
from the original names in place; however, in rare cases where you have a recipe
which sets LIC_FILES_CHKSUM to point to file(s) in
meta/files/common-licenses
(which in any case is not recommended) you will
need to update those.
4.19.4 New python3targetconfig
class
A new python3targetconfig class has been
created for situations where you would previously have inherited the
python3native
class but need access to target configuration data (such as
correct installation directories). Recipes where this situation applies should
be changed to inherit python3targetconfig
instead of python3native
. This
also adds a dependency on target python3
, so it should only be used where
appropriate in order to avoid unnecessarily lengthening builds.
Some example recipes where this change has been made: gpgme
, libcap-ng
,
python3-pycairo
.
4.19.5 setup.py
path for python modules
In a Python module, sometimes setup.py
can be buried deep in the
source tree. Previously this was handled in recipes by setting S to
point to the subdirectory within the source where setup.py
is located.
However with the recent pseudo
changes, some Python modules make changes to files beneath ${S}
, for
example:
S = "${WORKDIR}/git/python/pythonmodule"
then in setup.py
it works with source code in a relative fashion, such
as ../../src
. This causes pseudo to abort as it isn’t able to track
the paths properly. This release introduces a new DISTUTILS_SETUP_PATH
variable so that recipes can specify it explicitly, for example:
S = "${WORKDIR}/git"
DISTUTILS_SETUP_PATH = "${S}/python/pythonmodule"
Recipes that inherit from distutils3 (or setuptools3 which itself inherits distutils3) that also set S to point to a Python module within a subdirectory in the aforementioned manner should be changed to set DISTUTILS_SETUP_PATH instead.
4.19.6 BitBake changes
BitBake is now configured to use a default
umask
of022
for all tasks (specified via a new BB_DEFAULT_UMASK variable). If needed,umask
can still be set on a per-task basis via theumask
varflag on the task function, but that is unlikely to be necessary in most cases.If a version specified in PREFERRED_VERSION is not available this will now trigger a warning instead of just a note, making such issues more visible.
4.19.7 Packaging changes
The following packaging changes have been made; in all cases the main package still depends upon the split out packages so you should not need to do anything unless you want to take advantage of the improved granularity:
dbus
:-common
and-tools
split outiproute2
: splitip
binary to its own packagenet-tools
: splitmii-tool
into its own packageprocps
: splitps
andsysctl
into their own packagesrpm
: split build and extra functionality into separate packagessudo
: splitsudo
binary intosudo-sudo
and libs intosudo-lib
systemtap
: examples, python scripts and runtime material split oututil-linux
:libuuid
has been split out to its ownutil-linux-libuuid
recipe (and corresponding packages) to avoid circular dependencies iflibgcrypt
support is enabled inutil-linux
. (util-linux
depends uponutil-linux-libuuid
.)
4.19.8 Miscellaneous changes
The default poky DISTRO_VERSION value now uses the core metadata’s git hash (i.e. METADATA_REVISION) rather than the date (i.e. DATE) to reduce one small source of non-reproducibility. You can of course specify your own DISTRO_VERSION value as desired (particularly if you create your own custom distro configuration).
adwaita-icon-theme
version 3.34.3 has been added back, and is selected as the default via PREFERRED_VERSION inmeta/conf/distro/include/default-versions.inc
due to newer versions not working well withlibrsvg
2.40.librsvg
is not practically upgradeable at the moment as it has been ported to Rust, and Rust is not (yet) in OE-Core, but this will change in a future release.ffmpeg
is now configured to disable GPL-licensed portions by default to make it harder to accidentally violate the GPL. To explicitly enable GPL licensed portions, addgpl
to PACKAGECONFIG forffmpeg
using a bbappend (or usePACKAGECONFIG_append_pn-ffmpeg = " gpl"
in your configuration.)connman
is now set to conflict withsystemd-networkd
as they overlap functionally and may interfere with each other at runtime.Canonical SPDX license names are now used in image license manifests in order to avoid aliases of the same license from showing up together (e.g.
GPLv2
andGPL-2.0
)
5 Source Directory Structure
The Source Directory consists of numerous files, directories and subdirectories; understanding their locations and contents is key to using the Yocto Project effectively. This chapter describes the Source Directory and gives information about those files and directories.
For information on how to establish a local Source Directory on your development system, see the “Locating Yocto Project Source Files” section in the Yocto Project Development Tasks Manual.
Note
The OpenEmbedded build system does not support file or directory names that contain spaces. Be sure that the Source Directory you use does not contain these types of names.
5.1 Top-Level Core Components
This section describes the top-level components of the Source Directory.
5.1.1 bitbake/
This directory includes a copy of BitBake for ease of use. The copy usually matches the current stable BitBake release from the BitBake project. BitBake, a Metadata interpreter, reads the Yocto Project Metadata and runs the tasks defined by that data. Failures are usually caused by errors in your Metadata and not from BitBake itself; consequently, most users do not need to worry about BitBake.
When you run the bitbake
command, the main BitBake executable (which
resides in the bitbake/bin/
directory) starts. Sourcing the
environment setup script (i.e. oe-init-build-env) places
the scripts/
and bitbake/bin/
directories (in that order) into
the shell’s PATH
environment variable.
For more information on BitBake, see the BitBake User Manual.
5.1.2 build/
This directory contains user configuration files and the output
generated by the OpenEmbedded build system in its standard configuration
where the source tree is combined with the output. The Build Directory
is created initially when you source
the OpenEmbedded build environment setup script (i.e.
oe-init-build-env).
It is also possible to place output and configuration files in a
directory separate from the Source Directory by
providing a directory name when you source
the setup script. For
information on separating output from your local Source Directory files
(commonly described as an “out of tree” build), see the
“oe-init-build-env” section.
5.1.3 documentation/
This directory holds the source for the Yocto Project documentation as
well as templates and tools that allow you to generate PDF and HTML
versions of the manuals. Each manual is contained in its own sub-folder;
for example, the files for this reference manual reside in the
ref-manual/
directory.
5.1.4 meta/
This directory contains the minimal, underlying OpenEmbedded-Core
metadata. The directory holds recipes, common classes, and machine
configuration for strictly emulated targets (qemux86
, qemuarm
,
and so forth.)
5.1.5 meta-poky/
Designed above the meta/
content, this directory adds just enough
metadata to define the Poky reference distribution.
5.1.6 meta-yocto-bsp/
This directory contains the Yocto Project reference hardware Board Support Packages (BSPs). For more information on BSPs, see the Yocto Project Board Support Package Developer’s Guide.
5.1.7 meta-selftest/
This directory adds additional recipes and append files used by the
OpenEmbedded selftests to verify the behavior of the build system. You
do not have to add this layer to your bblayers.conf
file unless you
want to run the selftests.
5.1.8 meta-skeleton/
This directory contains template recipes for BSP and kernel development.
5.1.9 scripts/
This directory contains various integration scripts that implement extra
functionality in the Yocto Project environment (e.g. QEMU scripts). The
oe-init-build-env script prepends this directory to the
shell’s PATH
environment variable.
The scripts
directory has useful scripts that assist in contributing
back to the Yocto Project, such as create-pull-request
and
send-pull-request
.
5.1.10 oe-init-build-env
This script sets up the OpenEmbedded build environment. Running this
script with the source
command in a shell makes changes to PATH
and sets other core BitBake variables based on the current working
directory. You need to run an environment setup script before running
BitBake commands. The script uses other scripts within the scripts
directory to do the bulk of the work.
When you run this script, your Yocto Project environment is set up, a Build Directory is created, your working directory becomes the Build Directory, and you are presented with some simple suggestions as to what to do next, including a list of some possible targets to build. Here is an example:
$ source oe-init-build-env
### Shell environment set up for builds. ###
You can now run 'bitbake <target>'
Common targets are:
core-image-minimal
core-image-sato
meta-toolchain
meta-ide-support
You can also run generated QEMU images with a command like 'runqemu qemux86-64'
The default output of the oe-init-build-env
script is from the
conf-notes.txt
file, which is found in the meta-poky
directory
within the Source Directory. If you design a
custom distribution, you can include your own version of this
configuration file to mention the targets defined by your distribution.
See the
“Creating a Custom Template Configuration Directory”
section in the Yocto Project Development Tasks Manual for more
information.
By default, running this script without a Build Directory argument
creates the build/
directory in your current working directory. If
you provide a Build Directory argument when you source
the script,
you direct the OpenEmbedded build system to create a Build Directory of
your choice. For example, the following command creates a Build
Directory named mybuilds/
that is outside of the Source Directory:
$ source oe-init-build-env ~/mybuilds
The OpenEmbedded build system uses the template configuration files, which
are found by default in the meta-poky/conf/
directory in the Source
Directory. See the
“Creating a Custom Template Configuration Directory”
section in the Yocto Project Development Tasks Manual for more
information.
Note
The OpenEmbedded build system does not support file or directory
names that contain spaces. If you attempt to run the oe-init-build-env
script from a Source Directory that contains spaces in either the
filenames or directory names, the script returns an error indicating
no such file or directory. Be sure to use a Source Directory free of
names containing spaces.
5.1.11 LICENSE, README, and README.hardware
These files are standard top-level files.
5.2 The Build Directory - build/
The OpenEmbedded build system creates the Build Directory
when you run the build environment setup
script oe-init-build-env. If you do not give the Build
Directory a specific name when you run the setup script, the name
defaults to build/
.
For subsequent parsing and processing, the name of the Build directory is available via the TOPDIR variable.
5.2.1 build/buildhistory/
The OpenEmbedded build system creates this directory when you enable
build history via the buildhistory
class file. The directory
organizes build information into image, packages, and SDK
subdirectories. For information on the build history feature, see the
“Maintaining Build Output Quality”
section in the Yocto Project Development Tasks Manual.
5.2.2 build/conf/local.conf
This configuration file contains all the local user configurations for
your build environment. The local.conf
file contains documentation
on the various configuration options. Any variable set here overrides
any variable set elsewhere within the environment unless that variable
is hard-coded within a file (e.g. by using ‘=’ instead of ‘?=’). Some
variables are hard-coded for various reasons but such variables are
relatively rare.
At a minimum, you would normally edit this file to select the target
MACHINE
, which package types you wish to use
(PACKAGE_CLASSES), and the location from
which you want to access downloaded files (DL_DIR
).
If local.conf
is not present when you start the build, the
OpenEmbedded build system creates it from local.conf.sample
when you
source
the top-level build environment setup script
oe-init-build-env.
The source local.conf.sample
file used depends on the
$TEMPLATECONF
script variable, which defaults to meta-poky/conf/
when you are building from the Yocto Project development environment,
and to meta/conf/
when you are building from the OpenEmbedded-Core
environment. Because the script variable points to the source of the
local.conf.sample
file, this implies that you can configure your
build environment from any layer by setting the variable in the
top-level build environment setup script as follows:
TEMPLATECONF=your_layer/conf
Once the build process gets the sample
file, it uses sed
to substitute final
${
OEROOT}
values for all
##OEROOT##
values.
Note
You can see how the TEMPLATECONF
variable is used by looking at the
scripts/oe-setup-builddir`
script in the Source Directory.
You can find the Yocto Project version of the local.conf.sample
file in
the meta-poky/conf
directory.
5.2.3 build/conf/bblayers.conf
This configuration file defines
layers,
which are directory trees, traversed (or walked) by BitBake. The
bblayers.conf
file uses the BBLAYERS
variable to list the layers BitBake tries to find.
If bblayers.conf
is not present when you start the build, the
OpenEmbedded build system creates it from bblayers.conf.sample
when
you source
the top-level build environment setup script (i.e.
oe-init-build-env).
As with the local.conf
file, the source bblayers.conf.sample
file used depends on the $TEMPLATECONF
script variable, which
defaults to meta-poky/conf/
when you are building from the Yocto
Project development environment, and to meta/conf/
when you are
building from the OpenEmbedded-Core environment. Because the script
variable points to the source of the bblayers.conf.sample
file, this
implies that you can base your build from any layer by setting the
variable in the top-level build environment setup script as follows:
TEMPLATECONF=your_layer/conf
Once the build process gets the sample file, it uses sed
to substitute final
${
OEROOT}
values for all ##OEROOT##
values.
Note
You can see how the TEMPLATECONF
variable scripts/oe-setup-builddir
script in the Source Directory. You can find the Yocto Project
version of the bblayers.conf.sample
file in the meta-poky/conf/
directory.
5.2.4 build/cache/sanity_info
This file indicates the state of the sanity checks and is created during the build.
5.2.5 build/downloads/
This directory contains downloaded upstream source tarballs. You can
reuse the directory for multiple builds or move the directory to another
location. You can control the location of this directory through the
DL_DIR
variable.
5.2.6 build/sstate-cache/
This directory contains the shared state cache. You can reuse the
directory for multiple builds or move the directory to another location.
You can control the location of this directory through the
SSTATE_DIR
variable.
5.2.7 build/tmp/
The OpenEmbedded build system creates and uses this directory for all the build system’s output. The TMPDIR variable points to this directory.
BitBake creates this directory if it does not exist. As a last resort,
to clean up a build and start it from scratch (other than the
downloads), you can remove everything in the tmp
directory or get
rid of the directory completely. If you do, you should also completely
remove the build/sstate-cache
directory.
5.2.8 build/tmp/buildstats/
This directory stores the build statistics.
5.2.9 build/tmp/cache/
When BitBake parses the metadata (recipes and configuration files), it
caches the results in build/tmp/cache/
to speed up future builds.
The results are stored on a per-machine basis.
During subsequent builds, BitBake checks each recipe (together with, for example, any files included or appended to it) to see if they have been modified. Changes can be detected, for example, through file modification time (mtime) changes and hashing of file contents. If no changes to the file are detected, then the parsed result stored in the cache is reused. If the file has changed, it is reparsed.
5.2.10 build/tmp/deploy/
This directory contains any “end result” output from the OpenEmbedded
build process. The DEPLOY_DIR variable points
to this directory. For more detail on the contents of the deploy
directory, see the
“Images” and
“Application Development SDK” sections in the Yocto
Project Overview and Concepts Manual.
5.2.11 build/tmp/deploy/deb/
This directory receives any .deb
packages produced by the build
process. The packages are sorted into feeds for different architecture
types.
5.2.12 build/tmp/deploy/rpm/
This directory receives any .rpm
packages produced by the build
process. The packages are sorted into feeds for different architecture
types.
5.2.13 build/tmp/deploy/ipk/
This directory receives .ipk
packages produced by the build process.
5.2.14 build/tmp/deploy/licenses/
This directory receives package licensing information. For example, the
directory contains sub-directories for bash
, busybox
, and
glibc
(among others) that in turn contain appropriate COPYING
license files with other licensing information. For information on
licensing, see the
“Maintaining Open Source License Compliance During Your Product’s Lifecycle”
section in the Yocto Project Development Tasks Manual.
5.2.15 build/tmp/deploy/images/
This directory is populated with the basic output objects of the build (think of them as the “generated artifacts” of the build process), including things like the boot loader image, kernel, root filesystem and more. If you want to flash the resulting image from a build onto a device, look here for the necessary components.
Be careful when deleting files in this directory. You can safely delete
old images from this directory (e.g. core-image-*
). However, the
kernel (*zImage*
, *uImage*
, etc.), bootloader and other
supplementary files might be deployed here prior to building an image.
Because these files are not directly produced from the image, if you
delete them they will not be automatically re-created when you build the
image again.
If you do accidentally delete files here, you will need to force them to be re-created. In order to do that, you will need to know the target that produced them. For example, these commands rebuild and re-create the kernel files:
$ bitbake -c clean virtual/kernel
$ bitbake virtual/kernel
5.2.16 build/tmp/deploy/sdk/
The OpenEmbedded build system creates this directory to hold toolchain installer scripts which, when executed, install the sysroot that matches your target hardware. You can find out more about these installers in the “Building an SDK Installer” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
5.2.17 build/tmp/sstate-control/
The OpenEmbedded build system uses this directory for the shared state manifest files. The shared state code uses these files to record the files installed by each sstate task so that the files can be removed when cleaning the recipe or when a newer version is about to be installed. The build system also uses the manifests to detect and produce a warning when files from one task are overwriting those from another.
5.2.18 build/tmp/sysroots-components/
This directory is the location of the sysroot contents that the task
do_prepare_recipe_sysroot
links or copies into the recipe-specific sysroot for each recipe listed
in DEPENDS. Population of this directory is
handled through shared state, while the path is specified by the
COMPONENTS_DIR variable. Apart from a few
unusual circumstances, handling of the sysroots-components
directory
should be automatic, and recipes should not directly reference
build/tmp/sysroots-components
.
5.2.19 build/tmp/sysroots/
Previous versions of the OpenEmbedded build system used to create a
global shared sysroot per machine along with a native sysroot. Beginning
with the 2.3 version of the Yocto Project, sysroots exist in
recipe-specific WORKDIR directories. Thus, the
build/tmp/sysroots/
directory is unused.
Note
The build/tmp/sysroots/
directory can still be populated using the
bitbake build-sysroots
command and can be used for compatibility in some
cases. However, in general it is not recommended to populate this directory.
Individual recipe-specific sysroots should be used.
5.2.20 build/tmp/stamps/
This directory holds information that BitBake uses for accounting purposes to track what tasks have run and when they have run. The directory is sub-divided by architecture, package name, and version. Following is an example:
stamps/all-poky-linux/distcc-config/1.0-r0.do_build-2fdd....2do
Although the files in the directory are empty of data, BitBake uses the filenames and timestamps for tracking purposes.
For information on how BitBake uses stamp files to determine if a task should be rerun, see the “Stamp Files and the Rerunning of Tasks” section in the Yocto Project Overview and Concepts Manual.
5.2.21 build/tmp/log/
This directory contains general logs that are not otherwise placed using
the package’s WORKDIR
. Examples of logs are the output from the
do_check_pkg
or do_distro_check
tasks. Running a build does not
necessarily mean this directory is created.
5.2.22 build/tmp/work/
This directory contains architecture-specific work sub-directories for packages built by BitBake. All tasks execute from the appropriate work directory. For example, the source for a particular package is unpacked, patched, configured and compiled all within its own work directory. Within the work directory, organization is based on the package group and version for which the source is being compiled as defined by the WORKDIR.
It is worth considering the structure of a typical work directory. As an
example, consider linux-yocto-kernel-3.0
on the machine qemux86
built within the Yocto Project. For this package, a work directory of
tmp/work/qemux86-poky-linux/linux-yocto/3.0+git1+<.....>
, referred
to as the WORKDIR
, is created. Within this directory, the source is
unpacked to linux-qemux86-standard-build
and then patched by Quilt.
(See the “Using Quilt in Your Workflow” section in
the Yocto Project Development Tasks Manual for more information.) Within
the linux-qemux86-standard-build
directory, standard Quilt
directories linux-3.0/patches
and linux-3.0/.pc
are created, and
standard Quilt commands can be used.
There are other directories generated within WORKDIR
. The most
important directory is WORKDIR/temp/
, which has log files for each
task (log.do_*.pid
) and contains the scripts BitBake runs for each
task (run.do_*.pid
). The WORKDIR/image/
directory is where “make
install” places its output that is then split into sub-packages within
WORKDIR/packages-split/
.
5.2.23 build/tmp/work/tunearch/recipename/version/
The recipe work directory - ${WORKDIR}
.
As described earlier in the
“build/tmp/sysroots/” section,
beginning with the 2.3 release of the Yocto Project, the OpenEmbedded
build system builds each recipe in its own work directory (i.e.
WORKDIR). The path to the work directory is
constructed using the architecture of the given build (e.g.
TUNE_PKGARCH, MACHINE_ARCH, or “allarch”), the recipe
name, and the version of the recipe (i.e.
PE:
PV-
PR).
A number of key subdirectories exist within each recipe work directory:
${WORKDIR}/temp
: Contains the log files of each task executed for this recipe, the “run” files for each executed task, which contain the code run, and alog.task_order
file, which lists the order in which tasks were executed.${WORKDIR}/image
: Contains the output of the do_install task, which corresponds to the${
D}
variable in that task.${WORKDIR}/pseudo
: Contains the pseudo database and log for any tasks executed under pseudo for the recipe.${WORKDIR}/sysroot-destdir
: Contains the output of the do_populate_sysroot task.${WORKDIR}/package
: Contains the output of the do_package task before the output is split into individual packages.${WORKDIR}/packages-split
: Contains the output of thedo_package
task after the output has been split into individual packages. Subdirectories exist for each individual package created by the recipe.${WORKDIR}/recipe-sysroot
: A directory populated with the target dependencies of the recipe. This directory looks like the target filesystem and contains libraries that the recipe might need to link against (e.g. the C library).${WORKDIR}/recipe-sysroot-native
: A directory populated with the native dependencies of the recipe. This directory contains the tools the recipe needs to build (e.g. the compiler, Autoconf, libtool, and so forth).${WORKDIR}/build
: This subdirectory applies only to recipes that support builds where the source is separate from the build artifacts. The OpenEmbedded build system uses this directory as a separate build directory (i.e.${
B}
).
5.3 The Metadata - meta/
As mentioned previously, Metadata is the core of the Yocto Project. Metadata has several important subdivisions:
5.3.1 meta/classes/
This directory contains the *.bbclass
files. Class files are used to
abstract common code so it can be reused by multiple packages. Every
package inherits the base.bbclass
file. Examples of other important
classes are autotools.bbclass
, which in theory allows any
Autotool-enabled package to work with the Yocto Project with minimal
effort. Another example is kernel.bbclass
that contains common code
and functions for working with the Linux kernel. Functions like image
generation or packaging also have their specific class files such as
image.bbclass
, rootfs_*.bbclass
and package*.bbclass
.
For reference information on classes, see the “Classes” chapter.
5.3.2 meta/conf/
This directory contains the core set of configuration files that start
from bitbake.conf
and from which all other configuration files are
included. See the include statements at the end of the bitbake.conf
file and you will note that even local.conf
is loaded from there.
While bitbake.conf
sets up the defaults, you can often override
these by using the (local.conf
) file, machine file or the
distribution configuration file.
5.3.3 meta/conf/machine/
This directory contains all the machine configuration files. If you set
MACHINE = "qemux86"
, the OpenEmbedded build system looks for a
qemux86.conf
file in this directory. The include
directory
contains various data common to multiple machines. If you want to add
support for a new machine to the Yocto Project, look in this directory.
5.3.4 meta/conf/distro/
The contents of this directory controls any distribution-specific
configurations. For the Yocto Project, the defaultsetup.conf
is the
main file here. This directory includes the versions and the SRCDATE
definitions for applications that are configured here. An example of an
alternative configuration might be poky-bleeding.conf
. Although this
file mainly inherits its configuration from Poky.
5.3.5 meta/conf/machine-sdk/
The OpenEmbedded build system searches this directory for configuration files that correspond to the value of SDKMACHINE. By default, 32-bit and 64-bit x86 files ship with the Yocto Project that support some SDK hosts. However, it is possible to extend that support to other SDK hosts by adding additional configuration files in this subdirectory within another layer.
5.3.6 meta/files/
This directory contains common license files and several text files used by the build system. The text files contain minimal device information and lists of files and directories with known permissions.
5.3.7 meta/lib/
This directory contains OpenEmbedded Python library code used during the build process.
5.3.8 meta/recipes-bsp/
This directory contains anything linking to specific hardware or hardware configuration information such as “u-boot” and “grub”.
5.3.9 meta/recipes-connectivity/
This directory contains libraries and applications related to communication with other devices.
5.3.10 meta/recipes-core/
This directory contains what is needed to build a basic working Linux image including commonly used dependencies.
5.3.11 meta/recipes-devtools/
This directory contains tools that are primarily used by the build system. The tools, however, can also be used on targets.
5.3.12 meta/recipes-extended/
This directory contains non-essential applications that add features compared to the alternatives in core. You might need this directory for full tool functionality or for Linux Standard Base (LSB) compliance.
5.3.13 meta/recipes-gnome/
This directory contains all things related to the GTK+ application framework.
5.3.14 meta/recipes-graphics/
This directory contains X and other graphically related system libraries.
5.3.15 meta/recipes-kernel/
This directory contains the kernel and generic applications and libraries that have strong kernel dependencies.
5.3.16 meta/recipes-lsb4/
This directory contains recipes specifically added to support the Linux Standard Base (LSB) version 4.x.
5.3.17 meta/recipes-multimedia/
This directory contains codecs and support utilities for audio, images and video.
5.3.18 meta/recipes-rt/
This directory contains package and image recipes for using and testing
the PREEMPT_RT
kernel.
5.3.19 meta/recipes-sato/
This directory contains the Sato demo/reference UI/UX and its associated applications and configuration data.
5.3.20 meta/recipes-support/
This directory contains recipes used by other recipes, but that are not directly included in images (i.e. dependencies of other recipes).
5.3.21 meta/site/
This directory contains a list of cached results for various architectures. Because certain “autoconf” test results cannot be determined when cross-compiling due to the tests not able to run on a live system, the information in this directory is passed to “autoconf” for the various architectures.
5.3.22 meta/recipes.txt
This file is a description of the contents of recipes-*
.
6 Classes
Class files are used to abstract common functionality and share it
amongst multiple recipe (.bb
) files. To use a class file, you simply
make sure the recipe inherits the class. In most cases, when a recipe
inherits a class it is enough to enable its features. There are cases,
however, where in the recipe you might need to set variables or override
some default behavior.
Any Metadata usually found in a recipe can also be
placed in a class file. Class files are identified by the extension
.bbclass
and are usually placed in a classes/
directory beneath
the meta*/
directory found in the Source Directory.
Class files can also be pointed to by
BUILDDIR (e.g. build/
) in the same way as
.conf
files in the conf
directory. Class files are searched for
in BBPATH using the same method by which .conf
files are searched.
This chapter discusses only the most useful and important classes. Other
classes do exist within the meta/classes
directory in the Source
Directory. You can reference the .bbclass
files directly for more
information.
6.1 allarch.bbclass
The allarch
class is inherited by recipes that do not produce
architecture-specific output. The class disables functionality that is
normally needed for recipes that produce executable binaries (such as
building the cross-compiler and a C library as pre-requisites, and
splitting out of debug symbols during packaging).
Note
Unlike some distro recipes (e.g. Debian), OpenEmbedded recipes that
produce packages that depend on tunings through use of the
RDEPENDS and
TUNE_PKGARCH variables, should never be
configured for all architectures using allarch
. This is the case
even if the recipes do not produce architecture-specific output.
Configuring such recipes for all architectures causes the
do_package_write_*
tasks to
have different signatures for the machines with different tunings.
Additionally, unnecessary rebuilds occur every time an image for a
different MACHINE
is built even when the recipe never changes.
By default, all recipes inherit the base and
package classes, which enable
functionality needed for recipes that produce executable output. If your
recipe, for example, only produces packages that contain configuration
files, media files, or scripts (e.g. Python and Perl), then it should
inherit the allarch
class.
6.2 archiver.bbclass
The archiver
class supports releasing source code and other
materials with the binaries.
For more details on the source archiver, see the “Maintaining Open Source License Compliance During Your Product’s Lifecycle” section in the Yocto Project Development Tasks Manual. You can also see the ARCHIVER_MODE variable for information about the variable flags (varflags) that help control archive creation.
6.3 autotools*.bbclass
The autotools*
classes support Autotooled packages.
The autoconf
, automake
, and libtool
packages bring
standardization. This class defines a set of tasks (e.g. configure
,
compile
and so forth) that work for all Autotooled packages. It
should usually be enough to define a few standard variables and then
simply inherit autotools
. These classes can also work with software
that emulates Autotools. For more information, see the
“Autotooled Package” section
in the Yocto Project Development Tasks Manual.
By default, the autotools*
classes use out-of-tree builds (i.e.
autotools.bbclass
building with B != S
).
If the software being built by a recipe does not support using
out-of-tree builds, you should have the recipe inherit the
autotools-brokensep
class. The autotools-brokensep
class behaves
the same as the autotools
class but builds with B
== S. This method is useful when out-of-tree build
support is either not present or is broken.
Note
It is recommended that out-of-tree support be fixed and used if at all possible.
It’s useful to have some idea of how the tasks defined by the
autotools*
classes work and what they do behind the scenes.
do_configure - Regenerates the configure script (using
autoreconf
) and then launches it with a standard set of arguments used during cross-compilation. You can pass additional parameters toconfigure
through theEXTRA_OECONF
or PACKAGECONFIG_CONFARGS variables.do_compile - Runs
make
with arguments that specify the compiler and linker. You can pass additional arguments through theEXTRA_OEMAKE
variable.do_install - Runs
make install
and passes in${
D}
asDESTDIR
.
6.4 base.bbclass
The base
class is special in that every .bb
file implicitly
inherits the class. This class contains definitions for standard basic
tasks such as fetching, unpacking, configuring (empty by default),
compiling (runs any Makefile
present), installing (empty by default)
and packaging (empty by default). These classes are often overridden or
extended by other classes such as the
autotools class or the
package class.
The class also contains some commonly used functions such as
oe_runmake
, which runs make
with the arguments specified in
EXTRA_OEMAKE variable as well as the
arguments passed directly to oe_runmake
.
6.5 bash-completion.bbclass
Sets up packaging and dependencies appropriate for recipes that build software that includes bash-completion data.
6.6 bin_package.bbclass
The bin_package
class is a helper class for recipes that extract the
contents of a binary package (e.g. an RPM) and install those contents
rather than building the binary from source. The binary package is
extracted and new packages in the configured output package format are
created. Extraction and installation of proprietary binaries is a good
example use for this class.
Note
For RPMs and other packages that do not contain a subdirectory, you
should specify an appropriate fetcher parameter to point to the
subdirectory. For example, if BitBake is using the Git fetcher (git://
),
the “subpath” parameter limits the checkout to a specific subpath
of the tree. Here is an example where ${BP}
is used so that the files
are extracted into the subdirectory expected by the default value of
S
:
SRC_URI = "git://example.com/downloads/somepackage.rpm;subpath=${BP}"
See the “Fetchers” section in the BitBake User Manual for more information on supported BitBake Fetchers.
6.7 binconfig.bbclass
The binconfig
class helps to correct paths in shell scripts.
Before pkg-config
had become widespread, libraries shipped shell
scripts to give information about the libraries and include paths needed
to build software (usually named LIBNAME-config
). This class assists
any recipe using such scripts.
During staging, the OpenEmbedded build system installs such scripts into
the sysroots/
directory. Inheriting this class results in all paths
in these scripts being changed to point into the sysroots/
directory
so that all builds that use the script use the correct directories for
the cross compiling layout. See the
BINCONFIG_GLOB variable for more
information.
6.8 binconfig-disabled.bbclass
An alternative version of the binconfig
class, which disables binary configuration scripts by making them return
an error in favor of using pkg-config
to query the information. The
scripts to be disabled should be specified using the
BINCONFIG variable within the recipe inheriting
the class.
6.9 blacklist.bbclass
The blacklist
class prevents the OpenEmbedded build system from
building specific recipes (blacklists them). To use this class, inherit
the class globally and set PNBLACKLIST for
each recipe you wish to blacklist. Specify the PN
value as a variable flag (varflag) and provide a reason, which is
reported, if the package is requested to be built as the value. For
example, if you want to blacklist a recipe called “exoticware”, you add
the following to your local.conf
or distribution configuration:
INHERIT += "blacklist"
PNBLACKLIST[exoticware] = "Not supported by our organization."
6.10 buildhistory.bbclass
The buildhistory
class records a history of build output metadata,
which can be used to detect possible regressions as well as used for
analysis of the build output. For more information on using Build
History, see the
“Maintaining Build Output Quality”
section in the Yocto Project Development Tasks Manual.
6.11 buildstats.bbclass
The buildstats
class records performance statistics about each task
executed during the build (e.g. elapsed time, CPU usage, and I/O usage).
When you use this class, the output goes into the
BUILDSTATS_BASE directory, which defaults
to ${TMPDIR}/buildstats/
. You can analyze the elapsed time using
scripts/pybootchartgui/pybootchartgui.py
, which produces a cascading
chart of the entire build process and can be useful for highlighting
bottlenecks.
Collecting build statistics is enabled by default through the
USER_CLASSES variable from your
local.conf
file. Consequently, you do not have to do anything to
enable the class. However, if you want to disable the class, simply
remove “buildstats” from the USER_CLASSES
list.
6.12 buildstats-summary.bbclass
When inherited globally, prints statistics at the end of the build on sstate re-use. In order to function, this class requires the buildstats class be enabled.
6.13 ccache.bbclass
The ccache
class enables the C/C++ Compiler Cache for the build.
This class is used to give a minor performance boost during the build.
However, using the class can lead to unexpected side-effects. Thus, it
is recommended that you do not use this class. See
https://ccache.samba.org/ for information on the C/C++ Compiler
Cache.
6.14 chrpath.bbclass
The chrpath
class is a wrapper around the “chrpath” utility, which
is used during the build process for nativesdk
, cross
, and
cross-canadian
recipes to change RPATH
records within binaries
in order to make them relocatable.
6.15 clutter.bbclass
The clutter
class consolidates the major and minor version naming
and other common items used by Clutter and related recipes.
Note
Unlike some other classes related to specific libraries, recipes building other software that uses Clutter do not need to inherit this class unless they use the same recipe versioning scheme that the Clutter and related recipes do.
6.16 cmake.bbclass
The cmake
class allows for recipes that need to build software using
the CMake build system. You can use
the EXTRA_OECMAKE variable to specify
additional configuration options to be passed using the cmake
command line.
On the occasion that you would be installing custom CMake toolchain
files supplied by the application being built, you should install them
to the preferred CMake Module directory: ${D}${datadir}/cmake/
Modules during
do_install.
6.17 cml1.bbclass
The cml1
class provides basic support for the Linux kernel style
build configuration system.
6.18 compress_doc.bbclass
Enables compression for man pages and info pages. This class is intended to be inherited globally. The default compression mechanism is gz (gzip) but you can select an alternative mechanism by setting the DOC_COMPRESS variable.
6.19 copyleft_compliance.bbclass
The copyleft_compliance
class preserves source code for the purposes
of license compliance. This class is an alternative to the archiver
class and is still used by some users even though it has been deprecated
in favor of the archiver class.
6.20 copyleft_filter.bbclass
A class used by the archiver and
copyleft_compliance classes
for filtering licenses. The copyleft_filter
class is an internal
class and is not intended to be used directly.
6.21 core-image.bbclass
The core-image
class provides common definitions for the
core-image-*
image recipes, such as support for additional
IMAGE_FEATURES.
6.22 cpan*.bbclass
The cpan*
classes support Perl modules.
Recipes for Perl modules are simple. These recipes usually only need to point to the source’s archive and then inherit the proper class file. Building is split into two methods depending on which method the module authors used.
Modules that use old
Makefile.PL
-based build system requirecpan.bbclass
in their recipes.Modules that use
Build.PL
-based build system require usingcpan_build.bbclass
in their recipes.
Both build methods inherit the cpan-base
class for basic Perl
support.
6.23 cross.bbclass
The cross
class provides support for the recipes that build the
cross-compilation tools.
6.24 cross-canadian.bbclass
The cross-canadian
class provides support for the recipes that build
the Canadian Cross-compilation tools for SDKs. See the
“Cross-Development Toolchain Generation”
section in the Yocto Project Overview and Concepts Manual for more
discussion on these cross-compilation tools.
6.25 crosssdk.bbclass
The crosssdk
class provides support for the recipes that build the
cross-compilation tools used for building SDKs. See the
“Cross-Development Toolchain Generation”
section in the Yocto Project Overview and Concepts Manual for more
discussion on these cross-compilation tools.
6.26 debian.bbclass
The debian
class renames output packages so that they follow the
Debian naming policy (i.e. glibc
becomes libc6
and
glibc-devel
becomes libc6-dev
.) Renaming includes the library
name and version as part of the package name.
If a recipe creates packages for multiple libraries (shared object files
of .so
type), use the LEAD_SONAME
variable in the recipe to specify the library on which to apply the
naming scheme.
6.27 deploy.bbclass
The deploy
class handles deploying files to the
DEPLOY_DIR_IMAGE directory. The main
function of this class is to allow the deploy step to be accelerated by
shared state. Recipes that inherit this class should define their own
do_deploy function to copy the files to be
deployed to DEPLOYDIR, and use addtask
to
add the task at the appropriate place, which is usually after
do_compile or
do_install. The class then takes care of
staging the files from DEPLOYDIR
to DEPLOY_DIR_IMAGE
.
6.28 devshell.bbclass
The devshell
class adds the do_devshell
task. Distribution
policy dictates whether to include this class. See the “Using a Development Shell”
section in the Yocto Project Development Tasks Manual for more
information about using devshell
.
6.29 devupstream.bbclass
The devupstream
class uses
BBCLASSEXTEND to add a variant of the
recipe that fetches from an alternative URI (e.g. Git) instead of a
tarball. Following is an example:
BBCLASSEXTEND = "devupstream:target"
SRC_URI_class-devupstream = "git://git.example.com/example"
SRCREV_class-devupstream = "abcd1234"
Adding the above statements to your recipe creates a variant that has
DEFAULT_PREFERENCE set to “-1”.
Consequently, you need to select the variant of the recipe to use it.
Any development-specific adjustments can be done by using the
class-devupstream
override. Here is an example:
DEPENDS_append_class-devupstream = " gperf-native"
do_configure_prepend_class-devupstream() {
touch ${S}/README
}
The class
currently only supports creating a development variant of the target
recipe, not native
or nativesdk
variants.
The BBCLASSEXTEND
syntax (i.e. devupstream:target
) provides
support for native
and nativesdk
variants. Consequently, this
functionality can be added in a future release.
Support for other version control systems such as Subversion is limited
due to BitBake’s automatic fetch dependencies (e.g.
subversion-native
).
6.30 distutils3*.bbclass
The distutils3*
classes support recipes for Python version 3.x
extensions, which are simple. These recipes usually only need to point
to the source’s archive and then inherit the proper class. Building is
split into three methods depending on which method the module authors
used.
Extensions that use an Autotools-based build system require Autotools and
distutils
-based classes in their recipes.Extensions that use
distutils
-based build systems require thedistutils
class in their recipes.Extensions that use build systems based on
setuptools3
require the setuptools3 class in their recipes.
6.31 externalsrc.bbclass
The externalsrc
class supports building software from source code
that is external to the OpenEmbedded build system. Building software
from an external source tree means that the build system’s normal fetch,
unpack, and patch process is not used.
By default, the OpenEmbedded build system uses the S
and B variables to locate unpacked recipe source code
and to build it, respectively. When your recipe inherits the
externalsrc
class, you use the
EXTERNALSRC and
EXTERNALSRC_BUILD variables to
ultimately define S
and B
.
By default, this class expects the source code to support recipe builds
that use the B variable to point to the directory in
which the OpenEmbedded build system places the generated objects built
from the recipes. By default, the B
directory is set to the
following, which is separate from the source directory (S
):
${WORKDIR}/${BPN}/{PV}/
See these variables for more information: WORKDIR, BPN, and PV,
For more information on the externalsrc
class, see the comments in
meta/classes/externalsrc.bbclass
in the Source Directory.
For information on how to use the
externalsrc
class, see the
“Building Software from an External Source”
section in the Yocto Project Development Tasks Manual.
6.32 extrausers.bbclass
The extrausers
class allows additional user and group configuration
to be applied at the image level. Inheriting this class either globally
or from an image recipe allows additional user and group operations to
be performed using the
EXTRA_USERS_PARAMS variable.
Note
The user and group operations added using the extrausers class are not tied to a specific recipe outside of the recipe for the image. Thus, the operations can be performed across the image as a whole. Use the useradd class to add user and group configuration to a specific recipe.
Here is an example that uses this class in an image recipe:
inherit extrausers
EXTRA_USERS_PARAMS = "\
useradd -p '' tester; \
groupadd developers; \
userdel nobody; \
groupdel -g video; \
groupmod -g 1020 developers; \
usermod -s /bin/sh tester; \
"
Here is an example that adds two users named “tester-jim” and “tester-sue” and assigns passwords:
inherit extrausers
EXTRA_USERS_PARAMS = "\
useradd -P tester01 tester-jim; \
useradd -P tester01 tester-sue; \
"
Finally, here is an example that sets the root password to “1876*18”:
inherit extrausers
EXTRA_USERS_PARAMS = "\
usermod -P 1876*18 root; \
"
6.33 features_check.bbclass
The features_check
class allows individual recipes to check
for required and conflicting
DISTRO_FEATURES, MACHINE_FEATURES or COMBINED_FEATURES.
This class provides support for the following variables:
REQUIRED_MACHINE_FEATURES
CONFLICT_MACHINE_FEATURES
ANY_OF_MACHINE_FEATURES
REQUIRED_COMBINED_FEATURES
CONFLICT_COMBINED_FEATURES
ANY_OF_COMBINED_FEATURES
If any conditions specified in the recipe using the above variables are not met, the recipe will be skipped, and if the build system attempts to build the recipe then an error will be triggered.
6.34 fontcache.bbclass
The fontcache
class generates the proper post-install and
post-remove (postinst and postrm) scriptlets for font packages. These
scriptlets call fc-cache
(part of Fontconfig
) to add the fonts
to the font information cache. Since the cache files are
architecture-specific, fc-cache
runs using QEMU if the postinst
scriptlets need to be run on the build host during image creation.
If the fonts being installed are in packages other than the main package, set FONT_PACKAGES to specify the packages containing the fonts.
6.35 fs-uuid.bbclass
The fs-uuid
class extracts UUID from
${
ROOTFS}
, which must have been built
by the time that this function gets called. The fs-uuid
class only
works on ext
file systems and depends on tune2fs
.
6.36 gconf.bbclass
The gconf
class provides common functionality for recipes that need
to install GConf schemas. The schemas will be put into a separate
package (${
PN}-gconf
) that is created
automatically when this class is inherited. This package uses the
appropriate post-install and post-remove (postinst/postrm) scriptlets to
register and unregister the schemas in the target image.
6.37 gettext.bbclass
The gettext
class provides support for building software that uses
the GNU gettext
internationalization and localization system. All
recipes building software that use gettext
should inherit this
class.
6.38 gnomebase.bbclass
The gnomebase
class is the base class for recipes that build
software from the GNOME stack. This class sets
SRC_URI to download the source from the GNOME
mirrors as well as extending FILES with the typical
GNOME installation paths.
6.39 gobject-introspection.bbclass
Provides support for recipes building software that supports GObject introspection. This functionality is only enabled if the “gobject-introspection-data” feature is in DISTRO_FEATURES as well as “qemu-usermode” being in MACHINE_FEATURES.
Note
This functionality is backfilled by default and, if not applicable,
should be disabled through DISTRO_FEATURES_BACKFILL_CONSIDERED
or
MACHINE_FEATURES_BACKFILL_CONSIDERED
, respectively.
6.40 grub-efi.bbclass
The grub-efi
class provides grub-efi
-specific functions for
building bootable images.
This class supports several variables:
INITRD: Indicates list of filesystem images to concatenate and use as an initial RAM disk (initrd) (optional).
ROOTFS: Indicates a filesystem image to include as the root filesystem (optional).
GRUB_GFXSERIAL: Set this to “1” to have graphics and serial in the boot menu.
LABELS: A list of targets for the automatic configuration.
APPEND: An override list of append strings for each
LABEL
.GRUB_OPTS: Additional options to add to the configuration (optional). Options are delimited using semi-colon characters (
;
).GRUB_TIMEOUT: Timeout before executing the default
LABEL
(optional).
6.41 gsettings.bbclass
The gsettings
class provides common functionality for recipes that
need to install GSettings (glib) schemas. The schemas are assumed to be
part of the main package. Appropriate post-install and post-remove
(postinst/postrm) scriptlets are added to register and unregister the
schemas in the target image.
6.42 gtk-doc.bbclass
The gtk-doc
class is a helper class to pull in the appropriate
gtk-doc
dependencies and disable gtk-doc
.
6.43 gtk-icon-cache.bbclass
The gtk-icon-cache
class generates the proper post-install and
post-remove (postinst/postrm) scriptlets for packages that use GTK+ and
install icons. These scriptlets call gtk-update-icon-cache
to add
the fonts to GTK+’s icon cache. Since the cache files are
architecture-specific, gtk-update-icon-cache
is run using QEMU if
the postinst scriptlets need to be run on the build host during image
creation.
6.44 gtk-immodules-cache.bbclass
The gtk-immodules-cache
class generates the proper post-install and
post-remove (postinst/postrm) scriptlets for packages that install GTK+
input method modules for virtual keyboards. These scriptlets call
gtk-update-icon-cache
to add the input method modules to the cache.
Since the cache files are architecture-specific,
gtk-update-icon-cache
is run using QEMU if the postinst scriptlets
need to be run on the build host during image creation.
If the input method modules being installed are in packages other than the main package, set GTKIMMODULES_PACKAGES to specify the packages containing the modules.
6.45 gzipnative.bbclass
The gzipnative
class enables the use of different native versions of
gzip
and pigz
rather than the versions of these tools from the
build host.
6.46 icecc.bbclass
The icecc
class supports
Icecream, which facilitates
taking compile jobs and distributing them among remote machines.
The class stages directories with symlinks from gcc
and g++
to
icecc
, for both native and cross compilers. Depending on each
configure or compile, the OpenEmbedded build system adds the directories
at the head of the PATH
list and then sets the ICECC_CXX
and
ICEC_CC
variables, which are the paths to the g++
and gcc
compilers, respectively.
For the cross compiler, the class creates a tar.gz
file that
contains the Yocto Project toolchain and sets ICECC_VERSION
, which
is the version of the cross-compiler used in the cross-development
toolchain, accordingly.
The class handles all three different compile stages (i.e native
,cross-kernel and target) and creates the necessary environment
tar.gz
file to be used by the remote machines. The class also
supports SDK generation.
If ICECC_PATH is not set in your
local.conf
file, then the class tries to locate the icecc
binary
using which
. If ICECC_ENV_EXEC is set
in your local.conf
file, the variable should point to the
icecc-create-env
script provided by the user. If you do not point to
a user-provided script, the build system uses the default script
provided by the recipe icecc-create-env-native.bb
.
Note
This script is a modified version and not the one that comes with icecc.
If you do not want the Icecream distributed compile support to apply to
specific recipes or classes, you can effectively “blacklist” them by
listing the recipes and classes using the
ICECC_USER_PACKAGE_BL and
ICECC_USER_CLASS_BL, variables,
respectively, in your local.conf
file. Doing so causes the
OpenEmbedded build system to handle these compilations locally.
Additionally, you can list recipes using the
ICECC_USER_PACKAGE_WL variable in
your local.conf
file to force icecc
to be enabled for recipes
using an empty PARALLEL_MAKE variable.
Inheriting the icecc
class changes all sstate signatures.
Consequently, if a development team has a dedicated build system that
populates SSTATE_MIRRORS and they want to
reuse sstate from SSTATE_MIRRORS
, then all developers and the build
system need to either inherit the icecc
class or nobody should.
At the distribution level, you can inherit the icecc
class to be
sure that all builders start with the same sstate signatures. After
inheriting the class, you can then disable the feature by setting the
ICECC_DISABLED variable to “1” as follows:
INHERIT_DISTRO_append = " icecc"
ICECC_DISABLED ??= "1"
This practice
makes sure everyone is using the same signatures but also requires
individuals that do want to use Icecream to enable the feature
individually as follows in your local.conf
file:
ICECC_DISABLED = ""
6.47 image.bbclass
The image
class helps support creating images in different formats.
First, the root filesystem is created from packages using one of the
rootfs*.bbclass
files (depending on the package format used) and
then one or more image files are created.
The
IMAGE_FSTYPES
variable controls the types of images to generate.The
IMAGE_INSTALL
variable controls the list of packages to install into the image.
For information on customizing images, see the “Customizing Images” section in the Yocto Project Development Tasks Manual. For information on how images are created, see the “Images” section in the Yocto Project Overview and Concepts Manual.
6.48 image-buildinfo.bbclass
The image-buildinfo
class writes information to the target
filesystem on /etc/build
.
6.49 image_types.bbclass
The image_types
class defines all of the standard image output types
that you can enable through the
IMAGE_FSTYPES variable. You can use this
class as a reference on how to add support for custom image output
types.
By default, the image class automatically
enables the image_types
class. The image
class uses the
IMGCLASSES
variable as follows:
IMGCLASSES = "rootfs_${IMAGE_PKGTYPE} image_types ${IMAGE_CLASSES}"
IMGCLASSES += "${@['populate_sdk_base', 'populate_sdk_ext']['linux' in d.getVar("SDK_OS")]}"
IMGCLASSES += "${@bb.utils.contains_any('IMAGE_FSTYPES', 'live iso hddimg', 'image-live', '', d)}"
IMGCLASSES += "${@bb.utils.contains('IMAGE_FSTYPES', 'container', 'image-container', '', d)}"
IMGCLASSES += "image_types_wic"
IMGCLASSES += "rootfs-postcommands"
IMGCLASSES += "image-postinst-intercepts"
inherit ${IMGCLASSES}
The image_types
class also handles conversion and compression of images.
Note
To build a VMware VMDK image, you need to add “wic.vmdk” to
IMAGE_FSTYPES
. This would also be similar for Virtual Box Virtual Disk
Image (“vdi”) and QEMU Copy On Write Version 2 (“qcow2”) images.
6.50 image-live.bbclass
This class controls building “live” (i.e. HDDIMG and ISO) images. Live images contain syslinux for legacy booting, as well as the bootloader specified by EFI_PROVIDER if MACHINE_FEATURES contains “efi”.
Normally, you do not use this class directly. Instead, you add “live” to IMAGE_FSTYPES.
6.51 image-mklibs.bbclass
The image-mklibs
class enables the use of the mklibs
utility
during the do_rootfs task, which optimizes
the size of libraries contained in the image.
By default, the class is enabled in the local.conf.template
using
the USER_CLASSES variable as follows:
USER_CLASSES ?= "buildstats image-mklibs image-prelink"
6.52 image-prelink.bbclass
The image-prelink
class enables the use of the prelink
utility
during the do_rootfs task, which optimizes
the dynamic linking of shared libraries to reduce executable startup
time.
By default, the class is enabled in the local.conf.template
using
the USER_CLASSES variable as follows:
USER_CLASSES ?= "buildstats image-mklibs image-prelink"
6.53 insane.bbclass
The insane
class adds a step to the package generation process so
that output quality assurance checks are generated by the OpenEmbedded
build system. A range of checks are performed that check the build’s
output for common problems that show up during runtime. Distribution
policy usually dictates whether to include this class.
You can configure the sanity checks so that specific test failures either raise a warning or an error message. Typically, failures for new tests generate a warning. Subsequent failures for the same test would then generate an error message once the metadata is in a known and good condition. See the “QA Error and Warning Messages” Chapter for a list of all the warning and error messages you might encounter using a default configuration.
Use the WARN_QA and
ERROR_QA variables to control the behavior of
these checks at the global level (i.e. in your custom distro
configuration). However, to skip one or more checks in recipes, you
should use INSANE_SKIP. For example, to skip
the check for symbolic link .so
files in the main package of a
recipe, add the following to the recipe. You need to realize that the
package name override, in this example ${PN}
, must be used:
INSANE_SKIP_${PN} += "dev-so"
Please keep in mind that the QA checks exist in order to detect real or potential problems in the packaged output. So exercise caution when disabling these checks.
The following list shows the tests you can list with the WARN_QA
and
ERROR_QA
variables:
already-stripped:
Checks that produced binaries have not already been stripped prior to the build system extracting debug symbols. It is common for upstream software projects to default to stripping debug symbols for output binaries. In order for debugging to work on the target using-dbg
packages, this stripping must be disabled.arch:
Checks the Executable and Linkable Format (ELF) type, bit size, and endianness of any binaries to ensure they match the target architecture. This test fails if any binaries do not match the type since there would be an incompatibility. The test could indicate that the wrong compiler or compiler options have been used. Sometimes software, like bootloaders, might need to bypass this check.buildpaths:
Checks for paths to locations on the build host inside the output files. Currently, this test triggers too many false positives and thus is not normally enabled.build-deps:
Determines if a build-time dependency that is specified through DEPENDS, explicit RDEPENDS, or task-level dependencies exists to match any runtime dependency. This determination is particularly useful to discover where runtime dependencies are detected and added during packaging. If no explicit dependency has been specified within the metadata, at the packaging stage it is too late to ensure that the dependency is built, and thus you can end up with an error when the package is installed into the image during the do_rootfs task because the auto-detected dependency was not satisfied. An example of this would be where the update-rc.d class automatically adds a dependency on theinitscripts-functions
package to packages that install an initscript that refers to/etc/init.d/functions
. The recipe should really have an explicitRDEPENDS
for the package in question oninitscripts-functions
so that the OpenEmbedded build system is able to ensure that theinitscripts
recipe is actually built and thus theinitscripts-functions
package is made available.compile-host-path:
Checks the do_compile log for indications that paths to locations on the build host were used. Using such paths might result in host contamination of the build output.debug-deps:
Checks that all packages except-dbg
packages do not depend on-dbg
packages, which would cause a packaging bug.debug-files:
Checks for.debug
directories in anything but the-dbg
package. The debug files should all be in the-dbg
package. Thus, anything packaged elsewhere is incorrect packaging.dep-cmp:
Checks for invalid version comparison statements in runtime dependency relationships between packages (i.e. in RDEPENDS, RRECOMMENDS, RSUGGESTS, RPROVIDES, RREPLACES, and RCONFLICTS variable values). Any invalid comparisons might trigger failures or undesirable behavior when passed to the package manager.desktop:
Runs thedesktop-file-validate
program against any.desktop
files to validate their contents against the specification for.desktop
files.dev-deps:
Checks that all packages except-dev
or-staticdev
packages do not depend on-dev
packages, which would be a packaging bug.dev-so:
Checks that the.so
symbolic links are in the-dev
package and not in any of the other packages. In general, these symlinks are only useful for development purposes. Thus, the-dev
package is the correct location for them. Some very rare cases do exist for dynamically loaded modules where these symlinks are needed instead in the main package.file-rdeps:
Checks that file-level dependencies identified by the OpenEmbedded build system at packaging time are satisfied. For example, a shell script might start with the line#!/bin/bash
. This line would translate to a file dependency on/bin/bash
. Of the three package managers that the OpenEmbedded build system supports, only RPM directly handles file-level dependencies, resolving them automatically to packages providing the files. However, the lack of that functionality in the other two package managers does not mean the dependencies do not still need resolving. This QA check attempts to ensure that explicitly declared RDEPENDS exist to handle any file-level dependency detected in packaged files.files-invalid:
Checks for FILES variable values that contain “//”, which is invalid.host-user-contaminated:
Checks that no package produced by the recipe contains any files outside of/home
with a user or group ID that matches the user running BitBake. A match usually indicates that the files are being installed with an incorrect UID/GID, since target IDs are independent from host IDs. For additional information, see the section describing the do_install task.incompatible-license:
Report when packages are excluded from being created due to being marked with a license that is in INCOMPATIBLE_LICENSE.install-host-path:
Checks the do_install log for indications that paths to locations on the build host were used. Using such paths might result in host contamination of the build output.installed-vs-shipped:
Reports when files have been installed withindo_install
but have not been included in any package by way of the FILES variable. Files that do not appear in any package cannot be present in an image later on in the build process. Ideally, all installed files should be packaged or not installed at all. These files can be deleted at the end ofdo_install
if the files are not needed in any package.invalid-chars:
Checks that the recipe metadata variables DESCRIPTION, SUMMARY, LICENSE, and SECTION do not contain non-UTF-8 characters. Some package managers do not support such characters.invalid-packageconfig:
Checks that no undefined features are being added to PACKAGECONFIG. For example, any name “foo” for which the following form does not exist:PACKAGECONFIG[foo] = "..."
la:
Checks.la
files for anyTMPDIR
paths. Any.la
file containing these paths is incorrect sincelibtool
adds the correct sysroot prefix when using the files automatically itself.ldflags:
Ensures that the binaries were linked with the LDFLAGS options provided by the build system. If this test fails, check that theLDFLAGS
variable is being passed to the linker command.libdir:
Checks for libraries being installed into incorrect (possibly hardcoded) installation paths. For example, this test will catch recipes that install/lib/bar.so
when${base_libdir}
is “lib32”. Another example is when recipes install/usr/lib64/foo.so
when${libdir}
is “/usr/lib”.libexec:
Checks if a package contains files in/usr/libexec
. This check is not performed if thelibexecdir
variable has been set explicitly to/usr/libexec
.packages-list:
Checks for the same package being listed multiple times through the PACKAGES variable value. Installing the package in this manner can cause errors during packaging.perm-config:
Reports lines infs-perms.txt
that have an invalid format.perm-line:
Reports lines infs-perms.txt
that have an invalid format.perm-link:
Reports lines infs-perms.txt
that specify ‘link’ where the specified target already exists.perms:
Currently, this check is unused but reserved.pkgconfig:
Checks.pc
files for any TMPDIR/WORKDIR paths. Any.pc
file containing these paths is incorrect sincepkg-config
itself adds the correct sysroot prefix when the files are accessed.pkgname:
Checks that all packages in PACKAGES have names that do not contain invalid characters (i.e. characters other than 0-9, a-z, ., +, and -).pkgv-undefined:
Checks to see if thePKGV
variable is undefined during do_package.pkgvarcheck:
Checks through the variables RDEPENDS, RRECOMMENDS, RSUGGESTS, RCONFLICTS, RPROVIDES, RREPLACES, FILES, ALLOW_EMPTY,pkg_preinst
,pkg_postinst
,pkg_prerm
andpkg_postrm
, and reports if there are variable sets that are not package-specific. Using these variables without a package suffix is bad practice, and might unnecessarily complicate dependencies of other packages within the same recipe or have other unintended consequences.pn-overrides:
Checks that a recipe does not have a name (PN) value that appears in OVERRIDES. If a recipe is named such that itsPN
value matches something already inOVERRIDES
(e.g.PN
happens to be the same as MACHINE or DISTRO), it can have unexpected consequences. For example, assignments such asFILES_${PN} = "xyz"
effectively turn intoFILES = "xyz"
.rpaths:
Checks for rpaths in the binaries that contain build system paths such asTMPDIR
. If this test fails, bad-rpath
options are being passed to the linker commands and your binaries have potential security issues.split-strip:
Reports that splitting or stripping debug symbols from binaries has failed.staticdev:
Checks for static library files (*.a
) in non-staticdev
packages.symlink-to-sysroot:
Checks for symlinks in packages that point into TMPDIR on the host. Such symlinks will work on the host, but are clearly invalid when running on the target.textrel:
Checks for ELF binaries that contain relocations in their.text
sections, which can result in a performance impact at runtime. See the explanation for theELF binary
message in “QA Error and Warning Messages” for more information regarding runtime performance issues.unlisted-pkg-lics:
Checks that all declared licenses applying for a package are also declared on the recipe level (i.e. any license inLICENSE_*
should appear in LICENSE).useless-rpaths:
Checks for dynamic library load paths (rpaths) in the binaries that by default on a standard system are searched by the linker (e.g./lib
and/usr/lib
). While these paths will not cause any breakage, they do waste space and are unnecessary.var-undefined:
Reports when variables fundamental to packaging (i.e. WORKDIR, DEPLOY_DIR, D, PN, and PKGD) are undefined during do_package.version-going-backwards:
If Build History is enabled, reports when a package being written out has a lower version than the previously written package under the same name. If you are placing output packages into a feed and upgrading packages on a target system using that feed, the version of a package going backwards can result in the target system not correctly upgrading to the “new” version of the package.Note
If you are not using runtime package management on your target system, then you do not need to worry about this situation.
xorg-driver-abi:
Checks that all packages containing Xorg drivers have ABI dependencies. Thexserver-xorg
recipe provides driver ABI names. All drivers should depend on the ABI versions that they have been built against. Driver recipes that includexorg-driver-input.inc
orxorg-driver-video.inc
will automatically get these versions. Consequently, you should only need to explicitly add dependencies to binary driver recipes.
6.54 insserv.bbclass
The insserv
class uses the insserv
utility to update the order
of symbolic links in /etc/rc?.d/
within an image based on
dependencies specified by LSB headers in the init.d
scripts
themselves.
6.55 kernel.bbclass
The kernel
class handles building Linux kernels. The class contains
code to build all kernel trees. All needed headers are staged into the
STAGING_KERNEL_DIR
directory to allow out-of-tree module builds
using the module class.
This means that each built kernel module is packaged separately and
inter-module dependencies are created by parsing the modinfo
output.
If all modules are required, then installing the kernel-modules
package installs all packages with modules and various other kernel
packages such as kernel-vmlinux
.
The kernel
class contains logic that allows you to embed an initial
RAM filesystem (initramfs) image when you build the kernel image. For
information on how to build an initramfs, see the
“Building an Initial RAM Filesystem (initramfs) Image” section in
the Yocto Project Development Tasks Manual.
Various other classes are used by the kernel
and module
classes
internally including the kernel-arch,
module-base, and
linux-kernel-base classes.
6.56 kernel-arch.bbclass
The kernel-arch
class sets the ARCH
environment variable for
Linux kernel compilation (including modules).
6.57 kernel-devicetree.bbclass
The kernel-devicetree
class, which is inherited by the
kernel class, supports device tree
generation.
6.58 kernel-fitimage.bbclass
The kernel-fitimage
class provides support to pack a kernel image,
device trees, a U-boot script, a Initramfs bundle and a RAM disk
into a single FIT image. In theory, a FIT image can support any number
of kernels, U-boot scripts, Initramfs bundles, RAM disks and device-trees.
However, kernel-fitimage
currently only supports
limited usescases: just one kernel image, an optional U-boot script,
an optional Initramfs bundle, an optional RAM disk, and any number of
device tree.
To create a FIT image, it is required that KERNEL_CLASSES is set to include “kernel-fitimage” and KERNEL_IMAGETYPE is set to “fitImage”.
The options for the device tree compiler passed to mkimage -D
when creating the FIT image are specified using the
UBOOT_MKIMAGE_DTCOPTS variable.
Only a single kernel can be added to the FIT image created by
kernel-fitimage
and the kernel image in FIT is mandatory. The
address where the kernel image is to be loaded by U-Boot is
specified by UBOOT_LOADADDRESS and the entrypoint by
UBOOT_ENTRYPOINT.
Multiple device trees can be added to the FIT image created by
kernel-fitimage
and the device tree is optional.
The address where the device tree is to be loaded by U-Boot is
specified by UBOOT_DTBO_LOADADDRESS for device tree overlays
and by UBOOT_DTB_LOADADDRESS for device tree binaries.
Only a single RAM disk can be added to the FIT image created by
kernel-fitimage
and the RAM disk in FIT is optional.
The address where the RAM disk image is to be loaded by U-Boot
is specified by UBOOT_RD_LOADADDRESS and the entrypoint by
UBOOT_RD_ENTRYPOINT. The ramdisk is added to FIT image when
INITRAMFS_IMAGE is specified and that INITRAMFS_IMAGE_BUNDLE
is set to 0.
Only a single Initramfs bundle can be added to the FIT image created by
kernel-fitimage
and the Initramfs bundle in FIT is optional.
In case of Initramfs, the kernel is configured to be bundled with the rootfs
in the same binary (example: zImage-initramfs-MACHINE.bin).
When the kernel is copied to RAM and executed, it unpacks the Initramfs rootfs.
The Initramfs bundle can be enabled when INITRAMFS_IMAGE
is specified and that INITRAMFS_IMAGE_BUNDLE is set to 1.
The address where the Initramfs bundle is to be loaded by U-boot is specified
by UBOOT_LOADADDRESS and the entrypoint by UBOOT_ENTRYPOINT.
Only a single U-boot boot script can be added to the FIT image created by
kernel-fitimage
and the boot script is optional.
The boot script is specified in the ITS file as a text file containing
U-boot commands. When using a boot script the user should configure the
U-boot do_install
task to copy the script to sysroot.
So the script can be included in the FIT image by the kernel-fitimage
class. At run-time, U-boot CONFIG_BOOTCOMMAND define can be configured to
load the boot script from the FIT image and executes it.
The FIT image generated by kernel-fitimage
class is signed when the
variables UBOOT_SIGN_ENABLE, UBOOT_MKIMAGE_DTCOPTS,
UBOOT_SIGN_KEYDIR and UBOOT_SIGN_KEYNAME are set
appropriately. The default values used for FIT_HASH_ALG and
FIT_SIGN_ALG in kernel-fitimage
are “sha256” and
“rsa2048” respectively. The keys for signing fitImage can be generated using
the kernel-fitimage
class when both FIT_GENERATE_KEYS and
UBOOT_SIGN_ENABLE are set to “1”.
6.59 kernel-grub.bbclass
The kernel-grub
class updates the boot area and the boot menu with
the kernel as the priority boot mechanism while installing a RPM to
update the kernel on a deployed target.
6.60 kernel-module-split.bbclass
The kernel-module-split
class provides common functionality for
splitting Linux kernel modules into separate packages.
6.61 kernel-uboot.bbclass
The kernel-uboot
class provides support for building from
vmlinux-style kernel sources.
6.62 kernel-uimage.bbclass
The kernel-uimage
class provides support to pack uImage.
6.63 kernel-yocto.bbclass
The kernel-yocto
class provides common functionality for building
from linux-yocto style kernel source repositories.
6.64 kernelsrc.bbclass
The kernelsrc
class sets the Linux kernel source and version.
6.65 lib_package.bbclass
The lib_package
class supports recipes that build libraries and
produce executable binaries, where those binaries should not be
installed by default along with the library. Instead, the binaries are
added to a separate ${
PN}-bin
package to
make their installation optional.
6.66 libc*.bbclass
The libc*
classes support recipes that build packages with libc
:
The
libc-common
class provides common support for building withlibc
.The
libc-package
class supports packaging upglibc
andeglibc
.
6.67 license.bbclass
The license
class provides license manifest creation and license
exclusion. This class is enabled by default using the default value for
the INHERIT_DISTRO variable.
6.68 linux-kernel-base.bbclass
The linux-kernel-base
class provides common functionality for
recipes that build out of the Linux kernel source tree. These builds
goes beyond the kernel itself. For example, the Perf recipe also
inherits this class.
6.69 linuxloader.bbclass
Provides the function linuxloader()
, which gives the value of the
dynamic loader/linker provided on the platform. This value is used by a
number of other classes.
6.70 logging.bbclass
The logging
class provides the standard shell functions used to log
messages for various BitBake severity levels (i.e. bbplain
,
bbnote
, bbwarn
, bberror
, bbfatal
, and bbdebug
).
This class is enabled by default since it is inherited by the base
class.
6.71 meta.bbclass
The meta
class is inherited by recipes that do not build any output
packages themselves, but act as a “meta” target for building other
recipes.
6.72 metadata_scm.bbclass
The metadata_scm
class provides functionality for querying the
branch and revision of a Source Code Manager (SCM) repository.
The base class uses this class to print the
revisions of each layer before starting every build. The
metadata_scm
class is enabled by default because it is inherited by
the base
class.
6.73 migrate_localcount.bbclass
The migrate_localcount
class verifies a recipe’s localcount data and
increments it appropriately.
6.74 mime.bbclass
The mime
class generates the proper post-install and post-remove
(postinst/postrm) scriptlets for packages that install MIME type files.
These scriptlets call update-mime-database
to add the MIME types to
the shared database.
6.75 mirrors.bbclass
The mirrors
class sets up some standard
MIRRORS entries for source code mirrors. These
mirrors provide a fall-back path in case the upstream source specified
in SRC_URI within recipes is unavailable.
This class is enabled by default since it is inherited by the base class.
6.76 module.bbclass
The module
class provides support for building out-of-tree Linux
kernel modules. The class inherits the
module-base and
kernel-module-split classes,
and implements the do_compile and
do_install tasks. The class provides
everything needed to build and package a kernel module.
For general information on out-of-tree Linux kernel modules, see the “Incorporating Out-of-Tree Modules” section in the Yocto Project Linux Kernel Development Manual.
6.77 module-base.bbclass
The module-base
class provides the base functionality for building
Linux kernel modules. Typically, a recipe that builds software that
includes one or more kernel modules and has its own means of building
the module inherits this class as opposed to inheriting the
module class.
6.78 multilib*.bbclass
The multilib*
classes provide support for building libraries with
different target optimizations or target architectures and installing
them side-by-side in the same image.
For more information on using the Multilib feature, see the “Combining Multiple Versions of Library Files into One Image” section in the Yocto Project Development Tasks Manual.
6.79 native.bbclass
The native
class provides common functionality for recipes that
build tools to run on the Build Host (i.e. tools that use the compiler
or other tools from the build host).
You can create a recipe that builds tools that run natively on the host a couple different ways:
Create a myrecipe
-native.bb
recipe that inherits thenative
class. If you use this method, you must order the inherit statement in the recipe after all other inherit statements so that thenative
class is inherited last.Note
When creating a recipe this way, the recipe name must follow this naming convention:
myrecipe-native.bb
Not using this naming convention can lead to subtle problems caused by existing code that depends on that naming convention.
Create or modify a target recipe that contains the following:
BBCLASSEXTEND = "native"
Inside the recipe, use
_class-native
and_class-target
overrides to specify any functionality specific to the respective native or target case.
Although applied differently, the native
class is used with both
methods. The advantage of the second method is that you do not need to
have two separate recipes (assuming you need both) for native and
target. All common parts of the recipe are automatically shared.
6.80 nativesdk.bbclass
The nativesdk
class provides common functionality for recipes that
wish to build tools to run as part of an SDK (i.e. tools that run on
SDKMACHINE).
You can create a recipe that builds tools that run on the SDK machine a couple different ways:
Create a
nativesdk-
myrecipe.bb
recipe that inherits thenativesdk
class. If you use this method, you must order the inherit statement in the recipe after all other inherit statements so that thenativesdk
class is inherited last.Create a
nativesdk
variant of any recipe by adding the following:BBCLASSEXTEND = "nativesdk"
Inside the recipe, use
_class-nativesdk
and_class-target
overrides to specify any functionality specific to the respective SDK machine or target case.
Note
When creating a recipe, you must follow this naming convention:
nativesdk-myrecipe.bb
Not doing so can lead to subtle problems because code exists that depends on the naming convention.
Although applied differently, the nativesdk
class is used with both
methods. The advantage of the second method is that you do not need to
have two separate recipes (assuming you need both) for the SDK machine
and the target. All common parts of the recipe are automatically shared.
6.81 nopackages.bbclass
Disables packaging tasks for those recipes and classes where packaging is not needed.
6.82 npm.bbclass
Provides support for building Node.js software fetched using the node package manager (NPM).
Note
Currently, recipes inheriting this class must use the npm://
fetcher to have dependencies fetched and packaged automatically.
For information on how to create NPM packages, see the “Creating Node Package Manager (NPM) Packages” section in the Yocto Project Development Tasks Manual.
6.83 oelint.bbclass
The oelint
class is an obsolete lint checking tool that exists in
meta/classes
in the Source Directory.
A number of classes exist that could be generally useful in OE-Core but
are never actually used within OE-Core itself. The oelint
class is
one such example. However, being aware of this class can reduce the
proliferation of different versions of similar classes across multiple
layers.
6.84 own-mirrors.bbclass
The own-mirrors
class makes it easier to set up your own
PREMIRRORS from which to first fetch source
before attempting to fetch it from the upstream specified in
SRC_URI within each recipe.
To use this class, inherit it globally and specify SOURCE_MIRROR_URL. Here is an example:
INHERIT += "own-mirrors"
SOURCE_MIRROR_URL = "http://example.com/my-source-mirror"
You can specify only a single URL
in SOURCE_MIRROR_URL
.
6.85 package.bbclass
The package
class supports generating packages from a build’s
output. The core generic functionality is in package.bbclass
. The
code specific to particular package types resides in these
package-specific classes:
package_deb,
package_rpm,
package_ipk, and
package_tar.
Note
The package_tar class is broken and not supported. It is recommended that you do not use this class.
You can control the list of resulting package formats by using the
PACKAGE_CLASSES
variable defined in your conf/local.conf
configuration file, which is located in the Build Directory.
When defining the variable, you can
specify one or more package types. Since images are generated from
packages, a packaging class is needed to enable image generation. The
first class listed in this variable is used for image generation.
If you take the optional step to set up a repository (package feed) on the development host that can be used by DNF, you can install packages from the feed while you are running the image on the target (i.e. runtime installation of packages). For more information, see the “Using Runtime Package Management” section in the Yocto Project Development Tasks Manual.
The package-specific class you choose can affect build-time performance
and has space ramifications. In general, building a package with IPK
takes about thirty percent less time as compared to using RPM to build
the same or similar package. This comparison takes into account a
complete build of the package with all dependencies previously built.
The reason for this discrepancy is because the RPM package manager
creates and processes more Metadata than the IPK package
manager. Consequently, you might consider setting PACKAGE_CLASSES
to
“package_ipk” if you are building smaller systems.
Before making your package manager decision, however, you should consider some further things about using RPM:
RPM starts to provide more abilities than IPK due to the fact that it processes more Metadata. For example, this information includes individual file types, file checksum generation and evaluation on install, sparse file support, conflict detection and resolution for Multilib systems, ACID style upgrade, and repackaging abilities for rollbacks.
For smaller systems, the extra space used for the Berkeley Database and the amount of metadata when using RPM can affect your ability to perform on-device upgrades.
You can find additional information on the effects of the package class at these two Yocto Project mailing list links:
6.86 package_deb.bbclass
The package_deb
class provides support for creating packages that
use the Debian (i.e. .deb
) file format. The class ensures the
packages are written out in a .deb
file format to the
${
DEPLOY_DIR_DEB}
directory.
This class inherits the package class and
is enabled through the PACKAGE_CLASSES
variable in the local.conf
file.
6.87 package_ipk.bbclass
The package_ipk
class provides support for creating packages that
use the IPK (i.e. .ipk
) file format. The class ensures the packages
are written out in a .ipk
file format to the
${
DEPLOY_DIR_IPK}
directory.
This class inherits the package class and
is enabled through the PACKAGE_CLASSES
variable in the local.conf
file.
6.88 package_rpm.bbclass
The package_rpm
class provides support for creating packages that
use the RPM (i.e. .rpm
) file format. The class ensures the packages
are written out in a .rpm
file format to the
${
DEPLOY_DIR_RPM}
directory.
This class inherits the package class and
is enabled through the PACKAGE_CLASSES
variable in the local.conf
file.
6.89 package_tar.bbclass
The package_tar
class provides support for creating tarballs. The
class ensures the packages are written out in a tarball format to the
${
DEPLOY_DIR_TAR}
directory.
This class inherits the package class and
is enabled through the PACKAGE_CLASSES
variable in the local.conf
file.
Note
You cannot specify the package_tar
class first using the
PACKAGE_CLASSES
variable. You must use .deb
, .ipk
, or .rpm
file formats for your image or SDK.
6.90 packagedata.bbclass
The packagedata
class provides common functionality for reading
pkgdata
files found in PKGDATA_DIR. These
files contain information about each output package produced by the
OpenEmbedded build system.
This class is enabled by default because it is inherited by the package class.
6.91 packagegroup.bbclass
The packagegroup
class sets default values appropriate for package
group recipes (e.g. PACKAGES
, PACKAGE_ARCH
, ALLOW_EMPTY
, and
so forth). It is highly recommended that all package group recipes
inherit this class.
For information on how to use this class, see the “Customizing Images Using Custom Package Groups” section in the Yocto Project Development Tasks Manual.
Previously, this class was called the task
class.
6.92 patch.bbclass
The patch
class provides all functionality for applying patches
during the do_patch task.
This class is enabled by default because it is inherited by the base class.
6.93 perlnative.bbclass
When inherited by a recipe, the perlnative
class supports using the
native version of Perl built by the build system rather than using the
version provided by the build host.
6.94 pixbufcache.bbclass
The pixbufcache
class generates the proper post-install and
post-remove (postinst/postrm) scriptlets for packages that install
pixbuf loaders, which are used with gdk-pixbuf
. These scriptlets
call update_pixbuf_cache
to add the pixbuf loaders to the cache.
Since the cache files are architecture-specific, update_pixbuf_cache
is run using QEMU if the postinst scriptlets need to be run on the build
host during image creation.
If the pixbuf loaders being installed are in packages other than the recipe’s main package, set PIXBUF_PACKAGES to specify the packages containing the loaders.
6.95 pkgconfig.bbclass
The pkgconfig
class provides a standard way to get header and
library information by using pkg-config
. This class aims to smooth
integration of pkg-config
into libraries that use it.
During staging, BitBake installs pkg-config
data into the
sysroots/
directory. By making use of sysroot functionality within
pkg-config
, the pkgconfig
class no longer has to manipulate the
files.
6.96 populate_sdk.bbclass
The populate_sdk
class provides support for SDK-only recipes. For
information on advantages gained when building a cross-development
toolchain using the do_populate_sdk
task, see the “Building an SDK Installer”
section in the Yocto Project Application Development and the Extensible
Software Development Kit (eSDK) manual.
6.97 populate_sdk_*.bbclass
The populate_sdk_*
classes support SDK creation and consist of the
following classes:
populate_sdk_base
: The base class supporting SDK creation under all package managers (i.e. DEB, RPM, and opkg).populate_sdk_deb
: Supports creation of the SDK given the Debian package manager.populate_sdk_rpm
: Supports creation of the SDK given the RPM package manager.populate_sdk_ipk
: Supports creation of the SDK given the opkg (IPK format) package manager.populate_sdk_ext
: Supports extensible SDK creation under all package managers.
The populate_sdk_base
class inherits the appropriate
populate_sdk_*
(i.e. deb
, rpm
, and ipk
) based on
IMAGE_PKGTYPE.
The base class ensures all source and destination directories are
established and then populates the SDK. After populating the SDK, the
populate_sdk_base
class constructs two sysroots:
${
SDK_ARCH}-nativesdk
, which
contains the cross-compiler and associated tooling, and the target,
which contains a target root filesystem that is configured for the SDK
usage. These two images reside in SDK_OUTPUT,
which consists of the following:
${SDK_OUTPUT}/${SDK_ARCH}-nativesdk-pkgs
${SDK_OUTPUT}/${SDKTARGETSYSROOT}/target-pkgs
Finally, the base populate SDK class creates the toolchain environment setup script, the tarball of the SDK, and the installer.
The respective populate_sdk_deb
, populate_sdk_rpm
, and
populate_sdk_ipk
classes each support the specific type of SDK.
These classes are inherited by and used with the populate_sdk_base
class.
For more information on the cross-development toolchain generation, see the “Cross-Development Toolchain Generation” section in the Yocto Project Overview and Concepts Manual. For information on advantages gained when building a cross-development toolchain using the do_populate_sdk task, see the “Building an SDK Installer” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
6.98 prexport.bbclass
The prexport
class provides functionality for exporting
PR values.
Note
This class is not intended to be used directly. Rather, it is enabled
when using “bitbake-prserv-tool export
”.
6.99 primport.bbclass
The primport
class provides functionality for importing
PR values.
Note
This class is not intended to be used directly. Rather, it is enabled
when using “bitbake-prserv-tool import
”.
6.100 prserv.bbclass
The prserv
class provides functionality for using a PR
service in order to
automatically manage the incrementing of the PR
variable for each recipe.
This class is enabled by default because it is inherited by the package class. However, the OpenEmbedded build system will not enable the functionality of this class unless PRSERV_HOST has been set.
6.101 ptest.bbclass
The ptest
class provides functionality for packaging and installing
runtime tests for recipes that build software that provides these tests.
This class is intended to be inherited by individual recipes. However, the class’ functionality is largely disabled unless “ptest” appears in DISTRO_FEATURES. See the “Testing Packages With ptest” section in the Yocto Project Development Tasks Manual for more information on ptest.
6.102 ptest-gnome.bbclass
Enables package tests (ptests) specifically for GNOME packages, which
have tests intended to be executed with gnome-desktop-testing
.
For information on setting up and running ptests, see the “Testing Packages With ptest” section in the Yocto Project Development Tasks Manual.
6.103 python3-dir.bbclass
The python3-dir
class provides the base version, location, and site
package location for Python 3.
6.104 python3native.bbclass
The python3native
class supports using the native version of Python
3 built by the build system rather than support of the version provided
by the build host.
6.105 python3targetconfig.bbclass
The python3targetconfig
class supports using the native version of Python
3 built by the build system rather than support of the version provided
by the build host, except that the configuration for the target machine
is accessible (such as correct installation directories). This also adds a
dependency on target python3
, so should only be used where appropriate
in order to avoid unnecessarily lengthening builds.
6.106 qemu.bbclass
The qemu
class provides functionality for recipes that either need
QEMU or test for the existence of QEMU. Typically, this class is used to
run programs for a target system on the build host using QEMU’s
application emulation mode.
6.107 recipe_sanity.bbclass
The recipe_sanity
class checks for the presence of any host system
recipe prerequisites that might affect the build (e.g. variables that
are set or software that is present).
6.108 relocatable.bbclass
The relocatable
class enables relocation of binaries when they are
installed into the sysroot.
This class makes use of the chrpath class and is used by both the cross and native classes.
6.109 remove-libtool.bbclass
The remove-libtool
class adds a post function to the
do_install task to remove all .la
files
installed by libtool
. Removing these files results in them being
absent from both the sysroot and target packages.
If a recipe needs the .la
files to be installed, then the recipe can
override the removal by setting REMOVE_LIBTOOL_LA
to “0” as follows:
REMOVE_LIBTOOL_LA = "0"
Note
The remove-libtool
class is not enabled by default.
6.110 report-error.bbclass
The report-error
class supports enabling the error reporting
tool”,
which allows you to submit build error information to a central database.
The class collects debug information for recipe, recipe version, task,
machine, distro, build system, target system, host distro, branch,
commit, and log. From the information, report files using a JSON format
are created and stored in
${
LOG_DIR}/error-report
.
6.111 rm_work.bbclass
The rm_work
class supports deletion of temporary workspace, which
can ease your hard drive demands during builds.
The OpenEmbedded build system can use a substantial amount of disk space
during the build process. A portion of this space is the work files
under the ${TMPDIR}/work
directory for each recipe. Once the build
system generates the packages for a recipe, the work files for that
recipe are no longer needed. However, by default, the build system
preserves these files for inspection and possible debugging purposes. If
you would rather have these files deleted to save disk space as the
build progresses, you can enable rm_work
by adding the following to
your local.conf
file, which is found in the Build Directory.
INHERIT += "rm_work"
If you are
modifying and building source code out of the work directory for a
recipe, enabling rm_work
will potentially result in your changes to
the source being lost. To exclude some recipes from having their work
directories deleted by rm_work
, you can add the names of the recipe
or recipes you are working on to the RM_WORK_EXCLUDE
variable, which
can also be set in your local.conf
file. Here is an example:
RM_WORK_EXCLUDE += "busybox glibc"
6.112 rootfs*.bbclass
The rootfs*
classes support creating the root filesystem for an
image and consist of the following classes:
The
rootfs-postcommands
class, which defines filesystem post-processing functions for image recipes.The
rootfs_deb
class, which supports creation of root filesystems for images built using.deb
packages.The
rootfs_rpm
class, which supports creation of root filesystems for images built using.rpm
packages.The
rootfs_ipk
class, which supports creation of root filesystems for images built using.ipk
packages.The
rootfsdebugfiles
class, which installs additional files found on the build host directly into the root filesystem.
The root filesystem is created from packages using one of the
rootfs*.bbclass
files as determined by the
PACKAGE_CLASSES variable.
For information on how root filesystem images are created, see the “Image Generation” section in the Yocto Project Overview and Concepts Manual.
6.113 sanity.bbclass
The sanity
class checks to see if prerequisite software is present
on the host system so that users can be notified of potential problems
that might affect their build. The class also performs basic user
configuration checks from the local.conf
configuration file to
prevent common mistakes that cause build failures. Distribution policy
usually determines whether to include this class.
6.114 scons.bbclass
The scons
class supports recipes that need to build software that
uses the SCons build system. You can use the
EXTRA_OESCONS variable to specify
additional configuration options you want to pass SCons command line.
6.115 sdl.bbclass
The sdl
class supports recipes that need to build software that uses
the Simple DirectMedia Layer (SDL) library.
6.116 setuptools3.bbclass
The setuptools3
class supports Python version 3.x extensions that
use build systems based on setuptools
. If your recipe uses these
build systems, the recipe needs to inherit the setuptools3
class.
6.117 sign_rpm.bbclass
The sign_rpm
class supports generating signed RPM packages.
6.118 sip.bbclass
The sip
class supports recipes that build or package SIP-based
Python bindings.
6.119 siteconfig.bbclass
The siteconfig
class provides functionality for handling site
configuration. The class is used by the
autotools class to accelerate the
do_configure task.
6.120 siteinfo.bbclass
The siteinfo
class provides information about the targets that might
be needed by other classes or recipes.
As an example, consider Autotools, which can require tests that must
execute on the target hardware. Since this is not possible in general
when cross compiling, site information is used to provide cached test
results so these tests can be skipped over but still make the correct
values available. The meta/site directory
contains test results
sorted into different categories such as architecture, endianness, and
the libc
used. Site information provides a list of files containing
data relevant to the current build in the CONFIG_SITE
variable that
Autotools automatically picks up.
The class also provides variables like SITEINFO_ENDIANNESS
and
SITEINFO_BITS
that can be used elsewhere in the metadata.
6.121 sstate.bbclass
The sstate
class provides support for Shared State (sstate). By
default, the class is enabled through the
INHERIT_DISTRO variable’s default value.
For more information on sstate, see the “Shared State Cache” section in the Yocto Project Overview and Concepts Manual.
6.122 staging.bbclass
The staging
class installs files into individual recipe work
directories for sysroots. The class contains the following key tasks:
The do_populate_sysroot task, which is responsible for handing the files that end up in the recipe sysroots.
The do_prepare_recipe_sysroot task (a “partner” task to the
populate_sysroot
task), which installs the files into the individual recipe work directories (i.e. WORKDIR).
The code in the staging
class is complex and basically works in two
stages:
Stage One: The first stage addresses recipes that have files they want to share with other recipes that have dependencies on the originating recipe. Normally these dependencies are installed through the do_install task into
${
D}
. Thedo_populate_sysroot
task copies a subset of these files into${SYSROOT_DESTDIR}
. This subset of files is controlled by the SYSROOT_DIRS, SYSROOT_DIRS_NATIVE, and SYSROOT_DIRS_BLACKLIST variables.Note
Additionally, a recipe can customize the files further by declaring a processing function in the
SYSROOT_PREPROCESS_FUNCS
variable.A shared state (sstate) object is built from these files and the files are placed into a subdirectory of build/tmp/sysroots-components/. The files are scanned for hardcoded paths to the original installation location. If the location is found in text files, the hardcoded locations are replaced by tokens and a list of the files needing such replacements is created. These adjustments are referred to as “FIXMEs”. The list of files that are scanned for paths is controlled by the SSTATE_SCAN_FILES variable.
Stage Two: The second stage addresses recipes that want to use something from another recipe and declare a dependency on that recipe through the DEPENDS variable. The recipe will have a do_prepare_recipe_sysroot task and when this task executes, it creates the
recipe-sysroot
andrecipe-sysroot-native
in the recipe work directory (i.e. WORKDIR). The OpenEmbedded build system creates hard links to copies of the relevant files fromsysroots-components
into the recipe work directory.Note
If hard links are not possible, the build system uses actual copies.
The build system then addresses any “FIXMEs” to paths as defined from the list created in the first stage.
Finally, any files in
${bindir}
within the sysroot that have the prefix “postinst-
” are executed.Note
Although such sysroot post installation scripts are not recommended for general use, the files do allow some issues such as user creation and module indexes to be addressed.
Because recipes can have other dependencies outside of
DEPENDS
(e.g.do_unpack[depends] += "tar-native:do_populate_sysroot"
), the sysroot creation functionextend_recipe_sysroot
is also added as a pre-function for those tasks whose dependencies are not throughDEPENDS
but operate similarly.When installing dependencies into the sysroot, the code traverses the dependency graph and processes dependencies in exactly the same way as the dependencies would or would not be when installed from sstate. This processing means, for example, a native tool would have its native dependencies added but a target library would not have its dependencies traversed or installed. The same sstate dependency code is used so that builds should be identical regardless of whether sstate was used or not. For a closer look, see the
setscene_depvalid()
function in the sstate class.The build system is careful to maintain manifests of the files it installs so that any given dependency can be installed as needed. The sstate hash of the installed item is also stored so that if it changes, the build system can reinstall it.
6.123 syslinux.bbclass
The syslinux
class provides syslinux-specific functions for building
bootable images.
The class supports the following variables:
INITRD: Indicates list of filesystem images to concatenate and use as an initial RAM disk (initrd). This variable is optional.
ROOTFS: Indicates a filesystem image to include as the root filesystem. This variable is optional.
AUTO_SYSLINUXMENU: Enables creating an automatic menu when set to “1”.
LABELS: Lists targets for automatic configuration.
APPEND: Lists append string overrides for each label.
SYSLINUX_OPTS: Lists additional options to add to the syslinux file. Semicolon characters separate multiple options.
SYSLINUX_SPLASH: Lists a background for the VGA boot menu when you are using the boot menu.
SYSLINUX_DEFAULT_CONSOLE: Set to “console=ttyX” to change kernel boot default console.
SYSLINUX_SERIAL: Sets an alternate serial port. Or, turns off serial when the variable is set with an empty string.
SYSLINUX_SERIAL_TTY: Sets an alternate “console=tty…” kernel boot argument.
6.124 systemd.bbclass
The systemd
class provides support for recipes that install systemd
unit files.
The functionality for this class is disabled unless you have “systemd” in DISTRO_FEATURES.
Under this class, the recipe or Makefile (i.e. whatever the recipe is
calling during the do_install task)
installs unit files into
${
D}${systemd_unitdir}/system
. If the unit
files being installed go into packages other than the main package, you
need to set SYSTEMD_PACKAGES in your
recipe to identify the packages in which the files will be installed.
You should set SYSTEMD_SERVICE to the
name of the service file. You should also use a package name override to
indicate the package to which the value applies. If the value applies to
the recipe’s main package, use ${
PN}
. Here
is an example from the connman recipe:
SYSTEMD_SERVICE_${PN} = "connman.service"
Services are set up to start on boot automatically unless you have set SYSTEMD_AUTO_ENABLE to “disable”.
For more information on systemd
, see the
“Selecting an Initialization Manager”
section in the Yocto Project Development Tasks Manual.
6.125 systemd-boot.bbclass
The systemd-boot
class provides functions specific to the
systemd-boot bootloader for building bootable images. This is an
internal class and is not intended to be used directly.
Note
The systemd-boot
class is a result from merging the gummiboot
class
used in previous Yocto Project releases with the systemd
project.
Set the EFI_PROVIDER variable to “systemd-boot” to use this class. Doing so creates a standalone EFI bootloader that is not dependent on systemd.
For information on more variables used and supported in this class, see the SYSTEMD_BOOT_CFG, SYSTEMD_BOOT_ENTRIES, and SYSTEMD_BOOT_TIMEOUT variables.
You can also see the Systemd-boot documentation for more information.
6.126 terminal.bbclass
The terminal
class provides support for starting a terminal session.
The OE_TERMINAL variable controls which
terminal emulator is used for the session.
Other classes use the terminal
class anywhere a separate terminal
session needs to be started. For example, the
patch class assuming
PATCHRESOLVE is set to “user”, the
cml1 class, and the
devshell class all use the terminal
class.
6.127 testimage*.bbclass
The testimage*
classes support running automated tests against
images using QEMU and on actual hardware. The classes handle loading the
tests and starting the image. To use the classes, you need to perform
steps to set up the environment.
Note
Best practices include using IMAGE_CLASSES rather than
INHERIT to inherit the testimage
class for automated image
testing.
The tests are commands that run on the target system over ssh
. Each
test is written in Python and makes use of the unittest
module.
The testimage.bbclass
runs tests on an image when called using the
following:
$ bitbake -c testimage image
The testimage-auto
class
runs tests on an image after the image is constructed (i.e.
TESTIMAGE_AUTO must be set to “1”).
For information on how to enable, run, and create new tests, see the “Performing Automated Runtime Testing” section in the Yocto Project Development Tasks Manual.
6.128 testsdk.bbclass
This class supports running automated tests against software development
kits (SDKs). The testsdk
class runs tests on an SDK when called
using the following:
$ bitbake -c testsdk image
Note
Best practices include using IMAGE_CLASSES rather than
INHERIT to inherit the testsdk
class for automated SDK
testing.
6.129 texinfo.bbclass
This class should be inherited by recipes whose upstream packages invoke
the texinfo
utilities at build-time. Native and cross recipes are
made to use the dummy scripts provided by texinfo-dummy-native
, for
improved performance. Target architecture recipes use the genuine
Texinfo utilities. By default, they use the Texinfo utilities on the
host system.
Note
If you want to use the Texinfo recipe shipped with the build system, you can remove “texinfo-native” from ASSUME_PROVIDED and makeinfo from SANITY_REQUIRED_UTILITIES.
6.130 toaster.bbclass
The toaster
class collects information about packages and images and
sends them as events that the BitBake user interface can receive. The
class is enabled when the Toaster user interface is running.
This class is not intended to be used directly.
6.131 toolchain-scripts.bbclass
The toolchain-scripts
class provides the scripts used for setting up
the environment for installed SDKs.
6.132 typecheck.bbclass
The typecheck
class provides support for validating the values of
variables set at the configuration level against their defined types.
The OpenEmbedded build system allows you to define the type of a
variable using the “type” varflag. Here is an example:
IMAGE_FEATURES[type] = "list"
6.133 uboot-config.bbclass
The uboot-config
class provides support for U-Boot configuration for
a machine. Specify the machine in your recipe as follows:
UBOOT_CONFIG ??= <default>
UBOOT_CONFIG[foo] = "config,images"
You can also specify the machine using this method:
UBOOT_MACHINE = "config"
See the UBOOT_CONFIG and UBOOT_MACHINE variables for additional information.
6.134 uninative.bbclass
Attempts to isolate the build system from the host distribution’s C
library in order to make re-use of native shared state artifacts across
different host distributions practical. With this class enabled, a
tarball containing a pre-built C library is downloaded at the start of
the build. In the Poky reference distribution this is enabled by default
through meta/conf/distro/include/yocto-uninative.inc
. Other
distributions that do not derive from poky can also
“require conf/distro/include/yocto-uninative.inc
” to use this.
Alternatively if you prefer, you can build the uninative-tarball recipe
yourself, publish the resulting tarball (e.g. via HTTP) and set
UNINATIVE_URL
and UNINATIVE_CHECKSUM
appropriately. For an
example, see the meta/conf/distro/include/yocto-uninative.inc
.
The uninative
class is also used unconditionally by the extensible
SDK. When building the extensible SDK, uninative-tarball
is built
and the resulting tarball is included within the SDK.
6.135 update-alternatives.bbclass
The update-alternatives
class helps the alternatives system when
multiple sources provide the same command. This situation occurs when
several programs that have the same or similar function are installed
with the same name. For example, the ar
command is available from
the busybox
, binutils
and elfutils
packages. The
update-alternatives
class handles renaming the binaries so that
multiple packages can be installed without conflicts. The ar
command
still works regardless of which packages are installed or subsequently
removed. The class renames the conflicting binary in each package and
symlinks the highest priority binary during installation or removal of
packages.
To use this class, you need to define a number of variables:
These variables list alternative commands needed by a package, provide pathnames for links, default links for targets, and so forth. For details on how to use this class, see the comments in the update-alternatives.bbclass file.
Note
You can use the update-alternatives
command directly in your recipes.
However, this class simplifies things in most cases.
6.136 update-rc.d.bbclass
The update-rc.d
class uses update-rc.d
to safely install an
initialization script on behalf of the package. The OpenEmbedded build
system takes care of details such as making sure the script is stopped
before a package is removed and started when the package is installed.
Three variables control this class: INITSCRIPT_PACKAGES
,
INITSCRIPT_NAME
and INITSCRIPT_PARAMS
. See the variable links
for details.
6.137 useradd*.bbclass
The useradd*
classes support the addition of users or groups for
usage by the package on the target. For example, if you have packages
that contain system services that should be run under their own user or
group, you can use these classes to enable creation of the user or
group. The meta-skeleton/recipes-skeleton/useradd/useradd-example.bb
recipe in the Source Directory provides a simple
example that shows how to add three users and groups to two packages.
See the useradd-example.bb
recipe for more information on how to use
these classes.
The useradd_base
class provides basic functionality for user or
groups settings.
The useradd*
classes support the
USERADD_PACKAGES,
USERADD_PARAM,
GROUPADD_PARAM, and
GROUPMEMS_PARAM variables.
The useradd-staticids
class supports the addition of users or groups
that have static user identification (uid
) and group identification
(gid
) values.
The default behavior of the OpenEmbedded build system for assigning
uid
and gid
values when packages add users and groups during
package install time is to add them dynamically. This works fine for
programs that do not care what the values of the resulting users and
groups become. In these cases, the order of the installation determines
the final uid
and gid
values. However, if non-deterministic
uid
and gid
values are a problem, you can override the default,
dynamic application of these values by setting static values. When you
set static values, the OpenEmbedded build system looks in
BBPATH for files/passwd
and files/group
files for the values.
To use static uid
and gid
values, you need to set some
variables. See the USERADDEXTENSION,
USERADD_UID_TABLES,
USERADD_GID_TABLES, and
USERADD_ERROR_DYNAMIC variables.
You can also see the useradd class for
additional information.
Note
You do not use the useradd-staticids
class directly. You either enable
or disable the class by setting the USERADDEXTENSION
variable. If you
enable or disable the class in a configured system, TMPDIR might
contain incorrect uid
and gid
values. Deleting the TMPDIR
directory will correct this condition.
6.138 utility-tasks.bbclass
The utility-tasks
class provides support for various “utility” type
tasks that are applicable to all recipes, such as
do_clean and
do_listtasks.
This class is enabled by default because it is inherited by the base class.
6.139 utils.bbclass
The utils
class provides some useful Python functions that are
typically used in inline Python expressions (e.g. ${@...}
). One
example use is for bb.utils.contains()
.
This class is enabled by default because it is inherited by the base class.
6.140 vala.bbclass
The vala
class supports recipes that need to build software written
using the Vala programming language.
6.141 waf.bbclass
The waf
class supports recipes that need to build software that uses
the Waf build system. You can use the
EXTRA_OECONF or
PACKAGECONFIG_CONFARGS variables
to specify additional configuration options to be passed on the Waf
command line.
7 Tasks
Tasks are units of execution for BitBake. Recipes (.bb
files) use
tasks to complete configuring, compiling, and packaging software. This
chapter provides a reference of the tasks defined in the OpenEmbedded
build system.
7.1 Normal Recipe Build Tasks
The following sections describe normal tasks associated with building a recipe. For more information on tasks and dependencies, see the “Tasks” and “Dependencies” sections in the BitBake User Manual.
7.1.1 do_build
The default task for all recipes. This task depends on all other normal tasks required to build a recipe.
7.1.2 do_compile
Compiles the source code. This task runs with the current working
directory set to ${
B}
.
The default behavior of this task is to run the oe_runmake
function
if a makefile (Makefile
, makefile
, or GNUmakefile
) is found.
If no such file is found, the do_compile
task does nothing.
7.1.3 do_compile_ptest_base
Compiles the runtime test suite included in the software being built.
7.1.4 do_configure
Configures the source by enabling and disabling any build-time and
configuration options for the software being built. The task runs with
the current working directory set to ${
B}
.
The default behavior of this task is to run oe_runmake clean
if a
makefile (Makefile
, makefile
, or GNUmakefile
) is found and
CLEANBROKEN is not set to “1”. If no such
file is found or the CLEANBROKEN
variable is set to “1”, the
do_configure
task does nothing.
7.1.5 do_configure_ptest_base
Configures the runtime test suite included in the software being built.
7.1.6 do_deploy
Writes output files that are to be deployed to
${
DEPLOY_DIR_IMAGE}
. The
task runs with the current working directory set to
${
B}
.
Recipes implementing this task should inherit the
deploy class and should write the output
to ${
DEPLOYDIR}
, which is not to be
confused with ${DEPLOY_DIR}
. The deploy
class sets up
do_deploy
as a shared state (sstate) task that can be accelerated
through sstate use. The sstate mechanism takes care of copying the
output from ${DEPLOYDIR}
to ${DEPLOY_DIR_IMAGE}
.
Note
Do not write the output directly to ${DEPLOY_DIR_IMAGE}
, as this causes
the sstate mechanism to malfunction.
The do_deploy
task is not added as a task by default and
consequently needs to be added manually. If you want the task to run
after do_compile, you can add it by doing
the following:
addtask deploy after do_compile
Adding do_deploy
after other tasks works the same way.
Note
You do not need to add before do_build
to the addtask
command
(though it is harmless), because the base
class contains the following:
do_build[recrdeptask] += "do_deploy"
See the “Dependencies” section in the BitBake User Manual for more information.
If the do_deploy
task re-executes, any previous output is removed
(i.e. “cleaned”).
7.1.7 do_fetch
Fetches the source code. This task uses the SRC_URI variable and the argument’s prefix to determine the correct fetcher module.
7.1.8 do_image
Starts the image generation process. The do_image
task runs after
the OpenEmbedded build system has run the
do_rootfs task during which packages are
identified for installation into the image and the root filesystem is
created, complete with post-processing.
The do_image
task performs pre-processing on the image through the
IMAGE_PREPROCESS_COMMAND and
dynamically generates supporting do_image_*
tasks as needed.
For more information on image creation, see the “Image Generation” section in the Yocto Project Overview and Concepts Manual.
7.1.9 do_image_complete
Completes the image generation process. The do_image_complete
task
runs after the OpenEmbedded build system has run the
do_image task during which image
pre-processing occurs and through dynamically generated do_image_*
tasks the image is constructed.
The do_image_complete
task performs post-processing on the image
through the
IMAGE_POSTPROCESS_COMMAND.
For more information on image creation, see the “Image Generation” section in the Yocto Project Overview and Concepts Manual.
7.1.10 do_install
Copies files that are to be packaged into the holding area
${
D}
. This task runs with the current
working directory set to ${
B}
, which is the
compilation directory. The do_install
task, as well as other tasks
that either directly or indirectly depend on the installed files (e.g.
do_package, do_package_write_*
, and
do_rootfs), run under
fakeroot.
Note
When installing files, be careful not to set the owner and group IDs
of the installed files to unintended values. Some methods of copying
files, notably when using the recursive cp
command, can preserve
the UID and/or GID of the original file, which is usually not what
you want. The host-user-contaminated
QA check checks for files
that probably have the wrong ownership.
Safe methods for installing files include the following:
The
install
utility. This utility is the preferred method.The
cp
command with the “–no-preserve=ownership” option.The
tar
command with the “–no-same-owner” option. See thebin_package.bbclass
file in themeta/classes
directory of the Source Directory for an example.
7.1.11 do_install_ptest_base
Copies the runtime test suite files from the compilation directory to a holding area.
7.1.12 do_package
Analyzes the content of the holding area
${
D}
and splits the content into subsets
based on available packages and files. This task makes use of the
PACKAGES and FILES
variables.
The do_package
task, in conjunction with the
do_packagedata task, also saves some
important package metadata. For additional information, see the
PKGDESTWORK variable and the
“Automatically Added Runtime Dependencies”
section in the Yocto Project Overview and Concepts Manual.
7.1.13 do_package_qa
Runs QA checks on packaged files. For more information on these checks, see the insane class.
7.1.14 do_package_write_deb
Creates Debian packages (i.e. *.deb
files) and places them in the
${
DEPLOY_DIR_DEB}
directory in
the package feeds area. For more information, see the
“Package Feeds” section in
the Yocto Project Overview and Concepts Manual.
7.1.15 do_package_write_ipk
Creates IPK packages (i.e. *.ipk
files) and places them in the
${
DEPLOY_DIR_IPK}
directory in
the package feeds area. For more information, see the
“Package Feeds” section in
the Yocto Project Overview and Concepts Manual.
7.1.16 do_package_write_rpm
Creates RPM packages (i.e. *.rpm
files) and places them in the
${
DEPLOY_DIR_RPM}
directory in
the package feeds area. For more information, see the
“Package Feeds” section in
the Yocto Project Overview and Concepts Manual.
7.1.17 do_package_write_tar
Creates tarballs and places them in the
${
DEPLOY_DIR_TAR}
directory in
the package feeds area. For more information, see the
“Package Feeds” section in
the Yocto Project Overview and Concepts Manual.
7.1.18 do_packagedata
Saves package metadata generated by the do_package task in PKGDATA_DIR to make it available globally.
7.1.19 do_patch
Locates patch files and applies them to the source code.
After fetching and unpacking source files, the build system uses the recipe’s SRC_URI statements to locate and apply patch files to the source code.
Note
The build system uses the FILESPATH variable to determine the default set of directories when searching for patches.
Patch files, by default, are *.patch
and *.diff
files created
and kept in a subdirectory of the directory holding the recipe file. For
example, consider the
bluez5
recipe from the OE-Core layer (i.e. poky/meta
):
poky/meta/recipes-connectivity/bluez5
This recipe has two patch files located here:
poky/meta/recipes-connectivity/bluez5/bluez5
In the bluez5
recipe, the SRC_URI
statements point to the source
and patch files needed to build the package.
Note
In the case for the bluez5_5.48.bb
recipe, the SRC_URI
statements
are from an include file bluez5.inc
.
As mentioned earlier, the build system treats files whose file types are
.patch
and .diff
as patch files. However, you can use the
“apply=yes” parameter with the SRC_URI
statement to indicate any
file as a patch file:
SRC_URI = " \
git://path_to_repo/some_package \
file://file;apply=yes \
"
Conversely, if you have a directory full of patch files and you want to
exclude some so that the do_patch
task does not apply them during
the patch phase, you can use the “apply=no” parameter with the
SRC_URI
statement:
SRC_URI = " \
git://path_to_repo/some_package \
file://path_to_lots_of_patch_files \
file://path_to_lots_of_patch_files/patch_file5;apply=no \
"
In the
previous example, assuming all the files in the directory holding the
patch files end with either .patch
or .diff
, every file would be
applied as a patch by default except for the patch_file5
patch.
You can find out more about the patching process in the “Patching” section in the Yocto Project Overview and Concepts Manual and the “Patching Code” section in the Yocto Project Development Tasks Manual.
7.1.20 do_populate_lic
Writes license information for the recipe that is collected later when the image is constructed.
7.1.21 do_populate_sdk
Creates the file and directory structure for an installable SDK. See the “SDK Generation” section in the Yocto Project Overview and Concepts Manual for more information.
7.1.22 do_populate_sdk_ext
Creates the file and directory structure for an installable extensible SDK (eSDK). See the “SDK Generation” section in the Yocto Project Overview and Concepts Manual for more information.
7.1.23 do_populate_sysroot
Stages (copies) a subset of the files installed by the
do_install task into the appropriate
sysroot. For information on how to access these files from other
recipes, see the STAGING_DIR* variables.
Directories that would typically not be needed by other recipes at build
time (e.g. /etc
) are not copied by default.
For information on what directories are copied by default, see the SYSROOT_DIRS* variables. You can change these variables inside your recipe if you need to make additional (or fewer) directories available to other recipes at build time.
The do_populate_sysroot
task is a shared state (sstate) task, which
means that the task can be accelerated through sstate use. Realize also
that if the task is re-executed, any previous output is removed (i.e.
“cleaned”).
7.1.24 do_prepare_recipe_sysroot
Installs the files into the individual recipe specific sysroots (i.e.
recipe-sysroot
and recipe-sysroot-native
under
${
WORKDIR}
based upon the
dependencies specified by DEPENDS). See the
“staging” class for more information.
7.1.25 do_rm_work
Removes work files after the OpenEmbedded build system has finished with them. You can learn more by looking at the “rm_work.bbclass” section.
7.1.26 do_unpack
Unpacks the source code into a working directory pointed to by
${
WORKDIR}
. The S
variable also plays a role in where unpacked source files ultimately
reside. For more information on how source files are unpacked, see the
“Source Fetching”
section in the Yocto Project Overview and Concepts Manual and also see
the WORKDIR
and S
variable descriptions.
7.2 Manually Called Tasks
These tasks are typically manually triggered (e.g. by using the
bitbake -c
command-line option):
7.2.1 do_checkpkg
Provides information about the recipe including its upstream version and status. The upstream version and status reveals whether or not a version of the recipe exists upstream and a status of not updated, updated, or unknown.
To check the upstream version and status of a recipe, use the following devtool commands:
$ devtool latest-version
$ devtool check-upgrade-status
See the “devtool Quick Reference”
chapter for more information on
devtool
. See the “Checking on the Upgrade Status of a Recipe”
section for information on checking the upgrade status of a recipe.
To build the checkpkg
task, use the bitbake
command with the
“-c” option and task name:
$ bitbake core-image-minimal -c checkpkg
By default, the results are stored in $LOG_DIR (e.g.
$BUILD_DIR/tmp/log
).
7.2.2 do_checkuri
Validates the SRC_URI value.
7.2.3 do_clean
Removes all output files for a target from the
do_unpack task forward (i.e. do_unpack
,
do_configure,
do_compile,
do_install, and
do_package).
You can run this task using BitBake as follows:
$ bitbake -c clean recipe
Running this task does not remove the
sstate cache files.
Consequently, if no changes have been made and the recipe is rebuilt
after cleaning, output files are simply restored from the sstate cache.
If you want to remove the sstate cache files for the recipe, you need to
use the do_cleansstate task instead
(i.e. bitbake -c cleansstate
recipe).
7.2.4 do_cleanall
Removes all output files, shared state
(sstate) cache, and
downloaded source files for a target (i.e. the contents of
DL_DIR). Essentially, the do_cleanall
task is
identical to the do_cleansstate task
with the added removal of downloaded source files.
You can run this task using BitBake as follows:
$ bitbake -c cleanall recipe
Typically, you would not normally use the cleanall
task. Do so only
if you want to start fresh with the do_fetch
task.
7.2.5 do_cleansstate
Removes all output files and shared state
(sstate) cache for a
target. Essentially, the do_cleansstate
task is identical to the
do_clean task with the added removal of
shared state (sstate)
cache.
You can run this task using BitBake as follows:
$ bitbake -c cleansstate recipe
When you run the do_cleansstate
task, the OpenEmbedded build system
no longer uses any sstate. Consequently, building the recipe from
scratch is guaranteed.
Note
The do_cleansstate
task cannot remove sstate from a remote sstate
mirror. If you need to build a target from scratch using remote mirrors, use
the “-f” option as follows:
$ bitbake -f -c do_cleansstate target
7.2.6 do_devpyshell
Starts a shell in which an interactive Python interpreter allows you to
interact with the BitBake build environment. From within this shell, you
can directly examine and set bits from the data store and execute
functions as if within the BitBake environment. See the “Using a Development Python Shell” section in
the Yocto Project Development Tasks Manual for more information about
using devpyshell
.
7.2.7 do_devshell
Starts a shell whose environment is set up for development, debugging,
or both. See the “Using a Development Shell” section in the
Yocto Project Development Tasks Manual for more information about using
devshell
.
7.2.8 do_listtasks
Lists all defined tasks for a target.
7.2.9 do_package_index
Creates or updates the index in the Package Feeds area.
Note
This task is not triggered with the bitbake -c
command-line option as
are the other tasks in this section. Because this task is specifically for
the package-index
recipe, you run it using bitbake package-index
.
8 devtool
Quick Reference
The devtool
command-line tool provides a number of features that
help you build, test, and package software. This command is available
alongside the bitbake
command. Additionally, the devtool
command
is a key part of the extensible SDK.
This chapter provides a Quick Reference for the devtool
command. For
more information on how to apply the command when using the extensible
SDK, see the “Using the Extensible SDK” chapter in the Yocto
Project Application Development and the Extensible Software Development
Kit (eSDK) manual.
8.1 Getting Help
The devtool
command line is organized similarly to Git in that it
has a number of sub-commands for each function. You can run
devtool --help
to see all the commands:
$ devtool -h
NOTE: Starting bitbake server...
usage: devtool [--basepath BASEPATH] [--bbpath BBPATH] [-d] [-q] [--color COLOR] [-h] <subcommand> ...
OpenEmbedded development tool
options:
--basepath BASEPATH Base directory of SDK / build directory
--bbpath BBPATH Explicitly specify the BBPATH, rather than getting it from the metadata
-d, --debug Enable debug output
-q, --quiet Print only errors
--color COLOR Colorize output (where COLOR is auto, always, never)
-h, --help show this help message and exit
subcommands:
Beginning work on a recipe:
add Add a new recipe
modify Modify the source for an existing recipe
upgrade Upgrade an existing recipe
Getting information:
status Show workspace status
latest-version Report the latest version of an existing recipe
check-upgrade-status Report upgradability for multiple (or all) recipes
search Search available recipes
Working on a recipe in the workspace:
build Build a recipe
rename Rename a recipe file in the workspace
edit-recipe Edit a recipe file
find-recipe Find a recipe file
configure-help Get help on configure script options
update-recipe Apply changes from external source tree to recipe
reset Remove a recipe from your workspace
finish Finish working on a recipe in your workspace
Testing changes on target:
deploy-target Deploy recipe output files to live target machine
undeploy-target Undeploy recipe output files in live target machine
build-image Build image including workspace recipe packages
Advanced:
create-workspace Set up workspace in an alternative location
extract Extract the source for an existing recipe
sync Synchronize the source tree for an existing recipe
menuconfig Alter build-time configuration for a recipe
import Import exported tar archive into workspace
export Export workspace into a tar archive
other:
selftest-reverse Reverse value (for selftest)
pluginfile Print the filename of this plugin
bbdir Print the BBPATH directory of this plugin
count How many times have this plugin been registered.
multiloaded How many times have this plugin been initialized
Use devtool <subcommand> --help to get help on a specific command
As directed in the general help output, you can get more syntax on a specific command by providing the command name and using “–help”:
$ devtool add --help
NOTE: Starting bitbake server...
usage: devtool add [-h] [--same-dir | --no-same-dir] [--fetch URI] [--npm-dev] [--version VERSION] [--no-git] [--srcrev SRCREV | --autorev] [--srcbranch SRCBRANCH] [--binary] [--also-native] [--src-subdir SUBDIR] [--mirrors]
[--provides PROVIDES]
[recipename] [srctree] [fetchuri]
Adds a new recipe to the workspace to build a specified source tree. Can optionally fetch a remote URI and unpack it to create the source tree.
arguments:
recipename Name for new recipe to add (just name - no version, path or extension). If not specified, will attempt to auto-detect it.
srctree Path to external source tree. If not specified, a subdirectory of /media/build1/poky/build/workspace/sources will be used.
fetchuri Fetch the specified URI and extract it to create the source tree
options:
-h, --help show this help message and exit
--same-dir, -s Build in same directory as source
--no-same-dir Force build in a separate build directory
--fetch URI, -f URI Fetch the specified URI and extract it to create the source tree (deprecated - pass as positional argument instead)
--npm-dev For npm, also fetch devDependencies
--version VERSION, -V VERSION
Version to use within recipe (PV)
--no-git, -g If fetching source, do not set up source tree as a git repository
--srcrev SRCREV, -S SRCREV
Source revision to fetch if fetching from an SCM such as git (default latest)
--autorev, -a When fetching from a git repository, set SRCREV in the recipe to a floating revision instead of fixed
--srcbranch SRCBRANCH, -B SRCBRANCH
Branch in source repository if fetching from an SCM such as git (default master)
--binary, -b Treat the source tree as something that should be installed verbatim (no compilation, same directory structure). Useful with binary packages e.g. RPMs.
--also-native Also add native variant (i.e. support building recipe for the build host as well as the target machine)
--src-subdir SUBDIR Specify subdirectory within source tree to use
--mirrors Enable PREMIRRORS and MIRRORS for source tree fetching (disable by default).
--provides PROVIDES, -p PROVIDES
Specify an alias for the item provided by the recipe. E.g. virtual/libgl
8.2 The Workspace Layer Structure
devtool
uses a “Workspace” layer in which to accomplish builds. This
layer is not specific to any single devtool
command but is rather a
common working area used across the tool.
The following figure shows the workspace structure:
attic - A directory created if devtool believes it must preserve
anything when you run "devtool reset". For example, if you
run "devtool add", make changes to the recipe, and then
run "devtool reset", devtool takes notice that the file has
been changed and moves it into the attic should you still
want the recipe.
README - Provides information on what is in workspace layer and how to
manage it.
.devtool_md5 - A checksum file used by devtool.
appends - A directory that contains *.bbappend files, which point to
external source.
conf - A configuration directory that contains the layer.conf file.
recipes - A directory containing recipes. This directory contains a
folder for each directory added whose name matches that of the
added recipe. devtool places the recipe.bb file
within that sub-directory.
sources - A directory containing a working copy of the source files used
when building the recipe. This is the default directory used
as the location of the source tree when you do not provide a
source tree path. This directory contains a folder for each
set of source files matched to a corresponding recipe.
8.3 Adding a New Recipe to the Workspace Layer
Use the devtool add
command to add a new recipe to the workspace
layer. The recipe you add should not exist - devtool
creates it for
you. The source files the recipe uses should exist in an external area.
The following example creates and adds a new recipe named jackson
to
a workspace layer the tool creates. The source code built by the recipes
resides in /home/user/sources/jackson
:
$ devtool add jackson /home/user/sources/jackson
If you add a recipe and the workspace layer does not exist, the command creates the layer and populates it as described in “The Workspace Layer Structure” section.
Running devtool add
when the workspace layer exists causes the tool
to add the recipe, append files, and source files into the existing
workspace layer. The .bbappend
file is created to point to the
external source tree.
Note
If your recipe has runtime dependencies defined, you must be sure that these packages exist on the target hardware before attempting to run your application. If dependent packages (e.g. libraries) do not exist on the target, your application, when run, will fail to find those functions. For more information, see the “Deploying Your Software on the Target Machine” section.
By default, devtool add
uses the latest revision (i.e. master) when
unpacking files from a remote URI. In some cases, you might want to
specify a source revision by branch, tag, or commit hash. You can
specify these options when using the devtool add
command:
To specify a source branch, use the
--srcbranch
option:$ devtool add --srcbranch hardknott jackson /home/user/sources/jackson
In the previous example, you are checking out the hardknott branch.
To specify a specific tag or commit hash, use the
--srcrev
option:$ devtool add --srcrev yocto-3.3.1 jackson /home/user/sources/jackson $ devtool add --srcrev some_commit_hash /home/user/sources/jackson
The previous examples check out the yocto-3.3.1 tag and the commit associated with the some_commit_hash hash.
Note
If you prefer to use the latest revision every time the recipe is
built, use the options --autorev
or -a
.
8.4 Extracting the Source for an Existing Recipe
Use the devtool extract
command to extract the source for an
existing recipe. When you use this command, you must supply the root
name of the recipe (i.e. no version, paths, or extensions), and you must
supply the directory to which you want the source extracted.
Additional command options let you control the name of a development branch into which you can checkout the source and whether or not to keep a temporary directory, which is useful for debugging.
8.5 Synchronizing a Recipe’s Extracted Source Tree
Use the devtool sync
command to synchronize a previously extracted
source tree for an existing recipe. When you use this command, you must
supply the root name of the recipe (i.e. no version, paths, or
extensions), and you must supply the directory to which you want the
source extracted.
Additional command options let you control the name of a development branch into which you can checkout the source and whether or not to keep a temporary directory, which is useful for debugging.
8.6 Modifying an Existing Recipe
Use the devtool modify
command to begin modifying the source of an
existing recipe. This command is very similar to the
add command
except that it does not physically create the recipe in the workspace
layer because the recipe already exists in an another layer.
The devtool modify
command extracts the source for a recipe, sets it
up as a Git repository if the source had not already been fetched from
Git, checks out a branch for development, and applies any patches from
the recipe as commits on top. You can use the following command to
checkout the source files:
$ devtool modify recipe
Using the above command form, devtool
uses the existing recipe’s
SRC_URI statement to locate the upstream source,
extracts the source into the default sources location in the workspace.
The default development branch used is “devtool”.
8.7 Edit an Existing Recipe
Use the devtool edit-recipe
command to run the default editor, which
is identified using the EDITOR
variable, on the specified recipe.
When you use the devtool edit-recipe
command, you must supply the
root name of the recipe (i.e. no version, paths, or extensions). Also,
the recipe file itself must reside in the workspace as a result of the
devtool add
or devtool upgrade
commands. However, you can
override that requirement by using the “-a” or “–any-recipe” option.
Using either of these options allows you to edit any recipe regardless
of its location.
8.8 Updating a Recipe
Use the devtool update-recipe
command to update your recipe with
patches that reflect changes you make to the source files. For example,
if you know you are going to work on some code, you could first use the
devtool modify command to extract
the code and set up the workspace. After which, you could modify,
compile, and test the code.
When you are satisfied with the results and you have committed your
changes to the Git repository, you can then run the
devtool update-recipe
to create the patches and update the recipe:
$ devtool update-recipe recipe
If you run the devtool update-recipe
without committing your changes, the command ignores the changes.
Often, you might want to apply customizations made to your software in
your own layer rather than apply them to the original recipe. If so, you
can use the -a
or --append
option with the
devtool update-recipe
command. These options allow you to specify
the layer into which to write an append file:
$ devtool update-recipe recipe -a base-layer-directory
The *.bbappend
file is created at the
appropriate path within the specified layer directory, which may or may
not be in your bblayers.conf
file. If an append file already exists,
the command updates it appropriately.
8.9 Checking on the Upgrade Status of a Recipe
Upstream recipes change over time. Consequently, you might find that you need to determine if you can upgrade a recipe to a newer version.
To check on the upgrade status of a recipe, use the
devtool check-upgrade-status
command. The command displays a table
of your current recipe versions, the latest upstream versions, the email
address of the recipe’s maintainer, and any additional information such
as commit hash strings and reasons you might not be able to upgrade a
particular recipe.
Note
For the
oe-core
layer, recipe maintainers come from the maintainers.inc file.If the recipe is using the Git Fetcher (git://) rather than a tarball, the commit hash points to the commit that matches the recipe’s latest version tag.
As with all devtool
commands, you can get help on the individual
command:
$ devtool check-upgrade-status -h
NOTE: Starting bitbake server...
usage: devtool check-upgrade-status [-h] [--all] [recipe [recipe ...]]
Prints a table of recipes together with versions currently provided by recipes, and latest upstream versions, when there is a later version available
arguments:
recipe Name of the recipe to report (omit to report upgrade info for all recipes)
options:
-h, --help show this help message and exit
--all, -a Show all recipes, not just recipes needing upgrade
Unless you provide a specific recipe name on the command line, the command checks all recipes in all configured layers.
Following is a partial example table that reports on all the recipes.
Notice the reported reason for not upgrading the base-passwd
recipe.
In this example, while a new version is available upstream, you do not
want to use it because the dependency on cdebconf
is not easily
satisfied.
Note
When a reason for not upgrading displays, the reason is usually
written into the recipe using the RECIPE_NO_UPDATE_REASON
variable. See the
base-passwd.bb
recipe for an example.
$ devtool check-upgrade-status
...
NOTE: acpid 2.0.30 2.0.31 Ross Burton <ross.burton@intel.com>
NOTE: u-boot-fw-utils 2018.11 2019.01 Marek Vasut <marek.vasut@gmail.com> d3689267f92c5956e09cc7d1baa4700141662bff
NOTE: u-boot-tools 2018.11 2019.01 Marek Vasut <marek.vasut@gmail.com> d3689267f92c5956e09cc7d1baa4700141662bff
.
.
.
NOTE: base-passwd 3.5.29 3.5.45 Anuj Mittal <anuj.mittal@intel.com> cannot be updated due to: Version 3.5.38 requires cdebconf for update-passwd utility
NOTE: busybox 1.29.2 1.30.0 Andrej Valek <andrej.valek@siemens.com>
NOTE: dbus-test 1.12.10 1.12.12 Chen Qi <Qi.Chen@windriver.com>
8.10 Upgrading a Recipe
As software matures, upstream recipes are upgraded to newer versions. As
a developer, you need to keep your local recipes up-to-date with the
upstream version releases. Several methods exist by which you can
upgrade recipes. You can read about them in the “Upgrading Recipes”
section of the Yocto Project Development Tasks Manual. This section
overviews the devtool upgrade
command.
Before you upgrade a recipe, you can check on its upgrade status. See the “Checking on the Upgrade Status of a Recipe” section for more information.
The devtool upgrade
command upgrades an existing recipe to a more
recent version of the recipe upstream. The command puts the upgraded
recipe file along with any associated files into a “workspace” and, if
necessary, extracts the source tree to a specified location. During the
upgrade, patches associated with the recipe are rebased or added as
needed.
When you use the devtool upgrade
command, you must supply the root
name of the recipe (i.e. no version, paths, or extensions), and you must
supply the directory to which you want the source extracted. Additional
command options let you control things such as the version number to
which you want to upgrade (i.e. the PV), the source
revision to which you want to upgrade (i.e. the
SRCREV), whether or not to apply patches, and so
forth.
You can read more on the devtool upgrade
workflow in the
“Use devtool upgrade to Create a Version of the Recipe that Supports a Newer Version of the Software”
section in the Yocto Project Application Development and the Extensible
Software Development Kit (eSDK) manual. You can also see an example of
how to use devtool upgrade
in the “Using devtool upgrade”
section in the Yocto Project Development Tasks Manual.
8.11 Resetting a Recipe
Use the devtool reset
command to remove a recipe and its
configuration (e.g. the corresponding .bbappend
file) from the
workspace layer. Realize that this command deletes the recipe and the
append file. The command does not physically move them for you.
Consequently, you must be sure to physically relocate your updated
recipe and the append file outside of the workspace layer before running
the devtool reset
command.
If the devtool reset
command detects that the recipe or the append
files have been modified, the command preserves the modified files in a
separate “attic” subdirectory under the workspace layer.
Here is an example that resets the workspace directory that contains the
mtr
recipe:
$ devtool reset mtr
NOTE: Cleaning sysroot for recipe mtr...
NOTE: Leaving source tree /home/scottrif/poky/build/workspace/sources/mtr as-is; if you no longer need it then please delete it manually
$
8.12 Building Your Recipe
Use the devtool build
command to build your recipe. The
devtool build
command is equivalent to the
bitbake -c populate_sysroot
command.
When you use the devtool build
command, you must supply the root
name of the recipe (i.e. do not provide versions, paths, or extensions).
You can use either the “-s” or the “–disable-parallel-make” options to
disable parallel makes during the build. Here is an example:
$ devtool build recipe
8.13 Building Your Image
Use the devtool build-image
command to build an image, extending it
to include packages from recipes in the workspace. Using this command is
useful when you want an image that ready for immediate deployment onto a
device for testing. For proper integration into a final image, you need
to edit your custom image recipe appropriately.
When you use the devtool build-image
command, you must supply the
name of the image. This command has no command line options:
$ devtool build-image image
8.14 Deploying Your Software on the Target Machine
Use the devtool deploy-target
command to deploy the recipe’s build
output to the live target machine:
$ devtool deploy-target recipe target
The target is the address of the target machine, which must be running
an SSH server (i.e. user@hostname[:destdir]
).
This command deploys all files installed during the do_install task. Furthermore, you do not need to have package management enabled within the target machine. If you do, the package manager is bypassed.
Note
The deploy-target
functionality is for development only. You
should never use it to update an image that will be used in
production.
Some conditions exist that could prevent a deployed application from behaving as expected. When both of the following conditions exist, your application has the potential to not behave correctly when run on the target:
You are deploying a new application to the target and the recipe you used to build the application had correctly defined runtime dependencies.
The target does not physically have the packages on which the application depends installed.
If both of these conditions exist, your application will not behave as
expected. The reason for this misbehavior is because the
devtool deploy-target
command does not deploy the packages (e.g.
libraries) on which your new application depends. The assumption is that
the packages are already on the target. Consequently, when a runtime
call is made in the application for a dependent function (e.g. a library
call), the function cannot be found.
To be sure you have all the dependencies local to the target, you need to be sure that the packages are pre-deployed (installed) on the target before attempting to run your application.
8.15 Removing Your Software from the Target Machine
Use the devtool undeploy-target
command to remove deployed build
output from the target machine. For the devtool undeploy-target
command to work, you must have previously used the
“devtool deploy-target”
command.
$ devtool undeploy-target recipe target
The target is the
address of the target machine, which must be running an SSH server (i.e.
user@hostname
).
8.16 Creating the Workspace Layer in an Alternative Location
Use the devtool create-workspace
command to create a new workspace
layer in your Build Directory. When you create a
new workspace layer, it is populated with the README
file and the
conf
directory only.
The following example creates a new workspace layer in your current working and by default names the workspace layer “workspace”:
$ devtool create-workspace
You can create a workspace layer anywhere by supplying a pathname with the command. The following command creates a new workspace layer named “new-workspace”:
$ devtool create-workspace /home/scottrif/new-workspace
8.17 Get the Status of the Recipes in Your Workspace
Use the devtool status
command to list the recipes currently in your
workspace. Information includes the paths to their respective external
source trees.
The devtool status
command has no command-line options:
$ devtool status
Following is sample output after using
devtool add
to create and add the mtr_0.86.bb
recipe to the workspace
directory:
$ devtool status
mtr:/home/scottrif/poky/build/workspace/sources/mtr (/home/scottrif/poky/build/workspace/recipes/mtr/mtr_0.86.bb)
$
8.18 Search for Available Target Recipes
Use the devtool search
command to search for available target
recipes. The command matches the recipe name, package name, description,
and installed files. The command displays the recipe name as a result of
a match.
When you use the devtool search
command, you must supply a keyword.
The command uses the keyword when searching for a match.
9 OpenEmbedded Kickstart (.wks
) Reference
9.1 Introduction
The current Wic implementation supports only the basic kickstart
partitioning commands: partition
(or part
for short) and
bootloader
.
Note
Future updates will implement more commands and options. If you use anything that is not specifically supported, results can be unpredictable.
This chapter provides a reference on the available kickstart commands. The information lists the commands, their syntax, and meanings. Kickstart commands are based on the Fedora kickstart versions but with modifications to reflect Wic capabilities. You can see the original documentation for those commands at the following link: https://pykickstart.readthedocs.io/en/latest/kickstart-docs.html
9.2 Command: part or partition
Either of these commands creates a partition on the system and uses the following syntax:
part [mntpoint]
partition [mntpoint]
If you do not provide mntpoint, Wic creates a partition but does not mount it.
The mntpoint
is where the partition is mounted and must be in one of
the following forms:
/path
: For example, “/”, “/usr”, or “/home”swap
: The created partition is used as swap space
Specifying a mntpoint causes the partition to automatically be mounted.
Wic achieves this by adding entries to the filesystem table (fstab)
during image generation. In order for Wic to generate a valid fstab, you
must also provide one of the --ondrive
, --ondisk
, or
--use-uuid
partition options as part of the command.
Note
The mount program must understand the PARTUUID syntax you use with
--use-uuid
and non-root mountpoint, including swap. The BusyBox
versions of these application are currently excluded.
Here is an example that uses “/” as the mountpoint. The command uses
--ondisk
to force the partition onto the sdb
disk:
part / --source rootfs --ondisk sdb --fstype=ext3 --label platform --align 1024
Here is a list that describes other supported options you can use with
the part
and partition
commands:
--size
: The minimum partition size in MBytes. Specify an integer value such as 500. Do not append the number with “MB”. You do not need this option if you use--source
.--fixed-size
: The exact partition size in MBytes. You cannot specify with--size
. An error occurs when assembling the disk image if the partition data is larger than--fixed-size
.--source
: This option is a Wic-specific option that names the source of the data that populates the partition. The most common value for this option is “rootfs”, but you can use any value that maps to a valid source plugin. For information on the source plugins, see the “Using the Wic Plugin Interface” section in the Yocto Project Development Tasks Manual.If you use
--source rootfs
, Wic creates a partition as large as needed and fills it with the contents of the root filesystem pointed to by the-r
command-line option or the equivalent rootfs derived from the-e
command-line option. The filesystem type used to create the partition is driven by the value of the--fstype
option specified for the partition. See the entry on--fstype
that follows for more information.If you use
--source plugin-name
, Wic creates a partition as large as needed and fills it with the contents of the partition that is generated by the specified plugin name using the data pointed to by the-r
command-line option or the equivalent rootfs derived from the-e
command-line option. Exactly what those contents are and filesystem type used are dependent on the given plugin implementation.If you do not use the
--source
option, thewic
command creates an empty partition. Consequently, you must use the--size
option to specify the size of the empty partition.--ondisk
or--ondrive
: Forces the partition to be created on a particular disk.--fstype
: Sets the file system type for the partition. Valid values are:ext4
ext3
ext2
btrfs
squashfs
swap
--fsoptions
: Specifies a free-form string of options to be used when mounting the filesystem. This string is copied into the/etc/fstab
file of the installed system and should be enclosed in quotes. If not specified, the default string is “defaults”.--label label
: Specifies the label to give to the filesystem to be made on the partition. If the given label is already in use by another filesystem, a new label is created for the partition.--active
: Marks the partition as active.--align (in KBytes)
: This option is a Wic-specific option that says to start partitions on boundaries given x KBytes.--offset (in KBytes)
: This option is a Wic-specific option that says to place a partition at exactly the specified offset. If the partition cannot be placed at the specified offset, the image build will fail.--no-table
: This option is a Wic-specific option. Using the option reserves space for the partition and causes it to become populated. However, the partition is not added to the partition table.--exclude-path
: This option is a Wic-specific option that excludes the given relative path from the resulting image. This option is only effective with the rootfs source plugin.--extra-space
: This option is a Wic-specific option that adds extra space after the space filled by the content of the partition. The final size can exceed the size specified by the--size
option. The default value is 10 Mbytes.--overhead-factor
: This option is a Wic-specific option that multiplies the size of the partition by the option’s value. You must supply a value greater than or equal to “1”. The default value is “1.3”.--part-name
: This option is a Wic-specific option that specifies a name for GPT partitions.--part-type
: This option is a Wic-specific option that specifies the partition type globally unique identifier (GUID) for GPT partitions. You can find the list of partition type GUIDs at https://en.wikipedia.org/wiki/GUID_Partition_Table#Partition_type_GUIDs.--use-uuid
: This option is a Wic-specific option that causes Wic to generate a random GUID for the partition. The generated identifier is used in the bootloader configuration to specify the root partition.--uuid
: This option is a Wic-specific option that specifies the partition UUID.--fsuuid
: This option is a Wic-specific option that specifies the filesystem UUID. You can generate or modify WKS_FILE with this option if a preconfigured filesystem UUID is added to the kernel command line in the bootloader configuration before you run Wic.--system-id
: This option is a Wic-specific option that specifies the partition system ID, which is a one byte long, hexadecimal parameter with or without the 0x prefix.--mkfs-extraopts
: This option specifies additional options to pass to themkfs
utility. Some default options for certain filesystems do not take effect. See Wic’s help on kickstart (i.e.wic help kickstart
).
9.3 Command: bootloader
This command specifies how the bootloader should be configured and supports the following options:
Note
Bootloader functionality and boot partitions are implemented by the various –source plugins that implement bootloader functionality. The bootloader command essentially provides a means of modifying bootloader configuration.
--timeout
: Specifies the number of seconds before the bootloader times out and boots the default option.--append
: Specifies kernel parameters. These parameters will be added to the syslinuxAPPEND
orgrub
kernel command line.--configfile
: Specifies a user-defined configuration file for the bootloader. You can provide a full pathname for the file or a file that exists in thecanned-wks
folder. This option overrides all other bootloader options.
10 QA Error and Warning Messages
10.1 Introduction
When building a recipe, the OpenEmbedded build system performs various QA checks on the output to ensure that common issues are detected and reported. Sometimes when you create a new recipe to build new software, it will build with no problems. When this is not the case, or when you have QA issues building any software, it could take a little time to resolve them.
While it is tempting to ignore a QA message or even to disable QA checks, it is best to try and resolve any reported QA issues. This chapter provides a list of the QA messages and brief explanations of the issues you could encounter so that you can properly resolve problems.
The next section provides a list of all QA error and warning messages based on a default configuration. Each entry provides the message or error form along with an explanation.
Note
At the end of each message, the name of the associated QA test (as listed in the “insane.bbclass” section) appears within square brackets.
As mentioned, this list of error and warning messages is for QA checks only. The list does not cover all possible build errors or warnings you could encounter.
Because some QA checks are disabled by default, this list does not include all possible QA check errors and warnings.
10.2 Errors and Warnings
<packagename>: <path> is using libexec please relocate to <libexecdir> [libexec]
The specified package contains files in
/usr/libexec
when the distro configuration uses a different path for<libexecdir>
By default,<libexecdir>
is$prefix/libexec
. However, this default can be changed (e.g.${libdir}
).
package <packagename> contains bad RPATH <rpath> in file <file> [rpaths]
The specified binary produced by the recipe contains dynamic library load paths (rpaths) that contain build system paths such as TMPDIR, which are incorrect for the target and could potentially be a security issue. Check for bad
-rpath
options being passed to the linker in your do_compile log. Depending on the build system used by the software being built, there might be a configure option to disable rpath usage completely within the build of the software.
<packagename>: <file> contains probably-redundant RPATH <rpath> [useless-rpaths]
The specified binary produced by the recipe contains dynamic library load paths (rpaths) that on a standard system are searched by default by the linker (e.g.
/lib
and/usr/lib
). While these paths will not cause any breakage, they do waste space and are unnecessary. Depending on the build system used by the software being built, there might be a configure option to disable rpath usage completely within the build of the software.
<packagename> requires <files>, but no providers in its RDEPENDS [file-rdeps]
A file-level dependency has been identified from the specified package on the specified files, but there is no explicit corresponding entry in RDEPENDS. If particular files are required at runtime then
RDEPENDS
should be declared in the recipe to ensure the packages providing them are built.
<packagename1> rdepends on <packagename2>, but it isn't a build dependency? [build-deps]
A runtime dependency exists between the two specified packages, but there is nothing explicit within the recipe to enable the OpenEmbedded build system to ensure that dependency is satisfied. This condition is usually triggered by an RDEPENDS value being added at the packaging stage rather than up front, which is usually automatic based on the contents of the package. In most cases, you should change the recipe to add an explicit
RDEPENDS
for the dependency.
non -dev/-dbg/nativesdk- package contains symlink .so: <packagename> path '<path>' [dev-so]
Symlink
.so
files are for development only, and should therefore go into the-dev
package. This situation might occur if you add*.so*
rather than*.so.*
to a non-dev package. Change FILES (and possibly PACKAGES) such that the specified.so
file goes into an appropriate-dev
package.
non -staticdev package contains static .a library: <packagename> path '<path>' [staticdev]
Static
.a
library files should go into a-staticdev
package. Change FILES (and possibly PACKAGES) such that the specified.a
file goes into an appropriate-staticdev
package.
<packagename>: found library in wrong location [libdir]
The specified file may have been installed into an incorrect (possibly hardcoded) installation path. For example, this test will catch recipes that install
/lib/bar.so
when${base_libdir}
is “lib32”. Another example is when recipes install/usr/lib64/foo.so
when${libdir}
is “/usr/lib”. False positives occasionally exist. For these cases add “libdir” to INSANE_SKIP for the package.
non debug package contains .debug directory: <packagename> path <path> [debug-files]
The specified package contains a
.debug
directory, which should not appear in anything but the-dbg
package. This situation might occur if you add a path which contains a.debug
directory and do not explicitly add the.debug
directory to the-dbg
package. If this is the case, add the.debug
directory explicitly toFILES_${PN}-dbg
. See FILES for additional information onFILES
.
Architecture did not match (<file_arch>, expected <machine_arch>) in <file> [arch]
By default, the OpenEmbedded build system checks the Executable and Linkable Format (ELF) type, bit size, and endianness of any binaries to ensure they match the target architecture. This test fails if any binaries do not match the type since there would be an incompatibility. The test could indicate that the wrong compiler or compiler options have been used. Sometimes software, like bootloaders, might need to bypass this check. If the file you receive the error for is firmware that is not intended to be executed within the target operating system or is intended to run on a separate processor within the device, you can add “arch” to INSANE_SKIP for the package. Another option is to check the do_compile log and verify that the compiler options being used are correct.
Bit size did not match (<file_bits>, expected <machine_bits>) in <file> [arch]
By default, the OpenEmbedded build system checks the Executable and Linkable Format (ELF) type, bit size, and endianness of any binaries to ensure they match the target architecture. This test fails if any binaries do not match the type since there would be an incompatibility. The test could indicate that the wrong compiler or compiler options have been used. Sometimes software, like bootloaders, might need to bypass this check. If the file you receive the error for is firmware that is not intended to be executed within the target operating system or is intended to run on a separate processor within the device, you can add “arch” to INSANE_SKIP for the package. Another option is to check the do_compile log and verify that the compiler options being used are correct.
Endianness did not match (<file_endianness>, expected <machine_endianness>) in <file> [arch]
By default, the OpenEmbedded build system checks the Executable and Linkable Format (ELF) type, bit size, and endianness of any binaries to ensure they match the target architecture. This test fails if any binaries do not match the type since there would be an incompatibility. The test could indicate that the wrong compiler or compiler options have been used. Sometimes software, like bootloaders, might need to bypass this check. If the file you receive the error for is firmware that is not intended to be executed within the target operating system or is intended to run on a separate processor within the device, you can add “arch” to INSANE_SKIP for the package. Another option is to check the do_compile log and verify that the compiler options being used are correct.
ELF binary '<file>' has relocations in .text [textrel]
The specified ELF binary contains relocations in its
.text
sections. This situation can result in a performance impact at runtime.Typically, the way to solve this performance issue is to add “-fPIC” or “-fpic” to the compiler command-line options. For example, given software that reads CFLAGS when you build it, you could add the following to your recipe:
CFLAGS_append = " -fPIC "
For more information on text relocations at runtime, see https://www.akkadia.org/drepper/textrelocs.html.
File '<file>' in package '<package>' doesn't have GNU_HASH (didn't pass LDFLAGS?) [ldflags]
This indicates that binaries produced when building the recipe have not been linked with the LDFLAGS options provided by the build system. Check to be sure that the
LDFLAGS
variable is being passed to the linker command. A common workaround for this situation is to pass inLDFLAGS
using TARGET_CC_ARCH within the recipe as follows:TARGET_CC_ARCH += "${LDFLAGS}"
Package <packagename> contains Xorg driver (<driver>) but no xorg-abi- dependencies [xorg-driver-abi]
The specified package contains an Xorg driver, but does not have a corresponding ABI package dependency. The xserver-xorg recipe provides driver ABI names. All drivers should depend on the ABI versions that they have been built against. Driver recipes that include
xorg-driver-input.inc
orxorg-driver-video.inc
will automatically get these versions. Consequently, you should only need to explicitly add dependencies to binary driver recipes.
The /usr/share/info/dir file is not meant to be shipped in a particular package. [infodir]
The
/usr/share/info/dir
should not be packaged. Add the following line to your do_install task or to yourdo_install_append
within the recipe as follows:rm ${D}${infodir}/dir
Symlink <path> in <packagename> points to TMPDIR [symlink-to-sysroot]
The specified symlink points into TMPDIR on the host. Such symlinks will work on the host. However, they are clearly invalid when running on the target. You should either correct the symlink to use a relative path or remove the symlink.
<file> failed sanity test (workdir) in path <path> [la]
The specified
.la
file contains TMPDIR paths. Any.la
file containing these paths is incorrect sincelibtool
adds the correct sysroot prefix when using the files automatically itself.
<file> failed sanity test (tmpdir) in path <path> [pkgconfig]
The specified
.pc
file contains TMPDIR/
WORKDIR paths. Any.pc
file containing these paths is incorrect sincepkg-config
itself adds the correct sysroot prefix when the files are accessed.
<packagename> rdepends on <debug_packagename> [debug-deps]
A dependency exists between the specified non-dbg package (i.e. a package whose name does not end in
-dbg
) and a package that is adbg
package. Thedbg
packages contain debug symbols and are brought in using several different methods:Using the
dbg-pkgs
IMAGE_FEATURES value.Using IMAGE_INSTALL.
As a dependency of another
dbg
package that was brought in using one of the above methods.
The dependency might have been automatically added because the
dbg
package erroneously contains files that it should not contain (e.g. a non-symlink.so
file) or it might have been added manually (e.g. by adding to RDEPENDS).
<packagename> rdepends on <dev_packagename> [dev-deps]
A dependency exists between the specified non-dev package (a package whose name does not end in
-dev
) and a package that is adev
package. Thedev
packages contain development headers and are usually brought in using several different methods:Using the
dev-pkgs
IMAGE_FEATURES value.Using IMAGE_INSTALL.
As a dependency of another
dev
package that was brought in using one of the above methods.
The dependency might have been automatically added (because the
dev
package erroneously contains files that it should not have (e.g. a non-symlink.so
file) or it might have been added manually (e.g. by adding to RDEPENDS).
<var>_<packagename> is invalid: <comparison> (<value>) only comparisons <, =, >, <=, and >= are allowed [dep-cmp]
If you are adding a versioned dependency relationship to one of the dependency variables (RDEPENDS, RRECOMMENDS, RSUGGESTS, RPROVIDES, RREPLACES, or RCONFLICTS), you must only use the named comparison operators. Change the versioned dependency values you are adding to match those listed in the message.
<recipename>: The compile log indicates that host include and/or library paths were used. Please check the log '<logfile>' for more information. [compile-host-path]
The log for the do_compile task indicates that paths on the host were searched for files, which is not appropriate when cross-compiling. Look for “is unsafe for cross-compilation” or “CROSS COMPILE Badness” in the specified log file.
<recipename>: The install log indicates that host include and/or library paths were used. Please check the log '<logfile>' for more information. [install-host-path]
The log for the do_install task indicates that paths on the host were searched for files, which is not appropriate when cross-compiling. Look for “is unsafe for cross-compilation” or “CROSS COMPILE Badness” in the specified log file.
This autoconf log indicates errors, it looked at host include and/or library paths while determining system capabilities. Rerun configure task after fixing this. [configure-unsafe]
The log for the do_configure task indicates that paths on the host were searched for files, which is not appropriate when cross-compiling. Look for “is unsafe for cross-compilation” or “CROSS COMPILE Badness” in the specified log file.
<packagename> doesn't match the [a-z0-9.+-]+ regex [pkgname]
The convention within the OpenEmbedded build system (sometimes enforced by the package manager itself) is to require that package names are all lower case and to allow a restricted set of characters. If your recipe name does not match this, or you add packages to PACKAGES that do not conform to the convention, then you will receive this error. Rename your recipe. Or, if you have added a non-conforming package name to
PACKAGES
, change the package name appropriately.
<recipe>: configure was passed unrecognized options: <options> [unknown-configure-option]
The configure script is reporting that the specified options are unrecognized. This situation could be because the options were previously valid but have been removed from the configure script. Or, there was a mistake when the options were added and there is another option that should be used instead. If you are unsure, consult the upstream build documentation, the
./configure --help
output, and the upstream change log or release notes. Once you have worked out what the appropriate change is, you can update EXTRA_OECONF, PACKAGECONFIG_CONFARGS, or the individual PACKAGECONFIG option values accordingly.
Recipe <recipefile> has PN of "<recipename>" which is in OVERRIDES, this can result in unexpected behavior. [pn-overrides]
The specified recipe has a name (PN) value that appears in OVERRIDES. If a recipe is named such that its
PN
value matches something already inOVERRIDES
(e.g.PN
happens to be the same as MACHINE or DISTRO), it can have unexpected consequences. For example, assignments such asFILES_${PN} = "xyz"
effectively turn intoFILES = "xyz"
. Rename your recipe (or ifPN
is being set explicitly, change thePN
value) so that the conflict does not occur. See FILES for additional information.
<recipefile>: Variable <variable> is set as not being package specific, please fix this. [pkgvarcheck]
Certain variables (RDEPENDS, RRECOMMENDS, RSUGGESTS, RCONFLICTS, RPROVIDES, RREPLACES, FILES,
pkg_preinst
,pkg_postinst
,pkg_prerm
,pkg_postrm
, and ALLOW_EMPTY) should always be set specific to a package (i.e. they should be set with a package name override such asRDEPENDS_${PN} = "value"
rather thanRDEPENDS = "value"
). If you receive this error, correct any assignments to these variables within your recipe.recipe uses DEPENDS_${PN}, should use DEPENDS [pkgvarcheck]
This check looks for instances of setting
DEPENDS_${PN}
which is erroneous (DEPENDS is a recipe-wide variable and thus it is not correct to specify it for a particular package, nor will such an assignment actually work.) SetDEPENDS
instead.
File '<file>' from <recipename> was already stripped, this will prevent future debugging! [already-stripped]
Produced binaries have already been stripped prior to the build system extracting debug symbols. It is common for upstream software projects to default to stripping debug symbols for output binaries. In order for debugging to work on the target using
-dbg
packages, this stripping must be disabled.Depending on the build system used by the software being built, disabling this stripping could be as easy as specifying an additional configure option. If not, disabling stripping might involve patching the build scripts. In the latter case, look for references to “strip” or “STRIP”, or the “-s” or “-S” command-line options being specified on the linker command line (possibly through the compiler command line if preceded with “-Wl,”).
Note
Disabling stripping here does not mean that the final packaged binaries will be unstripped. Once the OpenEmbedded build system splits out debug symbols to the
-dbg
package, it will then strip the symbols from the binaries.
<packagename> is listed in PACKAGES multiple times, this leads to packaging errors. [packages-list]
Package names must appear only once in the PACKAGES variable. You might receive this error if you are attempting to add a package to
PACKAGES
that is already in the variable’s value.
FILES variable for package <packagename> contains '//' which is invalid. Attempting to fix this but you should correct the metadata. [files-invalid]
The string “//” is invalid in a Unix path. Correct all occurrences where this string appears in a FILES variable so that there is only a single “/”.
<recipename>: Files/directories were installed but not shipped in any package [installed-vs-shipped]
Files have been installed within the do_install task but have not been included in any package by way of the FILES variable. Files that do not appear in any package cannot be present in an image later on in the build process. You need to do one of the following:
Add the files to
FILES
for the package you want them to appear in (e.g.FILES_${
PN}
for the main package).Delete the files at the end of the
do_install
task if the files are not needed in any package.
<oldpackage>-<oldpkgversion> was registered as shlib provider for <library>, changing it to <newpackage>-<newpkgversion> because it was built later
This message means that both
<oldpackage>
and<newpackage>
provide the specified shared library. You can expect this message when a recipe has been renamed. However, if that is not the case, the message might indicate that a private version of a library is being erroneously picked up as the provider for a common library. If that is the case, you should add the library’s.so
file name to PRIVATE_LIBS in the recipe that provides the private version of the library.
LICENSE_<packagename> includes licenses (<licenses>) that are not listed in LICENSE [unlisted-pkg-lics]
The LICENSE of the recipe should be a superset of all the licenses of all packages produced by this recipe. In other words, any license in
LICENSE_*
should also appear in LICENSE.
AM_GNU_GETTEXT used but no inherit gettext [configure-gettext]
If a recipe is building something that uses automake and the automake files contain an
AM_GNU_GETTEXT
directive then this check will fail if there is noinherit gettext
statement in the recipe to ensure that gettext is available during the build. Addinherit gettext
to remove the warning.
package contains mime types but does not inherit mime: <packagename> path '<file>' [mime]
The specified package contains mime type files (
.xml
files in${datadir}/mime/packages
) and yet does not inherit the mime class which will ensure that these get properly installed. Either addinherit mime
to the recipe or remove the files at thedo_install
step if they are not needed.
package contains desktop file with key 'MimeType' but does not inhert mime-xdg: <packagename> path '<file>' [mime-xdg]
The specified package contains a .desktop file with a ‘MimeType’ key present, but does not inherit the mime-xdg class that is required in order for that to be activated. Either add
inherit mime
to the recipe or remove the files at thedo_install
step if they are not needed.
<recipename>: SRC_URI uses unstable GitHub archives [src-uri-bad]
GitHub provides “archive” tarballs, however these can be re-generated on the fly and thus the file’s signature will not necessarily match that in the SRC_URI checksums in future leading to build failures. It is recommended that you use an official release tarball or switch to pulling the corresponding revision in the actual git repository instead.
SRC_URI uses PN not BPN [src-uri-bad]
If some part of SRC_URI needs to reference the recipe name, it should do so using ${BPN} rather than ${PN} as the latter will change for different variants of the same recipe e.g. when BBCLASSEXTEND or multilib are being used. This check will fail if a reference to
${PN}
is found within theSRC_URI
value - change it to${BPN}
instead.
<recipename>: recipe doesn't inherit features_check [unhandled-features-check]
This check ensures that if one of the variables that the features_check class supports (e.g. REQUIRED_DISTRO_FEATURES) is used, then the recipe inherits
features_check
in order for the requirement to actually work. If you are seeing this message, either addinherit features_check
to your recipe or remove the reference to the variable if it is not needed.
<recipename>: recipe defines ALTERNATIVE_<packagename> but doesn't inherit update-alternatives. This might fail during do_rootfs later! [missing-update-alternatives]
This check ensures that if a recipe sets the ALTERNATIVE variable that the recipe also inherits update-alternatives such that the alternative will be correctly set up. If you are seeing this message, either add
inherit update-alternatives
to your recipe or remove the reference to the variable if it is not needed.
<packagename>: <file> maximum shebang size exceeded, the maximum size is 128. [shebang-size]
This check ensures that the shebang line (
#!
in the first line) for a script is not longer than 128 characters, which can cause an error at runtime depending on the operating system. If you are seeing this message then the specified script may need to be patched to have a shorter in order to avoid runtime problems.
<packagename> contains perllocal.pod (<files>), should not be installed [perllocalpod]
perllocal.pod
is an index file of locally installed modules and so shouldn’t be installed by any distribution packages. The cpan class already setsNO_PERLLOCAL
to stop this file being generated by most Perl recipes, but if a recipe is usingMakeMaker
directly then they might not be doing this correctly. This check ensures that perllocal.pod is not in any package in order to avoid multiple packages shipping this file and thus their packages conflicting if installed together.
<packagename> package is not obeying usrmerge distro feature. /<path> should be relocated to /usr. [usrmerge]
If
usrmerge
is in DISTRO_FEATURES, this check will ensure that no package installs files to root (/bin
,/sbin
,/lib
,/lib64
) directories. If you are seeing this message, it indicates that thedo_install
step (or perhaps the build process thatdo_install
is calling into, e.g.make install
is using hardcoded paths instead of the variables set up for this (bindir
,sbindir
, etc.), and should be changed so that it does.
Fuzz detected: <patch output> [patch-fuzz]
This check looks for evidence of “fuzz” when applying patches within the
do_patch
task. Patch fuzz is a situation when thepatch
tool ignores some of the context lines in order to apply the patch. Consider this example:Patch to be applied:
--- filename +++ filename context line 1 context line 2 context line 3 +newly added line context line 4 context line 5 context line 6
Original source code:
different context line 1 different context line 2 context line 3 context line 4 different context line 5 different context line 6
Outcome (after applying patch with fuzz):
different context line 1 different context line 2 context line 3 newly added line context line 4 different context line 5 different context line 6
Chances are, the newly added line was actually added in a completely wrong location, or it was already in the original source and was added for the second time. This is especially possible if the context line 3 and 4 are blank or have only generic things in them, such as
#endif
or}
. Depending on the patched code, it is entirely possible for an incorrectly patched file to still compile without errors.How to eliminate patch fuzz warnings
Use the
devtool
command as explained by the warning. First, unpack the source into devtool workspace:devtool modify <recipe>
This will apply all of the patches, and create new commits out of them in the workspace - with the patch context updated.
Then, replace the patches in the recipe layer:
devtool finish --force-patch-refresh <recipe> <layer_path>
The patch updates then need be reviewed (preferably with a side-by-side diff tool) to ensure they are indeed doing the right thing i.e.:
they are applied in the correct location within the file;
they do not introduce duplicate lines, or otherwise do things that are no longer necessary.
To confirm these things, you can also review the patched source code in devtool’s workspace, typically in
<build_dir>/workspace/sources/<recipe>/
Once the review is done, you can create and publish a layer commit with the patch updates that modify the context. Devtool may also refresh other things in the patches, those can be discarded.
10.3 Configuring and Disabling QA Checks
You can configure the QA checks globally so that specific check failures either raise a warning or an error message, using the WARN_QA and ERROR_QA variables, respectively. You can also disable checks within a particular recipe using INSANE_SKIP. For information on how to work with the QA checks, see the “insane.bbclass” section.
Note
Please keep in mind that the QA checks exist in order to detect real or potential problems in the packaged output. So exercise caution when disabling these checks.
11 Images
The OpenEmbedded build system provides several example images to satisfy
different needs. When you issue the bitbake
command you provide a
“top-level” recipe that essentially begins the build for the type of
image you want.
Note
Building an image without GNU General Public License Version 3
(GPLv3), GNU Lesser General Public License Version 3 (LGPLv3), and
the GNU Affero General Public License Version 3 (AGPL-3.0) components
is only supported for minimal and base images. Furthermore, if you
are going to build an image using non-GPLv3 and similarly licensed
components, you must make the following changes in the local.conf
file before using the BitBake command to build the minimal or base
image:
1. Comment out the EXTRA_IMAGE_FEATURES line
2. Set INCOMPATIBLE_LICENSE = "GPL-3.0 LGPL-3.0 AGPL-3.0"
From within the poky
Git repository, you can use the following
command to display the list of directories within the Source Directory
that contain image recipe files:
$ ls meta*/recipes*/images/*.bb
Following is a list of supported recipes:
build-appliance-image
: An example virtual machine that contains all the pieces required to run builds using the build system as well as the build system itself. You can boot and run the image using either the VMware Player or VMware Workstation. For more information on this image, see the Build Appliance page on the Yocto Project website.core-image-base
: A console-only image that fully supports the target device hardware.core-image-clutter
: An image with support for the Open GL-based toolkit Clutter, which enables development of rich and animated graphical user interfaces.core-image-full-cmdline
: A console-only image with more full-featured Linux system functionality installed.core-image-lsb
: An image that conforms to the Linux Standard Base (LSB) specification. This image requires a distribution configuration that enables LSB compliance (e.g.poky-lsb
). If you buildcore-image-lsb
without that configuration, the image will not be LSB-compliant.core-image-lsb-dev
: Acore-image-lsb
image that is suitable for development work using the host. The image includes headers and libraries you can use in a host development environment. This image requires a distribution configuration that enables LSB compliance (e.g.poky-lsb
). If you buildcore-image-lsb-dev
without that configuration, the image will not be LSB-compliant.core-image-lsb-sdk
: Acore-image-lsb
that includes everything in the cross-toolchain but also includes development headers and libraries to form a complete standalone SDK. This image requires a distribution configuration that enables LSB compliance (e.g.poky-lsb
). If you buildcore-image-lsb-sdk
without that configuration, the image will not be LSB-compliant. This image is suitable for development using the target.core-image-minimal
: A small image just capable of allowing a device to boot.core-image-minimal-dev
: Acore-image-minimal
image suitable for development work using the host. The image includes headers and libraries you can use in a host development environment.core-image-minimal-initramfs
: Acore-image-minimal
image that has the Minimal RAM-based Initial Root Filesystem (initramfs) as part of the kernel, which allows the system to find the first “init” program more efficiently. See the PACKAGE_INSTALL variable for additional information helpful when working with initramfs images.core-image-minimal-mtdutils
: Acore-image-minimal
image that has support for the Minimal MTD Utilities, which let the user interact with the MTD subsystem in the kernel to perform operations on flash devices.core-image-rt
: Acore-image-minimal
image plus a real-time test suite and tools appropriate for real-time use.core-image-rt-sdk
: Acore-image-rt
image that includes everything in the cross-toolchain. The image also includes development headers and libraries to form a complete stand-alone SDK and is suitable for development using the target.core-image-sato
: An image with Sato support, a mobile environment and visual style that works well with mobile devices. The image supports X11 with a Sato theme and applications such as a terminal, editor, file manager, media player, and so forth.core-image-sato-dev
: Acore-image-sato
image suitable for development using the host. The image includes libraries needed to build applications on the device itself, testing and profiling tools, and debug symbols. This image was formerlycore-image-sdk
.core-image-sato-sdk
: Acore-image-sato
image that includes everything in the cross-toolchain. The image also includes development headers and libraries to form a complete standalone SDK and is suitable for development using the target.core-image-testmaster
: A “master” image designed to be used for automated runtime testing. Provides a “known good” image that is deployed to a separate partition so that you can boot into it and use it to deploy a second image to be tested. You can find more information about runtime testing in the “Performing Automated Runtime Testing” section in the Yocto Project Development Tasks Manual.core-image-testmaster-initramfs
: A RAM-based Initial Root Filesystem (initramfs) image tailored for use with thecore-image-testmaster
image.core-image-weston
: A very basic Wayland image with a terminal. This image provides the Wayland protocol libraries and the reference Weston compositor. For more information, see the “Using Wayland and Weston” section in the Yocto Project Development Tasks Manual.core-image-x11
: A very basic X11 image with a terminal.
12 Features
This chapter provides a reference of shipped machine and distro features you can include as part of your image, a reference on image features you can select, and a reference on feature backfilling.
Features provide a mechanism for working out which packages should be
included in the generated images. Distributions can select which
features they want to support through the DISTRO_FEATURES
variable,
which is set or appended to in a distribution’s configuration file such
as poky.conf
, poky-tiny.conf
, poky-lsb.conf
and so forth.
Machine features are set in the MACHINE_FEATURES
variable, which is
set in the machine configuration file and specifies the hardware
features for a given machine.
These two variables combine to work out which kernel modules, utilities, and other packages to include. A given distribution can support a selected subset of features so some machine features might not be included if the distribution itself does not support them.
One method you can use to determine which recipes are checking to see if
a particular feature is contained or not is to grep
through the
Metadata for the feature. Here is an example that
discovers the recipes whose build is potentially changed based on a
given feature:
$ cd poky
$ git grep 'contains.*MACHINE_FEATURES.*feature'
12.1 Machine Features
The items below are features you can use with MACHINE_FEATURES. Features do not have a one-to-one correspondence to packages, and they can go beyond simply controlling the installation of a package or packages. Sometimes a feature can influence how certain recipes are built. For example, a feature might determine whether a particular configure option is specified within the do_configure task for a particular recipe.
This feature list only represents features as shipped with the Yocto Project metadata:
acpi: Hardware has ACPI (x86/x86_64 only)
alsa: Hardware has ALSA audio drivers
apm: Hardware uses APM (or APM emulation)
bluetooth: Hardware has integrated BT
efi: Support for booting through EFI
ext2: Hardware HDD or Microdrive
keyboard: Hardware has a keyboard
pcbios: Support for booting through BIOS
pci: Hardware has a PCI bus
pcmcia: Hardware has PCMCIA or CompactFlash sockets
phone: Mobile phone (voice) support
qvga: Machine has a QVGA (320x240) display
rtc: Machine has a Real-Time Clock
screen: Hardware has a screen
serial: Hardware has serial support (usually RS232)
touchscreen: Hardware has a touchscreen
usbgadget: Hardware is USB gadget device capable
usbhost: Hardware is USB Host capable
vfat: FAT file system support
wifi: Hardware has integrated WiFi
12.2 Distro Features
The items below are features you can use with DISTRO_FEATURES to enable features across your distribution. Features do not have a one-to-one correspondence to packages, and they can go beyond simply controlling the installation of a package or packages. In most cases, the presence or absence of a feature translates to the appropriate option supplied to the configure script during the do_configure task for the recipes that optionally support the feature.
Some distro features are also machine features. These select features make sense to be controlled both at the machine and distribution configuration level. See the COMBINED_FEATURES variable for more information.
This list only represents features as shipped with the Yocto Project metadata:
alsa: Include ALSA support (OSS compatibility kernel modules installed if available).
api-documentation: Enables generation of API documentation during recipe builds. The resulting documentation is added to SDK tarballs when the
bitbake -c populate_sdk
command is used. See the “Adding API Documentation to the Standard SDK” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.bluetooth: Include bluetooth support (integrated BT only).
cramfs: Include CramFS support.
directfb: Include DirectFB support.
ext2: Include tools for supporting for devices with internal HDD/Microdrive for storing files (instead of Flash only devices).
ipsec: Include IPSec support.
ipv6: Include IPv6 support.
keyboard: Include keyboard support (e.g. keymaps will be loaded during boot).
ldconfig: Include support for ldconfig and
ld.so.conf
on the target.nfs: Include NFS client support (for mounting NFS exports on device).
opengl: Include the Open Graphics Library, which is a cross-language, multi-platform application programming interface used for rendering two and three-dimensional graphics.
pci: Include PCI bus support.
pcmcia: Include PCMCIA/CompactFlash support.
ppp: Include PPP dialup support.
ptest: Enables building the package tests where supported by individual recipes. For more information on package tests, see the “Testing Packages With ptest” section in the Yocto Project Development Tasks Manual.
smbfs: Include SMB networks client support (for mounting Samba/Microsoft Windows shares on device).
systemd: Include support for this
init
manager, which is a full replacement of forinit
with parallel starting of services, reduced shell overhead, and other features. Thisinit
manager is used by many distributions.usbgadget: Include USB Gadget Device support (for USB networking/serial/storage).
usbhost: Include USB Host support (allows to connect external keyboard, mouse, storage, network etc).
usrmerge: Merges the
/bin
,/sbin
,/lib
, and/lib64
directories into their respective counterparts in the/usr
directory to provide better package and application compatibility.wayland: Include the Wayland display server protocol and the library that supports it.
wifi: Include WiFi support (integrated only).
x11: Include the X server and libraries.
12.3 Image Features
The contents of images generated by the OpenEmbedded build system can be controlled by the IMAGE_FEATURES and EXTRA_IMAGE_FEATURES variables that you typically configure in your image recipes. Through these variables, you can add several different predefined packages such as development utilities or packages with debug information needed to investigate application problems or profile applications.
The following image features are available for all images:
allow-empty-password: Allows Dropbear and OpenSSH to accept root logins and logins from accounts having an empty password string.
dbg-pkgs: Installs debug symbol packages for all packages installed in a given image.
debug-tweaks: Makes an image suitable for development (e.g. allows root logins without passwords and enables post-installation logging). See the ‘allow-empty-password’, ‘empty-root-password’, and ‘post-install-logging’ features in this list for additional information.
dev-pkgs: Installs development packages (headers and extra library links) for all packages installed in a given image.
doc-pkgs: Installs documentation packages for all packages installed in a given image.
empty-root-password: Sets the root password to an empty string, which allows logins with a blank password.
package-management: Installs package management tools and preserves the package manager database.
post-install-logging: Enables logging postinstall script runs to the
/var/log/postinstall.log
file on first boot of the image on the target system.Note
To make the
/var/log
directory on the target persistent, use the VOLATILE_LOG_DIR variable by setting it to “no”.ptest-pkgs: Installs ptest packages for all ptest-enabled recipes.
read-only-rootfs: Creates an image whose root filesystem is read-only. See the “Creating a Read-Only Root Filesystem” section in the Yocto Project Development Tasks Manual for more information.
splash: Enables showing a splash screen during boot. By default, this screen is provided by
psplash
, which does allow customization. If you prefer to use an alternative splash screen package, you can do so by setting theSPLASH
variable to a different package name (or names) within the image recipe or at the distro configuration level.staticdev-pkgs: Installs static development packages, which are static libraries (i.e.
*.a
files), for all packages installed in a given image.
Some image features are available only when you inherit the core-image class. The current list of these valid features is as follows:
hwcodecs: Installs hardware acceleration codecs.
nfs-server: Installs an NFS server.
perf: Installs profiling tools such as
perf
,systemtap
, andLTTng
. For general information on user-space tools, see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.ssh-server-dropbear: Installs the Dropbear minimal SSH server.
ssh-server-openssh: Installs the OpenSSH SSH server, which is more full-featured than Dropbear. Note that if both the OpenSSH SSH server and the Dropbear minimal SSH server are present in
IMAGE_FEATURES
, then OpenSSH will take precedence and Dropbear will not be installed.tools-debug: Installs debugging tools such as
strace
andgdb
. For information on GDB, see the “Debugging With the GNU Project Debugger (GDB) Remotely” section in the Yocto Project Development Tasks Manual. For information on tracing and profiling, see the Yocto Project Profiling and Tracing Manual.tools-sdk: Installs a full SDK that runs on the device.
tools-testapps: Installs device testing tools (e.g. touchscreen debugging).
x11: Installs the X server.
x11-base: Installs the X server with a minimal environment.
x11-sato: Installs the OpenedHand Sato environment.
12.4 Feature Backfilling
Sometimes it is necessary in the OpenEmbedded build system to extend
MACHINE_FEATURES or
DISTRO_FEATURES to control functionality
that was previously enabled and not able to be disabled. For these
cases, we need to add an additional feature item to appear in one of
these variables, but we do not want to force developers who have
existing values of the variables in their configuration to add the new
feature in order to retain the same overall level of functionality.
Thus, the OpenEmbedded build system has a mechanism to automatically
“backfill” these added features into existing distro or machine
configurations. You can see the list of features for which this is done
by finding the
DISTRO_FEATURES_BACKFILL and
MACHINE_FEATURES_BACKFILL
variables in the meta/conf/bitbake.conf
file.
Because such features are backfilled by default into all configurations as described in the previous paragraph, developers who wish to disable the new features need to be able to selectively prevent the backfilling from occurring. They can do this by adding the undesired feature or features to the DISTRO_FEATURES_BACKFILL_CONSIDERED or MACHINE_FEATURES_BACKFILL_CONSIDERED variables for distro features and machine features respectively.
Here are two examples to help illustrate feature backfilling:
The “pulseaudio” distro feature option: Previously, PulseAudio support was enabled within the Qt and GStreamer frameworks. Because of this, the feature is backfilled and thus enabled for all distros through the
DISTRO_FEATURES_BACKFILL
variable in themeta/conf/bitbake.conf
file. However, your distro needs to disable the feature. You can disable the feature without affecting other existing distro configurations that need PulseAudio support by adding “pulseaudio” toDISTRO_FEATURES_BACKFILL_CONSIDERED
in your distro’s.conf
file. Adding the feature to this variable when it also exists in theDISTRO_FEATURES_BACKFILL
variable prevents the build system from adding the feature to your configuration’sDISTRO_FEATURES
, effectively disabling the feature for that particular distro.The “rtc” machine feature option: Previously, real time clock (RTC) support was enabled for all target devices. Because of this, the feature is backfilled and thus enabled for all machines through the
MACHINE_FEATURES_BACKFILL
variable in themeta/conf/bitbake.conf
file. However, your target device does not have this capability. You can disable RTC support for your device without affecting other machines that need RTC support by adding the feature to your machine’sMACHINE_FEATURES_BACKFILL_CONSIDERED
list in the machine’s.conf
file. Adding the feature to this variable when it also exists in theMACHINE_FEATURES_BACKFILL
variable prevents the build system from adding the feature to your configuration’sMACHINE_FEATURES
, effectively disabling RTC support for that particular machine.
13 Variables Glossary
This chapter lists common variables used in the OpenEmbedded build system and gives an overview of their function and contents.
A B C D E F G H I K L M N O P R S T U V W X
- ABIEXTENSION
Extension to the Application Binary Interface (ABI) field of the GNU canonical architecture name (e.g. “eabi”).
ABI extensions are set in the machine include files. For example, the
meta/conf/machine/include/arm/arch-arm.inc
file sets the following extension:ABIEXTENSION = "eabi"
- ALLOW_EMPTY
Specifies whether to produce an output package even if it is empty. By default, BitBake does not produce empty packages. This default behavior can cause issues when there is an RDEPENDS or some other hard runtime requirement on the existence of the package.
Like all package-controlling variables, you must always use them in conjunction with a package name override, as in:
ALLOW_EMPTY_${PN} = "1" ALLOW_EMPTY_${PN}-dev = "1" ALLOW_EMPTY_${PN}-staticdev = "1"
- ALTERNATIVE
Lists commands in a package that need an alternative binary naming scheme. Sometimes the same command is provided in multiple packages. When this occurs, the OpenEmbedded build system needs to use the alternatives system to create a different binary naming scheme so the commands can co-exist.
To use the variable, list out the package’s commands that also exist as part of another package. For example, if the
busybox
package has four commands that also exist as part of another package, you identify them as follows:ALTERNATIVE_busybox = "sh sed test bracket"
For more information on the alternatives system, see the “update-alternatives.bbclass” section.
- ALTERNATIVE_LINK_NAME
Used by the alternatives system to map duplicated commands to actual locations. For example, if the
bracket
command provided by thebusybox
package is duplicated through another package, you must use theALTERNATIVE_LINK_NAME
variable to specify the actual location:ALTERNATIVE_LINK_NAME[bracket] = "/usr/bin/["
In this example, the binary for the
bracket
command (i.e.[
) from thebusybox
package resides in/usr/bin/
.Note
If
ALTERNATIVE_LINK_NAME
is not defined, it defaults to${bindir}/name
.For more information on the alternatives system, see the “update-alternatives.bbclass” section.
- ALTERNATIVE_PRIORITY
Used by the alternatives system to create default priorities for duplicated commands. You can use the variable to create a single default regardless of the command name or package, a default for specific duplicated commands regardless of the package, or a default for specific commands tied to particular packages. Here are the available syntax forms:
ALTERNATIVE_PRIORITY = "priority" ALTERNATIVE_PRIORITY[name] = "priority" ALTERNATIVE_PRIORITY_pkg[name] = "priority"
For more information on the alternatives system, see the “update-alternatives.bbclass” section.
- ALTERNATIVE_TARGET
Used by the alternatives system to create default link locations for duplicated commands. You can use the variable to create a single default location for all duplicated commands regardless of the command name or package, a default for specific duplicated commands regardless of the package, or a default for specific commands tied to particular packages. Here are the available syntax forms:
ALTERNATIVE_TARGET = "target" ALTERNATIVE_TARGET[name] = "target" ALTERNATIVE_TARGET_pkg[name] = "target"
Note
If
ALTERNATIVE_TARGET
is not defined, it inherits the value from the ALTERNATIVE_LINK_NAME variable.If
ALTERNATIVE_LINK_NAME
andALTERNATIVE_TARGET
are the same, the target forALTERNATIVE_TARGET
has “.{BPN}
” appended to it.Finally, if the file referenced has not been renamed, the alternatives system will rename it to avoid the need to rename alternative files in the do_install task while retaining support for the command if necessary.
For more information on the alternatives system, see the “update-alternatives.bbclass” section.
- ANY_OF_DISTRO_FEATURES
When inheriting the features_check class, this variable identifies a list of distribution features where at least one must be enabled in the current configuration in order for the OpenEmbedded build system to build the recipe. In other words, if none of the features listed in
ANY_OF_DISTRO_FEATURES
appear inDISTRO_FEATURES
within the current configuration, then the recipe will be skipped, and if the build system attempts to build the recipe then an error will be triggered.- APPEND
An override list of append strings for each target specified with LABELS.
See the grub-efi class for more information on how this variable is used.
- AR
The minimal command and arguments used to run
ar
.- ARCHIVER_MODE
When used with the archiver class, determines the type of information used to create a released archive. You can use this variable to create archives of patched source, original source, configured source, and so forth by employing the following variable flags (varflags):
ARCHIVER_MODE[src] = "original" # Uses original (unpacked) source files. ARCHIVER_MODE[src] = "patched" # Uses patched source files. This is the default. ARCHIVER_MODE[src] = "configured" # Uses configured source files. ARCHIVER_MODE[diff] = "1" # Uses patches between do_unpack and do_patch. ARCHIVER_MODE[diff-exclude] ?= "file file ..." # Lists files and directories to exclude from diff. ARCHIVER_MODE[dumpdata] = "1" # Uses environment data. ARCHIVER_MODE[recipe] = "1" # Uses recipe and include files. ARCHIVER_MODE[srpm] = "1" # Uses RPM package files.
For information on how the variable works, see the
meta/classes/archiver.bbclass
file in the Source Directory.- AS
Minimal command and arguments needed to run the assembler.
- ASSUME_PROVIDED
Lists recipe names (PN values) BitBake does not attempt to build. Instead, BitBake assumes these recipes have already been built.
In OpenEmbedded-Core,
ASSUME_PROVIDED
mostly specifies native tools that should not be built. An example isgit-native
, which when specified, allows for the Git binary from the host to be used rather than buildinggit-native
.- ASSUME_SHLIBS
Provides additional
shlibs
provider mapping information, which adds to or overwrites the information provided automatically by the system. Separate multiple entries using spaces.As an example, use the following form to add an
shlib
provider of shlibname in packagename with the optional version:shlibname:packagename[_version]
Here is an example that adds a shared library named
libEGL.so.1
as being provided by thelibegl-implementation
package:ASSUME_SHLIBS = "libEGL.so.1:libegl-implementation"
- AUTHOR
The email address used to contact the original author or authors in order to send patches and forward bugs.
- AUTO_LIBNAME_PKGS
When the debian class is inherited, which is the default behavior,
AUTO_LIBNAME_PKGS
specifies which packages should be checked for libraries and renamed according to Debian library package naming.The default value is “${PACKAGES}”, which causes the debian class to act on all packages that are explicitly generated by the recipe.
- AUTO_SYSLINUXMENU
Enables creating an automatic menu for the syslinux bootloader. You must set this variable in your recipe. The syslinux class checks this variable.
- AUTOREV
When
SRCREV
is set to the value of this variable, it specifies to use the latest source revision in the repository. Here is an example:SRCREV = "${AUTOREV}"
If you use the previous statement to retrieve the latest version of software, you need to be sure PV contains
${
SRCPV}
. For example, suppose you have a kernel recipe that inherits the kernel class and you use the previous statement. In this example,${SRCPV}
does not automatically get intoPV
. Consequently, you need to changePV
in your recipe so that it does contain${SRCPV}
.For more information see the “Automatically Incrementing a Package Version Number” section in the Yocto Project Development Tasks Manual.
- AVAILABLE_LICENSES
List of licenses found in the directories specified by COMMON_LICENSE_DIR and LICENSE_PATH.
Note
It is assumed that all changes to
COMMON_LICENSE_DIR
andLICENSE_PATH
have been done beforeAVAILABLE_LICENSES
is defined (in license.bbclass).- AVAILTUNES
The list of defined CPU and Application Binary Interface (ABI) tunings (i.e. “tunes”) available for use by the OpenEmbedded build system.
The list simply presents the tunes that are available. Not all tunes may be compatible with a particular machine configuration, or with each other in a Multilib configuration.
To add a tune to the list, be sure to append it with spaces using the “+=” BitBake operator. Do not simply replace the list by using the “=” operator. See the “Basic Syntax” section in the BitBake User Manual for more information.
- AZ_SAS
Azure Storage Shared Access Signature, when using the Azure Storage fetcher (az://) This variable can be defined to be used by the fetcher to authenticate and gain access to non-public artifacts.
AZ_SAS = ""se=2021-01-01&sp=r&sv=2018-11-09&sr=c&skoid=<skoid>&sig=<signature>""
For more information see Microsoft’s Azure Storage documentation at https://docs.microsoft.com/en-us/azure/storage/common/storage-sas-overview
- B
The directory within the Build Directory in which the OpenEmbedded build system places generated objects during a recipe’s build process. By default, this directory is the same as the S directory, which is defined as:
S = "${WORKDIR}/${BP}"
You can separate the (
S
) directory and the directory pointed to by theB
variable. Most Autotools-based recipes support separating these directories. The build system defaults to using separate directories forgcc
and some kernel recipes.- BAD_RECOMMENDATIONS
Lists “recommended-only” packages to not install. Recommended-only packages are packages installed only through the RRECOMMENDS variable. You can prevent any of these “recommended” packages from being installed by listing them with the
BAD_RECOMMENDATIONS
variable:BAD_RECOMMENDATIONS = "package_name package_name package_name ..."
You can set this variable globally in your
local.conf
file or you can attach it to a specific image recipe by using the recipe name override:BAD_RECOMMENDATIONS_pn-target_image = "package_name"
It is important to realize that if you choose to not install packages using this variable and some other packages are dependent on them (i.e. listed in a recipe’s RDEPENDS variable), the OpenEmbedded build system ignores your request and will install the packages to avoid dependency errors.
Support for this variable exists only when using the IPK and RPM packaging backend. Support does not exist for DEB.
See the NO_RECOMMENDATIONS and the PACKAGE_EXCLUDE variables for related information.
- BASE_LIB
The library directory name for the CPU or Application Binary Interface (ABI) tune. The
BASE_LIB
applies only in the Multilib context. See the “Combining Multiple Versions of Library Files into One Image” section in the Yocto Project Development Tasks Manual for information on Multilib.The
BASE_LIB
variable is defined in the machine include files in the Source Directory. If Multilib is not being used, the value defaults to “lib”.- BASE_WORKDIR
Points to the base of the work directory for all recipes. The default value is “${TMPDIR}/work”.
- BB_ALLOWED_NETWORKS
Specifies a space-delimited list of hosts that the fetcher is allowed to use to obtain the required source code. Following are considerations surrounding this variable:
This host list is only used if
BB_NO_NETWORK
is either not set or set to “0”.Limited support for wildcard matching against the beginning of host names exists. For example, the following setting matches
git.gnu.org
,ftp.gnu.org
, andfoo.git.gnu.org
.BB_ALLOWED_NETWORKS = "*.gnu.org"
Note
The use of the “
*
” character only works at the beginning of a host name and it must be isolated from the remainder of the host name. You cannot use the wildcard character in any other location of the name or combined with the front part of the name.For example,
*.foo.bar
is supported, while*aa.foo.bar
is not.Mirrors not in the host list are skipped and logged in debug.
Attempts to access networks not in the host list cause a failure.
Using
BB_ALLOWED_NETWORKS
in conjunction with PREMIRRORS is very useful. Adding the host you want to use toPREMIRRORS
results in the source code being fetched from an allowed location and avoids raising an error when a host that is not allowed is in a SRC_URI statement. This is because the fetcher does not attempt to use the host listed inSRC_URI
after a successful fetch from thePREMIRRORS
occurs.- BB_DANGLINGAPPENDS_WARNONLY
Defines how BitBake handles situations where an append file (
.bbappend
) has no corresponding recipe file (.bb
). This condition often occurs when layers get out of sync (e.g.oe-core
bumps a recipe version and the old recipe no longer exists and the other layer has not been updated to the new version of the recipe yet).The default fatal behavior is safest because it is the sane reaction given something is out of sync. It is important to realize when your changes are no longer being applied.
You can change the default behavior by setting this variable to “1”, “yes”, or “true” in your
local.conf
file, which is located in the Build Directory: Here is an example:BB_DANGLINGAPPENDS_WARNONLY = "1"
- BB_DISKMON_DIRS
Monitors disk space and available inodes during the build and allows you to control the build based on these parameters.
Disk space monitoring is disabled by default. To enable monitoring, add the
BB_DISKMON_DIRS
variable to yourconf/local.conf
file found in the Build Directory. Use the following form:BB_DISKMON_DIRS = "action,dir,threshold [...]" where: action is: ABORT: Immediately abort the build when a threshold is broken. STOPTASKS: Stop the build after the currently executing tasks have finished when a threshold is broken. WARN: Issue a warning but continue the build when a threshold is broken. Subsequent warnings are issued as defined by the BB_DISKMON_WARNINTERVAL variable, which must be defined in the conf/local.conf file. dir is: Any directory you choose. You can specify one or more directories to monitor by separating the groupings with a space. If two directories are on the same device, only the first directory is monitored. threshold is: Either the minimum available disk space, the minimum number of free inodes, or both. You must specify at least one. To omit one or the other, simply omit the value. Specify the threshold using G, M, K for Gbytes, Mbytes, and Kbytes, respectively. If you do not specify G, M, or K, Kbytes is assumed by default. Do not use GB, MB, or KB.
Here are some examples:
BB_DISKMON_DIRS = "ABORT,${TMPDIR},1G,100K WARN,${SSTATE_DIR},1G,100K" BB_DISKMON_DIRS = "STOPTASKS,${TMPDIR},1G" BB_DISKMON_DIRS = "ABORT,${TMPDIR},,100K"
The first example works only if you also provide the BB_DISKMON_WARNINTERVAL variable in the
conf/local.conf
. This example causes the build system to immediately abort when either the disk space in${TMPDIR}
drops below 1 Gbyte or the available free inodes drops below 100 Kbytes. Because two directories are provided with the variable, the build system also issue a warning when the disk space in the${SSTATE_DIR}
directory drops below 1 Gbyte or the number of free inodes drops below 100 Kbytes. Subsequent warnings are issued during intervals as defined by theBB_DISKMON_WARNINTERVAL
variable.The second example stops the build after all currently executing tasks complete when the minimum disk space in the
${TMPDIR}
directory drops below 1 Gbyte. No disk monitoring occurs for the free inodes in this case.The final example immediately aborts the build when the number of free inodes in the
${TMPDIR}
directory drops below 100 Kbytes. No disk space monitoring for the directory itself occurs in this case.- BB_DISKMON_WARNINTERVAL
Defines the disk space and free inode warning intervals. To set these intervals, define the variable in your
conf/local.conf
file in the Build Directory.If you are going to use the
BB_DISKMON_WARNINTERVAL
variable, you must also use the BB_DISKMON_DIRS variable and define its action as “WARN”. During the build, subsequent warnings are issued each time disk space or number of free inodes further reduces by the respective interval.If you do not provide a
BB_DISKMON_WARNINTERVAL
variable and you do useBB_DISKMON_DIRS
with the “WARN” action, the disk monitoring interval defaults to the following:BB_DISKMON_WARNINTERVAL = "50M,5K"
When specifying the variable in your configuration file, use the following form:
BB_DISKMON_WARNINTERVAL = "disk_space_interval,disk_inode_interval" where: disk_space_interval is: An interval of memory expressed in either G, M, or K for Gbytes, Mbytes, or Kbytes, respectively. You cannot use GB, MB, or KB. disk_inode_interval is: An interval of free inodes expressed in either G, M, or K for Gbytes, Mbytes, or Kbytes, respectively. You cannot use GB, MB, or KB.
Here is an example:
BB_DISKMON_DIRS = "WARN,${SSTATE_DIR},1G,100K" BB_DISKMON_WARNINTERVAL = "50M,5K"
These variables cause the OpenEmbedded build system to issue subsequent warnings each time the available disk space further reduces by 50 Mbytes or the number of free inodes further reduces by 5 Kbytes in the
${SSTATE_DIR}
directory. Subsequent warnings based on the interval occur each time a respective interval is reached beyond the initial warning (i.e. 1 Gbytes and 100 Kbytes).- BB_GENERATE_MIRROR_TARBALLS
Causes tarballs of the source control repositories (e.g. Git repositories), including metadata, to be placed in the DL_DIR directory.
For performance reasons, creating and placing tarballs of these repositories is not the default action by the OpenEmbedded build system.
BB_GENERATE_MIRROR_TARBALLS = "1"
Set this variable in your
local.conf
file in the Build Directory.Once you have the tarballs containing your source files, you can clean up your
DL_DIR
directory by deleting any Git or other source control work directories.- BB_NUMBER_THREADS
The maximum number of tasks BitBake should run in parallel at any one time. The OpenEmbedded build system automatically configures this variable to be equal to the number of cores on the build system. For example, a system with a dual core processor that also uses hyper-threading causes the
BB_NUMBER_THREADS
variable to default to “4”.For single socket systems (i.e. one CPU), you should not have to override this variable to gain optimal parallelism during builds. However, if you have very large systems that employ multiple physical CPUs, you might want to make sure the
BB_NUMBER_THREADS
variable is not set higher than “20”.For more information on speeding up builds, see the “Speeding Up a Build” section in the Yocto Project Development Tasks Manual.
- BB_SERVER_TIMEOUT
Specifies the time (in seconds) after which to unload the BitBake server due to inactivity. Set
BB_SERVER_TIMEOUT
to determine how long the BitBake server stays resident between invocations.For example, the following statement in your
local.conf
file instructs the server to be unloaded after 20 seconds of inactivity:BB_SERVER_TIMEOUT = "20"
If you want the server to never be unloaded, set
BB_SERVER_TIMEOUT
to “-1”.- BBCLASSEXTEND
Allows you to extend a recipe so that it builds variants of the software. Common variants for recipes exist such as “natives” like
quilt-native
, which is a copy of Quilt built to run on the build system; “crosses” such asgcc-cross
, which is a compiler built to run on the build machine but produces binaries that run on the target MACHINE; “nativesdk”, which targets the SDK machine instead ofMACHINE
; and “mulitlibs” in the form “multilib:
multilib_name”.To build a different variant of the recipe with a minimal amount of code, it usually is as simple as adding the following to your recipe:
BBCLASSEXTEND =+ "native nativesdk" BBCLASSEXTEND =+ "multilib:multilib_name"
Note
Internally, the
BBCLASSEXTEND
mechanism generates recipe variants by rewriting variable values and applying overrides such as_class-native
. For example, to generate a native version of a recipe, a DEPENDS on “foo” is rewritten to aDEPENDS
on “foo-native”.Even when using
BBCLASSEXTEND
, the recipe is only parsed once. Parsing once adds some limitations. For example, it is not possible to include a different file depending on the variant, sinceinclude
statements are processed when the recipe is parsed.- BBFILE_COLLECTIONS
Lists the names of configured layers. These names are used to find the other
BBFILE_*
variables. Typically, each layer will append its name to this variable in itsconf/layer.conf
file.- BBFILE_PATTERN
Variable that expands to match files from BBFILES in a particular layer. This variable is used in the
conf/layer.conf
file and must be suffixed with the name of the specific layer (e.g.BBFILE_PATTERN_emenlow
).- BBFILE_PRIORITY
Assigns the priority for recipe files in each layer.
This variable is useful in situations where the same recipe appears in more than one layer. Setting this variable allows you to prioritize a layer against other layers that contain the same recipe - effectively letting you control the precedence for the multiple layers. The precedence established through this variable stands regardless of a recipe’s version (PV variable). For example, a layer that has a recipe with a higher
PV
value but for which theBBFILE_PRIORITY
is set to have a lower precedence still has a lower precedence.A larger value for the
BBFILE_PRIORITY
variable results in a higher precedence. For example, the value 6 has a higher precedence than the value 5. If not specified, theBBFILE_PRIORITY
variable is set based on layer dependencies (see theLAYERDEPENDS
variable for more information. The default priority, if unspecified for a layer with no dependencies, is the lowest defined priority + 1 (or 1 if no priorities are defined).Tip
You can use the command
bitbake-layers show-layers
to list all configured layers along with their priorities.- BBFILES
A space-separated list of recipe files BitBake uses to build software.
When specifying recipe files, you can pattern match using Python’s glob syntax. For details on the syntax, see the documentation by following the previous link.
- BBFILES_DYNAMIC
Activates content when identified layers are present. You identify the layers by the collections that the layers define.
Use the
BBFILES_DYNAMIC
variable to avoid.bbappend
files whose corresponding.bb
file is in a layer that attempts to modify other layers through.bbappend
but does not want to introduce a hard dependency on those other layers.Use the following form for
BBFILES_DYNAMIC
: collection_name:filename_pattern The following example identifies two collection names and two filename patterns:BBFILES_DYNAMIC += " \ clang-layer:${LAYERDIR}/bbappends/meta-clang/*/*/*.bbappend \ core:${LAYERDIR}/bbappends/openembedded-core/meta/*/*/*.bbappend \ "
This next example shows an error message that occurs because invalid entries are found, which cause parsing to abort:
ERROR: BBFILES_DYNAMIC entries must be of the form <collection name>:<filename pattern>, not: /work/my-layer/bbappends/meta-security-isafw/*/*/*.bbappend /work/my-layer/bbappends/openembedded-core/meta/*/*/*.bbappend
- BBINCLUDELOGS
Variable that controls how BitBake displays logs on build failure.
- BBINCLUDELOGS_LINES
If BBINCLUDELOGS is set, specifies the maximum number of lines from the task log file to print when reporting a failed task. If you do not set
BBINCLUDELOGS_LINES
, the entire log is printed.- BBLAYERS
Lists the layers to enable during the build. This variable is defined in the
bblayers.conf
configuration file in the Build Directory. Here is an example:BBLAYERS = " \ /home/scottrif/poky/meta \ /home/scottrif/poky/meta-poky \ /home/scottrif/poky/meta-yocto-bsp \ /home/scottrif/poky/meta-mykernel \ "
This example enables four layers, one of which is a custom, user-defined layer named
meta-mykernel
.- BBMASK
Prevents BitBake from processing recipes and recipe append files.
You can use the
BBMASK
variable to “hide” these.bb
and.bbappend
files. BitBake ignores any recipe or recipe append files that match any of the expressions. It is as if BitBake does not see them at all. Consequently, matching files are not parsed or otherwise used by BitBake.The values you provide are passed to Python’s regular expression compiler. Consequently, the syntax follows Python’s Regular Expression (re) syntax. The expressions are compared against the full paths to the files. For complete syntax information, see Python’s documentation at https://docs.python.org/3/library/re.html#regular-expression-syntax.
The following example uses a complete regular expression to tell BitBake to ignore all recipe and recipe append files in the
meta-ti/recipes-misc/
directory:BBMASK = "meta-ti/recipes-misc/"
If you want to mask out multiple directories or recipes, you can specify multiple regular expression fragments. This next example masks out multiple directories and individual recipes:
BBMASK += "/meta-ti/recipes-misc/ meta-ti/recipes-ti/packagegroup/" BBMASK += "/meta-oe/recipes-support/" BBMASK += "/meta-foo/.*/openldap" BBMASK += "opencv.*\.bbappend" BBMASK += "lzma"
Note
When specifying a directory name, use the trailing slash character to ensure you match just that directory name.
- BBMULTICONFIG
Specifies each additional separate configuration when you are building targets with multiple configurations. Use this variable in your
conf/local.conf
configuration file. Specify a multiconfigname for each configuration file you are using. For example, the following line specifies three configuration files:BBMULTICONFIG = "configA configB configC"
Each configuration file you use must reside in the Build Directory
conf/multiconfig
directory (e.g. build_directory/conf/multiconfig/configA.conf
).For information on how to use
BBMULTICONFIG
in an environment that supports building targets with multiple configurations, see the “Building Images for Multiple Targets Using Multiple Configurations” section in the Yocto Project Development Tasks Manual.- BBPATH
Used by BitBake to locate
.bbclass
and configuration files. This variable is analogous to thePATH
variable.Note
If you run BitBake from a directory outside of the Build Directory, you must be sure to set
BBPATH
to point to the Build Directory. Set the variable as you would any environment variable and then run BitBake:$ BBPATH = "build_directory" $ export BBPATH $ bitbake target
- BBSERVER
If defined in the BitBake environment,
BBSERVER
points to the BitBake remote server.Use the following format to export the variable to the BitBake environment:
export BBSERVER=localhost:$port
By default,
BBSERVER
also appears in BB_HASHBASE_WHITELIST. Consequently,BBSERVER
is excluded from checksum and dependency data.- BINCONFIG
When inheriting the binconfig-disabled class, this variable specifies binary configuration scripts to disable in favor of using
pkg-config
to query the information. Thebinconfig-disabled
class will modify the specified scripts to return an error so that calls to them can be easily found and replaced.To add multiple scripts, separate them by spaces. Here is an example from the
libpng
recipe:BINCONFIG = "${bindir}/libpng-config ${bindir}/libpng16-config"
- BINCONFIG_GLOB
When inheriting the binconfig class, this variable specifies a wildcard for configuration scripts that need editing. The scripts are edited to correct any paths that have been set up during compilation so that they are correct for use when installed into the sysroot and called by the build processes of other recipes.
Note
The
BINCONFIG_GLOB
variable uses shell globbing, which is recognition and expansion of wildcards during pattern matching. Shell globbing is very similar to fnmatch and glob.For more information on how this variable works, see
meta/classes/binconfig.bbclass
in the Source Directory. You can also find general information on the class in the “binconfig.bbclass” section.- BP
The base recipe name and version but without any special recipe name suffix (i.e.
-native
,lib64-
, and so forth).BP
is comprised of the following:${BPN}-${PV}
- BPN
This variable is a version of the PN variable with common prefixes and suffixes removed, such as
nativesdk-
,-cross
,-native
, and multilib’slib64-
andlib32-
. The exact lists of prefixes and suffixes removed are specified by the MLPREFIX and SPECIAL_PKGSUFFIX variables, respectively.- BUGTRACKER
Specifies a URL for an upstream bug tracking website for a recipe. The OpenEmbedded build system does not use this variable. Rather, the variable is a useful pointer in case a bug in the software being built needs to be manually reported.
- BUILD_ARCH
Specifies the architecture of the build host (e.g.
i686
). The OpenEmbedded build system sets the value ofBUILD_ARCH
from the machine name reported by theuname
command.- BUILD_AS_ARCH
Specifies the architecture-specific assembler flags for the build host. By default, the value of
BUILD_AS_ARCH
is empty.- BUILD_CC_ARCH
Specifies the architecture-specific C compiler flags for the build host. By default, the value of
BUILD_CC_ARCH
is empty.- BUILD_CCLD
Specifies the linker command to be used for the build host when the C compiler is being used as the linker. By default,
BUILD_CCLD
points to GCC and passes as arguments the value of BUILD_CC_ARCH, assumingBUILD_CC_ARCH
is set.- BUILD_CFLAGS
Specifies the flags to pass to the C compiler when building for the build host. When building in the
-native
context, CFLAGS is set to the value of this variable by default.- BUILD_CPPFLAGS
Specifies the flags to pass to the C preprocessor (i.e. to both the C and the C++ compilers) when building for the build host. When building in the
-native
context, CPPFLAGS is set to the value of this variable by default.- BUILD_CXXFLAGS
Specifies the flags to pass to the C++ compiler when building for the build host. When building in the
-native
context, CXXFLAGS is set to the value of this variable by default.- BUILD_FC
Specifies the Fortran compiler command for the build host. By default,
BUILD_FC
points to Gfortran and passes as arguments the value of BUILD_CC_ARCH, assumingBUILD_CC_ARCH
is set.- BUILD_LD
Specifies the linker command for the build host. By default,
BUILD_LD
points to the GNU linker (ld) and passes as arguments the value of BUILD_LD_ARCH, assumingBUILD_LD_ARCH
is set.- BUILD_LD_ARCH
Specifies architecture-specific linker flags for the build host. By default, the value of
BUILD_LD_ARCH
is empty.- BUILD_LDFLAGS
Specifies the flags to pass to the linker when building for the build host. When building in the
-native
context, LDFLAGS is set to the value of this variable by default.- BUILD_OPTIMIZATION
Specifies the optimization flags passed to the C compiler when building for the build host or the SDK. The flags are passed through the BUILD_CFLAGS and BUILDSDK_CFLAGS default values.
The default value of the
BUILD_OPTIMIZATION
variable is “-O2 -pipe”.- BUILD_OS
Specifies the operating system in use on the build host (e.g. “linux”). The OpenEmbedded build system sets the value of
BUILD_OS
from the OS reported by theuname
command - the first word, converted to lower-case characters.- BUILD_PREFIX
The toolchain binary prefix used for native recipes. The OpenEmbedded build system uses the
BUILD_PREFIX
value to set the TARGET_PREFIX when building fornative
recipes.- BUILD_STRIP
Specifies the command to be used to strip debugging symbols from binaries produced for the build host. By default,
BUILD_STRIP
points to${
BUILD_PREFIX}strip
.- BUILD_SYS
Specifies the system, including the architecture and the operating system, to use when building for the build host (i.e. when building
native
recipes).The OpenEmbedded build system automatically sets this variable based on BUILD_ARCH, BUILD_VENDOR, and BUILD_OS. You do not need to set the
BUILD_SYS
variable yourself.- BUILD_VENDOR
Specifies the vendor name to use when building for the build host. The default value is an empty string (“”).
- BUILDDIR
Points to the location of the Build Directory. You can define this directory indirectly through the oe-init-build-env script by passing in a Build Directory path when you run the script. If you run the script and do not provide a Build Directory path, the
BUILDDIR
defaults tobuild
in the current directory.- BUILDHISTORY_COMMIT
When inheriting the buildhistory class, this variable specifies whether or not to commit the build history output in a local Git repository. If set to “1”, this local repository will be maintained automatically by the
buildhistory
class and a commit will be created on every build for changes to each top-level subdirectory of the build history output (images, packages, and sdk). If you want to track changes to build history over time, you should set this value to “1”.By default, the
buildhistory
class does not commit the build history output in a local Git repository:BUILDHISTORY_COMMIT ?= "0"
- BUILDHISTORY_COMMIT_AUTHOR
When inheriting the buildhistory class, this variable specifies the author to use for each Git commit. In order for the
BUILDHISTORY_COMMIT_AUTHOR
variable to work, the BUILDHISTORY_COMMIT variable must be set to “1”.Git requires that the value you provide for the
BUILDHISTORY_COMMIT_AUTHOR
variable takes the form of “name email@host”. Providing an email address or host that is not valid does not produce an error.By default, the
buildhistory
class sets the variable as follows:BUILDHISTORY_COMMIT_AUTHOR ?= "buildhistory <buildhistory@${DISTRO}>"
- BUILDHISTORY_DIR
When inheriting the buildhistory class, this variable specifies the directory in which build history information is kept. For more information on how the variable works, see the
buildhistory.class
.By default, the
buildhistory
class sets the directory as follows:BUILDHISTORY_DIR ?= "${TOPDIR}/buildhistory"
- BUILDHISTORY_FEATURES
When inheriting the buildhistory class, this variable specifies the build history features to be enabled. For more information on how build history works, see the “Maintaining Build Output Quality” section in the Yocto Project Development Tasks Manual.
You can specify these features in the form of a space-separated list:
image: Analysis of the contents of images, which includes the list of installed packages among other things.
package: Analysis of the contents of individual packages.
sdk: Analysis of the contents of the software development kit (SDK).
task: Save output file signatures for shared state (sstate) tasks. This saves one file per task and lists the SHA-256 checksums for each file staged (i.e. the output of the task).
By default, the
buildhistory
class enables the following features:BUILDHISTORY_FEATURES ?= "image package sdk"
- BUILDHISTORY_IMAGE_FILES
When inheriting the buildhistory class, this variable specifies a list of paths to files copied from the image contents into the build history directory under an “image-files” directory in the directory for the image, so that you can track the contents of each file. The default is to copy
/etc/passwd
and/etc/group
, which allows you to monitor for changes in user and group entries. You can modify the list to include any file. Specifying an invalid path does not produce an error. Consequently, you can include files that might not always be present.By default, the
buildhistory
class provides paths to the following files:BUILDHISTORY_IMAGE_FILES ?= "/etc/passwd /etc/group"
- BUILDHISTORY_PUSH_REPO
When inheriting the buildhistory class, this variable optionally specifies a remote repository to which build history pushes Git changes. In order for
BUILDHISTORY_PUSH_REPO
to work, BUILDHISTORY_COMMIT must be set to “1”.The repository should correspond to a remote address that specifies a repository as understood by Git, or alternatively to a remote name that you have set up manually using
git remote
within the local repository.By default, the
buildhistory
class sets the variable as follows:BUILDHISTORY_PUSH_REPO ?= ""
- BUILDSDK_CFLAGS
Specifies the flags to pass to the C compiler when building for the SDK. When building in the
nativesdk-
context, CFLAGS is set to the value of this variable by default.- BUILDSDK_CPPFLAGS
Specifies the flags to pass to the C pre-processor (i.e. to both the C and the C++ compilers) when building for the SDK. When building in the
nativesdk-
context, CPPFLAGS is set to the value of this variable by default.- BUILDSDK_CXXFLAGS
Specifies the flags to pass to the C++ compiler when building for the SDK. When building in the
nativesdk-
context, CXXFLAGS is set to the value of this variable by default.- BUILDSDK_LDFLAGS
Specifies the flags to pass to the linker when building for the SDK. When building in the
nativesdk-
context, LDFLAGS is set to the value of this variable by default.- BUILDSTATS_BASE
Points to the location of the directory that holds build statistics when you use and enable the buildstats class. The
BUILDSTATS_BASE
directory defaults to${
TMPDIR}/buildstats/
.- BUSYBOX_SPLIT_SUID
For the BusyBox recipe, specifies whether to split the output executable file into two parts: one for features that require
setuid root
, and one for the remaining features (i.e. those that do not requiresetuid root
).The
BUSYBOX_SPLIT_SUID
variable defaults to “1”, which results in splitting the output executable file. Set the variable to “0” to get a single output executable file.- CACHE
Specifies the directory BitBake uses to store a cache of the Metadata so it does not need to be parsed every time BitBake is started.
- CC
The minimal command and arguments used to run the C compiler.
- CFLAGS
Specifies the flags to pass to the C compiler. This variable is exported to an environment variable and thus made visible to the software being built during the compilation step.
Default initialization for
CFLAGS
varies depending on what is being built:TARGET_CFLAGS when building for the target
BUILD_CFLAGS when building for the build host (i.e.
-native
)BUILDSDK_CFLAGS when building for an SDK (i.e.
nativesdk-
)
- CLASSOVERRIDE
An internal variable specifying the special class override that should currently apply (e.g. “class-target”, “class-native”, and so forth). The classes that use this variable (e.g. native, nativesdk, and so forth) set the variable to appropriate values.
Note
CLASSOVERRIDE
gets its default “class-target” value from thebitbake.conf
file.As an example, the following override allows you to install extra files, but only when building for the target:
do_install_append_class-target() { install my-extra-file ${D}${sysconfdir} }
Here is an example where
FOO
is set to “native” when building for the build host, and to “other” when not building for the build host:FOO_class-native = "native" FOO = "other"
The underlying mechanism behind
CLASSOVERRIDE
is simply that it is included in the default value of OVERRIDES.- CLEANBROKEN
If set to “1” within a recipe,
CLEANBROKEN
specifies that themake clean
command does not work for the software being built. Consequently, the OpenEmbedded build system will not try to runmake clean
during the do_configure task, which is the default behavior.- COMBINED_FEATURES
Provides a list of hardware features that are enabled in both MACHINE_FEATURES and DISTRO_FEATURES. This select list of features contains features that make sense to be controlled both at the machine and distribution configuration level. For example, the “bluetooth” feature requires hardware support but should also be optional at the distribution level, in case the hardware supports Bluetooth but you do not ever intend to use it.
- COMMON_LICENSE_DIR
Points to
meta/files/common-licenses
in the Source Directory, which is where generic license files reside.- COMPATIBLE_HOST
A regular expression that resolves to one or more hosts (when the recipe is native) or one or more targets (when the recipe is non-native) with which a recipe is compatible. The regular expression is matched against HOST_SYS. You can use the variable to stop recipes from being built for classes of systems with which the recipes are not compatible. Stopping these builds is particularly useful with kernels. The variable also helps to increase parsing speed since the build system skips parsing recipes not compatible with the current system.
- COMPATIBLE_MACHINE
A regular expression that resolves to one or more target machines with which a recipe is compatible. The regular expression is matched against MACHINEOVERRIDES. You can use the variable to stop recipes from being built for machines with which the recipes are not compatible. Stopping these builds is particularly useful with kernels. The variable also helps to increase parsing speed since the build system skips parsing recipes not compatible with the current machine.
- COMPLEMENTARY_GLOB
Defines wildcards to match when installing a list of complementary packages for all the packages explicitly (or implicitly) installed in an image.
Note
The
COMPLEMENTARY_GLOB
variable uses Unix filename pattern matching (fnmatch), which is similar to the Unix style pathname pattern expansion (glob).The resulting list of complementary packages is associated with an item that can be added to IMAGE_FEATURES. An example usage of this is the “dev-pkgs” item that when added to
IMAGE_FEATURES
will install -dev packages (containing headers and other development files) for every package in the image.To add a new feature item pointing to a wildcard, use a variable flag to specify the feature item name and use the value to specify the wildcard. Here is an example:
COMPLEMENTARY_GLOB[dev-pkgs] = '*-dev'
- COMPONENTS_DIR
Stores sysroot components for each recipe. The OpenEmbedded build system uses
COMPONENTS_DIR
when constructing recipe-specific sysroots for other recipes.The default is “
${
STAGING_DIR}-components
.” (i.e. “${
TMPDIR}/sysroots-components
”).- CONF_VERSION
Tracks the version of the local configuration file (i.e.
local.conf
). The value forCONF_VERSION
increments each timebuild/conf/
compatibility changes.- CONFFILES
Identifies editable or configurable files that are part of a package. If the Package Management System (PMS) is being used to update packages on the target system, it is possible that configuration files you have changed after the original installation and that you now want to remain unchanged are overwritten. In other words, editable files might exist in the package that you do not want reset as part of the package update process. You can use the
CONFFILES
variable to list the files in the package that you wish to prevent the PMS from overwriting during this update process.To use the
CONFFILES
variable, provide a package name override that identifies the resulting package. Then, provide a space-separated list of files. Here is an example:CONFFILES_${PN} += "${sysconfdir}/file1 \ ${sysconfdir}/file2 ${sysconfdir}/file3"
A relationship exists between the
CONFFILES
andFILES
variables. The files listed withinCONFFILES
must be a subset of the files listed withinFILES
. Because the configuration files you provide withCONFFILES
are simply being identified so that the PMS will not overwrite them, it makes sense that the files must already be included as part of the package through theFILES
variable.Note
When specifying paths as part of the
CONFFILES
variable, it is good practice to use appropriate path variables. For example,${sysconfdir}
rather than/etc
or${bindir}
rather than/usr/bin
. You can find a list of these variables at the top of themeta/conf/bitbake.conf
file in the Source Directory.- CONFIG_INITRAMFS_SOURCE
Identifies the initial RAM filesystem (initramfs) source files. The OpenEmbedded build system receives and uses this kernel Kconfig variable as an environment variable. By default, the variable is set to null (“”).
The
CONFIG_INITRAMFS_SOURCE
can be either a single cpio archive with a.cpio
suffix or a space-separated list of directories and files for building the initramfs image. A cpio archive should contain a filesystem archive to be used as an initramfs image. Directories should contain a filesystem layout to be included in the initramfs image. Files should contain entries according to the format described by theusr/gen_init_cpio
program in the kernel tree.If you specify multiple directories and files, the initramfs image will be the aggregate of all of them.
For information on creating an initramfs, see the “Building an Initial RAM Filesystem (initramfs) Image” section in the Yocto Project Development Tasks Manual.
- CONFIG_SITE
A list of files that contains
autoconf
test results relevant to the current build. This variable is used by the Autotools utilities when runningconfigure
.- CONFIGURE_FLAGS
The minimal arguments for GNU configure.
- CONFLICT_DISTRO_FEATURES
When inheriting the features_check class, this variable identifies distribution features that would be in conflict should the recipe be built. In other words, if the
CONFLICT_DISTRO_FEATURES
variable lists a feature that also appears inDISTRO_FEATURES
within the current configuration, then the recipe will be skipped, and if the build system attempts to build the recipe then an error will be triggered.- COPYLEFT_LICENSE_EXCLUDE
A space-separated list of licenses to exclude from the source archived by the archiver class. In other words, if a license in a recipe’s LICENSE value is in the value of
COPYLEFT_LICENSE_EXCLUDE
, then its source is not archived by the class.Note
The
COPYLEFT_LICENSE_EXCLUDE
variable takes precedence over the COPYLEFT_LICENSE_INCLUDE variable.The default value, which is “CLOSED Proprietary”, for
COPYLEFT_LICENSE_EXCLUDE
is set by the copyleft_filter class, which is inherited by thearchiver
class.- COPYLEFT_LICENSE_INCLUDE
A space-separated list of licenses to include in the source archived by the archiver class. In other words, if a license in a recipe’s LICENSE value is in the value of
COPYLEFT_LICENSE_INCLUDE
, then its source is archived by the class.The default value is set by the copyleft_filter class, which is inherited by the
archiver
class. The default value includes “GPL*”, “LGPL*”, and “AGPL*”.- COPYLEFT_PN_EXCLUDE
A list of recipes to exclude in the source archived by the archiver class. The
COPYLEFT_PN_EXCLUDE
variable overrides the license inclusion and exclusion caused through the COPYLEFT_LICENSE_INCLUDE and COPYLEFT_LICENSE_EXCLUDE variables, respectively.The default value, which is “” indicating to not explicitly exclude any recipes by name, for
COPYLEFT_PN_EXCLUDE
is set by the copyleft_filter class, which is inherited by thearchiver
class.- COPYLEFT_PN_INCLUDE
A list of recipes to include in the source archived by the archiver class. The
COPYLEFT_PN_INCLUDE
variable overrides the license inclusion and exclusion caused through the COPYLEFT_LICENSE_INCLUDE and COPYLEFT_LICENSE_EXCLUDE variables, respectively.The default value, which is “” indicating to not explicitly include any recipes by name, for
COPYLEFT_PN_INCLUDE
is set by the copyleft_filter class, which is inherited by thearchiver
class.- COPYLEFT_RECIPE_TYPES
A space-separated list of recipe types to include in the source archived by the archiver class. Recipe types are
target
,native
,nativesdk
,cross
,crosssdk
, andcross-canadian
.The default value, which is “target*”, for
COPYLEFT_RECIPE_TYPES
is set by the copyleft_filter class, which is inherited by thearchiver
class.- COPY_LIC_DIRS
If set to “1” along with the COPY_LIC_MANIFEST variable, the OpenEmbedded build system copies into the image the license files, which are located in
/usr/share/common-licenses
, for each package. The license files are placed in directories within the image itself during build time.Note
The
COPY_LIC_DIRS
does not offer a path for adding licenses for newly installed packages to an image, which might be most suitable for read-only filesystems that cannot be upgraded. See the LICENSE_CREATE_PACKAGE variable for additional information. You can also reference the “Providing License Text” section in the Yocto Project Development Tasks Manual for information on providing license text.- COPY_LIC_MANIFEST
If set to “1”, the OpenEmbedded build system copies the license manifest for the image to
/usr/share/common-licenses/license.manifest
within the image itself during build time.Note
The
COPY_LIC_MANIFEST
does not offer a path for adding licenses for newly installed packages to an image, which might be most suitable for read-only filesystems that cannot be upgraded. See the LICENSE_CREATE_PACKAGE variable for additional information. You can also reference the “Providing License Text” section in the Yocto Project Development Tasks Manual for information on providing license text.- CORE_IMAGE_EXTRA_INSTALL
Specifies the list of packages to be added to the image. You should only set this variable in the
local.conf
configuration file found in the Build Directory.This variable replaces
POKY_EXTRA_INSTALL
, which is no longer supported.- COREBASE
Specifies the parent directory of the OpenEmbedded-Core Metadata layer (i.e.
meta
).It is an important distinction that
COREBASE
points to the parent of this layer and not the layer itself. Consider an example where you have cloned the Poky Git repository and retained thepoky
name for your local copy of the repository. In this case,COREBASE
points to thepoky
folder because it is the parent directory of thepoky/meta
layer.- COREBASE_FILES
Lists files from the COREBASE directory that should be copied other than the layers listed in the
bblayers.conf
file. TheCOREBASE_FILES
variable exists for the purpose of copying metadata from the OpenEmbedded build system into the extensible SDK.Explicitly listing files in
COREBASE
is needed because it typically contains build directories and other files that should not normally be copied into the extensible SDK. Consequently, the value ofCOREBASE_FILES
is used in order to only copy the files that are actually needed.- CPP
The minimal command and arguments used to run the C preprocessor.
- CPPFLAGS
Specifies the flags to pass to the C pre-processor (i.e. to both the C and the C++ compilers). This variable is exported to an environment variable and thus made visible to the software being built during the compilation step.
Default initialization for
CPPFLAGS
varies depending on what is being built:TARGET_CPPFLAGS when building for the target
BUILD_CPPFLAGS when building for the build host (i.e.
-native
)BUILDSDK_CPPFLAGS when building for an SDK (i.e.
nativesdk-
)
- CROSS_COMPILE
The toolchain binary prefix for the target tools. The
CROSS_COMPILE
variable is the same as the TARGET_PREFIX variable.Note
The OpenEmbedded build system sets the
CROSS_COMPILE
variable only in certain contexts (e.g. when building for kernel and kernel module recipes).- CVSDIR
The directory in which files checked out under the CVS system are stored.
- CXX
The minimal command and arguments used to run the C++ compiler.
- CXXFLAGS
Specifies the flags to pass to the C++ compiler. This variable is exported to an environment variable and thus made visible to the software being built during the compilation step.
Default initialization for
CXXFLAGS
varies depending on what is being built:TARGET_CXXFLAGS when building for the target
BUILD_CXXFLAGS when building for the build host (i.e.
-native
)BUILDSDK_CXXFLAGS when building for an SDK (i.e.
nativesdk-
)
- D
The destination directory. The location in the Build Directory where components are installed by the do_install task. This location defaults to:
${WORKDIR}/image
Note
Tasks that read from or write to this directory should run under fakeroot.
- DATE
The date the build was started. Dates appear using the year, month, and day (YMD) format (e.g. “20150209” for February 9th, 2015).
- DATETIME
The date and time on which the current build started. The format is suitable for timestamps.
- DEBIAN_NOAUTONAME
When the debian class is inherited, which is the default behavior,
DEBIAN_NOAUTONAME
specifies a particular package should not be renamed according to Debian library package naming. You must use the package name as an override when you set this variable. Here is an example from thefontconfig
recipe:DEBIAN_NOAUTONAME_fontconfig-utils = "1"
- DEBIANNAME
When the debian class is inherited, which is the default behavior,
DEBIANNAME
allows you to override the library name for an individual package. Overriding the library name in these cases is rare. You must use the package name as an override when you set this variable. Here is an example from thedbus
recipe:DEBIANNAME_${PN} = "dbus-1"
- DEBUGINFOD_URLS
Points to the URL of the “debuginfod” server. Such that for every debugging information lookup, the debuginfod client will query the server and return the requested information. You set this variable in your
local.conf
file.- DEBUG_BUILD
Specifies to build packages with debugging information. This influences the value of the
SELECTED_OPTIMIZATION
variable.- DEBUG_OPTIMIZATION
The options to pass in
TARGET_CFLAGS
andCFLAGS
when compiling a system for debugging. This variable defaults to “-O -fno-omit-frame-pointer ${DEBUG_FLAGS} -pipe”.- DEFAULT_PREFERENCE
Specifies a weak bias for recipe selection priority.
The most common usage of this is variable is to set it to “-1” within a recipe for a development version of a piece of software. Using the variable in this way causes the stable version of the recipe to build by default in the absence of
PREFERRED_VERSION
being used to build the development version.Note
The bias provided by
DEFAULT_PREFERENCE
is weak and is overridden by BBFILE_PRIORITY if that variable is different between two layers that contain different versions of the same recipe.- DEFAULTTUNE
The default CPU and Application Binary Interface (ABI) tunings (i.e. the “tune”) used by the OpenEmbedded build system. The
DEFAULTTUNE
helps define TUNE_FEATURES.The default tune is either implicitly or explicitly set by the machine (MACHINE). However, you can override the setting using available tunes as defined with AVAILTUNES.
- DEPENDS
Lists a recipe’s build-time dependencies. These are dependencies on other recipes whose contents (e.g. headers and shared libraries) are needed by the recipe at build time.
As an example, consider a recipe
foo
that contains the following assignment:DEPENDS = "bar"
The practical effect of the previous assignment is that all files installed by bar will be available in the appropriate staging sysroot, given by the STAGING_DIR* variables, by the time the do_configure task for
foo
runs. This mechanism is implemented by havingdo_configure
depend on the do_populate_sysroot task of each recipe listed inDEPENDS
, through a[
deptask]
declaration in the base class.Note
It seldom is necessary to reference, for example,
STAGING_DIR_HOST
explicitly. The standard classes and build-related variables are configured to automatically use the appropriate staging sysroots.As another example,
DEPENDS
can also be used to add utilities that run on the build machine during the build. For example, a recipe that makes use of a code generator built by the recipecodegen
might have the following:DEPENDS = "codegen-native"
For more information, see the native class and the EXTRANATIVEPATH variable.
Note
DEPENDS
is a list of recipe names. Or, to be more precise, it is a list of PROVIDES names, which usually match recipe names. Putting a package name such as “foo-dev” inDEPENDS
does not make sense. Use “foo” instead, as this will put files from all the packages that make upfoo
, which includes those fromfoo-dev
, into the sysroot.One recipe having another recipe in
DEPENDS
does not by itself add any runtime dependencies between the packages produced by the two recipes. However, as explained in the “Automatically Added Runtime Dependencies” section in the Yocto Project Overview and Concepts Manual, runtime dependencies will often be added automatically, meaningDEPENDS
alone is sufficient for most recipes.Counterintuitively,
DEPENDS
is often necessary even for recipes that install precompiled components. For example, iflibfoo
is a precompiled library that links againstlibbar
, then linking againstlibfoo
requires bothlibfoo
andlibbar
to be available in the sysroot. Without aDEPENDS
from the recipe that installslibfoo
to the recipe that installslibbar
, other recipes might fail to link againstlibfoo
.
For information on runtime dependencies, see the RDEPENDS variable. You can also see the “Tasks” and “Dependencies” sections in the BitBake User Manual for additional information on tasks and dependencies.
- DEPLOY_DIR
Points to the general area that the OpenEmbedded build system uses to place images, packages, SDKs, and other output files that are ready to be used outside of the build system. By default, this directory resides within the Build Directory as
${TMPDIR}/deploy
.For more information on the structure of the Build Directory, see “The Build Directory - build/” section. For more detail on the contents of the
deploy
directory, see the “Images”, “Package Feeds”, and “Application Development SDK” sections all in the Yocto Project Overview and Concepts Manual.- DEPLOY_DIR_DEB
Points to the area that the OpenEmbedded build system uses to place Debian packages that are ready to be used outside of the build system. This variable applies only when PACKAGE_CLASSES contains “package_deb”.
The BitBake configuration file initially defines the
DEPLOY_DIR_DEB
variable as a sub-folder of DEPLOY_DIR:DEPLOY_DIR_DEB = "${DEPLOY_DIR}/deb"
The package_deb class uses the
DEPLOY_DIR_DEB
variable to make sure the do_package_write_deb task writes Debian packages into the appropriate folder. For more information on how packaging works, see the “Package Feeds” section in the Yocto Project Overview and Concepts Manual.- DEPLOY_DIR_IMAGE
Points to the area that the OpenEmbedded build system uses to place images and other associated output files that are ready to be deployed onto the target machine. The directory is machine-specific as it contains the
${MACHINE}
name. By default, this directory resides within the Build Directory as${DEPLOY_DIR}/images/${MACHINE}/
.For more information on the structure of the Build Directory, see “The Build Directory - build/” section. For more detail on the contents of the
deploy
directory, see the “Images” and “Application Development SDK” sections both in the Yocto Project Overview and Concepts Manual.- DEPLOY_DIR_IPK
Points to the area that the OpenEmbedded build system uses to place IPK packages that are ready to be used outside of the build system. This variable applies only when PACKAGE_CLASSES contains “package_ipk”.
The BitBake configuration file initially defines this variable as a sub-folder of DEPLOY_DIR:
DEPLOY_DIR_IPK = "${DEPLOY_DIR}/ipk"
The package_ipk class uses the
DEPLOY_DIR_IPK
variable to make sure the do_package_write_ipk task writes IPK packages into the appropriate folder. For more information on how packaging works, see the “Package Feeds” section in the Yocto Project Overview and Concepts Manual.- DEPLOY_DIR_RPM
Points to the area that the OpenEmbedded build system uses to place RPM packages that are ready to be used outside of the build system. This variable applies only when PACKAGE_CLASSES contains “package_rpm”.
The BitBake configuration file initially defines this variable as a sub-folder of DEPLOY_DIR:
DEPLOY_DIR_RPM = "${DEPLOY_DIR}/rpm"
The package_rpm class uses the
DEPLOY_DIR_RPM
variable to make sure the do_package_write_rpm task writes RPM packages into the appropriate folder. For more information on how packaging works, see the “Package Feeds” section in the Yocto Project Overview and Concepts Manual.- DEPLOY_DIR_TAR
Points to the area that the OpenEmbedded build system uses to place tarballs that are ready to be used outside of the build system. This variable applies only when PACKAGE_CLASSES contains “package_tar”.
The BitBake configuration file initially defines this variable as a sub-folder of DEPLOY_DIR:
DEPLOY_DIR_TAR = "${DEPLOY_DIR}/tar"
The package_tar class uses the
DEPLOY_DIR_TAR
variable to make sure the do_package_write_tar task writes TAR packages into the appropriate folder. For more information on how packaging works, see the “Package Feeds” section in the Yocto Project Overview and Concepts Manual.- DEPLOYDIR
When inheriting the deploy class, the
DEPLOYDIR
points to a temporary work area for deployed files that is set in thedeploy
class as follows:DEPLOYDIR = "${WORKDIR}/deploy-${PN}"
Recipes inheriting the
deploy
class should copy files to be deployed intoDEPLOYDIR
, and the class will take care of copying them into DEPLOY_DIR_IMAGE afterwards.- DESCRIPTION
The package description used by package managers. If not set,
DESCRIPTION
takes the value of the SUMMARY variable.- DISTRO
The short name of the distribution. For information on the long name of the distribution, see the DISTRO_NAME variable.
The
DISTRO
variable corresponds to a distribution configuration file whose root name is the same as the variable’s argument and whose filename extension is.conf
. For example, the distribution configuration file for the Poky distribution is namedpoky.conf
and resides in themeta-poky/conf/distro
directory of the Source Directory.Within that
poky.conf
file, theDISTRO
variable is set as follows:DISTRO = "poky"
Distribution configuration files are located in a
conf/distro
directory within the Metadata that contains the distribution configuration. The value forDISTRO
must not contain spaces, and is typically all lower-case.Note
If the
DISTRO
variable is blank, a set of default configurations are used, which are specified withinmeta/conf/distro/defaultsetup.conf
also in the Source Directory.- DISTRO_CODENAME
Specifies a codename for the distribution being built.
- DISTRO_EXTRA_RDEPENDS
Specifies a list of distro-specific packages to add to all images. This variable takes affect through
packagegroup-base
so the variable only really applies to the more full-featured images that includepackagegroup-base
. You can use this variable to keep distro policy out of generic images. As with all other distro variables, you set this variable in the distro.conf
file.- DISTRO_EXTRA_RRECOMMENDS
Specifies a list of distro-specific packages to add to all images if the packages exist. The packages might not exist or be empty (e.g. kernel modules). The list of packages are automatically installed but you can remove them.
- DISTRO_FEATURES
The software support you want in your distribution for various features. You define your distribution features in the distribution configuration file.
In most cases, the presence or absence of a feature in
DISTRO_FEATURES
is translated to the appropriate option supplied to the configure script during the do_configure task for recipes that optionally support the feature. For example, specifying “x11” inDISTRO_FEATURES
, causes every piece of software built for the target that can optionally support X11 to have its X11 support enabled.Two more examples are Bluetooth and NFS support. For a more complete list of features that ships with the Yocto Project and that you can provide with this variable, see the “Distro Features” section.
- DISTRO_FEATURES_BACKFILL
Features to be added to
DISTRO_FEATURES
if not also present inDISTRO_FEATURES_BACKFILL_CONSIDERED
.This variable is set in the
meta/conf/bitbake.conf
file. It is not intended to be user-configurable. It is best to just reference the variable to see which distro features are being backfilled for all distro configurations. See the “Feature Backfilling” section for more information.- DISTRO_FEATURES_BACKFILL_CONSIDERED
Features from
DISTRO_FEATURES_BACKFILL
that should not be backfilled (i.e. added toDISTRO_FEATURES
) during the build. See the “Feature Backfilling” section for more information.- DISTRO_FEATURES_DEFAULT
A convenience variable that gives you the default list of distro features with the exception of any features specific to the C library (
libc
).When creating a custom distribution, you might find it useful to be able to reuse the default DISTRO_FEATURES options without the need to write out the full set. Here is an example that uses
DISTRO_FEATURES_DEFAULT
from a custom distro configuration file:DISTRO_FEATURES ?= "${DISTRO_FEATURES_DEFAULT} myfeature"
- DISTRO_FEATURES_FILTER_NATIVE
Specifies a list of features that if present in the target DISTRO_FEATURES value should be included in
DISTRO_FEATURES
when building native recipes. This variable is used in addition to the features filtered using the DISTRO_FEATURES_NATIVE variable.- DISTRO_FEATURES_FILTER_NATIVESDK
Specifies a list of features that if present in the target DISTRO_FEATURES value should be included in
DISTRO_FEATURES
when building nativesdk recipes. This variable is used in addition to the features filtered using the DISTRO_FEATURES_NATIVESDK variable.- DISTRO_FEATURES_NATIVE
Specifies a list of features that should be included in DISTRO_FEATURES when building native recipes. This variable is used in addition to the features filtered using the DISTRO_FEATURES_FILTER_NATIVE variable.
- DISTRO_FEATURES_NATIVESDK
Specifies a list of features that should be included in DISTRO_FEATURES when building nativesdk recipes. This variable is used in addition to the features filtered using the DISTRO_FEATURES_FILTER_NATIVESDK variable.
- DISTRO_NAME
The long name of the distribution. For information on the short name of the distribution, see the DISTRO variable.
The
DISTRO_NAME
variable corresponds to a distribution configuration file whose root name is the same as the variable’s argument and whose filename extension is.conf
. For example, the distribution configuration file for the Poky distribution is namedpoky.conf
and resides in themeta-poky/conf/distro
directory of the Source Directory.Within that
poky.conf
file, theDISTRO_NAME
variable is set as follows:DISTRO_NAME = "Poky (Yocto Project Reference Distro)"
Distribution configuration files are located in a
conf/distro
directory within the Metadata that contains the distribution configuration.Note
If the
DISTRO_NAME
variable is blank, a set of default configurations are used, which are specified withinmeta/conf/distro/defaultsetup.conf
also in the Source Directory.- DISTRO_VERSION
The version of the distribution.
- DISTROOVERRIDES
A colon-separated list of overrides specific to the current distribution. By default, this list includes the value of DISTRO.
You can extend
DISTROOVERRIDES
to add extra overrides that should apply to the distribution.The underlying mechanism behind
DISTROOVERRIDES
is simply that it is included in the default value of OVERRIDES.- DISTUTILS_SETUP_PATH
When used by recipes that inherit the distutils3 or setuptools3 class, this variable should be used to specify the directory in which the
setup.py
file is located if it is not at the root of the source tree (as specified by S). For example, in a recipe where the sources are fetched from a Git repository andsetup.py
is in apython/pythonmodule
subdirectory, you would have this:S = "${WORKDIR}/git" DISTUTILS_SETUP_PATH = "${S}/python/pythonmodule"
- DL_DIR
The central download directory used by the build process to store downloads. By default,
DL_DIR
gets files suitable for mirroring for everything except Git repositories. If you want tarballs of Git repositories, use the BB_GENERATE_MIRROR_TARBALLS variable.You can set this directory by defining the
DL_DIR
variable in theconf/local.conf
file. This directory is self-maintaining and you should not have to touch it. By default, the directory isdownloads
in the Build Directory.#DL_DIR ?= "${TOPDIR}/downloads"
To specify a different download directory, simply remove the comment from the line and provide your directory.
During a first build, the system downloads many different source code tarballs from various upstream projects. Downloading can take a while, particularly if your network connection is slow. Tarballs are all stored in the directory defined by
DL_DIR
and the build system looks there first to find source tarballs.Note
When wiping and rebuilding, you can preserve this directory to speed up this part of subsequent builds.
You can safely share this directory between multiple builds on the same development machine. For additional information on how the build process gets source files when working behind a firewall or proxy server, see this specific question in the “FAQ” chapter. You can also refer to the “Working Behind a Network Proxy” Wiki page.
- DOC_COMPRESS
When inheriting the compress_doc class, this variable sets the compression policy used when the OpenEmbedded build system compresses man pages and info pages. By default, the compression method used is gz (gzip). Other policies available are xz and bz2.
For information on policies and on how to use this variable, see the comments in the
meta/classes/compress_doc.bbclass
file.- EFI_PROVIDER
When building bootable images (i.e. where
hddimg
,iso
, orwic.vmdk
is in IMAGE_FSTYPES), theEFI_PROVIDER
variable specifies the EFI bootloader to use. The default is “grub-efi”, but “systemd-boot” can be used instead.See the systemd-boot and image-live classes for more information.
- ENABLE_BINARY_LOCALE_GENERATION
Variable that controls which locales for
glibc
are generated during the build (useful if the target device has 64Mbytes of RAM or less).- ERR_REPORT_DIR
When used with the report-error class, specifies the path used for storing the debug files created by the error reporting tool, which allows you to submit build errors you encounter to a central database. By default, the value of this variable is
${
LOG_DIR}/error-report
.You can set
ERR_REPORT_DIR
to the path you want the error reporting tool to store the debug files as follows in yourlocal.conf
file:ERR_REPORT_DIR = "path"
- ERROR_QA
Specifies the quality assurance checks whose failures are reported as errors by the OpenEmbedded build system. You set this variable in your distribution configuration file. For a list of the checks you can control with this variable, see the “insane.bbclass” section.
- EXCLUDE_FROM_SHLIBS
Triggers the OpenEmbedded build system’s shared libraries resolver to exclude an entire package when scanning for shared libraries.
Note
The shared libraries resolver’s functionality results in part from the internal function
package_do_shlibs
, which is part of the do_package task. You should be aware that the shared libraries resolver might implicitly define some dependencies between packages.The
EXCLUDE_FROM_SHLIBS
variable is similar to the PRIVATE_LIBS variable, which excludes a package’s particular libraries only and not the whole package.Use the
EXCLUDE_FROM_SHLIBS
variable by setting it to “1” for a particular package:EXCLUDE_FROM_SHLIBS = "1"
- EXCLUDE_FROM_WORLD
Directs BitBake to exclude a recipe from world builds (i.e.
bitbake world
). During world builds, BitBake locates, parses and builds all recipes found in every layer exposed in thebblayers.conf
configuration file.To exclude a recipe from a world build using this variable, set the variable to “1” in the recipe.
Note
Recipes added to
EXCLUDE_FROM_WORLD
may still be built during a world build in order to satisfy dependencies of other recipes. Adding a recipe toEXCLUDE_FROM_WORLD
only ensures that the recipe is not explicitly added to the list of build targets in a world build.- EXTENDPE
Used with file and pathnames to create a prefix for a recipe’s version based on the recipe’s PE value. If
PE
is set and greater than zero for a recipe,EXTENDPE
becomes that value (e.g ifPE
is equal to “1” thenEXTENDPE
becomes “1”). If a recipe’sPE
is not set (the default) or is equal to zero,EXTENDPE
becomes “”.See the STAMP variable for an example.
- EXTENDPKGV
The full package version specification as it appears on the final packages produced by a recipe. The variable’s value is normally used to fix a runtime dependency to the exact same version of another package in the same recipe:
RDEPENDS_${PN}-additional-module = "${PN} (= ${EXTENDPKGV})"
The dependency relationships are intended to force the package manager to upgrade these types of packages in lock-step.
- EXTERNAL_KERNEL_TOOLS
When set, the
EXTERNAL_KERNEL_TOOLS
variable indicates that these tools are not in the source tree.When kernel tools are available in the tree, they are preferred over any externally installed tools. Setting the
EXTERNAL_KERNEL_TOOLS
variable tells the OpenEmbedded build system to prefer the installed external tools. See the kernel-yocto class inmeta/classes
to see how the variable is used.- EXTERNALSRC
When inheriting the externalsrc class, this variable points to the source tree, which is outside of the OpenEmbedded build system. When set, this variable sets the S variable, which is what the OpenEmbedded build system uses to locate unpacked recipe source code.
For more information on
externalsrc.bbclass
, see the “externalsrc.bbclass” section. You can also find information on how to use this variable in the “Building Software from an External Source” section in the Yocto Project Development Tasks Manual.- EXTERNALSRC_BUILD
When inheriting the externalsrc class, this variable points to the directory in which the recipe’s source code is built, which is outside of the OpenEmbedded build system. When set, this variable sets the B variable, which is what the OpenEmbedded build system uses to locate the Build Directory.
For more information on
externalsrc.bbclass
, see the “externalsrc.bbclass” section. You can also find information on how to use this variable in the “Building Software from an External Source” section in the Yocto Project Development Tasks Manual.- EXTRA_AUTORECONF
For recipes inheriting the autotools class, you can use
EXTRA_AUTORECONF
to specify extra options to pass to theautoreconf
command that is executed during the do_configure task.The default value is “–exclude=autopoint”.
- EXTRA_IMAGE_FEATURES
A list of additional features to include in an image. When listing more than one feature, separate them with a space.
Typically, you configure this variable in your
local.conf
file, which is found in the Build Directory. Although you can use this variable from within a recipe, best practices dictate that you do not.Note
To enable primary features from within the image recipe, use the IMAGE_FEATURES variable.
Here are some examples of features you can add:
“dbg-pkgs” - Adds -dbg packages for all installed packages including symbol information for debugging and profiling.
“debug-tweaks” - Makes an image suitable for debugging. For example, allows root logins without passwords and enables post-installation logging. See the ‘allow-empty-password’ and ‘post-install-logging’ features in the “Image Features” section for more information.
“dev-pkgs” - Adds -dev packages for all installed packages. This is useful if you want to develop against the libraries in the image.
“read-only-rootfs” - Creates an image whose root filesystem is read-only. See the “Creating a Read-Only Root Filesystem” section in the Yocto Project Development Tasks Manual for more information
“tools-debug” - Adds debugging tools such as gdb and strace.
“tools-sdk” - Adds development tools such as gcc, make, pkgconfig and so forth.
“tools-testapps” - Adds useful testing tools such as ts_print, aplay, arecord and so forth.
For a complete list of image features that ships with the Yocto Project, see the “Image Features” section.
For an example that shows how to customize your image by using this variable, see the “Customizing Images Using Custom IMAGE_FEATURES and EXTRA_IMAGE_FEATURES” section in the Yocto Project Development Tasks Manual.
- EXTRA_IMAGECMD
Specifies additional options for the image creation command that has been specified in IMAGE_CMD. When setting this variable, use an override for the associated image type. Here is an example:
EXTRA_IMAGECMD_ext3 ?= "-i 4096"
- EXTRA_IMAGEDEPENDS
A list of recipes to build that do not provide packages for installing into the root filesystem.
Sometimes a recipe is required to build the final image but is not needed in the root filesystem. You can use the
EXTRA_IMAGEDEPENDS
variable to list these recipes and thus specify the dependencies. A typical example is a required bootloader in a machine configuration.Note
To add packages to the root filesystem, see the various RDEPENDS and RRECOMMENDS variables.
- EXTRANATIVEPATH
A list of subdirectories of
${
STAGING_BINDIR_NATIVE}
added to the beginning of the environment variablePATH
. As an example, the following prepends “${STAGING_BINDIR_NATIVE}/foo:${STAGING_BINDIR_NATIVE}/bar:” toPATH
:EXTRANATIVEPATH = "foo bar"
- EXTRA_OECMAKE
Additional CMake options. See the cmake class for additional information.
- EXTRA_OECONF
Additional
configure
script options. See PACKAGECONFIG_CONFARGS for additional information on passing configure script options.- EXTRA_OEMAKE
Additional GNU
make
options.Because the
EXTRA_OEMAKE
defaults to “”, you need to set the variable to specify any required GNU options.PARALLEL_MAKE and PARALLEL_MAKEINST also make use of
EXTRA_OEMAKE
to pass the required flags.- EXTRA_OESCONS
When inheriting the scons class, this variable specifies additional configuration options you want to pass to the
scons
command line.- EXTRA_USERS_PARAMS
When inheriting the extrausers class, this variable provides image level user and group operations. This is a more global method of providing user and group configuration as compared to using the useradd class, which ties user and group configurations to a specific recipe.
The set list of commands you can configure using the
EXTRA_USERS_PARAMS
is shown in theextrausers
class. These commands map to the normal Unix commands of the same names:# EXTRA_USERS_PARAMS = "\ # useradd -p '' tester; \ # groupadd developers; \ # userdel nobody; \ # groupdel -g video; \ # groupmod -g 1020 developers; \ # usermod -s /bin/sh tester; \ # "
Additionally there is a special
passwd-expire
command that will cause the password for a user to be expired and thus force changing it on first login, for example:EXTRA_USERS_PARAMS += " useradd myuser; passwd-expire myuser;"
Note
At present,
passwd-expire
may only work for remote logins when using OpenSSH and not dropbear as an SSH server.- FEATURE_PACKAGES
Defines one or more packages to include in an image when a specific item is included in IMAGE_FEATURES. When setting the value,
FEATURE_PACKAGES
should have the name of the feature item as an override. Here is an example:FEATURE_PACKAGES_widget = "package1 package2"
In this example, if “widget” were added to
IMAGE_FEATURES
, package1 and package2 would be included in the image.Note
Packages installed by features defined through
FEATURE_PACKAGES
are often package groups. While similarly named, you should not confuse theFEATURE_PACKAGES
variable with package groups, which are discussed elsewhere in the documentation.- FEED_DEPLOYDIR_BASE_URI
Points to the base URL of the server and location within the document-root that provides the metadata and packages required by OPKG to support runtime package management of IPK packages. You set this variable in your
local.conf
file.Consider the following example:
FEED_DEPLOYDIR_BASE_URI = "http://192.168.7.1/BOARD-dir"
This example assumes you are serving your packages over HTTP and your databases are located in a directory named
BOARD-dir
, which is underneath your HTTP server’s document-root. In this case, the OpenEmbedded build system generates a set of configuration files for you in your target that work with the feed.- FILES
The list of files and directories that are placed in a package. The PACKAGES variable lists the packages generated by a recipe.
To use the
FILES
variable, provide a package name override that identifies the resulting package. Then, provide a space-separated list of files or paths that identify the files you want included as part of the resulting package. Here is an example:FILES_${PN} += "${bindir}/mydir1 ${bindir}/mydir2/myfile"
Note
When specifying files or paths, you can pattern match using Python’s glob syntax. For details on the syntax, see the documentation by following the previous link.
When specifying paths as part of the
FILES
variable, it is good practice to use appropriate path variables. For example, use${sysconfdir}
rather than/etc
, or${bindir}
rather than/usr/bin
. You can find a list of these variables at the top of themeta/conf/bitbake.conf
file in the Source Directory. You will also find the default values of the variousFILES_*
variables in this file.
If some of the files you provide with the
FILES
variable are editable and you know they should not be overwritten during the package update process by the Package Management System (PMS), you can identify these files so that the PMS will not overwrite them. See the CONFFILES variable for information on how to identify these files to the PMS.- FILES_SOLIBSDEV
Defines the file specification to match SOLIBSDEV. In other words,
FILES_SOLIBSDEV
defines the full path name of the development symbolic link (symlink) for shared libraries on the target platform.The following statement from the
bitbake.conf
shows how it is set:FILES_SOLIBSDEV ?= "${base_libdir}/lib*${SOLIBSDEV} ${libdir}/lib*${SOLIBSDEV}"
- FILESEXTRAPATHS
Extends the search path the OpenEmbedded build system uses when looking for files and patches as it processes recipes and append files. The default directories BitBake uses when it processes recipes are initially defined by the FILESPATH variable. You can extend
FILESPATH
variable by usingFILESEXTRAPATHS
.Best practices dictate that you accomplish this by using
FILESEXTRAPATHS
from within a.bbappend
file and that you prepend paths as follows:FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:"
In the above example, the build system first looks for files in a directory that has the same name as the corresponding append file.
Note
When extending
FILESEXTRAPATHS
, be sure to use the immediate expansion (:=
) operator. Immediate expansion makes sure that BitBake evaluates THISDIR at the time the directive is encountered rather than at some later time when expansion might result in a directory that does not contain the files you need.Also, include the trailing separating colon character if you are prepending. The trailing colon character is necessary because you are directing BitBake to extend the path by prepending directories to the search path.
Here is another common use:
FILESEXTRAPATHS_prepend := "${THISDIR}/files:"
In this example, the build system extends the
FILESPATH
variable to include a directory namedfiles
that is in the same directory as the corresponding append file.This next example specifically adds three paths:
FILESEXTRAPATHS_prepend := "path_1:path_2:path_3:"
A final example shows how you can extend the search path and include a MACHINE-specific override, which is useful in a BSP layer:
FILESEXTRAPATHS_prepend_intel-x86-common := "${THISDIR}/${PN}:"
The previous statement appears in the
linux-yocto-dev.bbappend
file, which is found in the Yocto Project Source Repositories inmeta-intel/common/recipes-kernel/linux
. Here, the machine override is a special PACKAGE_ARCH definition for multiplemeta-intel
machines.Note
For a layer that supports a single BSP, the override could just be the value of
MACHINE
.By prepending paths in
.bbappend
files, you allow multiple append files that reside in different layers but are used for the same recipe to correctly extend the path.- FILESOVERRIDES
A subset of OVERRIDES used by the OpenEmbedded build system for creating FILESPATH. The
FILESOVERRIDES
variable uses overrides to automatically extend the FILESPATH variable. For an example of how that works, see the FILESPATH variable description. Additionally, you find more information on how overrides are handled in the “Conditional Syntax (Overrides)” section of the BitBake User Manual.By default, the
FILESOVERRIDES
variable is defined as:FILESOVERRIDES = "${TRANSLATED_TARGET_ARCH}:${MACHINEOVERRIDES}:${DISTROOVERRIDES}"
Note
Do not hand-edit the
FILESOVERRIDES
variable. The values match up with expected overrides and are used in an expected manner by the build system.- FILESPATH
The default set of directories the OpenEmbedded build system uses when searching for patches and files.
During the build process, BitBake searches each directory in
FILESPATH
in the specified order when looking for files and patches specified by eachfile://
URI in a recipe’s SRC_URI statements.The default value for the
FILESPATH
variable is defined in thebase.bbclass
class found inmeta/classes
in the Source Directory:FILESPATH = "${@base_set_filespath(["${FILE_DIRNAME}/${BP}", \ "${FILE_DIRNAME}/${BPN}", "${FILE_DIRNAME}/files"], d)}"
The
FILESPATH
variable is automatically extended using the overrides from the FILESOVERRIDES variable.Note
Do not hand-edit the
FILESPATH
variable. If you want the build system to look in directories other than the defaults, extend theFILESPATH
variable by using the FILESEXTRAPATHS variable.Be aware that the default
FILESPATH
directories do not map to directories in custom layers where append files (.bbappend
) are used. If you want the build system to find patches or files that reside with your append files, you need to extend theFILESPATH
variable by using theFILESEXTRAPATHS
variable.
You can take advantage of this searching behavior in useful ways. For example, consider a case where the following directory structure exists for general and machine-specific configurations:
files/defconfig files/MACHINEA/defconfig files/MACHINEB/defconfig
Also in the example, the
SRC_URI
statement contains “file://defconfig”. Given this scenario, you can set MACHINE to “MACHINEA” and cause the build system to use files fromfiles/MACHINEA
. SetMACHINE
to “MACHINEB” and the build system uses files fromfiles/MACHINEB
. Finally, for any machine other than “MACHINEA” and “MACHINEB”, the build system uses files fromfiles/defconfig
.You can find out more about the patching process in the “Patching” section in the Yocto Project Overview and Concepts Manual and the “Patching Code” section in the Yocto Project Development Tasks Manual. See the do_patch task as well.
- FILESYSTEM_PERMS_TABLES
Allows you to define your own file permissions settings table as part of your configuration for the packaging process. For example, suppose you need a consistent set of custom permissions for a set of groups and users across an entire work project. It is best to do this in the packages themselves but this is not always possible.
By default, the OpenEmbedded build system uses the
fs-perms.txt
, which is located in themeta/files
folder in the Source Directory. If you create your own file permissions setting table, you should place it in your layer or the distro’s layer.You define the
FILESYSTEM_PERMS_TABLES
variable in theconf/local.conf
file, which is found in the Build Directory, to point to your customfs-perms.txt
. You can specify more than a single file permissions setting table. The paths you specify to these files must be defined within the BBPATH variable.For guidance on how to create your own file permissions settings table file, examine the existing
fs-perms.txt
.- FIT_DESC
Specifies the description string encoded into a fitImage. The default value is set by the kernel-fitimage class as follows:
FIT_DESC ?= "U-Boot fitImage for ${DISTRO_NAME}/${PV}/${MACHINE}"
- FIT_GENERATE_KEYS
Decides whether to generate the keys for signing fitImage if they don’t already exist. The keys are created in
UBOOT_SIGN_KEYDIR
. The default value is 0.- FIT_HASH_ALG
Specifies the hash algorithm used in creating the FIT Image. For e.g. sha256.
- FIT_KERNEL_COMP_ALG
Compression algorithm to use for the kernel image inside the FIT Image. At present, the only supported values are “gzip” (default) or “none” If you set this variable to anything other than “none” you may also need to set FIT_KERNEL_COMP_ALG_EXTENSION.
- FIT_KERNEL_COMP_ALG_EXTENSION
File extension corresponding to FIT_KERNEL_COMP_ALG. The default value is “.gz”.
- FIT_KEY_GENRSA_ARGS
Arguments to openssl genrsa for generating RSA private key for signing fitImage. The default value is “-F4”. i.e. the public exponent 65537 to use.
- FIT_KEY_REQ_ARGS
Arguments to openssl req for generating certificate for signing fitImage. The default value is “-batch -new”. batch for non interactive mode and new for generating new keys.
- FIT_KEY_SIGN_PKCS
Format for public key certificate used in signing fitImage. The default value is “x509”.
- FIT_SIGN_ALG
Specifies the signature algorithm used in creating the FIT Image. For e.g. rsa2048.
- FIT_SIGN_NUMBITS
Size of private key in number of bits used in fitImage. The default value is “2048”.
- FIT_SIGN_INDIVIDUAL
If set to “1”, then the kernel-fitimage class will sign the kernel, dtb and ramdisk images individually in addition to signing the fitImage itself. This could be useful if you are intending to verify signatures in another context than booting via U-Boot.
- FONT_EXTRA_RDEPENDS
When inheriting the fontcache class, this variable specifies the runtime dependencies for font packages. By default, the
FONT_EXTRA_RDEPENDS
is set to “fontconfig-utils”.- FONT_PACKAGES
When inheriting the fontcache class, this variable identifies packages containing font files that need to be cached by Fontconfig. By default, the
fontcache
class assumes that fonts are in the recipe’s main package (i.e.${
PN}
). Use this variable if fonts you need are in a package other than that main package.- FORCE_RO_REMOVE
Forces the removal of the packages listed in
ROOTFS_RO_UNNEEDED
during the generation of the root filesystem.Set the variable to “1” to force the removal of these packages.
- FULL_OPTIMIZATION
The options to pass in
TARGET_CFLAGS
andCFLAGS
when compiling an optimized system. This variable defaults to “-O2 -pipe ${DEBUG_FLAGS}”.- GCCPIE
Enables Position Independent Executables (PIE) within the GNU C Compiler (GCC). Enabling PIE in the GCC makes Return Oriented Programming (ROP) attacks much more difficult to execute.
By default the
security_flags.inc
file enables PIE by setting the variable as follows:GCCPIE ?= "--enable-default-pie"
- GCCVERSION
Specifies the default version of the GNU C Compiler (GCC) used for compilation. By default,
GCCVERSION
is set to “8.x” in themeta/conf/distro/include/tcmode-default.inc
include file:GCCVERSION ?= "8.%"
You can override this value by setting it in a configuration file such as the
local.conf
.- GDB
The minimal command and arguments to run the GNU Debugger.
- GITDIR
The directory in which a local copy of a Git repository is stored when it is cloned.
- GLIBC_GENERATE_LOCALES
Specifies the list of GLIBC locales to generate should you not wish to generate all LIBC locals, which can be time consuming.
Note
If you specifically remove the locale
en_US.UTF-8
, you must set IMAGE_LINGUAS appropriately.You can set
GLIBC_GENERATE_LOCALES
in yourlocal.conf
file. By default, all locales are generated.GLIBC_GENERATE_LOCALES = "en_GB.UTF-8 en_US.UTF-8"
- GROUPADD_PARAM
When inheriting the useradd class, this variable specifies for a package what parameters should be passed to the
groupadd
command if you wish to add a group to the system when the package is installed.Here is an example from the
dbus
recipe:GROUPADD_PARAM_${PN} = "-r netdev"
For information on the standard Linux shell command
groupadd
, see https://linux.die.net/man/8/groupadd.- GROUPMEMS_PARAM
When inheriting the useradd class, this variable specifies for a package what parameters should be passed to the
groupmems
command if you wish to modify the members of a group when the package is installed.For information on the standard Linux shell command
groupmems
, see https://linux.die.net/man/8/groupmems.- GRUB_GFXSERIAL
Configures the GNU GRand Unified Bootloader (GRUB) to have graphics and serial in the boot menu. Set this variable to “1” in your
local.conf
or distribution configuration file to enable graphics and serial in the menu.See the grub-efi class for more information on how this variable is used.
- GRUB_OPTS
Additional options to add to the GNU GRand Unified Bootloader (GRUB) configuration. Use a semi-colon character (
;
) to separate multiple options.The
GRUB_OPTS
variable is optional. See the grub-efi class for more information on how this variable is used.- GRUB_TIMEOUT
Specifies the timeout before executing the default
LABEL
in the GNU GRand Unified Bootloader (GRUB).The
GRUB_TIMEOUT
variable is optional. See the grub-efi class for more information on how this variable is used.- GTKIMMODULES_PACKAGES
When inheriting the gtk-immodules-cache class, this variable specifies the packages that contain the GTK+ input method modules being installed when the modules are in packages other than the main package.
- HOMEPAGE
Website where more information about the software the recipe is building can be found.
- HOST_ARCH
The name of the target architecture, which is normally the same as TARGET_ARCH. The OpenEmbedded build system supports many architectures. Here is an example list of architectures supported. This list is by no means complete as the architecture is configurable:
arm
i586
x86_64
powerpc
powerpc64
mips
mipsel
- HOST_CC_ARCH
Specifies architecture-specific compiler flags that are passed to the C compiler.
Default initialization for
HOST_CC_ARCH
varies depending on what is being built:TARGET_CC_ARCH when building for the target
BUILD_CC_ARCH when building for the build host (i.e.
-native
)BUILDSDK_CC_ARCH
when building for an SDK (i.e.nativesdk-
)
- HOST_OS
Specifies the name of the target operating system, which is normally the same as the TARGET_OS. The variable can be set to “linux” for
glibc
-based systems and to “linux-musl” formusl
. For ARM/EABI targets, there are also “linux-gnueabi” and “linux-musleabi” values possible.- HOST_PREFIX
Specifies the prefix for the cross-compile toolchain.
HOST_PREFIX
is normally the same as TARGET_PREFIX.- HOST_SYS
Specifies the system, including the architecture and the operating system, for which the build is occurring in the context of the current recipe.
The OpenEmbedded build system automatically sets this variable based on HOST_ARCH, HOST_VENDOR, and HOST_OS variables.
Note
You do not need to set the variable yourself.
Consider these two examples:
Given a native recipe on a 32-bit x86 machine running Linux, the value is “i686-linux”.
Given a recipe being built for a little-endian MIPS target running Linux, the value might be “mipsel-linux”.
- HOSTTOOLS
A space-separated list (filter) of tools on the build host that should be allowed to be called from within build tasks. Using this filter helps reduce the possibility of host contamination. If a tool specified in the value of
HOSTTOOLS
is not found on the build host, the OpenEmbedded build system produces an error and the build is not started.For additional information, see HOSTTOOLS_NONFATAL.
- HOSTTOOLS_NONFATAL
A space-separated list (filter) of tools on the build host that should be allowed to be called from within build tasks. Using this filter helps reduce the possibility of host contamination. Unlike HOSTTOOLS, the OpenEmbedded build system does not produce an error if a tool specified in the value of
HOSTTOOLS_NONFATAL
is not found on the build host. Thus, you can useHOSTTOOLS_NONFATAL
to filter optional host tools.- HOST_VENDOR
Specifies the name of the vendor.
HOST_VENDOR
is normally the same as TARGET_VENDOR.- ICECC_DISABLED
Disables or enables the
icecc
(Icecream) function. For more information on this function and best practices for using this variable, see the “icecc.bbclass” section.Setting this variable to “1” in your
local.conf
disables the function:ICECC_DISABLED ??= "1"
To enable the function, set the variable as follows:
ICECC_DISABLED = ""
- ICECC_ENV_EXEC
Points to the
icecc-create-env
script that you provide. This variable is used by the icecc class. You set this variable in yourlocal.conf
file.If you do not point to a script that you provide, the OpenEmbedded build system uses the default script provided by the
icecc-create-env.bb
recipe, which is a modified version and not the one that comes withicecc
.- ICECC_PARALLEL_MAKE
Extra options passed to the
make
command during the do_compile task that specify parallel compilation. This variable usually takes the form of “-j x”, where x represents the maximum number of parallel threadsmake
can run.Note
The options passed affect builds on all enabled machines on the network, which are machines running the
iceccd
daemon.If your enabled machines support multiple cores, coming up with the maximum number of parallel threads that gives you the best performance could take some experimentation since machine speed, network lag, available memory, and existing machine loads can all affect build time. Consequently, unlike the PARALLEL_MAKE variable, there is no rule-of-thumb for setting
ICECC_PARALLEL_MAKE
to achieve optimal performance.If you do not set
ICECC_PARALLEL_MAKE
, the build system does not use it (i.e. the system does not detect and assign the number of cores as is done withPARALLEL_MAKE
).- ICECC_PATH
The location of the
icecc
binary. You can set this variable in yourlocal.conf
file. If yourlocal.conf
file does not define this variable, the icecc class attempts to define it by locatingicecc
usingwhich
.- ICECC_USER_CLASS_BL
Identifies user classes that you do not want the Icecream distributed compile support to consider. This variable is used by the icecc class. You set this variable in your
local.conf
file.When you list classes using this variable, you are “blacklisting” them from distributed compilation across remote hosts. Any classes you list will be distributed and compiled locally.
- ICECC_USER_PACKAGE_BL
Identifies user recipes that you do not want the Icecream distributed compile support to consider. This variable is used by the icecc class. You set this variable in your
local.conf
file.When you list packages using this variable, you are “blacklisting” them from distributed compilation across remote hosts. Any packages you list will be distributed and compiled locally.
- ICECC_USER_PACKAGE_WL
Identifies user recipes that use an empty PARALLEL_MAKE variable that you want to force remote distributed compilation on using the Icecream distributed compile support. This variable is used by the icecc class. You set this variable in your
local.conf
file.- IMAGE_BASENAME
The base name of image output files. This variable defaults to the recipe name (
${
PN}
).- IMAGE_EFI_BOOT_FILES
A space-separated list of files installed into the boot partition when preparing an image using the Wic tool with the
bootimg-efi
source plugin. By default, the files are installed under the same name as the source files. To change the installed name, separate it from the original name with a semi-colon (;). Source files need to be located in DEPLOY_DIR_IMAGE. Here are two examples:IMAGE_EFI_BOOT_FILES = "${KERNEL_IMAGETYPE};bz2" IMAGE_EFI_BOOT_FILES = "${KERNEL_IMAGETYPE} microcode.cpio"
Alternatively, source files can be picked up using a glob pattern. In this case, the destination file must have the same name as the base name of the source file path. To install files into a directory within the target location, pass its name after a semi-colon (;). Here are two examples:
IMAGE_EFI_BOOT_FILES = "boot/loader/*" IMAGE_EFI_BOOT_FILES = "boot/loader/*;boot/"
The first example installs all files from
${DEPLOY_DIR_IMAGE}/boot/loader/
into the root of the target partition. The second example installs the same files into aboot
directory within the target partition.You can find information on how to use the Wic tool in the “Creating Partitioned Images Using Wic” section of the Yocto Project Development Tasks Manual. Reference material for Wic is located in the “OpenEmbedded Kickstart (.wks) Reference” chapter.
- IMAGE_BOOT_FILES
A space-separated list of files installed into the boot partition when preparing an image using the Wic tool with the
bootimg-partition
source plugin. By default, the files are installed under the same name as the source files. To change the installed name, separate it from the original name with a semi-colon (;). Source files need to be located in DEPLOY_DIR_IMAGE. Here are two examples:IMAGE_BOOT_FILES = "u-boot.img uImage;kernel" IMAGE_BOOT_FILES = "u-boot.${UBOOT_SUFFIX} ${KERNEL_IMAGETYPE}"
Alternatively, source files can be picked up using a glob pattern. In this case, the destination file must have the same name as the base name of the source file path. To install files into a directory within the target location, pass its name after a semi-colon (;). Here are two examples:
IMAGE_BOOT_FILES = "bcm2835-bootfiles/*" IMAGE_BOOT_FILES = "bcm2835-bootfiles/*;boot/"
The first example installs all files from
${DEPLOY_DIR_IMAGE}/bcm2835-bootfiles
into the root of the target partition. The second example installs the same files into aboot
directory within the target partition.You can find information on how to use the Wic tool in the “Creating Partitioned Images Using Wic” section of the Yocto Project Development Tasks Manual. Reference material for Wic is located in the “OpenEmbedded Kickstart (.wks) Reference” chapter.
- IMAGE_CLASSES
A list of classes that all images should inherit. You typically use this variable to specify the list of classes that register the different types of images the OpenEmbedded build system creates.
The default value for
IMAGE_CLASSES
isimage_types
. You can set this variable in yourlocal.conf
or in a distribution configuration file.For more information, see
meta/classes/image_types.bbclass
in the Source Directory.- IMAGE_CMD
Specifies the command to create the image file for a specific image type, which corresponds to the value set in IMAGE_FSTYPES, (e.g.
ext3
,btrfs
, and so forth). When setting this variable, you should use an override for the associated type. Here is an example:IMAGE_CMD_jffs2 = "mkfs.jffs2 --root=${IMAGE_ROOTFS} \ --faketime --output=${DEPLOY_DIR_IMAGE}/${IMAGE_NAME}.rootfs.jffs2 \ ${EXTRA_IMAGECMD}"
You typically do not need to set this variable unless you are adding support for a new image type. For more examples on how to set this variable, see the image_types class file, which is
meta/classes/image_types.bbclass
.- IMAGE_DEVICE_TABLES
Specifies one or more files that contain custom device tables that are passed to the
makedevs
command as part of creating an image. These files list basic device nodes that should be created under/dev
within the image. IfIMAGE_DEVICE_TABLES
is not set,files/device_table-minimal.txt
is used, which is located by BBPATH. For details on how you should write device table files, seemeta/files/device_table-minimal.txt
as an example.- IMAGE_FEATURES
The primary list of features to include in an image. Typically, you configure this variable in an image recipe. Although you can use this variable from your
local.conf
file, which is found in the Build Directory, best practices dictate that you do not.Note
To enable extra features from outside the image recipe, use the EXTRA_IMAGE_FEATURES variable.
For a list of image features that ships with the Yocto Project, see the “Image Features” section.
For an example that shows how to customize your image by using this variable, see the “Customizing Images Using Custom IMAGE_FEATURES and EXTRA_IMAGE_FEATURES” section in the Yocto Project Development Tasks Manual.
- IMAGE_FSTYPES
Specifies the formats the OpenEmbedded build system uses during the build when creating the root filesystem. For example, setting
IMAGE_FSTYPES
as follows causes the build system to create root filesystems using two formats:.ext3
and.tar.bz2
:IMAGE_FSTYPES = "ext3 tar.bz2"
For the complete list of supported image formats from which you can choose, see IMAGE_TYPES.
Note
If an image recipe uses the “inherit image” line and you are setting
IMAGE_FSTYPES
inside the recipe, you must setIMAGE_FSTYPES
prior to using the “inherit image” line.Due to the way the OpenEmbedded build system processes this variable, you cannot update its contents by using
_append
or_prepend
. You must use the+=
operator to add one or more options to theIMAGE_FSTYPES
variable.
- IMAGE_INSTALL
Used by recipes to specify the packages to install into an image through the image class. Use the
IMAGE_INSTALL
variable with care to avoid ordering issues.Image recipes set
IMAGE_INSTALL
to specify the packages to install into an image throughimage.bbclass
. Additionally, “helper” classes such as the core-image class exist that can take lists used withIMAGE_FEATURES
and turn them into auto-generated entries inIMAGE_INSTALL
in addition to its default contents.When you use this variable, it is best to use it as follows:
IMAGE_INSTALL_append = " package-name"
Be sure to include the space between the quotation character and the start of the package name or names.
Note
When working with a core-image-minimal-initramfs image, do not use the
IMAGE_INSTALL
variable to specify packages for installation. Instead, use the PACKAGE_INSTALL variable, which allows the initial RAM filesystem (initramfs) recipe to use a fixed set of packages and not be affected byIMAGE_INSTALL
. For information on creating an initramfs, see the “Building an Initial RAM Filesystem (initramfs) Image” section in the Yocto Project Development Tasks Manual.Using
IMAGE_INSTALL
with the += BitBake operator within the/conf/local.conf
file or from within an image recipe is not recommended. Use of this operator in these ways can cause ordering issues. Sincecore-image.bbclass
setsIMAGE_INSTALL
to a default value using the ?= operator, using a+=
operation againstIMAGE_INSTALL
results in unexpected behavior when used withinconf/local.conf
. Furthermore, the same operation from within an image recipe may or may not succeed depending on the specific situation. In both these cases, the behavior is contrary to how most users expect the+=
operator to work.
- IMAGE_LINGUAS
Specifies the list of locales to install into the image during the root filesystem construction process. The OpenEmbedded build system automatically splits locale files, which are used for localization, into separate packages. Setting the
IMAGE_LINGUAS
variable ensures that any locale packages that correspond to packages already selected for installation into the image are also installed. Here is an example:IMAGE_LINGUAS = "pt-br de-de"
In this example, the build system ensures any Brazilian Portuguese and German locale files that correspond to packages in the image are installed (i.e.
*-locale-pt-br
and*-locale-de-de
as well as*-locale-pt
and*-locale-de
, since some software packages only provide locale files by language and not by country-specific language).See the GLIBC_GENERATE_LOCALES variable for information on generating GLIBC locales.
- IMAGE_LINK_NAME
The name of the output image symlink (which does not include the version part as IMAGE_NAME does). The default value is derived using the IMAGE_BASENAME and MACHINE variables:
IMAGE_LINK_NAME ?= "${IMAGE_BASENAME}-${MACHINE}"
- IMAGE_MANIFEST
The manifest file for the image. This file lists all the installed packages that make up the image. The file contains package information on a line-per-package basis as follows:
packagename packagearch version
The image class defines the manifest file as follows:
IMAGE_MANIFEST ="${DEPLOY_DIR_IMAGE}/${IMAGE_NAME}.rootfs.manifest"
The location is derived using the DEPLOY_DIR_IMAGE and IMAGE_NAME variables. You can find information on how the image is created in the “Image Generation” section in the Yocto Project Overview and Concepts Manual.
- IMAGE_NAME
The name of the output image files minus the extension. This variable is derived using the IMAGE_BASENAME, MACHINE, and IMAGE_VERSION_SUFFIX variables:
IMAGE_NAME ?= "${IMAGE_BASENAME}-${MACHINE}${IMAGE_VERSION_SUFFIX}"
- IMAGE_NAME_SUFFIX
Suffix used for the image output file name - defaults to
".rootfs"
to distinguish the image file from other files created during image building; however if this suffix is redundant or not desired you can clear the value of this variable (set the value to “”). For example, this is typically cleared in initramfs image recipes.- IMAGE_OVERHEAD_FACTOR
Defines a multiplier that the build system applies to the initial image size for cases when the multiplier times the returned disk usage value for the image is greater than the sum of
IMAGE_ROOTFS_SIZE
andIMAGE_ROOTFS_EXTRA_SPACE
. The result of the multiplier applied to the initial image size creates free disk space in the image as overhead. By default, the build process uses a multiplier of 1.3 for this variable. This default value results in 30% free disk space added to the image when this method is used to determine the final generated image size. You should be aware that post install scripts and the package management system uses disk space inside this overhead area. Consequently, the multiplier does not produce an image with all the theoretical free disk space. SeeIMAGE_ROOTFS_SIZE
for information on how the build system determines the overall image size.The default 30% free disk space typically gives the image enough room to boot and allows for basic post installs while still leaving a small amount of free disk space. If 30% free space is inadequate, you can increase the default value. For example, the following setting gives you 50% free space added to the image:
IMAGE_OVERHEAD_FACTOR = "1.5"
Alternatively, you can ensure a specific amount of free disk space is added to the image by using the
IMAGE_ROOTFS_EXTRA_SPACE
variable.- IMAGE_PKGTYPE
Defines the package type (i.e. DEB, RPM, IPK, or TAR) used by the OpenEmbedded build system. The variable is defined appropriately by the package_deb, package_rpm, package_ipk, or package_tar class.
Note
The
package_tar
class is broken and is not supported. It is recommended that you do not use it.The populate_sdk_* and image classes use the
IMAGE_PKGTYPE
for packaging up images and SDKs.You should not set the
IMAGE_PKGTYPE
manually. Rather, the variable is set indirectly through the appropriate package_* class using the PACKAGE_CLASSES variable. The OpenEmbedded build system uses the first package type (e.g. DEB, RPM, or IPK) that appears with the variableNote
Files using the
.tar
format are never used as a substitute packaging format for DEB, RPM, and IPK formatted files for your image or SDK.- IMAGE_POSTPROCESS_COMMAND
Specifies a list of functions to call once the OpenEmbedded build system creates the final image output files. You can specify functions separated by semicolons:
IMAGE_POSTPROCESS_COMMAND += "function; ... "
If you need to pass the root filesystem path to a command within the function, you can use
${IMAGE_ROOTFS}
, which points to the directory that becomes the root filesystem image. See the IMAGE_ROOTFS variable for more information.- IMAGE_PREPROCESS_COMMAND
Specifies a list of functions to call before the OpenEmbedded build system creates the final image output files. You can specify functions separated by semicolons:
IMAGE_PREPROCESS_COMMAND += "function; ... "
If you need to pass the root filesystem path to a command within the function, you can use
${IMAGE_ROOTFS}
, which points to the directory that becomes the root filesystem image. See the IMAGE_ROOTFS variable for more information.- IMAGE_ROOTFS
The location of the root filesystem while it is under construction (i.e. during the do_rootfs task). This variable is not configurable. Do not change it.
- IMAGE_ROOTFS_ALIGNMENT
Specifies the alignment for the output image file in Kbytes. If the size of the image is not a multiple of this value, then the size is rounded up to the nearest multiple of the value. The default value is “1”. See IMAGE_ROOTFS_SIZE for additional information.
- IMAGE_ROOTFS_EXTRA_SPACE
Defines additional free disk space created in the image in Kbytes. By default, this variable is set to “0”. This free disk space is added to the image after the build system determines the image size as described in
IMAGE_ROOTFS_SIZE
.This variable is particularly useful when you want to ensure that a specific amount of free disk space is available on a device after an image is installed and running. For example, to be sure 5 Gbytes of free disk space is available, set the variable as follows:
IMAGE_ROOTFS_EXTRA_SPACE = "5242880"
For example, the Yocto Project Build Appliance specifically requests 40 Gbytes of extra space with the line:
IMAGE_ROOTFS_EXTRA_SPACE = "41943040"
- IMAGE_ROOTFS_SIZE
Defines the size in Kbytes for the generated image. The OpenEmbedded build system determines the final size for the generated image using an algorithm that takes into account the initial disk space used for the generated image, a requested size for the image, and requested additional free disk space to be added to the image. Programatically, the build system determines the final size of the generated image as follows:
if (image-du * overhead) < rootfs-size: internal-rootfs-size = rootfs-size + xspace else: internal-rootfs-size = (image-du * overhead) + xspace where: image-du = Returned value of the du command on the image. overhead = IMAGE_OVERHEAD_FACTOR rootfs-size = IMAGE_ROOTFS_SIZE internal-rootfs-size = Initial root filesystem size before any modifications. xspace = IMAGE_ROOTFS_EXTRA_SPACE
See the IMAGE_OVERHEAD_FACTOR and IMAGE_ROOTFS_EXTRA_SPACE variables for related information.
- IMAGE_TYPEDEP
Specifies a dependency from one image type on another. Here is an example from the image-live class:
IMAGE_TYPEDEP_live = "ext3"
In the previous example, the variable ensures that when “live” is listed with the IMAGE_FSTYPES variable, the OpenEmbedded build system produces an
ext3
image first since one of the components of the live image is anext3
formatted partition containing the root filesystem.- IMAGE_TYPES
Specifies the complete list of supported image types by default:
btrfs
container
cpio
cpio.gz
cpio.lz4
cpio.lzma
cpio.xz
cramfs
ext2
ext2.bz2
ext2.gz
ext2.lzma
ext3
ext3.gz
ext4
ext4.gz
f2fs
hddimg
iso
jffs2
jffs2.sum
multiubi
squashfs
squashfs-lz4
squashfs-lzo
squashfs-xz
tar
tar.bz2
tar.gz
tar.lz4
tar.xz
tar.zst
ubi
ubifs
wic
wic.bz2
wic.gz
wic.lzma
For more information about these types of images, see
meta/classes/image_types*.bbclass
in the Source Directory.- IMAGE_VERSION_SUFFIX
Version suffix that is part of the default IMAGE_NAME and KERNEL_ARTIFACT_NAME values. Defaults to
"-${DATETIME}"
, however you could set this to a version string that comes from your external build environment if desired, and this suffix would then be used consistently across the build artifacts.- INC_PR
Helps define the recipe revision for recipes that share a common
include
file. You can think of this variable as part of the recipe revision as set from within an include file.Suppose, for example, you have a set of recipes that are used across several projects. And, within each of those recipes the revision (its PR value) is set accordingly. In this case, when the revision of those recipes changes, the burden is on you to find all those recipes and be sure that they get changed to reflect the updated version of the recipe. In this scenario, it can get complicated when recipes that are used in many places and provide common functionality are upgraded to a new revision.
A more efficient way of dealing with this situation is to set the
INC_PR
variable inside theinclude
files that the recipes share and then expand theINC_PR
variable within the recipes to help define the recipe revision.The following provides an example that shows how to use the
INC_PR
variable given a commoninclude
file that defines the variable. Once the variable is defined in theinclude
file, you can use the variable to set thePR
values in each recipe. You will notice that when you set a recipe’sPR
you can provide more granular revisioning by appending values to theINC_PR
variable:recipes-graphics/xorg-font/xorg-font-common.inc:INC_PR = "r2" recipes-graphics/xorg-font/encodings_1.0.4.bb:PR = "${INC_PR}.1" recipes-graphics/xorg-font/font-util_1.3.0.bb:PR = "${INC_PR}.0" recipes-graphics/xorg-font/font-alias_1.0.3.bb:PR = "${INC_PR}.3"
The first line of the example establishes the baseline revision to be used for all recipes that use the
include
file. The remaining lines in the example are from individual recipes and show how thePR
value is set.- INCOMPATIBLE_LICENSE
Specifies a space-separated list of license names (as they would appear in LICENSE) that should be excluded from the build. Recipes that provide no alternatives to listed incompatible licenses are not built. Packages that are individually licensed with the specified incompatible licenses will be deleted.
Note
This functionality is only regularly tested using the following setting:
INCOMPATIBLE_LICENSE = "GPL-3.0 LGPL-3.0 AGPL-3.0"
Although you can use other settings, you might be required to remove dependencies on or provide alternatives to components that are required to produce a functional system image.
Note
It is possible to define a list of licenses that are allowed to be used instead of the licenses that are excluded. To do this, define a variable
COMPATIBLE_LICENSES
with the names of the licenses that are allowed. Then defineINCOMPATIBLE_LICENSE
as:INCOMPATIBLE_LICENSE = "${@' '.join(sorted(set(d.getVar('AVAILABLE_LICENSES').split()) - set(d.getVar('COMPATIBLE_LICENSES').split())))}"
This will result in
INCOMPATIBLE_LICENSE
containing the names of all licenses from AVAILABLE_LICENSES except the ones specified inCOMPATIBLE_LICENSES
, thus only allowing the latter licenses to be used.- INHERIT
Causes the named class or classes to be inherited globally. Anonymous functions in the class or classes are not executed for the base configuration and in each individual recipe. The OpenEmbedded build system ignores changes to
INHERIT
in individual recipes.For more information on
INHERIT
, see the INHERIT Configuration Directive” section in the Bitbake User Manual.- INHERIT_DISTRO
Lists classes that will be inherited at the distribution level. It is unlikely that you want to edit this variable.
The default value of the variable is set as follows in the
meta/conf/distro/defaultsetup.conf
file:INHERIT_DISTRO ?= "debian devshell sstate license"
- INHIBIT_DEFAULT_DEPS
Prevents the default dependencies, namely the C compiler and standard C library (libc), from being added to DEPENDS. This variable is usually used within recipes that do not require any compilation using the C compiler.
Set the variable to “1” to prevent the default dependencies from being added.
- INHIBIT_PACKAGE_DEBUG_SPLIT
Prevents the OpenEmbedded build system from splitting out debug information during packaging. By default, the build system splits out debugging information during the do_package task. For more information on how debug information is split out, see the PACKAGE_DEBUG_SPLIT_STYLE variable.
To prevent the build system from splitting out debug information during packaging, set the
INHIBIT_PACKAGE_DEBUG_SPLIT
variable as follows:INHIBIT_PACKAGE_DEBUG_SPLIT = "1"
- INHIBIT_PACKAGE_STRIP
If set to “1”, causes the build to not strip binaries in resulting packages and prevents the
-dbg
package from containing the source files.By default, the OpenEmbedded build system strips binaries and puts the debugging symbols into
${
PN}-dbg
. Consequently, you should not setINHIBIT_PACKAGE_STRIP
when you plan to debug in general.- INHIBIT_SYSROOT_STRIP
If set to “1”, causes the build to not strip binaries in the resulting sysroot.
By default, the OpenEmbedded build system strips binaries in the resulting sysroot. When you specifically set the
INHIBIT_SYSROOT_STRIP
variable to “1” in your recipe, you inhibit this stripping.If you want to use this variable, include the staging class. This class uses a
sys_strip()
function to test for the variable and acts accordingly.Note
Use of the
INHIBIT_SYSROOT_STRIP
variable occurs in rare and special circumstances. For example, suppose you are building bare-metal firmware by using an external GCC toolchain. Furthermore, even if the toolchain’s binaries are strippable, other files exist that are needed for the build that are not strippable.- INITRAMFS_FSTYPES
Defines the format for the output image of an initial RAM filesystem (initramfs), which is used during boot. Supported formats are the same as those supported by the IMAGE_FSTYPES variable.
The default value of this variable, which is set in the
meta/conf/bitbake.conf
configuration file in the Source Directory, is “cpio.gz”. The Linux kernel’s initramfs mechanism, as opposed to the initial RAM filesystem initrd mechanism, expects an optionally compressed cpio archive.- INITRAMFS_IMAGE
Specifies the PROVIDES name of an image recipe that is used to build an initial RAM filesystem (initramfs) image. In other words, the
INITRAMFS_IMAGE
variable causes an additional recipe to be built as a dependency to whatever root filesystem recipe you might be using (e.g.core-image-sato
). The initramfs image recipe you provide should set IMAGE_FSTYPES to INITRAMFS_FSTYPES.An initramfs image provides a temporary root filesystem used for early system initialization (e.g. loading of modules needed to locate and mount the “real” root filesystem).
Note
See the
meta/recipes-core/images/core-image-minimal-initramfs.bb
recipe in the Source Directory for an example initramfs recipe. To select this sample recipe as the one built to provide the initramfs image, setINITRAMFS_IMAGE
to “core-image-minimal-initramfs”.You can also find more information by referencing the
meta-poky/conf/local.conf.sample.extended
configuration file in the Source Directory, the image class, and the kernel class to see how to use theINITRAMFS_IMAGE
variable.If
INITRAMFS_IMAGE
is empty, which is the default, then no initramfs image is built.For more information, you can also see the INITRAMFS_IMAGE_BUNDLE variable, which allows the generated image to be bundled inside the kernel image. Additionally, for information on creating an initramfs image, see the “Building an Initial RAM Filesystem (initramfs) Image” section in the Yocto Project Development Tasks Manual.
- INITRAMFS_IMAGE_BUNDLE
Controls whether or not the image recipe specified by INITRAMFS_IMAGE is run through an extra pass (do_bundle_initramfs) during kernel compilation in order to build a single binary that contains both the kernel image and the initial RAM filesystem (initramfs) image. This makes use of the CONFIG_INITRAMFS_SOURCE kernel feature.
Note
Using an extra compilation pass to bundle the initramfs avoids a circular dependency between the kernel recipe and the initramfs recipe should the initramfs include kernel modules. Should that be the case, the initramfs recipe depends on the kernel for the kernel modules, and the kernel depends on the initramfs recipe since the initramfs is bundled inside the kernel image.
The combined binary is deposited into the
tmp/deploy
directory, which is part of the Build Directory.Setting the variable to “1” in a configuration file causes the OpenEmbedded build system to generate a kernel image with the initramfs specified in
INITRAMFS_IMAGE
bundled within:INITRAMFS_IMAGE_BUNDLE = "1"
By default, the kernel class sets this variable to a null string as follows:
INITRAMFS_IMAGE_BUNDLE ?= ""
Note
You must set the
INITRAMFS_IMAGE_BUNDLE
variable in a configuration file. You cannot set the variable in a recipe file.See the local.conf.sample.extended file for additional information. Also, for information on creating an initramfs, see the “Building an Initial RAM Filesystem (initramfs) Image” section in the Yocto Project Development Tasks Manual.
- INITRAMFS_LINK_NAME
The link name of the initial RAM filesystem image. This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:INITRAMFS_LINK_NAME ?= "initramfs-${KERNEL_ARTIFACT_LINK_NAME}"
The value of the
KERNEL_ARTIFACT_LINK_NAME
variable, which is set in the same file, has the following value:KERNEL_ARTIFACT_LINK_NAME ?= "${MACHINE}"
See the MACHINE variable for additional information.
- INITRAMFS_NAME
The base name of the initial RAM filesystem image. This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:INITRAMFS_NAME ?= "initramfs-${KERNEL_ARTIFACT_NAME}"
The value of the KERNEL_ARTIFACT_NAME variable, which is set in the same file, has the following value:
KERNEL_ARTIFACT_NAME ?= "${PKGE}-${PKGV}-${PKGR}-${MACHINE}${IMAGE_VERSION_SUFFIX}"
- INITRD
Indicates list of filesystem images to concatenate and use as an initial RAM disk (
initrd
).The
INITRD
variable is an optional variable used with the image-live class.- INITRD_IMAGE
When building a “live” bootable image (i.e. when IMAGE_FSTYPES contains “live”),
INITRD_IMAGE
specifies the image recipe that should be built to provide the initial RAM disk image. The default value is “core-image-minimal-initramfs”.See the image-live class for more information.
- INITSCRIPT_NAME
The filename of the initialization script as installed to
${sysconfdir}/init.d
.This variable is used in recipes when using
update-rc.d.bbclass
. The variable is mandatory.- INITSCRIPT_PACKAGES
A list of the packages that contain initscripts. If multiple packages are specified, you need to append the package name to the other
INITSCRIPT_*
as an override.This variable is used in recipes when using
update-rc.d.bbclass
. The variable is optional and defaults to the PN variable.- INITSCRIPT_PARAMS
Specifies the options to pass to
update-rc.d
. Here is an example:INITSCRIPT_PARAMS = "start 99 5 2 . stop 20 0 1 6 ."
In this example, the script has a runlevel of 99, starts the script in initlevels 2 and 5, and stops the script in levels 0, 1 and 6.
The variable’s default value is “defaults”, which is set in the update-rc.d class.
The value in
INITSCRIPT_PARAMS
is passed through to theupdate-rc.d
command. For more information on valid parameters, please see theupdate-rc.d
manual page at https://manpages.debian.org/buster/init-system-helpers/update-rc.d.8.en.html- INSANE_SKIP
Specifies the QA checks to skip for a specific package within a recipe. For example, to skip the check for symbolic link
.so
files in the main package of a recipe, add the following to the recipe. The package name override must be used, which in this example is${PN}
:INSANE_SKIP_${PN} += "dev-so"
See the “insane.bbclass” section for a list of the valid QA checks you can specify using this variable.
- INSTALL_TIMEZONE_FILE
By default, the
tzdata
recipe packages an/etc/timezone
file. Set theINSTALL_TIMEZONE_FILE
variable to “0” at the configuration level to disable this behavior.- IPK_FEED_URIS
When the IPK backend is in use and package management is enabled on the target, you can use this variable to set up
opkg
in the target image to point to package feeds on a nominated server. Once the feed is established, you can perform installations or upgrades using the package manager at runtime.- KARCH
Defines the kernel architecture used when assembling the configuration. Architectures supported for this release are:
powerpc
i386
x86_64
arm
qemu
mips
You define the
KARCH
variable in the BSP Descriptions.- KBRANCH
A regular expression used by the build process to explicitly identify the kernel branch that is validated, patched, and configured during a build. You must set this variable to ensure the exact kernel branch you want is being used by the build process.
Values for this variable are set in the kernel’s recipe file and the kernel’s append file. For example, if you are using the
linux-yocto_4.12
kernel, the kernel recipe file is themeta/recipes-kernel/linux/linux-yocto_4.12.bb
file.KBRANCH
is set as follows in that kernel recipe file:KBRANCH ?= "standard/base"
This variable is also used from the kernel’s append file to identify the kernel branch specific to a particular machine or target hardware. Continuing with the previous kernel example, the kernel’s append file (i.e.
linux-yocto_4.12.bbappend
) is located in the BSP layer for a given machine. For example, the append file for the Beaglebone, EdgeRouter, and generic versions of both 32 and 64-bit IA machines (meta-yocto-bsp
) is namedmeta-yocto-bsp/recipes-kernel/linux/linux-yocto_4.12.bbappend
. Here are the related statements from that append file:KBRANCH_genericx86 = "standard/base" KBRANCH_genericx86-64 = "standard/base" KBRANCH_edgerouter = "standard/edgerouter" KBRANCH_beaglebone = "standard/beaglebone"
The
KBRANCH
statements identify the kernel branch to use when building for each supported BSP.- KBUILD_DEFCONFIG
When used with the kernel-yocto class, specifies an “in-tree” kernel configuration file for use during a kernel build.
Typically, when using a
defconfig
to configure a kernel during a build, you place the file in your layer in the same manner as you would place patch files and configuration fragment files (i.e. “out-of-tree”). However, if you want to use adefconfig
file that is part of the kernel tree (i.e. “in-tree”), you can use theKBUILD_DEFCONFIG
variable and append the KMACHINE variable to point to thedefconfig
file.To use the variable, set it in the append file for your kernel recipe using the following form:
KBUILD_DEFCONFIG_KMACHINE ?= defconfig_file
Here is an example from a “raspberrypi2”
KMACHINE
build that uses adefconfig
file named “bcm2709_defconfig”:KBUILD_DEFCONFIG_raspberrypi2 = "bcm2709_defconfig"
As an alternative, you can use the following within your append file:
KBUILD_DEFCONFIG_pn-linux-yocto ?= defconfig_file
For more information on how to use the
KBUILD_DEFCONFIG
variable, see the “Using an “In-Tree” defconfig File” section in the Yocto Project Linux Kernel Development Manual.- KERNEL_ALT_IMAGETYPE
Specifies an alternate kernel image type for creation in addition to the kernel image type specified using the KERNEL_IMAGETYPE variable.
- KERNEL_ARTIFACT_NAME
Specifies the name of all of the build artifacts. You can change the name of the artifacts by changing the
KERNEL_ARTIFACT_NAME
variable.The value of
KERNEL_ARTIFACT_NAME
, which is set in themeta/classes/kernel-artifact-names.bbclass
file, has the following default value:KERNEL_ARTIFACT_NAME ?= "${PKGE}-${PKGV}-${PKGR}-${MACHINE}${IMAGE_VERSION_SUFFIX}"
See the PKGE, PKGV, PKGR, MACHINE and IMAGE_VERSION_SUFFIX variables for additional information.
- KERNEL_CLASSES
A list of classes defining kernel image types that the kernel class should inherit. You typically append this variable to enable extended image types. An example is the “kernel-fitimage”, which enables fitImage support and resides in
meta/classes/kernel-fitimage.bbclass
. You can register custom kernel image types with thekernel
class using this variable.- KERNEL_DEVICETREE
Specifies the name of the generated Linux kernel device tree (i.e. the
.dtb
) file.Note
Legacy support exists for specifying the full path to the device tree. However, providing just the
.dtb
file is preferred.In order to use this variable, the kernel-devicetree class must be inherited.
- KERNEL_DTB_LINK_NAME
The link name of the kernel device tree binary (DTB). This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:KERNEL_DTB_LINK_NAME ?= "${KERNEL_ARTIFACT_LINK_NAME}"
The value of the
KERNEL_ARTIFACT_LINK_NAME
variable, which is set in the same file, has the following value:KERNEL_ARTIFACT_LINK_NAME ?= "${MACHINE}"
See the MACHINE variable for additional information.
- KERNEL_DTB_NAME
The base name of the kernel device tree binary (DTB). This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:KERNEL_DTB_NAME ?= "${KERNEL_ARTIFACT_NAME}"
The value of the KERNEL_ARTIFACT_NAME variable, which is set in the same file, has the following value:
KERNEL_ARTIFACT_NAME ?= "${PKGE}-${PKGV}-${PKGR}-${MACHINE}${IMAGE_VERSION_SUFFIX}"
- KERNEL_DTC_FLAGS
Specifies the
dtc
flags that are passed to the Linux kernel build system when generating the device trees (viaDTC_FLAGS
environment variable).In order to use this variable, the kernel-devicetree class must be inherited.
- KERNEL_EXTRA_ARGS
Specifies additional
make
command-line arguments the OpenEmbedded build system passes on when compiling the kernel.- KERNEL_FEATURES
Includes additional kernel metadata. In the OpenEmbedded build system, the default Board Support Packages (BSPs) Metadata is provided through the KMACHINE and KBRANCH variables. You can use the
KERNEL_FEATURES
variable from within the kernel recipe or kernel append file to further add metadata for all BSPs or specific BSPs.The metadata you add through this variable includes config fragments and features descriptions, which usually includes patches as well as config fragments. You typically override the
KERNEL_FEATURES
variable for a specific machine. In this way, you can provide validated, but optional, sets of kernel configurations and features.For example, the following example from the
linux-yocto-rt_4.12
kernel recipe adds “netfilter” and “taskstats” features to all BSPs as well as “virtio” configurations to all QEMU machines. The last two statements add specific configurations to targeted machine types:KERNEL_EXTRA_FEATURES ?= "features/netfilter/netfilter.scc features/taskstats/taskstats.scc" KERNEL_FEATURES_append = "${KERNEL_EXTRA_FEATURES}" KERNEL_FEATURES_append_qemuall = "cfg/virtio.scc" KERNEL_FEATURES_append_qemux86 = " cfg/sound.scc cfg/paravirt_kvm.scc" KERNEL_FEATURES_append_qemux86-64 = "cfg/sound.scc"
- KERNEL_FIT_LINK_NAME
The link name of the kernel flattened image tree (FIT) image. This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:KERNEL_FIT_LINK_NAME ?= "${KERNEL_ARTIFACT_LINK_NAME}"
The value of the
KERNEL_ARTIFACT_LINK_NAME
variable, which is set in the same file, has the following value:KERNEL_ARTIFACT_LINK_NAME ?= "${MACHINE}"
See the MACHINE variable for additional information.
- KERNEL_FIT_NAME
The base name of the kernel flattened image tree (FIT) image. This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:KERNEL_FIT_NAME ?= "${KERNEL_ARTIFACT_NAME}"
The value of the KERNEL_ARTIFACT_NAME variable, which is set in the same file, has the following value:
KERNEL_ARTIFACT_NAME ?= "${PKGE}-${PKGV}-${PKGR}-${MACHINE}${IMAGE_VERSION_SUFFIX}"
- KERNEL_IMAGE_LINK_NAME
The link name for the kernel image. This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:KERNEL_IMAGE_LINK_NAME ?= "${KERNEL_ARTIFACT_LINK_NAME}"
The value of the
KERNEL_ARTIFACT_LINK_NAME
variable, which is set in the same file, has the following value:KERNEL_ARTIFACT_LINK_NAME ?= "${MACHINE}"
See the MACHINE variable for additional information.
- KERNEL_IMAGE_MAXSIZE
Specifies the maximum size of the kernel image file in kilobytes. If
KERNEL_IMAGE_MAXSIZE
is set, the size of the kernel image file is checked against the set value during the do_sizecheck task. The task fails if the kernel image file is larger than the setting.KERNEL_IMAGE_MAXSIZE
is useful for target devices that have a limited amount of space in which the kernel image must be stored.By default, this variable is not set, which means the size of the kernel image is not checked.
- KERNEL_IMAGE_NAME
The base name of the kernel image. This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:KERNEL_IMAGE_NAME ?= "${KERNEL_ARTIFACT_NAME}"
The value of the KERNEL_ARTIFACT_NAME variable, which is set in the same file, has the following value:
KERNEL_ARTIFACT_NAME ?= "${PKGE}-${PKGV}-${PKGR}-${MACHINE}${IMAGE_VERSION_SUFFIX}"
- KERNEL_IMAGETYPE
The type of kernel to build for a device, usually set by the machine configuration files and defaults to “zImage”. This variable is used when building the kernel and is passed to
make
as the target to build.If you want to build an alternate kernel image type in addition to that specified by
KERNEL_IMAGETYPE
, use the KERNEL_ALT_IMAGETYPE variable.- KERNEL_MODULE_AUTOLOAD
Lists kernel modules that need to be auto-loaded during boot.
Note
This variable replaces the deprecated module_autoload variable.
You can use the
KERNEL_MODULE_AUTOLOAD
variable anywhere that it can be recognized by the kernel recipe or by an out-of-tree kernel module recipe (e.g. a machine configuration file, a distribution configuration file, an append file for the recipe, or the recipe itself).Specify it as follows:
KERNEL_MODULE_AUTOLOAD += "module_name1 module_name2 module_name3"
Including
KERNEL_MODULE_AUTOLOAD
causes the OpenEmbedded build system to populate the/etc/modules-load.d/modname.conf
file with the list of modules to be auto-loaded on boot. The modules appear one-per-line in the file. Here is an example of the most common use case:KERNEL_MODULE_AUTOLOAD += "module_name"
For information on how to populate the
modname.conf
file withmodprobe.d
syntax lines, see the KERNEL_MODULE_PROBECONF variable.- KERNEL_MODULE_PROBECONF
Provides a list of modules for which the OpenEmbedded build system expects to find
module_conf_
modname values that specify configuration for each of the modules. For information on how to provide those module configurations, see the module_conf_* variable.- KERNEL_PATH
The location of the kernel sources. This variable is set to the value of the STAGING_KERNEL_DIR within the module class. For information on how this variable is used, see the “Incorporating Out-of-Tree Modules” section in the Yocto Project Linux Kernel Development Manual.
To help maximize compatibility with out-of-tree drivers used to build modules, the OpenEmbedded build system also recognizes and uses the KERNEL_SRC variable, which is identical to the
KERNEL_PATH
variable. Both variables are common variables used by external Makefiles to point to the kernel source directory.- KERNEL_SRC
The location of the kernel sources. This variable is set to the value of the STAGING_KERNEL_DIR within the module class. For information on how this variable is used, see the “Incorporating Out-of-Tree Modules” section in the Yocto Project Linux Kernel Development Manual.
To help maximize compatibility with out-of-tree drivers used to build modules, the OpenEmbedded build system also recognizes and uses the KERNEL_PATH variable, which is identical to the
KERNEL_SRC
variable. Both variables are common variables used by external Makefiles to point to the kernel source directory.- KERNEL_VERSION
Specifies the version of the kernel as extracted from
version.h
orutsrelease.h
within the kernel sources. Effects of setting this variable do not take affect until the kernel has been configured. Consequently, attempting to refer to this variable in contexts prior to configuration will not work.- KERNELDEPMODDEPEND
Specifies whether the data referenced through PKGDATA_DIR is needed or not. The
KERNELDEPMODDEPEND
does not control whether or not that data exists, but simply whether or not it is used. If you do not need to use the data, set theKERNELDEPMODDEPEND
variable in yourinitramfs
recipe. Setting the variable there when the data is not needed avoids a potential dependency loop.- KFEATURE_DESCRIPTION
Provides a short description of a configuration fragment. You use this variable in the
.scc
file that describes a configuration fragment file. Here is the variable used in a file namedsmp.scc
to describe SMP being enabled:define KFEATURE_DESCRIPTION "Enable SMP"
- KMACHINE
The machine as known by the kernel. Sometimes the machine name used by the kernel does not match the machine name used by the OpenEmbedded build system. For example, the machine name that the OpenEmbedded build system understands as
core2-32-intel-common
goes by a different name in the Linux Yocto kernel. The kernel understands that machine asintel-core2-32
. For cases like these, theKMACHINE
variable maps the kernel machine name to the OpenEmbedded build system machine name.These mappings between different names occur in the Yocto Linux Kernel’s
meta
branch. As an example take a look in thecommon/recipes-kernel/linux/linux-yocto_3.19.bbappend
file:LINUX_VERSION_core2-32-intel-common = "3.19.0" COMPATIBLE_MACHINE_core2-32-intel-common = "${MACHINE}" SRCREV_meta_core2-32-intel-common = "8897ef68b30e7426bc1d39895e71fb155d694974" SRCREV_machine_core2-32-intel-common = "43b9eced9ba8a57add36af07736344dcc383f711" KMACHINE_core2-32-intel-common = "intel-core2-32" KBRANCH_core2-32-intel-common = "standard/base" KERNEL_FEATURES_append_core2-32-intel-common = "${KERNEL_FEATURES_INTEL_COMMON}"
The
KMACHINE
statement says that the kernel understands the machine name as “intel-core2-32”. However, the OpenEmbedded build system understands the machine as “core2-32-intel-common”.- KTYPE
Defines the kernel type to be used in assembling the configuration. The linux-yocto recipes define “standard”, “tiny”, and “preempt-rt” kernel types. See the “Kernel Types” section in the Yocto Project Linux Kernel Development Manual for more information on kernel types.
You define the
KTYPE
variable in the BSP Descriptions. The value you use must match the value used for the LINUX_KERNEL_TYPE value used by the kernel recipe.- LABELS
Provides a list of targets for automatic configuration.
See the grub-efi class for more information on how this variable is used.
- LAYERDEPENDS
Lists the layers, separated by spaces, on which this recipe depends. Optionally, you can specify a specific layer version for a dependency by adding it to the end of the layer name. Here is an example:
LAYERDEPENDS_mylayer = "anotherlayer (=3)"
In this previous example, version 3 of “anotherlayer” is compared against LAYERVERSION
_anotherlayer
.An error is produced if any dependency is missing or the version numbers (if specified) do not match exactly. This variable is used in the
conf/layer.conf
file and must be suffixed with the name of the specific layer (e.g.LAYERDEPENDS_mylayer
).- LAYERDIR
When used inside the
layer.conf
configuration file, this variable provides the path of the current layer. This variable is not available outside oflayer.conf
and references are expanded immediately when parsing of the file completes.- LAYERRECOMMENDS
Lists the layers, separated by spaces, recommended for use with this layer.
Optionally, you can specify a specific layer version for a recommendation by adding the version to the end of the layer name. Here is an example:
LAYERRECOMMENDS_mylayer = "anotherlayer (=3)"
In this previous example, version 3 of “anotherlayer” is compared against
LAYERVERSION_anotherlayer
.This variable is used in the
conf/layer.conf
file and must be suffixed with the name of the specific layer (e.g.LAYERRECOMMENDS_mylayer
).- LAYERSERIES_COMPAT
Lists the versions of the OpenEmbedded-Core (OE-Core) for which a layer is compatible. Using the
LAYERSERIES_COMPAT
variable allows the layer maintainer to indicate which combinations of the layer and OE-Core can be expected to work. The variable gives the system a way to detect when a layer has not been tested with new releases of OE-Core (e.g. the layer is not maintained).To specify the OE-Core versions for which a layer is compatible, use this variable in your layer’s
conf/layer.conf
configuration file. For the list, use the Yocto Project Release Name (e.g. hardknott). To specify multiple OE-Core versions for the layer, use a space-separated list:LAYERSERIES_COMPAT_layer_root_name = "hardknott gatesgarth"
Note
Setting
LAYERSERIES_COMPAT
is required by the Yocto Project Compatible version 2 standard. The OpenEmbedded build system produces a warning if the variable is not set for any given layer.See the “Creating Your Own Layer” section in the Yocto Project Development Tasks Manual.
- LAYERVERSION
Optionally specifies the version of a layer as a single number. You can use this within LAYERDEPENDS for another layer in order to depend on a specific version of the layer. This variable is used in the
conf/layer.conf
file and must be suffixed with the name of the specific layer (e.g.LAYERVERSION_mylayer
).- LD
The minimal command and arguments used to run the linker.
- LDFLAGS
Specifies the flags to pass to the linker. This variable is exported to an environment variable and thus made visible to the software being built during the compilation step.
Default initialization for
LDFLAGS
varies depending on what is being built:TARGET_LDFLAGS when building for the target
BUILD_LDFLAGS when building for the build host (i.e.
-native
)BUILDSDK_LDFLAGS when building for an SDK (i.e.
nativesdk-
)
- LEAD_SONAME
Specifies the lead (or primary) compiled library file (i.e.
.so
) that the debian class applies its naming policy to given a recipe that packages multiple libraries.This variable works in conjunction with the
debian
class.- LIC_FILES_CHKSUM
Checksums of the license text in the recipe source code.
This variable tracks changes in license text of the source code files. If the license text is changed, it will trigger a build failure, which gives the developer an opportunity to review any license change.
This variable must be defined for all recipes (unless LICENSE is set to “CLOSED”).
For more information, see the “Tracking License Changes” section in the Yocto Project Development Tasks Manual.
- LICENSE
The list of source licenses for the recipe. Follow these rules:
Do not use spaces within individual license names.
Separate license names using | (pipe) when there is a choice between licenses.
Separate license names using & (ampersand) when multiple licenses exist that cover different parts of the source.
You can use spaces between license names.
For standard licenses, use the names of the files in
meta/files/common-licenses/
or the SPDXLICENSEMAP flag names defined inmeta/conf/licenses.conf
.
Here are some examples:
LICENSE = "LGPLv2.1 | GPLv3" LICENSE = "MPL-1 & LGPLv2.1" LICENSE = "GPLv2+"
The first example is from the recipes for Qt, which the user may choose to distribute under either the LGPL version 2.1 or GPL version 3. The second example is from Cairo where two licenses cover different parts of the source code. The final example is from
sysstat
, which presents a single license.You can also specify licenses on a per-package basis to handle situations where components of the output have different licenses. For example, a piece of software whose code is licensed under GPLv2 but has accompanying documentation licensed under the GNU Free Documentation License 1.2 could be specified as follows:
LICENSE = "GFDL-1.2 & GPLv2" LICENSE_${PN} = "GPLv2" LICENSE_${PN}-doc = "GFDL-1.2"
- LICENSE_CREATE_PACKAGE
Setting
LICENSE_CREATE_PACKAGE
to “1” causes the OpenEmbedded build system to create an extra package (i.e.${
PN}-lic
) for each recipe and to add those packages to the RRECOMMENDS_${PN}
.The
${PN}-lic
package installs a directory in/usr/share/licenses
named${PN}
, which is the recipe’s base name, and installs files in that directory that contain license and copyright information (i.e. copies of the appropriate license files frommeta/common-licenses
that match the licenses specified in the LICENSE variable of the recipe metadata and copies of files marked in LIC_FILES_CHKSUM as containing license text).For related information on providing license text, see the COPY_LIC_DIRS variable, the COPY_LIC_MANIFEST variable, and the “Providing License Text” section in the Yocto Project Development Tasks Manual.
- LICENSE_FLAGS
Specifies additional flags for a recipe you must whitelist through LICENSE_FLAGS_WHITELIST in order to allow the recipe to be built. When providing multiple flags, separate them with spaces.
This value is independent of LICENSE and is typically used to mark recipes that might require additional licenses in order to be used in a commercial product. For more information, see the “Enabling Commercially Licensed Recipes” section in the Yocto Project Development Tasks Manual.
- LICENSE_FLAGS_WHITELIST
Lists license flags that when specified in LICENSE_FLAGS within a recipe should not prevent that recipe from being built. This practice is otherwise known as “whitelisting” license flags. For more information, see the “Enabling Commercially Licensed Recipes” section in the Yocto Project Development Tasks Manual.
- LICENSE_PATH
Path to additional licenses used during the build. By default, the OpenEmbedded build system uses
COMMON_LICENSE_DIR
to define the directory that holds common license text used during the build. TheLICENSE_PATH
variable allows you to extend that location to other areas that have additional licenses:LICENSE_PATH += "path-to-additional-common-licenses"
- LINUX_KERNEL_TYPE
Defines the kernel type to be used in assembling the configuration. The linux-yocto recipes define “standard”, “tiny”, and “preempt-rt” kernel types. See the “Kernel Types” section in the Yocto Project Linux Kernel Development Manual for more information on kernel types.
If you do not specify a
LINUX_KERNEL_TYPE
, it defaults to “standard”. Together with KMACHINE, theLINUX_KERNEL_TYPE
variable defines the search arguments used by the kernel tools to find the appropriate description within the kernel Metadata with which to build out the sources and configuration.- LINUX_VERSION
The Linux version from
kernel.org
on which the Linux kernel image being built using the OpenEmbedded build system is based. You define this variable in the kernel recipe. For example, thelinux-yocto-3.4.bb
kernel recipe found inmeta/recipes-kernel/linux
defines the variables as follows:LINUX_VERSION ?= "3.4.24"
The
LINUX_VERSION
variable is used to define PV for the recipe:PV = "${LINUX_VERSION}+git${SRCPV}"
- LINUX_VERSION_EXTENSION
A string extension compiled into the version string of the Linux kernel built with the OpenEmbedded build system. You define this variable in the kernel recipe. For example, the linux-yocto kernel recipes all define the variable as follows:
LINUX_VERSION_EXTENSION ?= "-yocto-${LINUX_KERNEL_TYPE}"
Defining this variable essentially sets the Linux kernel configuration item
CONFIG_LOCALVERSION
, which is visible through theuname
command. Here is an example that shows the extension assuming it was set as previously shown:$ uname -r 3.7.0-rc8-custom
- LOG_DIR
Specifies the directory to which the OpenEmbedded build system writes overall log files. The default directory is
${TMPDIR}/log
.For the directory containing logs specific to each task, see the T variable.
- MACHINE
Specifies the target device for which the image is built. You define
MACHINE
in thelocal.conf
file found in the Build Directory. By default,MACHINE
is set to “qemux86”, which is an x86-based architecture machine to be emulated using QEMU:MACHINE ?= "qemux86"
The variable corresponds to a machine configuration file of the same name, through which machine-specific configurations are set. Thus, when
MACHINE
is set to “qemux86” there exists the correspondingqemux86.conf
machine configuration file, which can be found in the Source Directory inmeta/conf/machine
.The list of machines supported by the Yocto Project as shipped include the following:
MACHINE ?= "qemuarm" MACHINE ?= "qemuarm64" MACHINE ?= "qemumips" MACHINE ?= "qemumips64" MACHINE ?= "qemuppc" MACHINE ?= "qemux86" MACHINE ?= "qemux86-64" MACHINE ?= "genericx86" MACHINE ?= "genericx86-64" MACHINE ?= "beaglebone" MACHINE ?= "edgerouter"
The last five are Yocto Project reference hardware boards, which are provided in the
meta-yocto-bsp
layer.Note
Adding additional Board Support Package (BSP) layers to your configuration adds new possible settings for
MACHINE
.- MACHINE_ARCH
Specifies the name of the machine-specific architecture. This variable is set automatically from MACHINE or TUNE_PKGARCH. You should not hand-edit the
MACHINE_ARCH
variable.- MACHINE_ESSENTIAL_EXTRA_RDEPENDS
A list of required machine-specific packages to install as part of the image being built. The build process depends on these packages being present. Furthermore, because this is a “machine-essential” variable, the list of packages are essential for the machine to boot. The impact of this variable affects images based on
packagegroup-core-boot
, including thecore-image-minimal
image.This variable is similar to the
MACHINE_ESSENTIAL_EXTRA_RRECOMMENDS
variable with the exception that the image being built has a build dependency on the variable’s list of packages. In other words, the image will not build if a file in this list is not found.As an example, suppose the machine for which you are building requires
example-init
to be run during boot to initialize the hardware. In this case, you would use the following in the machine’s.conf
configuration file:MACHINE_ESSENTIAL_EXTRA_RDEPENDS += "example-init"
- MACHINE_ESSENTIAL_EXTRA_RRECOMMENDS
A list of recommended machine-specific packages to install as part of the image being built. The build process does not depend on these packages being present. However, because this is a “machine-essential” variable, the list of packages are essential for the machine to boot. The impact of this variable affects images based on
packagegroup-core-boot
, including thecore-image-minimal
image.This variable is similar to the
MACHINE_ESSENTIAL_EXTRA_RDEPENDS
variable with the exception that the image being built does not have a build dependency on the variable’s list of packages. In other words, the image will still build if a package in this list is not found. Typically, this variable is used to handle essential kernel modules, whose functionality may be selected to be built into the kernel rather than as a module, in which case a package will not be produced.Consider an example where you have a custom kernel where a specific touchscreen driver is required for the machine to be usable. However, the driver can be built as a module or into the kernel depending on the kernel configuration. If the driver is built as a module, you want it to be installed. But, when the driver is built into the kernel, you still want the build to succeed. This variable sets up a “recommends” relationship so that in the latter case, the build will not fail due to the missing package. To accomplish this, assuming the package for the module was called
kernel-module-ab123
, you would use the following in the machine’s.conf
configuration file:MACHINE_ESSENTIAL_EXTRA_RRECOMMENDS += "kernel-module-ab123"
Note
In this example, the
kernel-module-ab123
recipe needs to explicitly set its PACKAGES variable to ensure that BitBake does not use the kernel recipe’s PACKAGES_DYNAMIC variable to satisfy the dependency.Some examples of these machine essentials are flash, screen, keyboard, mouse, or touchscreen drivers (depending on the machine).
- MACHINE_EXTRA_RDEPENDS
A list of machine-specific packages to install as part of the image being built that are not essential for the machine to boot. However, the build process for more fully-featured images depends on the packages being present.
This variable affects all images based on
packagegroup-base
, which does not include thecore-image-minimal
orcore-image-full-cmdline
images.The variable is similar to the
MACHINE_EXTRA_RRECOMMENDS
variable with the exception that the image being built has a build dependency on the variable’s list of packages. In other words, the image will not build if a file in this list is not found.An example is a machine that has WiFi capability but is not essential for the machine to boot the image. However, if you are building a more fully-featured image, you want to enable the WiFi. The package containing the firmware for the WiFi hardware is always expected to exist, so it is acceptable for the build process to depend upon finding the package. In this case, assuming the package for the firmware was called
wifidriver-firmware
, you would use the following in the.conf
file for the machine:MACHINE_EXTRA_RDEPENDS += "wifidriver-firmware"
- MACHINE_EXTRA_RRECOMMENDS
A list of machine-specific packages to install as part of the image being built that are not essential for booting the machine. The image being built has no build dependency on this list of packages.
This variable affects only images based on
packagegroup-base
, which does not include thecore-image-minimal
orcore-image-full-cmdline
images.This variable is similar to the
MACHINE_EXTRA_RDEPENDS
variable with the exception that the image being built does not have a build dependency on the variable’s list of packages. In other words, the image will build if a file in this list is not found.An example is a machine that has WiFi capability but is not essential For the machine to boot the image. However, if you are building a more fully-featured image, you want to enable WiFi. In this case, the package containing the WiFi kernel module will not be produced if the WiFi driver is built into the kernel, in which case you still want the build to succeed instead of failing as a result of the package not being found. To accomplish this, assuming the package for the module was called
kernel-module-examplewifi
, you would use the following in the.conf
file for the machine:MACHINE_EXTRA_RRECOMMENDS += "kernel-module-examplewifi"
- MACHINE_FEATURES
Specifies the list of hardware features the MACHINE is capable of supporting. For related information on enabling features, see the DISTRO_FEATURES, COMBINED_FEATURES, and IMAGE_FEATURES variables.
For a list of hardware features supported by the Yocto Project as shipped, see the “Machine Features” section.
- MACHINE_FEATURES_BACKFILL
Features to be added to
MACHINE_FEATURES
if not also present inMACHINE_FEATURES_BACKFILL_CONSIDERED
.This variable is set in the
meta/conf/bitbake.conf
file. It is not intended to be user-configurable. It is best to just reference the variable to see which machine features are being backfilled for all machine configurations. See the “Feature Backfilling” section for more information.- MACHINE_FEATURES_BACKFILL_CONSIDERED
Features from
MACHINE_FEATURES_BACKFILL
that should not be backfilled (i.e. added toMACHINE_FEATURES
) during the build. See the “Feature Backfilling” section for more information.- MACHINEOVERRIDES
A colon-separated list of overrides that apply to the current machine. By default, this list includes the value of MACHINE.
You can extend
MACHINEOVERRIDES
to add extra overrides that should apply to a machine. For example, all machines emulated in QEMU (e.g.qemuarm
,qemux86
, and so forth) include a file namedmeta/conf/machine/include/qemu.inc
that prepends the following override toMACHINEOVERRIDES
:MACHINEOVERRIDES =. "qemuall:"
This override allows variables to be overridden for all machines emulated in QEMU, like in the following example from the
connman-conf
recipe:SRC_URI_append_qemuall = " file://wired.config \ file://wired-setup \ "
The underlying mechanism behind
MACHINEOVERRIDES
is simply that it is included in the default value of OVERRIDES.- MAINTAINER
The email address of the distribution maintainer.
- METADATA_BRANCH
The branch currently checked out for the OpenEmbedded-Core layer (path determined by COREBASE).
- METADATA_REVISION
The revision currently checked out for the OpenEmbedded-Core layer (path determined by COREBASE).
- MIRRORS
Specifies additional paths from which the OpenEmbedded build system gets source code. When the build system searches for source code, it first tries the local download directory. If that location fails, the build system tries locations defined by PREMIRRORS, the upstream source, and then locations specified by
MIRRORS
in that order.Assuming your distribution (DISTRO) is “poky”, the default value for
MIRRORS
is defined in theconf/distro/poky.conf
file in themeta-poky
Git repository.- MLPREFIX
Specifies a prefix has been added to PN to create a special version of a recipe or package (i.e. a Multilib version). The variable is used in places where the prefix needs to be added to or removed from a the name (e.g. the BPN variable).
MLPREFIX
gets set when a prefix has been added toPN
.Note
The “ML” in
MLPREFIX
stands for “MultiLib”. This representation is historical and comes from a time whennativesdk
was a suffix rather than a prefix on the recipe name. Whennativesdk
was turned into a prefix, it made sense to setMLPREFIX
for it as well.To help understand when
MLPREFIX
might be needed, consider when BBCLASSEXTEND is used to provide anativesdk
version of a recipe in addition to the target version. If that recipe declares build-time dependencies on tasks in other recipes by using DEPENDS, then a dependency on “foo” will automatically get rewritten to a dependency on “nativesdk-foo”. However, dependencies like the following will not get rewritten automatically:do_foo[depends] += "recipe:do_foo"
If you want such a dependency to also get transformed, you can do the following:
do_foo[depends] += "${MLPREFIX}recipe:do_foo"
- module_autoload
This variable has been replaced by the
KERNEL_MODULE_AUTOLOAD
variable. You should replace all occurrences ofmodule_autoload
with additions toKERNEL_MODULE_AUTOLOAD
, for example:module_autoload_rfcomm = "rfcomm"
should now be replaced with:
KERNEL_MODULE_AUTOLOAD += "rfcomm"
See the KERNEL_MODULE_AUTOLOAD variable for more information.
- module_conf
Specifies modprobe.d syntax lines for inclusion in the
/etc/modprobe.d/modname.conf
file.You can use this variable anywhere that it can be recognized by the kernel recipe or out-of-tree kernel module recipe (e.g. a machine configuration file, a distribution configuration file, an append file for the recipe, or the recipe itself). If you use this variable, you must also be sure to list the module name in the KERNEL_MODULE_AUTOLOAD variable.
Here is the general syntax:
module_conf_module_name = "modprobe.d-syntax"
You must use the kernel module name override.
Run
man modprobe.d
in the shell to find out more information on the exact syntax you want to provide withmodule_conf
.Including
module_conf
causes the OpenEmbedded build system to populate the/etc/modprobe.d/modname.conf
file withmodprobe.d
syntax lines. Here is an example that adds the optionsarg1
andarg2
to a module namedmymodule
:module_conf_mymodule = "options mymodule arg1=val1 arg2=val2"
For information on how to specify kernel modules to auto-load on boot, see the KERNEL_MODULE_AUTOLOAD variable.
- MODULE_TARBALL_DEPLOY
Controls creation of the
modules-*.tgz
file. Set this variable to “0” to disable creation of this file, which contains all of the kernel modules resulting from a kernel build.- MODULE_TARBALL_LINK_NAME
The link name of the kernel module tarball. This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:MODULE_TARBALL_LINK_NAME ?= "${KERNEL_ARTIFACT_LINK_NAME}"
The value of the
KERNEL_ARTIFACT_LINK_NAME
variable, which is set in the same file, has the following value:KERNEL_ARTIFACT_LINK_NAME ?= "${MACHINE}"
See the MACHINE variable for additional information.
- MODULE_TARBALL_NAME
The base name of the kernel module tarball. This variable is set in the
meta/classes/kernel-artifact-names.bbclass
file as follows:MODULE_TARBALL_NAME ?= "${KERNEL_ARTIFACT_NAME}"
The value of the KERNEL_ARTIFACT_NAME variable, which is set in the same file, has the following value:
KERNEL_ARTIFACT_NAME ?= "${PKGE}-${PKGV}-${PKGR}-${MACHINE}${IMAGE_VERSION_SUFFIX}"
- MULTIMACH_TARGET_SYS
Uniquely identifies the type of the target system for which packages are being built. This variable allows output for different types of target systems to be put into different subdirectories of the same output directory.
The default value of this variable is:
${PACKAGE_ARCH}${TARGET_VENDOR}-${TARGET_OS}
Some classes (e.g. cross-canadian) modify the
MULTIMACH_TARGET_SYS
value.See the STAMP variable for an example. See the STAGING_DIR_TARGET variable for more information.
- NATIVELSBSTRING
A string identifying the host distribution. Strings consist of the host distributor ID followed by the release, as reported by the
lsb_release
tool or as read from/etc/lsb-release
. For example, when running a build on Ubuntu 12.10, the value is “Ubuntu-12.10”. If this information is unable to be determined, the value resolves to “Unknown”.This variable is used by default to isolate native shared state packages for different distributions (e.g. to avoid problems with
glibc
version incompatibilities). Additionally, the variable is checked against SANITY_TESTED_DISTROS if that variable is set.- NM
The minimal command and arguments to run
nm
.- NO_GENERIC_LICENSE
Avoids QA errors when you use a non-common, non-CLOSED license in a recipe. Packages exist, such as the linux-firmware package, with many licenses that are not in any way common. Also, new licenses are added occasionally to avoid introducing a lot of common license files, which are only applicable to a specific package.
NO_GENERIC_LICENSE
is used to allow copying a license that does not exist in common licenses.The following example shows how to add
NO_GENERIC_LICENSE
to a recipe:NO_GENERIC_LICENSE[license_name] = "license_file_in_fetched_source"
The following is an example that uses the
LICENSE.Abilis.txt
file as the license from the fetched source:NO_GENERIC_LICENSE[Firmware-Abilis] = "LICENSE.Abilis.txt"
- NO_RECOMMENDATIONS
Prevents installation of all “recommended-only” packages. Recommended-only packages are packages installed only through the RRECOMMENDS variable). Setting the
NO_RECOMMENDATIONS
variable to “1” turns this feature on:NO_RECOMMENDATIONS = "1"
You can set this variable globally in your
local.conf
file or you can attach it to a specific image recipe by using the recipe name override:NO_RECOMMENDATIONS_pn-target_image = "1"
It is important to realize that if you choose to not install packages using this variable and some other packages are dependent on them (i.e. listed in a recipe’s RDEPENDS variable), the OpenEmbedded build system ignores your request and will install the packages to avoid dependency errors.
Note
Some recommended packages might be required for certain system functionality, such as kernel modules. It is up to you to add packages with the IMAGE_INSTALL variable.
Support for this variable exists only when using the IPK and RPM packaging backend. Support does not exist for DEB.
See the BAD_RECOMMENDATIONS and the PACKAGE_EXCLUDE variables for related information.
- NOAUTOPACKAGEDEBUG
Disables auto package from splitting
.debug
files. If a recipe requiresFILES_${PN}-dbg
to be set manually, theNOAUTOPACKAGEDEBUG
can be defined allowing you to define the content of the debug package. For example:NOAUTOPACKAGEDEBUG = "1" FILES_${PN}-dev = "${includedir}/${QT_DIR_NAME}/Qt/*" FILES_${PN}-dbg = "/usr/src/debug/" FILES_${QT_BASE_NAME}-demos-doc = "${docdir}/${QT_DIR_NAME}/qch/qt.qch"
- OBJCOPY
The minimal command and arguments to run
objcopy
.- OBJDUMP
The minimal command and arguments to run
objdump
.- OE_BINCONFIG_EXTRA_MANGLE
When inheriting the binconfig class, this variable specifies additional arguments passed to the “sed” command. The sed command alters any paths in configuration scripts that have been set up during compilation. Inheriting this class results in all paths in these scripts being changed to point into the
sysroots/
directory so that all builds that use the script will use the correct directories for the cross compiling layout.See the
meta/classes/binconfig.bbclass
in the Source Directory for details on how this class applies these additional sed command arguments. For general information on thebinconfig
class, see the “binconfig.bbclass” section.- OE_IMPORTS
An internal variable used to tell the OpenEmbedded build system what Python modules to import for every Python function run by the system.
Note
Do not set this variable. It is for internal use only.
- OE_INIT_ENV_SCRIPT
The name of the build environment setup script for the purposes of setting up the environment within the extensible SDK. The default value is “oe-init-build-env”.
If you use a custom script to set up your build environment, set the
OE_INIT_ENV_SCRIPT
variable to its name.- OE_TERMINAL
Controls how the OpenEmbedded build system spawns interactive terminals on the host development system (e.g. using the BitBake command with the
-c devshell
command-line option). For more information, see the “Using a Development Shell” section in the Yocto Project Development Tasks Manual.You can use the following values for the
OE_TERMINAL
variable:auto
gnome
xfce
rxvt
screen
konsole
none
- OEROOT
The directory from which the top-level build environment setup script is sourced. The Yocto Project provides a top-level build environment setup script: oe-init-build-env. When you run this script, the
OEROOT
variable resolves to the directory that contains the script.For additional information on how this variable is used, see the initialization script.
- OLDEST_KERNEL
Declares the oldest version of the Linux kernel that the produced binaries must support. This variable is passed into the build of the Embedded GNU C Library (
glibc
).The default for this variable comes from the
meta/conf/bitbake.conf
configuration file. You can override this default by setting the variable in a custom distribution configuration file.- OVERRIDES
A colon-separated list of overrides that currently apply. Overrides are a BitBake mechanism that allows variables to be selectively overridden at the end of parsing. The set of overrides in
OVERRIDES
represents the “state” during building, which includes the current recipe being built, the machine for which it is being built, and so forth.As an example, if the string “an-override” appears as an element in the colon-separated list in
OVERRIDES
, then the following assignment will overrideFOO
with the value “overridden” at the end of parsing:FOO_an-override = "overridden"
See the “Conditional Syntax (Overrides)” section in the BitBake User Manual for more information on the overrides mechanism.
The default value of
OVERRIDES
includes the values of the CLASSOVERRIDE, MACHINEOVERRIDES, and DISTROOVERRIDES variables. Another important override included by default ispn-${PN}
. This override allows variables to be set for a single recipe within configuration (.conf
) files. Here is an example:FOO_pn-myrecipe = "myrecipe-specific value"
Note
An easy way to see what overrides apply is to search for
OVERRIDES
in the output of thebitbake -e
command. See the “Viewing Variable Values” section in the Yocto Project Development Tasks Manual for more information.- P
The recipe name and version.
P
is comprised of the following:${PN}-${PV}
- PACKAGE_ADD_METADATA
This variable defines additional metadata to add to packages.
You may find you need to inject additional metadata into packages. This variable allows you to do that by setting the injected data as the value. Multiple fields can be added by splitting the content with the literal separator “n”.
The suffixes ‘_IPK’, ‘_DEB’, or ‘_RPM’ can be applied to the variable to do package type specific settings. It can also be made package specific by using the package name as a suffix.
You can find out more about applying this variable in the “Adding custom metadata to packages” section in the Yocto Project Development Tasks Manual.
- PACKAGE_ARCH
The architecture of the resulting package or packages.
By default, the value of this variable is set to TUNE_PKGARCH when building for the target, BUILD_ARCH when building for the build host, and “${SDK_ARCH}-${SDKPKGSUFFIX}” when building for the SDK.
Note
See SDK_ARCH for more information.
However, if your recipe’s output packages are built specific to the target machine rather than generally for the architecture of the machine, you should set
PACKAGE_ARCH
to the value of MACHINE_ARCH in the recipe as follows:PACKAGE_ARCH = "${MACHINE_ARCH}"
- PACKAGE_ARCHS
Specifies a list of architectures compatible with the target machine. This variable is set automatically and should not normally be hand-edited. Entries are separated using spaces and listed in order of priority. The default value for
PACKAGE_ARCHS
is “all any noarch ${PACKAGE_EXTRA_ARCHS} ${MACHINE_ARCH}”.- PACKAGE_BEFORE_PN
Enables easily adding packages to
PACKAGES
before${PN}
so that those added packages can pick up files that would normally be included in the default package.- PACKAGE_CLASSES
This variable, which is set in the
local.conf
configuration file found in theconf
folder of the Build Directory, specifies the package manager the OpenEmbedded build system uses when packaging data.You can provide one or more of the following arguments for the variable: PACKAGE_CLASSES ?= “package_rpm package_deb package_ipk package_tar”
Note
While it is a legal option, the
package_tar
class has limited functionality due to no support for package dependencies by that backend. Therefore, it is recommended that you do not use it.The build system uses only the first argument in the list as the package manager when creating your image or SDK. However, packages will be created using any additional packaging classes you specify. For example, if you use the following in your
local.conf
file:PACKAGE_CLASSES ?= "package_ipk"
The OpenEmbedded build system uses the IPK package manager to create your image or SDK.
For information on packaging and build performance effects as a result of the package manager in use, see the “package.bbclass” section.
- PACKAGE_DEBUG_SPLIT_STYLE
Determines how to split up the binary and debug information when creating
*-dbg
packages to be used with the GNU Project Debugger (GDB).With the
PACKAGE_DEBUG_SPLIT_STYLE
variable, you can control where debug information, which can include or exclude source files, is stored:“.debug”: Debug symbol files are placed next to the binary in a
.debug
directory on the target. For example, if a binary is installed into/bin
, the corresponding debug symbol files are installed in/bin/.debug
. Source files are placed in/usr/src/debug
.“debug-file-directory”: Debug symbol files are placed under
/usr/lib/debug
on the target, and separated by the path from where the binary is installed. For example, if a binary is installed in/bin
, the corresponding debug symbols are installed in/usr/lib/debug/bin
. Source files are placed in/usr/src/debug
.“debug-without-src”: The same behavior as “.debug” previously described with the exception that no source files are installed.
“debug-with-srcpkg”: The same behavior as “.debug” previously described with the exception that all source files are placed in a separate
*-src
pkg. This is the default behavior.
You can find out more about debugging using GDB by reading the “Debugging With the GNU Project Debugger (GDB) Remotely” section in the Yocto Project Development Tasks Manual.
- PACKAGE_EXCLUDE_COMPLEMENTARY
Prevents specific packages from being installed when you are installing complementary packages.
You might find that you want to prevent installing certain packages when you are installing complementary packages. For example, if you are using IMAGE_FEATURES to install
dev-pkgs
, you might not want to install all packages from a particular multilib. If you find yourself in this situation, you can use thePACKAGE_EXCLUDE_COMPLEMENTARY
variable to specify regular expressions to match the packages you want to exclude.- PACKAGE_EXCLUDE
Lists packages that should not be installed into an image. For example:
PACKAGE_EXCLUDE = "package_name package_name package_name ..."
You can set this variable globally in your
local.conf
file or you can attach it to a specific image recipe by using the recipe name override:PACKAGE_EXCLUDE_pn-target_image = "package_name"
If you choose to not install a package using this variable and some other package is dependent on it (i.e. listed in a recipe’s RDEPENDS variable), the OpenEmbedded build system generates a fatal installation error. Because the build system halts the process with a fatal error, you can use the variable with an iterative development process to remove specific components from a system.
Support for this variable exists only when using the IPK and RPM packaging backend. Support does not exist for DEB.
See the NO_RECOMMENDATIONS and the BAD_RECOMMENDATIONS variables for related information.
- PACKAGE_EXTRA_ARCHS
Specifies the list of architectures compatible with the device CPU. This variable is useful when you build for several different devices that use miscellaneous processors such as XScale and ARM926-EJS.
- PACKAGE_FEED_ARCHS
Optionally specifies the package architectures used as part of the package feed URIs during the build. When used, the
PACKAGE_FEED_ARCHS
variable is appended to the final package feed URI, which is constructed using the PACKAGE_FEED_URIS and PACKAGE_FEED_BASE_PATHS variables.Note
You can use the
PACKAGE_FEED_ARCHS
variable to whitelist specific package architectures. If you do not need to whitelist specific architectures, which is a common case, you can omit this variable. Omitting the variable results in all available architectures for the current machine being included into remote package feeds.Consider the following example where the
PACKAGE_FEED_URIS
,PACKAGE_FEED_BASE_PATHS
, andPACKAGE_FEED_ARCHS
variables are defined in yourlocal.conf
file:PACKAGE_FEED_URIS = "https://example.com/packagerepos/release \ https://example.com/packagerepos/updates" PACKAGE_FEED_BASE_PATHS = "rpm rpm-dev" PACKAGE_FEED_ARCHS = "all core2-64"
Given these settings, the resulting package feeds are as follows:
https://example.com/packagerepos/release/rpm/all https://example.com/packagerepos/release/rpm/core2-64 https://example.com/packagerepos/release/rpm-dev/all https://example.com/packagerepos/release/rpm-dev/core2-64 https://example.com/packagerepos/updates/rpm/all https://example.com/packagerepos/updates/rpm/core2-64 https://example.com/packagerepos/updates/rpm-dev/all https://example.com/packagerepos/updates/rpm-dev/core2-64
- PACKAGE_FEED_BASE_PATHS
Specifies the base path used when constructing package feed URIs. The
PACKAGE_FEED_BASE_PATHS
variable makes up the middle portion of a package feed URI used by the OpenEmbedded build system. The base path lies between the PACKAGE_FEED_URIS and PACKAGE_FEED_ARCHS variables.Consider the following example where the
PACKAGE_FEED_URIS
,PACKAGE_FEED_BASE_PATHS
, andPACKAGE_FEED_ARCHS
variables are defined in yourlocal.conf
file:PACKAGE_FEED_URIS = "https://example.com/packagerepos/release \ https://example.com/packagerepos/updates" PACKAGE_FEED_BASE_PATHS = "rpm rpm-dev" PACKAGE_FEED_ARCHS = "all core2-64"
Given these settings, the resulting package feeds are as follows:
https://example.com/packagerepos/release/rpm/all https://example.com/packagerepos/release/rpm/core2-64 https://example.com/packagerepos/release/rpm-dev/all https://example.com/packagerepos/release/rpm-dev/core2-64 https://example.com/packagerepos/updates/rpm/all https://example.com/packagerepos/updates/rpm/core2-64 https://example.com/packagerepos/updates/rpm-dev/all https://example.com/packagerepos/updates/rpm-dev/core2-64
- PACKAGE_FEED_URIS
Specifies the front portion of the package feed URI used by the OpenEmbedded build system. Each final package feed URI is comprised of
PACKAGE_FEED_URIS
, PACKAGE_FEED_BASE_PATHS, and PACKAGE_FEED_ARCHS variables.Consider the following example where the
PACKAGE_FEED_URIS
,PACKAGE_FEED_BASE_PATHS
, andPACKAGE_FEED_ARCHS
variables are defined in yourlocal.conf
file:PACKAGE_FEED_URIS = "https://example.com/packagerepos/release \ https://example.com/packagerepos/updates" PACKAGE_FEED_BASE_PATHS = "rpm rpm-dev" PACKAGE_FEED_ARCHS = "all core2-64"
Given these settings, the resulting package feeds are as follows:
https://example.com/packagerepos/release/rpm/all https://example.com/packagerepos/release/rpm/core2-64 https://example.com/packagerepos/release/rpm-dev/all https://example.com/packagerepos/release/rpm-dev/core2-64 https://example.com/packagerepos/updates/rpm/all https://example.com/packagerepos/updates/rpm/core2-64 https://example.com/packagerepos/updates/rpm-dev/all https://example.com/packagerepos/updates/rpm-dev/core2-64
- PACKAGE_INSTALL
The final list of packages passed to the package manager for installation into the image.
Because the package manager controls actual installation of all packages, the list of packages passed using
PACKAGE_INSTALL
is not the final list of packages that are actually installed. This variable is internal to the image construction code. Consequently, in general, you should use the IMAGE_INSTALL variable to specify packages for installation. The exception to this is when working with the core-image-minimal-initramfs image. When working with an initial RAM filesystem (initramfs) image, use thePACKAGE_INSTALL
variable. For information on creating an initramfs, see the “Building an Initial RAM Filesystem (initramfs) Image” section in the Yocto Project Development Tasks Manual.- PACKAGE_INSTALL_ATTEMPTONLY
Specifies a list of packages the OpenEmbedded build system attempts to install when creating an image. If a listed package fails to install, the build system does not generate an error. This variable is generally not user-defined.
- PACKAGE_PREPROCESS_FUNCS
Specifies a list of functions run to pre-process the PKGD directory prior to splitting the files out to individual packages.
- PACKAGE_WRITE_DEPS
Specifies a list of dependencies for post-installation and pre-installation scripts on native/cross tools. If your post-installation or pre-installation script can execute at rootfs creation time rather than on the target but depends on a native tool in order to execute, you need to list the tools in
PACKAGE_WRITE_DEPS
.For information on running post-installation scripts, see the “Post-Installation Scripts” section in the Yocto Project Development Tasks Manual.
- PACKAGECONFIG
This variable provides a means of enabling or disabling features of a recipe on a per-recipe basis.
PACKAGECONFIG
blocks are defined in recipes when you specify features and then arguments that define feature behaviors. Here is the basic block structure (broken over multiple lines for readability):PACKAGECONFIG ??= "f1 f2 f3 ..." PACKAGECONFIG[f1] = "\ --with-f1, \ --without-f1, \ build-deps-for-f1, \ runtime-deps-for-f1, \ runtime-recommends-for-f1, \ packageconfig-conflicts-for-f1" PACKAGECONFIG[f2] = "\ ... and so on and so on ...
The
PACKAGECONFIG
variable itself specifies a space-separated list of the features to enable. Following the features, you can determine the behavior of each feature by providing up to six order-dependent arguments, which are separated by commas. You can omit any argument you like but must retain the separating commas. The order is important and specifies the following:Extra arguments that should be added to the configure script argument list (EXTRA_OECONF or PACKAGECONFIG_CONFARGS) if the feature is enabled.
Extra arguments that should be added to
EXTRA_OECONF
orPACKAGECONFIG_CONFARGS
if the feature is disabled.Additional build dependencies (DEPENDS) that should be added if the feature is enabled.
Additional runtime dependencies (RDEPENDS) that should be added if the feature is enabled.
Additional runtime recommendations (RRECOMMENDS) that should be added if the feature is enabled.
Any conflicting (that is, mutually exclusive)
PACKAGECONFIG
settings for this feature.
Consider the following
PACKAGECONFIG
block taken from thelibrsvg
recipe. In this example the feature isgtk
, which has three arguments that determine the feature’s behavior.PACKAGECONFIG[gtk] = "--with-gtk3,--without-gtk3,gtk+3"
The
--with-gtk3
andgtk+3
arguments apply only if the feature is enabled. In this case,--with-gtk3
is added to the configure script argument list andgtk+3
is added toDEPENDS
. On the other hand, if the feature is disabled say through a.bbappend
file in another layer, then the second argument--without-gtk3
is added to the configure script instead.The basic
PACKAGECONFIG
structure previously described holds true regardless of whether you are creating a block or changing a block. When creating a block, use the structure inside your recipe.If you want to change an existing
PACKAGECONFIG
block, you can do so one of two ways:Append file: Create an append file named recipename
.bbappend
in your layer and override the value ofPACKAGECONFIG
. You can either completely override the variable:PACKAGECONFIG = "f4 f5"
Or, you can just append the variable:
PACKAGECONFIG_append = " f4"
Configuration file: This method is identical to changing the block through an append file except you edit your
local.conf
ormydistro.conf
file. As with append files previously described, you can either completely override the variable:PACKAGECONFIG_pn-recipename = "f4 f5"
Or, you can just amend the variable:
PACKAGECONFIG_append_pn-recipename = " f4"
- PACKAGECONFIG_CONFARGS
A space-separated list of configuration options generated from the PACKAGECONFIG setting.
Classes such as autotools and cmake use
PACKAGECONFIG_CONFARGS
to passPACKAGECONFIG
options toconfigure
andcmake
, respectively. If you are usingPACKAGECONFIG
but not a class that handles thedo_configure
task, then you need to usePACKAGECONFIG_CONFARGS
appropriately.- PACKAGEGROUP_DISABLE_COMPLEMENTARY
For recipes inheriting the packagegroup class, setting
PACKAGEGROUP_DISABLE_COMPLEMENTARY
to “1” specifies that the normal complementary packages (i.e.-dev
,-dbg
, and so forth) should not be automatically created by thepackagegroup
recipe, which is the default behavior.- PACKAGES
The list of packages the recipe creates. The default value is the following:
${PN}-dbg ${PN}-staticdev ${PN}-dev ${PN}-doc ${PN}-locale ${PACKAGE_BEFORE_PN} ${PN}
During packaging, the do_package task goes through
PACKAGES
and uses the FILES variable corresponding to each package to assign files to the package. If a file matches theFILES
variable for more than one package inPACKAGES
, it will be assigned to the earliest (leftmost) package.Packages in the variable’s list that are empty (i.e. where none of the patterns in
FILES_
pkg match any files installed by the do_install task) are not generated, unless generation is forced through the ALLOW_EMPTY variable.- PACKAGES_DYNAMIC
A promise that your recipe satisfies runtime dependencies for optional modules that are found in other recipes.
PACKAGES_DYNAMIC
does not actually satisfy the dependencies, it only states that they should be satisfied. For example, if a hard, runtime dependency (RDEPENDS) of another package is satisfied at build time through thePACKAGES_DYNAMIC
variable, but a package with the module name is never actually produced, then the other package will be broken. Thus, if you attempt to include that package in an image, you will get a dependency failure from the packaging system during the do_rootfs task.Typically, if there is a chance that such a situation can occur and the package that is not created is valid without the dependency being satisfied, then you should use RRECOMMENDS (a soft runtime dependency) instead of
RDEPENDS
.For an example of how to use the
PACKAGES_DYNAMIC
variable when you are splitting packages, see the “Handling Optional Module Packaging” section in the Yocto Project Development Tasks Manual.- PACKAGESPLITFUNCS
Specifies a list of functions run to perform additional splitting of files into individual packages. Recipes can either prepend to this variable or prepend to the
populate_packages
function in order to perform additional package splitting. In either case, the function should set PACKAGES, FILES, RDEPENDS and other packaging variables appropriately in order to perform the desired splitting.- PARALLEL_MAKE
Extra options passed to the
make
command during the do_compile task in order to specify parallel compilation on the local build host. This variable is usually in the form “-j x”, where x represents the maximum number of parallel threadsmake
can run.Note
In order for
PARALLEL_MAKE
to be effective,make
must be called with${
EXTRA_OEMAKE}
. An easy way to ensure this is to use theoe_runmake
function.By default, the OpenEmbedded build system automatically sets this variable to be equal to the number of cores the build system uses.
Note
If the software being built experiences dependency issues during the
do_compile
task that result in race conditions, you can clear thePARALLEL_MAKE
variable within the recipe as a workaround. For information on addressing race conditions, see the “Debugging Parallel Make Races” section in the Yocto Project Development Tasks Manual.For single socket systems (i.e. one CPU), you should not have to override this variable to gain optimal parallelism during builds. However, if you have very large systems that employ multiple physical CPUs, you might want to make sure the
PARALLEL_MAKE
variable is not set higher than “-j 20”.For more information on speeding up builds, see the “Speeding Up a Build” section in the Yocto Project Development Tasks Manual.
- PARALLEL_MAKEINST
Extra options passed to the
make install
command during the do_install task in order to specify parallel installation. This variable defaults to the value of PARALLEL_MAKE.Note
In order for
PARALLEL_MAKEINST
to be effective,make
must be called with${
EXTRA_OEMAKE}
. An easy way to ensure this is to use theoe_runmake
function.If the software being built experiences dependency issues during the
do_install
task that result in race conditions, you can clear thePARALLEL_MAKEINST
variable within the recipe as a workaround. For information on addressing race conditions, see the “Debugging Parallel Make Races” section in the Yocto Project Development Tasks Manual.- PATCHRESOLVE
Determines the action to take when a patch fails. You can set this variable to one of two values: “noop” and “user”.
The default value of “noop” causes the build to simply fail when the OpenEmbedded build system cannot successfully apply a patch. Setting the value to “user” causes the build system to launch a shell and places you in the right location so that you can manually resolve the conflicts.
Set this variable in your
local.conf
file.- PATCHTOOL
Specifies the utility used to apply patches for a recipe during the do_patch task. You can specify one of three utilities: “patch”, “quilt”, or “git”. The default utility used is “quilt” except for the quilt-native recipe itself. Because the quilt tool is not available at the time quilt-native is being patched, it uses “patch”.
If you wish to use an alternative patching tool, set the variable in the recipe using one of the following:
PATCHTOOL = "patch" PATCHTOOL = "quilt" PATCHTOOL = "git"
- PE
The epoch of the recipe. By default, this variable is unset. The variable is used to make upgrades possible when the versioning scheme changes in some backwards incompatible way.
PE
is the default value of the PKGE variable.- PF
Specifies the recipe or package name and includes all version and revision numbers (i.e.
glibc-2.13-r20+svnr15508/
andbash-4.2-r1/
). This variable is comprised of the following: ${PN}-${EXTENDPE}${PV}-${PR}- PIXBUF_PACKAGES
When inheriting the pixbufcache class, this variable identifies packages that contain the pixbuf loaders used with
gdk-pixbuf
. By default, thepixbufcache
class assumes that the loaders are in the recipe’s main package (i.e.${
PN}
). Use this variable if the loaders you need are in a package other than that main package.- PKG
The name of the resulting package created by the OpenEmbedded build system.
Note
When using the
PKG
variable, you must use a package name override.For example, when the debian class renames the output package, it does so by setting
PKG_packagename
.- PKG_CONFIG_PATH
The path to
pkg-config
files for the current build context.pkg-config
reads this variable from the environment.- PKGD
Points to the destination directory for files to be packaged before they are split into individual packages. This directory defaults to the following:
${WORKDIR}/package
Do not change this default.
- PKGDATA_DIR
Points to a shared, global-state directory that holds data generated during the packaging process. During the packaging process, the do_packagedata task packages data for each recipe and installs it into this temporary, shared area. This directory defaults to the following, which you should not change:
${STAGING_DIR_HOST}/pkgdata
For examples of how this data is used, see the “Automatically Added Runtime Dependencies” section in the Yocto Project Overview and Concepts Manual and the “Viewing Package Information with oe-pkgdata-util” section in the Yocto Project Development Tasks Manual. For more information on the shared, global-state directory, see STAGING_DIR_HOST.
- PKGDEST
Points to the parent directory for files to be packaged after they have been split into individual packages. This directory defaults to the following:
${WORKDIR}/packages-split
Under this directory, the build system creates directories for each package specified in PACKAGES. Do not change this default.
- PKGDESTWORK
Points to a temporary work area where the do_package task saves package metadata. The
PKGDESTWORK
location defaults to the following:${WORKDIR}/pkgdata
Do not change this default.
The do_packagedata task copies the package metadata from
PKGDESTWORK
to PKGDATA_DIR to make it available globally.- PKGE
The epoch of the package(s) built by the recipe. By default,
PKGE
is set to PE.- PKGR
The revision of the package(s) built by the recipe. By default,
PKGR
is set to PR.- PKGV
The version of the package(s) built by the recipe. By default,
PKGV
is set to PV.- PN
This variable can have two separate functions depending on the context: a recipe name or a resulting package name.
PN
refers to a recipe name in the context of a file used by the OpenEmbedded build system as input to create a package. The name is normally extracted from the recipe file name. For example, if the recipe is namedexpat_2.0.1.bb
, then the default value ofPN
will be “expat”.The variable refers to a package name in the context of a file created or produced by the OpenEmbedded build system.
If applicable, the
PN
variable also contains any special suffix or prefix. For example, usingbash
to build packages for the native machine,PN
isbash-native
. Usingbash
to build packages for the target and for Multilib,PN
would bebash
andlib64-bash
, respectively.- PNBLACKLIST
Lists recipes you do not want the OpenEmbedded build system to build. This variable works in conjunction with the blacklist class, which is inherited globally.
To prevent a recipe from being built, use the
PNBLACKLIST
variable in yourlocal.conf
file. Here is an example that preventsmyrecipe
from being built:PNBLACKLIST[myrecipe] = "Not supported by our organization."
- POPULATE_SDK_POST_HOST_COMMAND
Specifies a list of functions to call once the OpenEmbedded build system has created the host part of the SDK. You can specify functions separated by semicolons:
POPULATE_SDK_POST_HOST_COMMAND += "function; ... "
If you need to pass the SDK path to a command within a function, you can use
${SDK_DIR}
, which points to the parent directory used by the OpenEmbedded build system when creating SDK output. See the SDK_DIR variable for more information.- POPULATE_SDK_POST_TARGET_COMMAND
Specifies a list of functions to call once the OpenEmbedded build system has created the target part of the SDK. You can specify functions separated by semicolons:
POPULATE_SDK_POST_TARGET_COMMAND += "function; ... "
If you need to pass the SDK path to a command within a function, you can use
${SDK_DIR}
, which points to the parent directory used by the OpenEmbedded build system when creating SDK output. See the SDK_DIR variable for more information.- PR
The revision of the recipe. The default value for this variable is “r0”. Subsequent revisions of the recipe conventionally have the values “r1”, “r2”, and so forth. When PV increases,
PR
is conventionally reset to “r0”.Note
The OpenEmbedded build system does not need the aid of
PR
to know when to rebuild a recipe. The build system uses the task input checksums along with the stamp and Shared State Cache mechanisms.The
PR
variable primarily becomes significant when a package manager dynamically installs packages on an already built image. In this case,PR
, which is the default value of PKGR, helps the package manager distinguish which package is the most recent one in cases where many packages have the samePV
(i.e.PKGV
). A component having many packages with the samePV
usually means that the packages all install the same upstream version, but with later (PR
) version packages including packaging fixes.Note
PR
does not need to be increased for changes that do not change the package contents or metadata.Because manually managing
PR
can be cumbersome and error-prone, an automated solution exists. See the “Working With a PR Service” section in the Yocto Project Development Tasks Manual for more information.- PREFERRED_PROVIDER
If multiple recipes provide the same item, this variable determines which recipe is preferred and thus provides the item (i.e. the preferred provider). You should always suffix this variable with the name of the provided item. And, you should define the variable using the preferred recipe’s name (PN). Here is a common example:
PREFERRED_PROVIDER_virtual/kernel ?= "linux-yocto"
In the previous example, multiple recipes are providing “virtual/kernel”. The
PREFERRED_PROVIDER
variable is set with the name (PN
) of the recipe you prefer to provide “virtual/kernel”.Following are more examples:
PREFERRED_PROVIDER_virtual/xserver = "xserver-xf86" PREFERRED_PROVIDER_virtual/libgl ?= "mesa"
For more information, see the “Using Virtual Providers” section in the Yocto Project Development Tasks Manual.
Note
If you use a
virtual/\*
item withPREFERRED_PROVIDER
, then any recipe that PROVIDES that item but is not selected (defined) byPREFERRED_PROVIDER
is prevented from building, which is usually desirable since this mechanism is designed to select between mutually exclusive alternative providers.- PREFERRED_VERSION
If there are multiple versions of a recipe available, this variable determines which version should be given preference. You must always suffix the variable with the PN you want to select (python in the first example below), and you should specify the PV accordingly (3.4.0 in the example).
The
PREFERRED_VERSION
variable supports limited wildcard use through the “%
” character. You can use the character to match any number of characters, which can be useful when specifying versions that contain long revision numbers that potentially change. Here are two examples:PREFERRED_VERSION_python = "3.4.0" PREFERRED_VERSION_linux-yocto = "5.0%"
Note
The use of the “%” character is limited in that it only works at the end of the string. You cannot use the wildcard character in any other location of the string.
The specified version is matched against PV, which does not necessarily match the version part of the recipe’s filename. For example, consider two recipes
foo_1.2.bb
andfoo_git.bb
wherefoo_git.bb
contains the following assignment:PV = "1.1+git${SRCPV}"
In this case, the correct way to select
foo_git.bb
is by using an assignment such as the following:PREFERRED_VERSION_foo = "1.1+git%"
Compare that previous example against the following incorrect example, which does not work:
PREFERRED_VERSION_foo = "git"
Sometimes the
PREFERRED_VERSION
variable can be set by configuration files in a way that is hard to change. You can use OVERRIDES to set a machine-specific override. Here is an example:PREFERRED_VERSION_linux-yocto_qemux86 = "5.0%"
Although not recommended, worst case, you can also use the “forcevariable” override, which is the strongest override possible. Here is an example:
PREFERRED_VERSION_linux-yocto_forcevariable = "5.0%"
Note
The
\_forcevariable
override is not handled specially. This override only works because the default value ofOVERRIDES
includes “forcevariable”.If a recipe with the specified version is not available, a warning message will be shown. See REQUIRED_VERSION if you want this to be an error instead.
- PREMIRRORS
Specifies additional paths from which the OpenEmbedded build system gets source code. When the build system searches for source code, it first tries the local download directory. If that location fails, the build system tries locations defined by
PREMIRRORS
, the upstream source, and then locations specified by MIRRORS in that order.Assuming your distribution (DISTRO) is “poky”, the default value for
PREMIRRORS
is defined in theconf/distro/poky.conf
file in themeta-poky
Git repository.Typically, you could add a specific server for the build system to attempt before any others by adding something like the following to the
local.conf
configuration file in the Build Directory:PREMIRRORS_prepend = "\ git://.*/.* http://www.yoctoproject.org/sources/ \n \ ftp://.*/.* http://www.yoctoproject.org/sources/ \n \ http://.*/.* http://www.yoctoproject.org/sources/ \n \ https://.*/.* http://www.yoctoproject.org/sources/ \n"
These changes cause the build system to intercept Git, FTP, HTTP, and HTTPS requests and direct them to the
http://
sources mirror. You can usefile://
URLs to point to local directories or network shares as well.- PRIORITY
Indicates the importance of a package.
PRIORITY
is considered to be part of the distribution policy because the importance of any given recipe depends on the purpose for which the distribution is being produced. Thus,PRIORITY
is not normally set within recipes.You can set
PRIORITY
to “required”, “standard”, “extra”, and “optional”, which is the default.- PRIVATE_LIBS
Specifies libraries installed within a recipe that should be ignored by the OpenEmbedded build system’s shared library resolver. This variable is typically used when software being built by a recipe has its own private versions of a library normally provided by another recipe. In this case, you would not want the package containing the private libraries to be set as a dependency on other unrelated packages that should instead depend on the package providing the standard version of the library.
Libraries specified in this variable should be specified by their file name. For example, from the Firefox recipe in meta-browser:
PRIVATE_LIBS = "libmozjs.so \ libxpcom.so \ libnspr4.so \ libxul.so \ libmozalloc.so \ libplc4.so \ libplds4.so"
For more information, see the “Automatically Added Runtime Dependencies” section in the Yocto Project Overview and Concepts Manual.
- PROVIDES
A list of aliases by which a particular recipe can be known. By default, a recipe’s own
PN
is implicitly already in itsPROVIDES
list and therefore does not need to mention that it provides itself. If a recipe usesPROVIDES
, the additional aliases are synonyms for the recipe and can be useful for satisfying dependencies of other recipes during the build as specified byDEPENDS
.Consider the following example
PROVIDES
statement from the recipe fileeudev_3.2.9.bb
:PROVIDES += "udev"
The
PROVIDES
statement results in the “eudev” recipe also being available as simply “udev”.Note
A recipe’s own recipe name (PN) is always implicitly prepended to PROVIDES, so while using “+=” in the above example may not be strictly necessary it is recommended to avoid confusion.
In addition to providing recipes under alternate names, the
PROVIDES
mechanism is also used to implement virtual targets. A virtual target is a name that corresponds to some particular functionality (e.g. a Linux kernel). Recipes that provide the functionality in question list the virtual target inPROVIDES
. Recipes that depend on the functionality in question can include the virtual target inDEPENDS
to leave the choice of provider open.Conventionally, virtual targets have names on the form “virtual/function” (e.g. “virtual/kernel”). The slash is simply part of the name and has no syntactical significance.
The PREFERRED_PROVIDER variable is used to select which particular recipe provides a virtual target.
Note
A corresponding mechanism for virtual runtime dependencies (packages) exists. However, the mechanism does not depend on any special functionality beyond ordinary variable assignments. For example,
VIRTUAL-RUNTIME_dev_manager
refers to the package of the component that manages the/dev
directory.Setting the “preferred provider” for runtime dependencies is as simple as using the following assignment in a configuration file:
VIRTUAL-RUNTIME_dev_manager = "udev"
- PRSERV_HOST
The network based PR service host and port.
The
conf/local.conf.sample.extended
configuration file in the Source Directory shows how thePRSERV_HOST
variable is set:PRSERV_HOST = "localhost:0"
You must set the variable if you want to automatically start a local PR service. You can set
PRSERV_HOST
to other values to use a remote PR service.- PSEUDO_IGNORE_PATHS
A comma-separated (without spaces) list of path prefixes that should be ignored by pseudo when monitoring and recording file operations, in order to avoid problems with files being written to outside of the pseudo context and reduce pseudo’s overhead. A path is ignored if it matches any prefix in the list and can include partial directory (or file) names.
- PTEST_ENABLED
Specifies whether or not Package Test (ptest) functionality is enabled when building a recipe. You should not set this variable directly. Enabling and disabling building Package Tests at build time should be done by adding “ptest” to (or removing it from) DISTRO_FEATURES.
- PV
The version of the recipe. The version is normally extracted from the recipe filename. For example, if the recipe is named
expat_2.0.1.bb
, then the default value ofPV
will be “2.0.1”.PV
is generally not overridden within a recipe unless it is building an unstable (i.e. development) version from a source code repository (e.g. Git or Subversion).PV
is the default value of the PKGV variable.- PYTHON_ABI
When used by recipes that inherit the distutils3, setuptools3 classes, denotes the Application Binary Interface (ABI) currently in use for Python. By default, the ABI is “m”. You do not have to set this variable as the OpenEmbedded build system sets it for you.
The OpenEmbedded build system uses the ABI to construct directory names used when installing the Python headers and libraries in sysroot (e.g.
.../python3.3m/...
).Recipes that inherit the
distutils3
class during cross-builds also use this variable to locate the headers and libraries of the appropriate Python that the extension is targeting.- PYTHON_PN
When used by recipes that inherit the distutils3 <ref-classes-distutils3>, setuptools3 classes, specifies the major Python version being built. For Python 3.x,
PYTHON_PN
would be “python3”. You do not have to set this variable as the OpenEmbedded build system automatically sets it for you.The variable allows recipes to use common infrastructure such as the following:
DEPENDS += "${PYTHON_PN}-native"
In the previous example, the version of the dependency is
PYTHON_PN
.- RANLIB
The minimal command and arguments to run
ranlib
.- RCONFLICTS
The list of packages that conflict with packages. Note that packages will not be installed if conflicting packages are not first removed.
Like all package-controlling variables, you must always use them in conjunction with a package name override. Here is an example:
RCONFLICTS_${PN} = "another_conflicting_package_name"
BitBake, which the OpenEmbedded build system uses, supports specifying versioned dependencies. Although the syntax varies depending on the packaging format, BitBake hides these differences from you. Here is the general syntax to specify versions with the
RCONFLICTS
variable:RCONFLICTS_${PN} = "package (operator version)"
For
operator
, you can specify the following:=
<
>
<=
>=
For example, the following sets up a dependency on version 1.2 or greater of the package
foo
:RCONFLICTS_${PN} = "foo (>= 1.2)"
- RDEPENDS
Lists runtime dependencies of a package. These dependencies are other packages that must be installed in order for the package to function correctly. As an example, the following assignment declares that the package
foo
needs the packagesbar
andbaz
to be installed:RDEPENDS_foo = "bar baz"
The most common types of package runtime dependencies are automatically detected and added. Therefore, most recipes do not need to set
RDEPENDS
. For more information, see the “Automatically Added Runtime Dependencies” section in the Yocto Project Overview and Concepts Manual.The practical effect of the above
RDEPENDS
assignment is thatbar
andbaz
will be declared as dependencies inside the packagefoo
when it is written out by one of the do_package_write_* tasks. Exactly how this is done depends on which package format is used, which is determined by PACKAGE_CLASSES. When the corresponding package manager installs the package, it will know to also install the packages on which it depends.To ensure that the packages
bar
andbaz
get built, the previousRDEPENDS
assignment also causes a task dependency to be added. This dependency is from the recipe’s do_build (not to be confused with do_compile) task to thedo_package_write_*
task of the recipes that buildbar
andbaz
.The names of the packages you list within
RDEPENDS
must be the names of other packages - they cannot be recipe names. Although package names and recipe names usually match, the important point here is that you are providing package names within theRDEPENDS
variable. For an example of the default list of packages created from a recipe, see the PACKAGES variable.Because the
RDEPENDS
variable applies to packages being built, you should always use the variable in a form with an attached package name (remember that a single recipe can build multiple packages). For example, suppose you are building a development package that depends on theperl
package. In this case, you would use the followingRDEPENDS
statement:RDEPENDS_${PN}-dev += "perl"
In the example, the development package depends on the
perl
package. Thus, theRDEPENDS
variable has the${PN}-dev
package name as part of the variable.Note
RDEPENDS_${PN}-dev
includes${
PN}
by default. This default is set in the BitBake configuration file (meta/conf/bitbake.conf
). Be careful not to accidentally remove${PN}
when modifyingRDEPENDS_${PN}-dev
. Use the “+=” operator rather than the “=” operator.The package names you use with
RDEPENDS
must appear as they would in thePACKAGES
variable. The PKG variable allows a different name to be used for the final package (e.g. the debian class uses this to rename packages), but this final package name cannot be used withRDEPENDS
, which makes sense asRDEPENDS
is meant to be independent of the package format used.BitBake, which the OpenEmbedded build system uses, supports specifying versioned dependencies. Although the syntax varies depending on the packaging format, BitBake hides these differences from you. Here is the general syntax to specify versions with the
RDEPENDS
variable:RDEPENDS_${PN} = "package (operator version)"
For
operator
, you can specify the following:=
<
>
<=
>=
For version, provide the version number.
Note
You can use
EXTENDPKGV
to provide a full package version specification.For example, the following sets up a dependency on version 1.2 or greater of the package
foo
:RDEPENDS_${PN} = "foo (>= 1.2)"
For information on build-time dependencies, see the DEPENDS variable. You can also see the “Tasks” and “Dependencies” sections in the BitBake User Manual for additional information on tasks and dependencies.
- REQUIRED_DISTRO_FEATURES
When inheriting the features_check class, this variable identifies distribution features that must exist in the current configuration in order for the OpenEmbedded build system to build the recipe. In other words, if the
REQUIRED_DISTRO_FEATURES
variable lists a feature that does not appear inDISTRO_FEATURES
within the current configuration, then the recipe will be skipped, and if the build system attempts to build the recipe then an error will be triggered.- REQUIRED_VERSION
If there are multiple versions of a recipe available, this variable determines which version should be given preference. REQUIRED_VERSION works in exactly the same manner as PREFERRED_VERSION, except that if the specified version is not available then an error message is shown and the build fails immediately.
If both REQUIRED_VERSION and PREFERRED_VERSION are set for the same recipe, the REQUIRED_VERSION value applies.
- RM_WORK_EXCLUDE
With
rm_work
enabled, this variable specifies a list of recipes whose work directories should not be removed. See the “rm_work.bbclass” section for more details.- ROOT_HOME
Defines the root home directory. By default, this directory is set as follows in the BitBake configuration file:
ROOT_HOME ??= "/home/root"
Note
This default value is likely used because some embedded solutions prefer to have a read-only root filesystem and prefer to keep writeable data in one place.
You can override the default by setting the variable in any layer or in the
local.conf
file. Because the default is set using a “weak” assignment (i.e. “??=”), you can use either of the following forms to define your override:ROOT_HOME = "/root" ROOT_HOME ?= "/root"
These override examples use
/root
, which is probably the most commonly used override.- ROOTFS
Indicates a filesystem image to include as the root filesystem.
The
ROOTFS
variable is an optional variable used with the image-live class.- ROOTFS_POSTINSTALL_COMMAND
Specifies a list of functions to call after the OpenEmbedded build system has installed packages. You can specify functions separated by semicolons:
ROOTFS_POSTINSTALL_COMMAND += "function; ... "
If you need to pass the root filesystem path to a command within a function, you can use
${IMAGE_ROOTFS}
, which points to the directory that becomes the root filesystem image. See the IMAGE_ROOTFS variable for more information.- ROOTFS_POSTPROCESS_COMMAND
Specifies a list of functions to call once the OpenEmbedded build system has created the root filesystem. You can specify functions separated by semicolons:
ROOTFS_POSTPROCESS_COMMAND += "function; ... "
If you need to pass the root filesystem path to a command within a function, you can use
${IMAGE_ROOTFS}
, which points to the directory that becomes the root filesystem image. See the IMAGE_ROOTFS variable for more information.- ROOTFS_POSTUNINSTALL_COMMAND
Specifies a list of functions to call after the OpenEmbedded build system has removed unnecessary packages. When runtime package management is disabled in the image, several packages are removed including
base-passwd
,shadow
, andupdate-alternatives
. You can specify functions separated by semicolons:ROOTFS_POSTUNINSTALL_COMMAND += "function; ... "
If you need to pass the root filesystem path to a command within a function, you can use
${IMAGE_ROOTFS}
, which points to the directory that becomes the root filesystem image. See the IMAGE_ROOTFS variable for more information.- ROOTFS_PREPROCESS_COMMAND
Specifies a list of functions to call before the OpenEmbedded build system has created the root filesystem. You can specify functions separated by semicolons:
ROOTFS_PREPROCESS_COMMAND += "function; ... "
If you need to pass the root filesystem path to a command within a function, you can use
${IMAGE_ROOTFS}
, which points to the directory that becomes the root filesystem image. See the IMAGE_ROOTFS variable for more information.- RPROVIDES
A list of package name aliases that a package also provides. These aliases are useful for satisfying runtime dependencies of other packages both during the build and on the target (as specified by
RDEPENDS
).Note
A package’s own name is implicitly already in its
RPROVIDES
list.As with all package-controlling variables, you must always use the variable in conjunction with a package name override. Here is an example:
RPROVIDES_${PN} = "widget-abi-2"
- RRECOMMENDS
A list of packages that extends the usability of a package being built. The package being built does not depend on this list of packages in order to successfully build, but rather uses them for extended usability. To specify runtime dependencies for packages, see the
RDEPENDS
variable.The package manager will automatically install the
RRECOMMENDS
list of packages when installing the built package. However, you can prevent listed packages from being installed by using the BAD_RECOMMENDATIONS, NO_RECOMMENDATIONS, and PACKAGE_EXCLUDE variables.Packages specified in
RRECOMMENDS
need not actually be produced. However, a recipe must exist that provides each package, either through the PACKAGES or PACKAGES_DYNAMIC variables or the RPROVIDES variable, or an error will occur during the build. If such a recipe does exist and the package is not produced, the build continues without error.Because the
RRECOMMENDS
variable applies to packages being built, you should always attach an override to the variable to specify the particular package whose usability is being extended. For example, suppose you are building a development package that is extended to support wireless functionality. In this case, you would use the following:RRECOMMENDS_${PN}-dev += "wireless_package_name"
In the example, the package name (
${PN}-dev
) must appear as it would in thePACKAGES
namespace before any renaming of the output package by classes such asdebian.bbclass
.BitBake, which the OpenEmbedded build system uses, supports specifying versioned recommends. Although the syntax varies depending on the packaging format, BitBake hides these differences from you. Here is the general syntax to specify versions with the
RRECOMMENDS
variable:RRECOMMENDS_${PN} = "package (operator version)"
For
operator
, you can specify the following:=
<
>
<=
>=
For example, the following sets up a recommend on version 1.2 or greater of the package
foo
:RRECOMMENDS_${PN} = "foo (>= 1.2)"
- RREPLACES
A list of packages replaced by a package. The package manager uses this variable to determine which package should be installed to replace other package(s) during an upgrade. In order to also have the other package(s) removed at the same time, you must add the name of the other package to the
RCONFLICTS
variable.As with all package-controlling variables, you must use this variable in conjunction with a package name override. Here is an example:
RREPLACES_${PN} = "other_package_being_replaced"
BitBake, which the OpenEmbedded build system uses, supports specifying versioned replacements. Although the syntax varies depending on the packaging format, BitBake hides these differences from you. Here is the general syntax to specify versions with the
RREPLACES
variable:RREPLACES_${PN} = "package (operator version)"
For
operator
, you can specify the following:=
<
>
<=
>=
For example, the following sets up a replacement using version 1.2 or greater of the package
foo
:RREPLACES_${PN} = "foo (>= 1.2)"
- RSUGGESTS
A list of additional packages that you can suggest for installation by the package manager at the time a package is installed. Not all package managers support this functionality.
As with all package-controlling variables, you must always use this variable in conjunction with a package name override. Here is an example:
RSUGGESTS_${PN} = "useful_package another_package"
- S
The location in the Build Directory where unpacked recipe source code resides. By default, this directory is
${
WORKDIR}/${
BPN}-${
PV}
, where${BPN}
is the base recipe name and${PV}
is the recipe version. If the source tarball extracts the code to a directory named anything other than${BPN}-${PV}
, or if the source code is fetched from an SCM such as Git or Subversion, then you must setS
in the recipe so that the OpenEmbedded build system knows where to find the unpacked source.As an example, assume a Source Directory top-level folder named
poky
and a default Build Directory atpoky/build
. In this case, the work directory the build system uses to keep the unpacked recipe fordb
is the following:poky/build/tmp/work/qemux86-poky-linux/db/5.1.19-r3/db-5.1.19
The unpacked source code resides in the
db-5.1.19
folder.This next example assumes a Git repository. By default, Git repositories are cloned to
${WORKDIR}/git
during do_fetch. Since this path is different from the default value ofS
, you must set it specifically so the source can be located:SRC_URI = "git://path/to/repo.git" S = "${WORKDIR}/git"
- SANITY_REQUIRED_UTILITIES
Specifies a list of command-line utilities that should be checked for during the initial sanity checking process when running BitBake. If any of the utilities are not installed on the build host, then BitBake immediately exits with an error.
- SANITY_TESTED_DISTROS
A list of the host distribution identifiers that the build system has been tested against. Identifiers consist of the host distributor ID followed by the release, as reported by the
lsb_release
tool or as read from/etc/lsb-release
. Separate the list items with explicit newline characters (\n
). IfSANITY_TESTED_DISTROS
is not empty and the current value of NATIVELSBSTRING does not appear in the list, then the build system reports a warning that indicates the current host distribution has not been tested as a build host.- SDK_ARCH
The target architecture for the SDK. Typically, you do not directly set this variable. Instead, use SDKMACHINE.
- SDK_CUSTOM_TEMPLATECONF
When building the extensible SDK, if
SDK_CUSTOM_TEMPLATECONF
is set to “1” and aconf/templateconf.conf
file exists in the build directory (TOPDIR) then this will be copied into the SDK.- SDK_DEPLOY
The directory set up and used by the populate_sdk_base class to which the SDK is deployed. The
populate_sdk_base
class definesSDK_DEPLOY
as follows:SDK_DEPLOY = "${TMPDIR}/deploy/sdk"
- SDK_DIR
The parent directory used by the OpenEmbedded build system when creating SDK output. The populate_sdk_base class defines the variable as follows:
SDK_DIR = "${WORKDIR}/sdk"
Note
The
SDK_DIR
directory is a temporary directory as it is part ofWORKDIR
. The final output directory is SDK_DEPLOY.- SDK_EXT_TYPE
Controls whether or not shared state artifacts are copied into the extensible SDK. The default value of “full” copies all of the required shared state artifacts into the extensible SDK. The value “minimal” leaves these artifacts out of the SDK.
Note
If you set the variable to “minimal”, you need to ensure SSTATE_MIRRORS is set in the SDK’s configuration to enable the artifacts to be fetched as needed.
- SDK_HOST_MANIFEST
The manifest file for the host part of the SDK. This file lists all the installed packages that make up the host part of the SDK. The file contains package information on a line-per-package basis as follows:
packagename packagearch version
The populate_sdk_base class defines the manifest file as follows:
SDK_HOST_MANIFEST = "${SDK_DEPLOY}/${TOOLCHAIN_OUTPUTNAME}.host.manifest"
The location is derived using the SDK_DEPLOY and TOOLCHAIN_OUTPUTNAME variables.
- SDK_INCLUDE_PKGDATA
When set to “1”, specifies to include the packagedata for all recipes in the “world” target in the extensible SDK. Including this data allows the
devtool search
command to find these recipes in search results, as well as allows thedevtool add
command to map dependencies more effectively.Note
Enabling the
SDK_INCLUDE_PKGDATA
variable significantly increases build time because all of world needs to be built. Enabling the variable also slightly increases the size of the extensible SDK.- SDK_INCLUDE_TOOLCHAIN
When set to “1”, specifies to include the toolchain in the extensible SDK. Including the toolchain is useful particularly when SDK_EXT_TYPE is set to “minimal” to keep the SDK reasonably small but you still want to provide a usable toolchain. For example, suppose you want to use the toolchain from an IDE or from other tools and you do not want to perform additional steps to install the toolchain.
The
SDK_INCLUDE_TOOLCHAIN
variable defaults to “0” ifSDK_EXT_TYPE
is set to “minimal”, and defaults to “1” ifSDK_EXT_TYPE
is set to “full”.- SDK_INHERIT_BLACKLIST
A list of classes to remove from the INHERIT value globally within the extensible SDK configuration. The populate-sdk-ext class sets the default value:
SDK_INHERIT_BLACKLIST ?= "buildhistory icecc"
Some classes are not generally applicable within the extensible SDK context. You can use this variable to disable those classes.
For additional information on how to customize the extensible SDK’s configuration, see the “Configuring the Extensible SDK” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
- SDK_LOCAL_CONF_BLACKLIST
A list of variables not allowed through from the OpenEmbedded build system configuration into the extensible SDK configuration. Usually, these are variables that are specific to the machine on which the build system is running and thus would be potentially problematic within the extensible SDK.
By default,
SDK_LOCAL_CONF_BLACKLIST
is set in the populate-sdk-ext class and excludes the following variables:For additional information on how to customize the extensible SDK’s configuration, see the “Configuring the Extensible SDK” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
- SDK_LOCAL_CONF_WHITELIST
A list of variables allowed through from the OpenEmbedded build system configuration into the extensible SDK configuration. By default, the list of variables is empty and is set in the populate-sdk-ext class.
This list overrides the variables specified using the SDK_LOCAL_CONF_BLACKLIST variable as well as any variables identified by automatic blacklisting due to the “/” character being found at the start of the value, which is usually indicative of being a path and thus might not be valid on the system where the SDK is installed.
For additional information on how to customize the extensible SDK’s configuration, see the “Configuring the Extensible SDK” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
- SDK_NAME
The base name for SDK output files. The name is derived from the DISTRO, TCLIBC, SDK_ARCH, IMAGE_BASENAME, and TUNE_PKGARCH variables:
SDK_NAME = "${DISTRO}-${TCLIBC}-${SDK_ARCH}-${IMAGE_BASENAME}-${TUNE_PKGARCH}"
- SDK_OS
Specifies the operating system for which the SDK will be built. The default value is the value of BUILD_OS.
- SDK_OUTPUT
The location used by the OpenEmbedded build system when creating SDK output. The populate_sdk_base class defines the variable as follows:
SDK_DIR = "${WORKDIR}/sdk" SDK_OUTPUT = "${SDK_DIR}/image" SDK_DEPLOY = "${DEPLOY_DIR}/sdk"
Note
The
SDK_OUTPUT
directory is a temporary directory as it is part of WORKDIR by way of SDK_DIR. The final output directory is SDK_DEPLOY.- SDK_PACKAGE_ARCHS
Specifies a list of architectures compatible with the SDK machine. This variable is set automatically and should not normally be hand-edited. Entries are separated using spaces and listed in order of priority. The default value for
SDK_PACKAGE_ARCHS
is “all any noarch ${SDK_ARCH}-${SDKPKGSUFFIX}”.- SDK_POSTPROCESS_COMMAND
Specifies a list of functions to call once the OpenEmbedded build system creates the SDK. You can specify functions separated by semicolons: SDK_POSTPROCESS_COMMAND += “function; … “
If you need to pass an SDK path to a command within a function, you can use
${SDK_DIR}
, which points to the parent directory used by the OpenEmbedded build system when creating SDK output. See the SDK_DIR variable for more information.- SDK_PREFIX
The toolchain binary prefix used for
nativesdk
recipes. The OpenEmbedded build system uses theSDK_PREFIX
value to set the TARGET_PREFIX when buildingnativesdk
recipes. The default value is “${SDK_SYS}-“.- SDK_RECRDEP_TASKS
A list of shared state tasks added to the extensible SDK. By default, the following tasks are added:
do_populate_lic
do_package_qa
do_populate_sysroot
do_deploy
Despite the default value of “” for the
SDK_RECRDEP_TASKS
variable, the above four tasks are always added to the SDK. To specify tasks beyond these four, you need to use theSDK_RECRDEP_TASKS
variable (e.g. you are defining additional tasks that are needed in order to build SDK_TARGETS).- SDK_SYS
Specifies the system, including the architecture and the operating system, for which the SDK will be built.
The OpenEmbedded build system automatically sets this variable based on SDK_ARCH, SDK_VENDOR, and SDK_OS. You do not need to set the
SDK_SYS
variable yourself.- SDK_TARGET_MANIFEST
The manifest file for the target part of the SDK. This file lists all the installed packages that make up the target part of the SDK. The file contains package information on a line-per-package basis as follows:
packagename packagearch version
The populate_sdk_base class defines the manifest file as follows:
SDK_TARGET_MANIFEST = "${SDK_DEPLOY}/${TOOLCHAIN_OUTPUTNAME}.target.manifest"
The location is derived using the SDK_DEPLOY and TOOLCHAIN_OUTPUTNAME variables.
- SDK_TARGETS
A list of targets to install from shared state as part of the standard or extensible SDK installation. The default value is “${PN}” (i.e. the image from which the SDK is built).
The
SDK_TARGETS
variable is an internal variable and typically would not be changed.- SDK_TITLE
The title to be printed when running the SDK installer. By default, this title is based on the DISTRO_NAME or DISTRO variable and is set in the populate_sdk_base class as follows:
SDK_TITLE ??= "${@d.getVar('DISTRO_NAME') or d.getVar('DISTRO')} SDK"
For the default distribution “poky”,
SDK_TITLE
is set to “Poky (Yocto Project Reference Distro)”.For information on how to change this default title, see the “Changing the Extensible SDK Installer Title” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
- SDK_UPDATE_URL
An optional URL for an update server for the extensible SDK. If set, the value is used as the default update server when running
devtool sdk-update
within the extensible SDK.- SDK_VENDOR
Specifies the name of the SDK vendor.
- SDK_VERSION
Specifies the version of the SDK. The Poky distribution configuration file (
/meta-poky/conf/distro/poky.conf
) sets the defaultSDK_VERSION
as follows:SDK_VERSION = "${@d.getVar('DISTRO_VERSION').replace('snapshot-${METADATA_REVISION}', 'snapshot')}"
For additional information, see the DISTRO_VERSION and METADATA_REVISION variables.
- SDKEXTPATH
The default installation directory for the Extensible SDK. By default, this directory is based on the DISTRO variable and is set in the populate_sdk_base class as follows:
SDKEXTPATH ??= "~/${@d.getVar('DISTRO')}_sdk"
For the default distribution “poky”, the
SDKEXTPATH
is set to “poky_sdk”.For information on how to change this default directory, see the “Changing the Default SDK Installation Directory” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
- SDKIMAGE_FEATURES
Equivalent to
IMAGE_FEATURES
. However, this variable applies to the SDK generated from an image using the following command:$ bitbake -c populate_sdk imagename
- SDKMACHINE
The machine for which the SDK is built. In other words, the SDK is built such that it runs on the target you specify with the
SDKMACHINE
value. The value points to a corresponding.conf
file underconf/machine-sdk/
.You can use “i686” and “x86_64” as possible values for this variable. The variable defaults to “i686” and is set in the local.conf file in the Build Directory.
SDKMACHINE ?= "i686"
Note
You cannot set the
SDKMACHINE
variable in your distribution configuration file. If you do, the configuration will not take affect.- SDKPATH
Defines the path offered to the user for installation of the SDK that is generated by the OpenEmbedded build system. The path appears as the default location for installing the SDK when you run the SDK’s installation script. You can override the offered path when you run the script.
- SDKTARGETSYSROOT
The full path to the sysroot used for cross-compilation within an SDK as it will be when installed into the default SDKPATH.
- SECTION
The section in which packages should be categorized. Package management utilities can make use of this variable.
- SELECTED_OPTIMIZATION
Specifies the optimization flags passed to the C compiler when building for the target. The flags are passed through the default value of the TARGET_CFLAGS variable.
The
SELECTED_OPTIMIZATION
variable takes the value ofFULL_OPTIMIZATION
unlessDEBUG_BUILD
= “1”. If that is the case, the value ofDEBUG_OPTIMIZATION
is used.- SERIAL_CONSOLE
Defines a serial console (TTY) to enable using getty. Provide a value that specifies the baud rate followed by the TTY device name separated by a space. You cannot specify more than one TTY device:
SERIAL_CONSOLE = "115200 ttyS0"
Note
The
SERIAL_CONSOLE
variable is deprecated. Please use the SERIAL_CONSOLES variable.- SERIAL_CONSOLES
Defines a serial console (TTY) to enable using getty. Provide a value that specifies the baud rate followed by the TTY device name separated by a semicolon. Use spaces to separate multiple devices:
SERIAL_CONSOLES = "115200;ttyS0 115200;ttyS1"
- SERIAL_CONSOLES_CHECK
Specifies serial consoles, which must be listed in SERIAL_CONSOLES, to check against
/proc/console
before enabling them using getty. This variable allows aliasing in the format: <device>:<alias>. If a device was listed as “sclp_line0” in/dev/
and “ttyS0” was listed in/proc/console
, you would do the following:SERIAL_CONSOLES_CHECK = "slcp_line0:ttyS0"
This variable is currently only supported with SysVinit (i.e. not with systemd).
- SIGGEN_EXCLUDE_SAFE_RECIPE_DEPS
A list of recipe dependencies that should not be used to determine signatures of tasks from one recipe when they depend on tasks from another recipe. For example:
SIGGEN_EXCLUDE_SAFE_RECIPE_DEPS += "intone->mplayer2"
In the previous example,
intone
depends onmplayer2
.You can use the special token
"*"
on the left-hand side of the dependency to match all recipes except the one on the right-hand side. Here is an example:SIGGEN_EXCLUDE_SAFE_RECIPE_DEPS += "*->quilt-native"
In the previous example, all recipes except
quilt-native
ignore task signatures from thequilt-native
recipe when determining their task signatures.Use of this variable is one mechanism to remove dependencies that affect task signatures and thus force rebuilds when a recipe changes.
Note
If you add an inappropriate dependency for a recipe relationship, the software might break during runtime if the interface of the second recipe was changed after the first recipe had been built.
- SIGGEN_EXCLUDERECIPES_ABISAFE
A list of recipes that are completely stable and will never change. The ABI for the recipes in the list are presented by output from the tasks run to build the recipe. Use of this variable is one way to remove dependencies from one recipe on another that affect task signatures and thus force rebuilds when the recipe changes.
Note
If you add an inappropriate variable to this list, the software might break at runtime if the interface of the recipe was changed after the other had been built.
- SITEINFO_BITS
Specifies the number of bits for the target system CPU. The value should be either “32” or “64”.
- SITEINFO_ENDIANNESS
Specifies the endian byte order of the target system. The value should be either “le” for little-endian or “be” for big-endian.
- SKIP_FILEDEPS
Enables removal of all files from the “Provides” section of an RPM package. Removal of these files is required for packages containing prebuilt binaries and libraries such as
libstdc++
andglibc
.To enable file removal, set the variable to “1” in your
conf/local.conf
configuration file in your: Build Directory.SKIP_FILEDEPS = "1"
- SOC_FAMILY
Groups together machines based upon the same family of SOC (System On Chip). You typically set this variable in a common
.inc
file that you include in the configuration files of all the machines.Note
You must include
conf/machine/include/soc-family.inc
for this variable to appear in MACHINEOVERRIDES.- SOLIBS
Defines the suffix for shared libraries used on the target platform. By default, this suffix is “.so.*” for all Linux-based systems and is defined in the
meta/conf/bitbake.conf
configuration file.You will see this variable referenced in the default values of
FILES_${PN}
.- SOLIBSDEV
Defines the suffix for the development symbolic link (symlink) for shared libraries on the target platform. By default, this suffix is “.so” for Linux-based systems and is defined in the
meta/conf/bitbake.conf
configuration file.You will see this variable referenced in the default values of
FILES_${PN}-dev
.- SOURCE_MIRROR_FETCH
When you are fetching files to create a mirror of sources (i.e. creating a source mirror), setting
SOURCE_MIRROR_FETCH
to “1” in yourlocal.conf
configuration file ensures the source for all recipes are fetched regardless of whether or not a recipe is compatible with the configuration. A recipe is considered incompatible with the currently configured machine when either or both the COMPATIBLE_MACHINE variable and COMPATIBLE_HOST variables specify compatibility with a machine other than that of the current machine or host.Note
Do not set the
SOURCE_MIRROR_FETCH
variable unless you are creating a source mirror. In other words, do not set the variable during a normal build.- SOURCE_MIRROR_URL
Defines your own PREMIRRORS from which to first fetch source before attempting to fetch from the upstream specified in SRC_URI.
To use this variable, you must globally inherit the own-mirrors class and then provide the URL to your mirrors. Here is the general syntax:
INHERIT += "own-mirrors" SOURCE_MIRROR_URL = "http://example.com/my_source_mirror"
Note
You can specify only a single URL in
SOURCE_MIRROR_URL
.- SPDXLICENSEMAP
Maps commonly used license names to their SPDX counterparts found in
meta/files/common-licenses/
. For the defaultSPDXLICENSEMAP
mappings, see themeta/conf/licenses.conf
file.For additional information, see the LICENSE variable.
- SPECIAL_PKGSUFFIX
A list of prefixes for PN used by the OpenEmbedded build system to create variants of recipes or packages. The list specifies the prefixes to strip off during certain circumstances such as the generation of the BPN variable.
- SPL_BINARY
The file type for the Secondary Program Loader (SPL). Some devices use an SPL from which to boot (e.g. the BeagleBone development board). For such cases, you can declare the file type of the SPL binary in the
u-boot.inc
include file, which is used in the U-Boot recipe.The SPL file type is set to “null” by default in the
u-boot.inc
file as follows:# Some versions of u-boot build an SPL (Second Program Loader) image that # should be packaged along with the u-boot binary as well as placed in the # deploy directory. For those versions they can set the following variables # to allow packaging the SPL. SPL_BINARY ?= "" SPL_BINARYNAME ?= "${@os.path.basename(d.getVar("SPL_BINARY"))}" SPL_IMAGE ?= "${SPL_BINARYNAME}-${MACHINE}-${PV}-${PR}" SPL_SYMLINK ?= "${SPL_BINARYNAME}-${MACHINE}"
The
SPL_BINARY
variable helps form variousSPL_*
variables used by the OpenEmbedded build system.See the BeagleBone machine configuration example in the “Adding a Layer Using the bitbake-layers Script” section in the Yocto Project Board Support Package Developer’s Guide for additional information.
- SRC_URI
The list of source files - local or remote. This variable tells the OpenEmbedded build system which bits to pull in for the build and how to pull them in. For example, if the recipe or append file only needs to fetch a tarball from the Internet, the recipe or append file uses a single
SRC_URI
entry. On the other hand, if the recipe or append file needs to fetch a tarball, apply two patches, and include a custom file, the recipe or append file would include four instances of the variable.The following list explains the available URI protocols. URI protocols are highly dependent on particular BitBake Fetcher submodules. Depending on the fetcher BitBake uses, various URL parameters are employed. For specifics on the supported Fetchers, see the “Fetchers” section in the BitBake User Manual.
file://
- Fetches files, which are usually files shipped with the Metadata, from the local machine (e.g. patch files). The path is relative to the FILESPATH variable. Thus, the build system searches, in order, from the following directories, which are assumed to be a subdirectories of the directory in which the recipe file (.bb
) or append file (.bbappend
) resides:${BPN}
- The base recipe name without any special suffix or version numbers.${BP}
-${BPN}-${PV}
. The base recipe name and version but without any special package name suffix.files - Files within a directory, which is named
files
and is also alongside the recipe or append file.
Note
If you want the build system to pick up files specified through a SRC_URI statement from your append file, you need to be sure to extend the FILESPATH variable by also using the FILESEXTRAPATHS variable from within your append file.
bzr://
- Fetches files from a Bazaar revision control repository.git://
- Fetches files from a Git revision control repository.osc://
- Fetches files from an OSC (openSUSE Build service) revision control repository.repo://
- Fetches files from a repo (Git) repository.ccrc://
- Fetches files from a ClearCase repository.http://
- Fetches files from the Internet usinghttp
.https://
- Fetches files from the Internet usinghttps
.ftp://
- Fetches files from the Internet usingftp
.cvs://
- Fetches files from a CVS revision control repository.hg://
- Fetches files from a Mercurial (hg
) revision control repository.p4://
- Fetches files from a Perforce (p4
) revision control repository.ssh://
- Fetches files from a secure shell.svn://
- Fetches files from a Subversion (svn
) revision control repository.npm://
- Fetches JavaScript modules from a registry.az://
- Fetches files from an Azure Storage account.
Standard and recipe-specific options for
SRC_URI
exist. Here are standard options:apply
- Whether to apply the patch or not. The default action is to apply the patch.striplevel
- Which striplevel to use when applying the patch. The default level is 1.patchdir
- Specifies the directory in which the patch should be applied. The default is${
S}
.
Here are options specific to recipes building code from a revision control system:
mindate
- Apply the patch only if SRCDATE is equal to or greater thanmindate
.maxdate
- Apply the patch only ifSRCDATE
is not later thanmaxdate
.minrev
- Apply the patch only ifSRCREV
is equal to or greater thanminrev
.maxrev
- Apply the patch only ifSRCREV
is not later thanmaxrev
.rev
- Apply the patch only ifSRCREV
is equal torev
.notrev
- Apply the patch only ifSRCREV
is not equal torev
.
Here are some additional options worth mentioning:
unpack
- Controls whether or not to unpack the file if it is an archive. The default action is to unpack the file.destsuffix
- Places the file (or extracts its contents) into the specified subdirectory of WORKDIR when the Git fetcher is used.subdir
- Places the file (or extracts its contents) into the specified subdirectory ofWORKDIR
when the local (file://
) fetcher is used.localdir
- Places the file (or extracts its contents) into the specified subdirectory ofWORKDIR
when the CVS fetcher is used.subpath
- Limits the checkout to a specific subpath of the tree when using the Git fetcher is used.name
- Specifies a name to be used for association withSRC_URI
checksums or SRCREV when you have more than one file or git repository specified inSRC_URI
. For example:SRC_URI = "git://example.com/foo.git;name=first \ git://example.com/bar.git;name=second \ http://example.com/file.tar.gz;name=third" SRCREV_first = "f1d2d2f924e986ac86fdf7b36c94bcdf32beec15" SRCREV_second = "e242ed3bffccdf271b7fbaf34ed72d089537b42f" SRC_URI[third.sha256sum] = "13550350a8681c84c861aac2e5b440161c2b33a3e4f302ac680ca5b686de48de"
downloadfilename
- Specifies the filename used when storing the downloaded file.
- SRC_URI_OVERRIDES_PACKAGE_ARCH
By default, the OpenEmbedded build system automatically detects whether
SRC_URI
contains files that are machine-specific. If so, the build system automatically changesPACKAGE_ARCH
. Setting this variable to “0” disables this behavior.- SRCDATE
The date of the source code used to build the package. This variable applies only if the source was fetched from a Source Code Manager (SCM).
- SRCPV
Returns the version string of the current package. This string is used to help define the value of PV.
The
SRCPV
variable is defined in themeta/conf/bitbake.conf
configuration file in the Source Directory as follows:SRCPV = "${@bb.fetch2.get_srcrev(d)}"
Recipes that need to define
PV
do so with the help of theSRCPV
. For example, theofono
recipe (ofono_git.bb
) located inmeta/recipes-connectivity
in the Source Directory definesPV
as follows:PV = "0.12-git${SRCPV}"
- SRCREV
The revision of the source code used to build the package. This variable applies to Subversion, Git, Mercurial, and Bazaar only. Note that if you want to build a fixed revision and you want to avoid performing a query on the remote repository every time BitBake parses your recipe, you should specify a
SRCREV
that is a full revision identifier and not just a tag.Note
For information on limitations when inheriting the latest revision of software using
SRCREV
, see the AUTOREV variable description and the “Automatically Incrementing a Package Version Number” section, which is in the Yocto Project Development Tasks Manual.- SSTATE_DIR
The directory for the shared state cache.
- SSTATE_MIRROR_ALLOW_NETWORK
If set to “1”, allows fetches from mirrors that are specified in SSTATE_MIRRORS to work even when fetching from the network is disabled by setting
BB_NO_NETWORK
to “1”. Using theSSTATE_MIRROR_ALLOW_NETWORK
variable is useful if you have setSSTATE_MIRRORS
to point to an internal server for your shared state cache, but you want to disable any other fetching from the network.- SSTATE_MIRRORS
Configures the OpenEmbedded build system to search other mirror locations for prebuilt cache data objects before building out the data. This variable works like fetcher MIRRORS and PREMIRRORS and points to the cache locations to check for the shared state (sstate) objects.
You can specify a filesystem directory or a remote URL such as HTTP or FTP. The locations you specify need to contain the shared state cache (sstate-cache) results from previous builds. The sstate-cache you point to can also be from builds on other machines.
When pointing to sstate build artifacts on another machine that uses a different GCC version for native builds, you must configure
SSTATE_MIRRORS
with a regular expression that maps local search paths to server paths. The paths need to take into account NATIVELSBSTRING set by the uninative class. For example, the following maps the local search pathuniversal-4.9
to the server-provided path server_url_sstate_path:SSTATE_MIRRORS ?= "file://universal-4.9/(.*) http://server_url_sstate_path/universal-4.8/\1 \n"
If a mirror uses the same structure as SSTATE_DIR, you need to add “PATH” at the end as shown in the examples below. The build system substitutes the correct path within the directory structure.
SSTATE_MIRRORS ?= "\ file://.* http://someserver.tld/share/sstate/PATH;downloadfilename=PATH \n \ file://.* file:///some-local-dir/sstate/PATH"
- SSTATE_SCAN_FILES
Controls the list of files the OpenEmbedded build system scans for hardcoded installation paths. The variable uses a space-separated list of filenames (not paths) with standard wildcard characters allowed.
During a build, the OpenEmbedded build system creates a shared state (sstate) object during the first stage of preparing the sysroots. That object is scanned for hardcoded paths for original installation locations. The list of files that are scanned for paths is controlled by the
SSTATE_SCAN_FILES
variable. Typically, recipes add files they want to be scanned to the value ofSSTATE_SCAN_FILES
rather than the variable being comprehensively set. The sstate class specifies the default list of files.For details on the process, see the staging class.
- STAGING_BASE_LIBDIR_NATIVE
Specifies the path to the
/lib
subdirectory of the sysroot directory for the build host.- STAGING_BASELIBDIR
Specifies the path to the
/lib
subdirectory of the sysroot directory for the target for which the current recipe is being built (STAGING_DIR_HOST).- STAGING_BINDIR
Specifies the path to the
/usr/bin
subdirectory of the sysroot directory for the target for which the current recipe is being built (STAGING_DIR_HOST).- STAGING_BINDIR_CROSS
Specifies the path to the directory containing binary configuration scripts. These scripts provide configuration information for other software that wants to make use of libraries or include files provided by the software associated with the script.
Note
This style of build configuration has been largely replaced by
pkg-config
. Consequently, ifpkg-config
is supported by the library to which you are linking, it is recommended you usepkg-config
instead of a provided configuration script.- STAGING_BINDIR_NATIVE
Specifies the path to the
/usr/bin
subdirectory of the sysroot directory for the build host.- STAGING_DATADIR
Specifies the path to the
/usr/share
subdirectory of the sysroot directory for the target for which the current recipe is being built (STAGING_DIR_HOST).- STAGING_DATADIR_NATIVE
Specifies the path to the
/usr/share
subdirectory of the sysroot directory for the build host.- STAGING_DIR
Helps construct the
recipe-sysroots
directory, which is used during packaging.For information on how staging for recipe-specific sysroots occurs, see the do_populate_sysroot task, the “Sharing Files Between Recipes” section in the Yocto Project Development Tasks Manual, the “Configuration, Compilation, and Staging” section in the Yocto Project Overview and Concepts Manual, and the SYSROOT_DIRS variable.
Note
Recipes should never write files directly under the
STAGING_DIR
directory because the OpenEmbedded build system manages the directory automatically. Instead, files should be installed to${
D}
within your recipe’s do_install task and then the OpenEmbedded build system will stage a subset of those files into the sysroot.- STAGING_DIR_HOST
Specifies the path to the sysroot directory for the system on which the component is built to run (the system that hosts the component). For most recipes, this sysroot is the one in which that recipe’s do_populate_sysroot task copies files. Exceptions include
-native
recipes, where thedo_populate_sysroot
task instead uses STAGING_DIR_NATIVE. Depending on the type of recipe and the build target,STAGING_DIR_HOST
can have the following values:For recipes building for the target machine, the value is “${STAGING_DIR}/${MACHINE}”.
For native recipes building for the build host, the value is empty given the assumption that when building for the build host, the build host’s own directories should be used.
Note
-native
recipes are not installed into host paths like such as/usr
. Rather, these recipes are installed intoSTAGING_DIR_NATIVE
. When compiling-native
recipes, standard build environment variables such as CPPFLAGS and CFLAGS are set up so that both host paths andSTAGING_DIR_NATIVE
are searched for libraries and headers using, for example, GCC’s-isystem
option.Thus, the emphasis is that the
STAGING_DIR*
variables should be viewed as input variables by tasks such as do_configure, do_compile, and do_install. Having the real system root correspond toSTAGING_DIR_HOST
makes conceptual sense for-native
recipes, as they make use of host headers and libraries.
- STAGING_DIR_NATIVE
Specifies the path to the sysroot directory used when building components that run on the build host itself.
- STAGING_DIR_TARGET
Specifies the path to the sysroot used for the system for which the component generates code. For components that do not generate code, which is the majority,
STAGING_DIR_TARGET
is set to match STAGING_DIR_HOST.Some recipes build binaries that can run on the target system but those binaries in turn generate code for another different system (e.g. cross-canadian recipes). Using terminology from GNU, the primary system is referred to as the “HOST” and the secondary, or different, system is referred to as the “TARGET”. Thus, the binaries run on the “HOST” system and generate binaries for the “TARGET” system. The
STAGING_DIR_HOST
variable points to the sysroot used for the “HOST” system, whileSTAGING_DIR_TARGET
points to the sysroot used for the “TARGET” system.- STAGING_ETCDIR_NATIVE
Specifies the path to the
/etc
subdirectory of the sysroot directory for the build host.- STAGING_EXECPREFIXDIR
Specifies the path to the
/usr
subdirectory of the sysroot directory for the target for which the current recipe is being built (STAGING_DIR_HOST).- STAGING_INCDIR
Specifies the path to the
/usr/include
subdirectory of the sysroot directory for the target for which the current recipe being built (STAGING_DIR_HOST).- STAGING_INCDIR_NATIVE
Specifies the path to the
/usr/include
subdirectory of the sysroot directory for the build host.- STAGING_KERNEL_BUILDDIR
Points to the directory containing the kernel build artifacts. Recipes building software that needs to access kernel build artifacts (e.g.
systemtap-uprobes
) can look in the directory specified with theSTAGING_KERNEL_BUILDDIR
variable to find these artifacts after the kernel has been built.- STAGING_KERNEL_DIR
The directory with kernel headers that are required to build out-of-tree modules.
- STAGING_LIBDIR
Specifies the path to the
/usr/lib
subdirectory of the sysroot directory for the target for which the current recipe is being built (STAGING_DIR_HOST).- STAGING_LIBDIR_NATIVE
Specifies the path to the
/usr/lib
subdirectory of the sysroot directory for the build host.- STAMP
Specifies the base path used to create recipe stamp files. The path to an actual stamp file is constructed by evaluating this string and then appending additional information. Currently, the default assignment for
STAMP
as set in themeta/conf/bitbake.conf
file is:STAMP = "${STAMPS_DIR}/${MULTIMACH_TARGET_SYS}/${PN}/${EXTENDPE}${PV}-${PR}"
For information on how BitBake uses stamp files to determine if a task should be rerun, see the “Stamp Files and the Rerunning of Tasks” section in the Yocto Project Overview and Concepts Manual.
See STAMPS_DIR, MULTIMACH_TARGET_SYS, PN, EXTENDPE, PV, and PR for related variable information.
- STAMPS_DIR
Specifies the base directory in which the OpenEmbedded build system places stamps. The default directory is
${TMPDIR}/stamps
.- STRIP
The minimal command and arguments to run
strip
, which is used to strip symbols.- SUMMARY
The short (72 characters or less) summary of the binary package for packaging systems such as
opkg
,rpm
, ordpkg
. By default,SUMMARY
is used to define the DESCRIPTION variable ifDESCRIPTION
is not set in the recipe.- SVNDIR
The directory in which files checked out of a Subversion system are stored.
- SYSLINUX_DEFAULT_CONSOLE
Specifies the kernel boot default console. If you want to use a console other than the default, set this variable in your recipe as follows where “X” is the console number you want to use:
SYSLINUX_DEFAULT_CONSOLE = "console=ttyX"
The syslinux class initially sets this variable to null but then checks for a value later.
- SYSLINUX_OPTS
Lists additional options to add to the syslinux file. You need to set this variable in your recipe. If you want to list multiple options, separate the options with a semicolon character (
;
).The syslinux class uses this variable to create a set of options.
- SYSLINUX_SERIAL
Specifies the alternate serial port or turns it off. To turn off serial, set this variable to an empty string in your recipe. The variable’s default value is set in the syslinux class as follows:
SYSLINUX_SERIAL ?= "0 115200"
The class checks for and uses the variable as needed.
- SYSLINUX_SERIAL_TTY
Specifies the alternate console=tty… kernel boot argument. The variable’s default value is set in the syslinux class as follows:
SYSLINUX_SERIAL_TTY ?= "console=ttyS0,115200"
The class checks for and uses the variable as needed.
- SYSLINUX_SPLASH
An
.LSS
file used as the background for the VGA boot menu when you use the boot menu. You need to set this variable in your recipe.The syslinux class checks for this variable and if found, the OpenEmbedded build system installs the splash screen.
- SYSROOT_DESTDIR
Points to the temporary directory under the work directory (default “
${
WORKDIR}/sysroot-destdir
”) where the files populated into the sysroot are assembled during the do_populate_sysroot task.- SYSROOT_DIRS
Directories that are staged into the sysroot by the do_populate_sysroot task. By default, the following directories are staged:
SYSROOT_DIRS = " \ ${includedir} \ ${libdir} \ ${base_libdir} \ ${nonarch_base_libdir} \ ${datadir} \ "
- SYSROOT_DIRS_BLACKLIST
Directories that are not staged into the sysroot by the do_populate_sysroot task. You can use this variable to exclude certain subdirectories of directories listed in SYSROOT_DIRS from staging. By default, the following directories are not staged:
SYSROOT_DIRS_BLACKLIST = " \ ${mandir} \ ${docdir} \ ${infodir} \ ${datadir}/locale \ ${datadir}/applications \ ${datadir}/fonts \ ${datadir}/pixmaps \ "
- SYSROOT_DIRS_NATIVE
Extra directories staged into the sysroot by the do_populate_sysroot task for
-native
recipes, in addition to those specified in SYSROOT_DIRS. By default, the following extra directories are staged:SYSROOT_DIRS_NATIVE = " \ ${bindir} \ ${sbindir} \ ${base_bindir} \ ${base_sbindir} \ ${libexecdir} \ ${sysconfdir} \ ${localstatedir} \ "
Note
Programs built by
-native
recipes run directly from the sysroot (STAGING_DIR_NATIVE), which is why additional directories containing program executables and supporting files need to be staged.- SYSROOT_PREPROCESS_FUNCS
A list of functions to execute after files are staged into the sysroot. These functions are usually used to apply additional processing on the staged files, or to stage additional files.
- SYSTEMD_AUTO_ENABLE
When inheriting the systemd class, this variable specifies whether the specified service in SYSTEMD_SERVICE should start automatically or not. By default, the service is enabled to automatically start at boot time. The default setting is in the systemd class as follows:
SYSTEMD_AUTO_ENABLE ??= "enable"
You can disable the service by setting the variable to “disable”.
- SYSTEMD_BOOT_CFG
When EFI_PROVIDER is set to “systemd-boot”, the
SYSTEMD_BOOT_CFG
variable specifies the configuration file that should be used. By default, the systemd-boot class sets theSYSTEMD_BOOT_CFG
as follows:SYSTEMD_BOOT_CFG ?= "${:term:`S`}/loader.conf"
For information on Systemd-boot, see the Systemd-boot documentation.
- SYSTEMD_BOOT_ENTRIES
When EFI_PROVIDER is set to “systemd-boot”, the
SYSTEMD_BOOT_ENTRIES
variable specifies a list of entry files (*.conf
) to install that contain one boot entry per file. By default, the systemd-boot class sets theSYSTEMD_BOOT_ENTRIES
as follows:SYSTEMD_BOOT_ENTRIES ?= ""
For information on Systemd-boot, see the Systemd-boot documentation.
- SYSTEMD_BOOT_TIMEOUT
When EFI_PROVIDER is set to “systemd-boot”, the
SYSTEMD_BOOT_TIMEOUT
variable specifies the boot menu timeout in seconds. By default, the systemd-boot class sets theSYSTEMD_BOOT_TIMEOUT
as follows:SYSTEMD_BOOT_TIMEOUT ?= "10"
For information on Systemd-boot, see the Systemd-boot documentation.
- SYSTEMD_PACKAGES
When inheriting the systemd class, this variable locates the systemd unit files when they are not found in the main recipe’s package. By default, the
SYSTEMD_PACKAGES
variable is set such that the systemd unit files are assumed to reside in the recipes main package:SYSTEMD_PACKAGES ?= "${PN}"
If these unit files are not in this recipe’s main package, you need to use
SYSTEMD_PACKAGES
to list the package or packages in which the build system can find the systemd unit files.- SYSTEMD_SERVICE
When inheriting the systemd class, this variable specifies the systemd service name for a package.
When you specify this file in your recipe, use a package name override to indicate the package to which the value applies. Here is an example from the connman recipe:
SYSTEMD_SERVICE_${PN} = "connman.service"
- SYSVINIT_ENABLED_GETTYS
When using SysVinit, specifies a space-separated list of the virtual terminals that should run a getty (allowing login), assuming USE_VT is not set to “0”.
The default value for
SYSVINIT_ENABLED_GETTYS
is “1” (i.e. only run a getty on the first virtual terminal).- T
This variable points to a directory were BitBake places temporary files, which consist mostly of task logs and scripts, when building a particular recipe. The variable is typically set as follows:
T = "${WORKDIR}/temp"
The WORKDIR is the directory into which BitBake unpacks and builds the recipe. The default
bitbake.conf
file sets this variable.The
T
variable is not to be confused with the TMPDIR variable, which points to the root of the directory tree where BitBake places the output of an entire build.- TARGET_ARCH
The target machine’s architecture. The OpenEmbedded build system supports many architectures. Here is an example list of architectures supported. This list is by no means complete as the architecture is configurable:
arm
i586
x86_64
powerpc
powerpc64
mips
mipsel
For additional information on machine architectures, see the TUNE_ARCH variable.
- TARGET_AS_ARCH
Specifies architecture-specific assembler flags for the target system.
TARGET_AS_ARCH
is initialized from TUNE_ASARGS by default in the BitBake configuration file (meta/conf/bitbake.conf
):TARGET_AS_ARCH = "${TUNE_ASARGS}"
- TARGET_CC_ARCH
Specifies architecture-specific C compiler flags for the target system.
TARGET_CC_ARCH
is initialized from TUNE_CCARGS by default.Note
It is a common workaround to append LDFLAGS to
TARGET_CC_ARCH
in recipes that build software for the target that would not otherwise respect the exportedLDFLAGS
variable.- TARGET_CC_KERNEL_ARCH
This is a specific kernel compiler flag for a CPU or Application Binary Interface (ABI) tune. The flag is used rarely and only for cases where a userspace TUNE_CCARGS is not compatible with the kernel compilation. The
TARGET_CC_KERNEL_ARCH
variable allows the kernel (and associated modules) to use a different configuration. See themeta/conf/machine/include/arm/feature-arm-thumb.inc
file in the Source Directory for an example.- TARGET_CFLAGS
Specifies the flags to pass to the C compiler when building for the target. When building in the target context, CFLAGS is set to the value of this variable by default.
Additionally, the SDK’s environment setup script sets the
CFLAGS
variable in the environment to theTARGET_CFLAGS
value so that executables built using the SDK also have the flags applied.- TARGET_CPPFLAGS
Specifies the flags to pass to the C pre-processor (i.e. to both the C and the C++ compilers) when building for the target. When building in the target context, CPPFLAGS is set to the value of this variable by default.
Additionally, the SDK’s environment setup script sets the
CPPFLAGS
variable in the environment to theTARGET_CPPFLAGS
value so that executables built using the SDK also have the flags applied.- TARGET_CXXFLAGS
Specifies the flags to pass to the C++ compiler when building for the target. When building in the target context, CXXFLAGS is set to the value of this variable by default.
Additionally, the SDK’s environment setup script sets the
CXXFLAGS
variable in the environment to theTARGET_CXXFLAGS
value so that executables built using the SDK also have the flags applied.- TARGET_FPU
Specifies the method for handling FPU code. For FPU-less targets, which include most ARM CPUs, the variable must be set to “soft”. If not, the kernel emulation gets used, which results in a performance penalty.
- TARGET_LD_ARCH
Specifies architecture-specific linker flags for the target system.
TARGET_LD_ARCH
is initialized from TUNE_LDARGS by default in the BitBake configuration file (meta/conf/bitbake.conf
):TARGET_LD_ARCH = "${TUNE_LDARGS}"
- TARGET_LDFLAGS
Specifies the flags to pass to the linker when building for the target. When building in the target context, LDFLAGS is set to the value of this variable by default.
Additionally, the SDK’s environment setup script sets the LDFLAGS variable in the environment to the
TARGET_LDFLAGS
value so that executables built using the SDK also have the flags applied.- TARGET_OS
Specifies the target’s operating system. The variable can be set to “linux” for glibc-based systems (GNU C Library) and to “linux-musl” for musl libc. For ARM/EABI targets, “linux-gnueabi” and “linux-musleabi” possible values exist.
- TARGET_PREFIX
Specifies the prefix used for the toolchain binary target tools.
Depending on the type of recipe and the build target,
TARGET_PREFIX
is set as follows:For recipes building for the target machine, the value is “${TARGET_SYS}-“.
For native recipes, the build system sets the variable to the value of
BUILD_PREFIX
.For native SDK recipes (
nativesdk
), the build system sets the variable to the value ofSDK_PREFIX
.
- TARGET_SYS
Specifies the system, including the architecture and the operating system, for which the build is occurring in the context of the current recipe.
The OpenEmbedded build system automatically sets this variable based on TARGET_ARCH, TARGET_VENDOR, and TARGET_OS variables.
Note
You do not need to set the
TARGET_SYS
variable yourself.Consider these two examples:
Given a native recipe on a 32-bit, x86 machine running Linux, the value is “i686-linux”.
Given a recipe being built for a little-endian, MIPS target running Linux, the value might be “mipsel-linux”.
- TARGET_VENDOR
Specifies the name of the target vendor.
- TCLIBC
Specifies the GNU standard C library (
libc
) variant to use during the build process. This variable replacesPOKYLIBC
, which is no longer supported.You can select “glibc”, “musl”, “newlib”, or “baremetal”
- TCLIBCAPPEND
Specifies a suffix to be appended onto the TMPDIR value. The suffix identifies the
libc
variant for building. When you are building for multiple variants with the same Build Directory, this mechanism ensures that output for differentlibc
variants is kept separate to avoid potential conflicts.In the
defaultsetup.conf
file, the default value ofTCLIBCAPPEND
is “-${TCLIBC}”. However, distros such as poky, which normally only support onelibc
variant, setTCLIBCAPPEND
to “” in their distro configuration file resulting in no suffix being applied.- TCMODE
Specifies the toolchain selector.
TCMODE
controls the characteristics of the generated packages and images by telling the OpenEmbedded build system which toolchain profile to use. By default, the OpenEmbedded build system builds its own internal toolchain. The variable’s default value is “default”, which uses that internal toolchain.Note
If
TCMODE
is set to a value other than “default”, then it is your responsibility to ensure that the toolchain is compatible with the default toolchain. Using older or newer versions of these components might cause build problems. See the Release Notes for the Yocto Project release for the specific components with which the toolchain must be compatible. To access the Release Notes, go to the Downloads page on the Yocto Project website and click on the “RELEASE INFORMATION” link for the appropriate release.The
TCMODE
variable is similar to TCLIBC, which controls the variant of the GNU standard C library (libc
) used during the build process:glibc
ormusl
.With additional layers, it is possible to use a pre-compiled external toolchain. One example is the Sourcery G++ Toolchain. The support for this toolchain resides in the separate Mentor Graphics
meta-sourcery
layer at https://github.com/MentorEmbedded/meta-sourcery/.The layer’s
README
file contains information on how to use the Sourcery G++ Toolchain as an external toolchain. In summary, you must be sure to add the layer to yourbblayers.conf
file in front of themeta
layer and then set theEXTERNAL_TOOLCHAIN
variable in yourlocal.conf
file to the location in which you installed the toolchain.The fundamentals used for this example apply to any external toolchain. You can use
meta-sourcery
as a template for adding support for other external toolchains.- TEST_EXPORT_DIR
The location the OpenEmbedded build system uses to export tests when the TEST_EXPORT_ONLY variable is set to “1”.
The
TEST_EXPORT_DIR
variable defaults to"${TMPDIR}/testimage/${PN}"
.- TEST_EXPORT_ONLY
Specifies to export the tests only. Set this variable to “1” if you do not want to run the tests but you want them to be exported in a manner that you to run them outside of the build system.
- TEST_LOG_DIR
Holds the SSH log and the boot log for QEMU machines. The
TEST_LOG_DIR
variable defaults to"${WORKDIR}/testimage"
.Note
Actual test results reside in the task log (
log.do_testimage
), which is in the${WORKDIR}/temp/
directory.- TEST_POWERCONTROL_CMD
For automated hardware testing, specifies the command to use to control the power of the target machine under test. Typically, this command would point to a script that performs the appropriate action (e.g. interacting with a web-enabled power strip). The specified command should expect to receive as the last argument “off”, “on” or “cycle” specifying to power off, on, or cycle (power off and then power on) the device, respectively.
- TEST_POWERCONTROL_EXTRA_ARGS
For automated hardware testing, specifies additional arguments to pass through to the command specified in TEST_POWERCONTROL_CMD. Setting
TEST_POWERCONTROL_EXTRA_ARGS
is optional. You can use it if you wish, for example, to separate the machine-specific and non-machine-specific parts of the arguments.- TEST_QEMUBOOT_TIMEOUT
The time in seconds allowed for an image to boot before automated runtime tests begin to run against an image. The default timeout period to allow the boot process to reach the login prompt is 500 seconds. You can specify a different value in the
local.conf
file.For more information on testing images, see the “Performing Automated Runtime Testing” section in the Yocto Project Development Tasks Manual.
- TEST_SERIALCONTROL_CMD
For automated hardware testing, specifies the command to use to connect to the serial console of the target machine under test. This command simply needs to connect to the serial console and forward that connection to standard input and output as any normal terminal program does.
For example, to use the Picocom terminal program on serial device
/dev/ttyUSB0
at 115200bps, you would set the variable as follows:TEST_SERIALCONTROL_CMD = "picocom /dev/ttyUSB0 -b 115200"
- TEST_SERIALCONTROL_EXTRA_ARGS
For automated hardware testing, specifies additional arguments to pass through to the command specified in TEST_SERIALCONTROL_CMD. Setting
TEST_SERIALCONTROL_EXTRA_ARGS
is optional. You can use it if you wish, for example, to separate the machine-specific and non-machine-specific parts of the command.- TEST_SERVER_IP
The IP address of the build machine (host machine). This IP address is usually automatically detected. However, if detection fails, this variable needs to be set to the IP address of the build machine (i.e. where the build is taking place).
Note
The
TEST_SERVER_IP
variable is only used for a small number of tests such as the “dnf” test suite, which needs to download packages fromWORKDIR/oe-rootfs-repo
.- TEST_SUITES
An ordered list of tests (modules) to run against an image when performing automated runtime testing.
The OpenEmbedded build system provides a core set of tests that can be used against images.
Note
Currently, there is only support for running these tests under QEMU.
Tests include
ping
,ssh
,df
among others. You can add your own tests to the list of tests by appendingTEST_SUITES
as follows:TEST_SUITES_append = " mytest"
Alternatively, you can provide the “auto” option to have all applicable tests run against the image.
TEST_SUITES_append = " auto"
Using this option causes the build system to automatically run tests that are applicable to the image. Tests that are not applicable are skipped.
The order in which tests are run is important. Tests that depend on another test must appear later in the list than the test on which they depend. For example, if you append the list of tests with two tests (
test_A
andtest_B
) wheretest_B
is dependent ontest_A
, then you must order the tests as follows:TEST_SUITES = "test_A test_B"
For more information on testing images, see the “Performing Automated Runtime Testing” section in the Yocto Project Development Tasks Manual.
- TEST_TARGET
Specifies the target controller to use when running tests against a test image. The default controller to use is “qemu”:
TEST_TARGET = "qemu"
A target controller is a class that defines how an image gets deployed on a target and how a target is started. A layer can extend the controllers by adding a module in the layer’s
/lib/oeqa/controllers
directory and by inheriting theBaseTarget
class, which is an abstract class that cannot be used as a value ofTEST_TARGET
.You can provide the following arguments with
TEST_TARGET
:“qemu”: Boots a QEMU image and runs the tests. See the “Enabling Runtime Tests on QEMU” section in the Yocto Project Development Tasks Manual for more information.
“simpleremote”: Runs the tests on target hardware that is already up and running. The hardware can be on the network or it can be a device running an image on QEMU. You must also set TEST_TARGET_IP when you use “simpleremote”.
Note
This argument is defined in
meta/lib/oeqa/controllers/simpleremote.py
.
For information on running tests on hardware, see the “Enabling Runtime Tests on Hardware” section in the Yocto Project Development Tasks Manual.
- TEST_TARGET_IP
The IP address of your hardware under test. The
TEST_TARGET_IP
variable has no effect when TEST_TARGET is set to “qemu”.When you specify the IP address, you can also include a port. Here is an example:
TEST_TARGET_IP = "192.168.1.4:2201"
Specifying a port is useful when SSH is started on a non-standard port or in cases when your hardware under test is behind a firewall or network that is not directly accessible from your host and you need to do port address translation.
- TESTIMAGE_AUTO
Automatically runs the series of automated tests for images when an image is successfully built. Setting
TESTIMAGE_AUTO
to “1” causes any image that successfully builds to automatically boot under QEMU. Using the variable also adds in dependencies so that any SDK for which testing is requested is automatically built first.These tests are written in Python making use of the
unittest
module, and the majority of them run commands on the target system overssh
. You can set this variable to “1” in yourlocal.conf
file in the Build Directory to have the OpenEmbedded build system automatically run these tests after an image successfully builds:TESTIMAGE_AUTO = “1”
For more information on enabling, running, and writing these tests, see the “Performing Automated Runtime Testing” section in the Yocto Project Development Tasks Manual and the “testimage*.bbclass” section.
- THISDIR
The directory in which the file BitBake is currently parsing is located. Do not manually set this variable.
- TIME
The time the build was started. Times appear using the hour, minute, and second (HMS) format (e.g. “140159” for one minute and fifty-nine seconds past 1400 hours).
- TMPDIR
This variable is the base directory the OpenEmbedded build system uses for all build output and intermediate files (other than the shared state cache). By default, the
TMPDIR
variable points totmp
within the Build Directory.If you want to establish this directory in a location other than the default, you can uncomment and edit the following statement in the
conf/local.conf
file in the Source Directory:#TMPDIR = "${TOPDIR}/tmp"
An example use for this scenario is to set
TMPDIR
to a local disk, which does not use NFS, while having the Build Directory use NFS.The filesystem used by
TMPDIR
must have standard filesystem semantics (i.e. mixed-case files are unique, POSIX file locking, and persistent inodes). Due to various issues with NFS and bugs in some implementations, NFS does not meet this minimum requirement. Consequently,TMPDIR
cannot be on NFS.- TOOLCHAIN_HOST_TASK
This variable lists packages the OpenEmbedded build system uses when building an SDK, which contains a cross-development environment. The packages specified by this variable are part of the toolchain set that runs on the SDKMACHINE, and each package should usually have the prefix
nativesdk-
. For example, consider the following command when building an SDK:$ bitbake -c populate_sdk imagename
In this case, a default list of packages is set in this variable, but you can add additional packages to the list. See the “Adding Individual Packages to the Standard SDK” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual for more information.
For background information on cross-development toolchains in the Yocto Project development environment, see the “The Cross-Development Toolchain” section in the Yocto Project Overview and Concepts Manual. For information on setting up a cross-development environment, see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
- TOOLCHAIN_OUTPUTNAME
This variable defines the name used for the toolchain output. The populate_sdk_base class sets the
TOOLCHAIN_OUTPUTNAME
variable as follows:TOOLCHAIN_OUTPUTNAME ?= "${SDK_NAME}-toolchain-${SDK_VERSION}"
See the SDK_NAME and SDK_VERSION variables for additional information.
- TOOLCHAIN_TARGET_TASK
This variable lists packages the OpenEmbedded build system uses when it creates the target part of an SDK (i.e. the part built for the target hardware), which includes libraries and headers. Use this variable to add individual packages to the part of the SDK that runs on the target. See the “Adding Individual Packages to the Standard SDK” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual for more information.
For background information on cross-development toolchains in the Yocto Project development environment, see the “The Cross-Development Toolchain” section in the Yocto Project Overview and Concepts Manual. For information on setting up a cross-development environment, see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
- TOPDIR
The top-level Build Directory. BitBake automatically sets this variable when you initialize your build environment using oe-init-build-env.
- TRANSLATED_TARGET_ARCH
A sanitized version of TARGET_ARCH. This variable is used where the architecture is needed in a value where underscores are not allowed, for example within package filenames. In this case, dash characters replace any underscore characters used in
TARGET_ARCH
.Do not edit this variable.
- TUNE_ARCH
The GNU canonical architecture for a specific architecture (i.e.
arm
,armeb
,mips
,mips64
, and so forth). BitBake uses this value to setup configuration.TUNE_ARCH
definitions are specific to a given architecture. The definitions can be a single static definition, or can be dynamically adjusted. You can see details for a given CPU family by looking at the architecture’sREADME
file. For example, themeta/conf/machine/include/mips/README
file in the Source Directory provides information forTUNE_ARCH
specific to themips
architecture.TUNE_ARCH
is tied closely to TARGET_ARCH, which defines the target machine’s architecture. The BitBake configuration file (meta/conf/bitbake.conf
) setsTARGET_ARCH
as follows:TARGET_ARCH = "${TUNE_ARCH}"
The following list, which is by no means complete since architectures are configurable, shows supported machine architectures:
arm
i586
x86_64
powerpc
powerpc64
mips
mipsel
- TUNE_ASARGS
Specifies architecture-specific assembler flags for the target system. The set of flags is based on the selected tune features.
TUNE_ASARGS
is set using the tune include files, which are typically undermeta/conf/machine/include/
and are influenced through TUNE_FEATURES. For example, themeta/conf/machine/include/x86/arch-x86.inc
file defines the flags for the x86 architecture as follows:TUNE_ASARGS += "${@bb.utils.contains("TUNE_FEATURES", "mx32", "-x32", "", d)}"
Note
Board Support Packages (BSPs) select the tune. The selected tune, in turn, affects the tune variables themselves (i.e. the tune can supply its own set of flags).
- TUNE_CCARGS
Specifies architecture-specific C compiler flags for the target system. The set of flags is based on the selected tune features.
TUNE_CCARGS
is set using the tune include files, which are typically undermeta/conf/machine/include/
and are influenced through TUNE_FEATURES.Note
Board Support Packages (BSPs) select the tune. The selected tune, in turn, affects the tune variables themselves (i.e. the tune can supply its own set of flags).
- TUNE_FEATURES
Features used to “tune” a compiler for optimal use given a specific processor. The features are defined within the tune files and allow arguments (i.e.
TUNE_*ARGS
) to be dynamically generated based on the features.The OpenEmbedded build system verifies the features to be sure they are not conflicting and that they are supported.
The BitBake configuration file (
meta/conf/bitbake.conf
) definesTUNE_FEATURES
as follows:TUNE_FEATURES ??= "${TUNE_FEATURES_tune-${DEFAULTTUNE}}"
See the DEFAULTTUNE variable for more information.
- TUNE_LDARGS
Specifies architecture-specific linker flags for the target system. The set of flags is based on the selected tune features.
TUNE_LDARGS
is set using the tune include files, which are typically undermeta/conf/machine/include/
and are influenced through TUNE_FEATURES. For example, themeta/conf/machine/include/x86/arch-x86.inc
file defines the flags for the x86 architecture as follows:TUNE_LDARGS += "${@bb.utils.contains("TUNE_FEATURES", "mx32", "-m elf32_x86_64", "", d)}"
Note
Board Support Packages (BSPs) select the tune. The selected tune, in turn, affects the tune variables themselves (i.e. the tune can supply its own set of flags).
- TUNE_PKGARCH
The package architecture understood by the packaging system to define the architecture, ABI, and tuning of output packages. The specific tune is defined using the “_tune” override as follows:
TUNE_PKGARCH_tune-tune = "tune"
These tune-specific package architectures are defined in the machine include files. Here is an example of the “core2-32” tuning as used in the
meta/conf/machine/include/tune-core2.inc
file:TUNE_PKGARCH_tune-core2-32 = "core2-32"
- TUNEABI
An underlying Application Binary Interface (ABI) used by a particular tuning in a given toolchain layer. Providers that use prebuilt libraries can use the
TUNEABI
, TUNEABI_OVERRIDE, and TUNEABI_WHITELIST variables to check compatibility of tunings against their selection of libraries.If
TUNEABI
is undefined, then every tuning is allowed. See the sanity class to see how the variable is used.- TUNEABI_OVERRIDE
If set, the OpenEmbedded system ignores the TUNEABI_WHITELIST variable. Providers that use prebuilt libraries can use the
TUNEABI_OVERRIDE
,TUNEABI_WHITELIST
, and TUNEABI variables to check compatibility of a tuning against their selection of libraries.See the sanity class to see how the variable is used.
- TUNEABI_WHITELIST
A whitelist of permissible TUNEABI values. If
TUNEABI_WHITELIST
is not set, all tunes are allowed. Providers that use prebuilt libraries can use theTUNEABI_WHITELIST
, TUNEABI_OVERRIDE, andTUNEABI
variables to check compatibility of a tuning against their selection of libraries.See the sanity class to see how the variable is used.
- TUNECONFLICTS[feature]
Specifies CPU or Application Binary Interface (ABI) tuning features that conflict with feature.
Known tuning conflicts are specified in the machine include files in the Source Directory. Here is an example from the
meta/conf/machine/include/mips/arch-mips.inc
include file that lists the “o32” and “n64” features as conflicting with the “n32” feature:TUNECONFLICTS[n32] = "o32 n64"
- TUNEVALID[feature]
Specifies a valid CPU or Application Binary Interface (ABI) tuning feature. The specified feature is stored as a flag. Valid features are specified in the machine include files (e.g.
meta/conf/machine/include/arm/arch-arm.inc
). Here is an example from that file:TUNEVALID[bigendian] = "Enable big-endian mode."
See the machine include files in the Source Directory for these features.
- UBOOT_CONFIG
Configures the UBOOT_MACHINE and can also define IMAGE_FSTYPES for individual cases.
Following is an example from the
meta-fsl-arm
layer.UBOOT_CONFIG ??= "sd" UBOOT_CONFIG[sd] = "mx6qsabreauto_config,sdcard" UBOOT_CONFIG[eimnor] = "mx6qsabreauto_eimnor_config" UBOOT_CONFIG[nand] = "mx6qsabreauto_nand_config,ubifs" UBOOT_CONFIG[spinor] = "mx6qsabreauto_spinor_config"
In this example, “sd” is selected as the configuration of the possible four for the
UBOOT_MACHINE
. The “sd” configuration defines “mx6qsabreauto_config” as the value forUBOOT_MACHINE
, while the “sdcard” specifies theIMAGE_FSTYPES
to use for the U-Boot image.For more information on how the
UBOOT_CONFIG
is handled, see the uboot-config class.- UBOOT_DTB_LOADADDRESS
Specifies the load address for the dtb image used by U-Boot. During FIT image creation, the
UBOOT_DTB_LOADADDRESS
variable is used in kernel-fitimage class to specify the load address to be used in creating the dtb sections of Image Tree Source for the FIT image.- UBOOT_DTBO_LOADADDRESS
Specifies the load address for the dtbo image used by U-Boot. During FIT image creation, the
UBOOT_DTBO_LOADADDRESS
variable is used in kernel-fitimage class to specify the load address to be used in creating the dtbo sections of Image Tree Source for the FIT image.- UBOOT_ENTRYPOINT
Specifies the entry point for the U-Boot image. During U-Boot image creation, the
UBOOT_ENTRYPOINT
variable is passed as a command-line parameter to theuboot-mkimage
utility.- UBOOT_LOADADDRESS
Specifies the load address for the U-Boot image. During U-Boot image creation, the
UBOOT_LOADADDRESS
variable is passed as a command-line parameter to theuboot-mkimage
utility.- UBOOT_LOCALVERSION
Appends a string to the name of the local version of the U-Boot image. For example, assuming the version of the U-Boot image built was “2013.10”, the full version string reported by U-Boot would be “2013.10-yocto” given the following statement:
UBOOT_LOCALVERSION = "-yocto"
- UBOOT_MACHINE
Specifies the value passed on the
make
command line when building a U-Boot image. The value indicates the target platform configuration. You typically set this variable from the machine configuration file (i.e.conf/machine/machine_name.conf
).Please see the “Selection of Processor Architecture and Board Type” section in the U-Boot README for valid values for this variable.
- UBOOT_MAKE_TARGET
Specifies the target called in the
Makefile
. The default target is “all”.- UBOOT_MKIMAGE
Specifies the name of the mkimage command as used by the kernel-fitimage class to assemble the FIT image. This can be used to substitute an alternative command, wrapper script or function if desired. The default is “uboot-mkimage”.
- UBOOT_MKIMAGE_DTCOPTS
Options for the device tree compiler passed to mkimage ‘-D’ feature while creating FIT image in kernel-fitimage class. If
UBOOT_MKIMAGE_DTCOPTS
is not set then kernel-fitimage will not pass the-D
option to mkimage.- UBOOT_MKIMAGE_SIGN
Specifies the name of the mkimage command as used by the kernel-fitimage class to sign the FIT image after it has been assembled (if enabled). This can be used to substitute an alternative command, wrapper script or function if desired. The default is “${UBOOT_MKIMAGE}”.
- UBOOT_MKIMAGE_SIGN_ARGS
Optionally specifies additional arguments for the kernel-fitimage class to pass to the mkimage command when signing the FIT image.
- UBOOT_RD_ENTRYPOINT
Specifies the entrypoint for the RAM disk image. During FIT image creation, the
UBOOT_RD_ENTRYPOINT
variable is used in kernel-fitimage class to specify the entrypoint to be used in creating the Image Tree Source for the FIT image.- UBOOT_RD_LOADADDRESS
Specifies the load address for the RAM disk image. During FIT image creation, the
UBOOT_RD_LOADADDRESS
variable is used in kernel-fitimage class to specify the load address to be used in creating the Image Tree Source for the FIT image.- UBOOT_SIGN_ENABLE
Enable signing of FIT image. The default value is “0”.
- UBOOT_SIGN_KEYDIR
Location of the directory containing the RSA key and certificate used for signing FIT image.
- UBOOT_SIGN_KEYNAME
The name of keys used for signing U-Boot FIT image stored in UBOOT_SIGN_KEYDIR directory. For e.g. dev.key key and dev.crt certificate stored in UBOOT_SIGN_KEYDIR directory will have UBOOT_SIGN_KEYNAME set to “dev”.
- UBOOT_SUFFIX
Points to the generated U-Boot extension. For example,
u-boot.sb
has a.sb
extension.The default U-Boot extension is
.bin
- UBOOT_TARGET
Specifies the target used for building U-Boot. The target is passed directly as part of the “make” command (e.g. SPL and AIS). If you do not specifically set this variable, the OpenEmbedded build process passes and uses “all” for the target during the U-Boot building process.
- UNKNOWN_CONFIGURE_WHITELIST
Specifies a list of options that, if reported by the configure script as being invalid, should not generate a warning during the do_configure task. Normally, invalid configure options are simply not passed to the configure script (e.g. should be removed from EXTRA_OECONF or PACKAGECONFIG_CONFARGS). However, common options, for example, exist that are passed to all configure scripts at a class level that might not be valid for some configure scripts. It follows that no benefit exists in seeing a warning about these options. For these cases, the options are added to
UNKNOWN_CONFIGURE_WHITELIST
.The configure arguments check that uses
UNKNOWN_CONFIGURE_WHITELIST
is part of the insane class and is only enabled if the recipe inherits the autotools class.- UPDATERCPN
For recipes inheriting the update-rc.d class,
UPDATERCPN
specifies the package that contains the initscript that is enabled.The default value is “${PN}”. Given that almost all recipes that install initscripts package them in the main package for the recipe, you rarely need to set this variable in individual recipes.
- UPSTREAM_CHECK_GITTAGREGEX
You can perform a per-recipe check for what the latest upstream source code version is by calling
bitbake -c checkpkg
recipe. If the recipe source code is provided from Git repositories, the OpenEmbedded build system determines the latest upstream version by picking the latest tag from the list of all repository tags.You can use the
UPSTREAM_CHECK_GITTAGREGEX
variable to provide a regular expression to filter only the relevant tags should the default filter not work correctly.UPSTREAM_CHECK_GITTAGREGEX = "git_tag_regex"
- UPSTREAM_CHECK_REGEX
Use the
UPSTREAM_CHECK_REGEX
variable to specify a different regular expression instead of the default one when the package checking system is parsing the page found using UPSTREAM_CHECK_URI.UPSTREAM_CHECK_REGEX = "package_regex"
- UPSTREAM_CHECK_URI
You can perform a per-recipe check for what the latest upstream source code version is by calling
bitbake -c checkpkg
recipe. If the source code is provided from tarballs, the latest version is determined by fetching the directory listing where the tarball is and attempting to find a later tarball. When this approach does not work, you can useUPSTREAM_CHECK_URI
to provide a different URI that contains the link to the latest tarball.UPSTREAM_CHECK_URI = "recipe_url"
- USE_DEVFS
Determines if
devtmpfs
is used for/dev
population. The default value used forUSE_DEVFS
is “1” when no value is specifically set. Typically, you would setUSE_DEVFS
to “0” for a statically populated/dev
directory.See the “Selecting a Device Manager” section in the Yocto Project Development Tasks Manual for information on how to use this variable.
- USE_VT
When using SysVinit, determines whether or not to run a getty on any virtual terminals in order to enable logging in through those terminals.
The default value used for
USE_VT
is “1” when no default value is specifically set. Typically, you would setUSE_VT
to “0” in the machine configuration file for machines that do not have a graphical display attached and therefore do not need virtual terminal functionality.- USER_CLASSES
A list of classes to globally inherit. These classes are used by the OpenEmbedded build system to enable extra features (e.g.
buildstats
,image-mklibs
, and so forth).The default list is set in your
local.conf
file:USER_CLASSES ?= "buildstats image-mklibs image-prelink"
For more information, see
meta-poky/conf/local.conf.sample
in the Source Directory.- USERADD_ERROR_DYNAMIC
If set to
error
, forces the OpenEmbedded build system to produce an error if the user identification (uid
) and group identification (gid
) values are not defined in any of the files listed in USERADD_UID_TABLES and USERADD_GID_TABLES. If set towarn
, a warning will be issued instead.The default behavior for the build system is to dynamically apply
uid
andgid
values. Consequently, theUSERADD_ERROR_DYNAMIC
variable is by default not set. If you plan on using statically assignedgid
anduid
values, you should set theUSERADD_ERROR_DYNAMIC
variable in yourlocal.conf
file as follows:USERADD_ERROR_DYNAMIC = "error"
Overriding the default behavior implies you are going to also take steps to set static
uid
andgid
values through use of the USERADDEXTENSION, USERADD_UID_TABLES, and USERADD_GID_TABLES variables.Note
There is a difference in behavior between setting
USERADD_ERROR_DYNAMIC
toerror
and setting it towarn
. When it is set towarn
, the build system will report a warning for every undefineduid
andgid
in any recipe. But when it is set toerror
, it will only report errors for recipes that are actually built. This saves you from having to add static IDs for recipes that you know will never be built.- USERADD_GID_TABLES
Specifies a password file to use for obtaining static group identification (
gid
) values when the OpenEmbedded build system adds a group to the system during package installation.When applying static group identification (
gid
) values, the OpenEmbedded build system looks in BBPATH for afiles/group
file and then applies thoseuid
values. Set the variable as follows in yourlocal.conf
file:USERADD_GID_TABLES = "files/group"
Note
Setting the USERADDEXTENSION variable to “useradd-staticids” causes the build system to use static
gid
values.- USERADD_PACKAGES
When inheriting the useradd class, this variable specifies the individual packages within the recipe that require users and/or groups to be added.
You must set this variable if the recipe inherits the class. For example, the following enables adding a user for the main package in a recipe:
USERADD_PACKAGES = "${PN}"
Note
It follows that if you are going to use the
USERADD_PACKAGES
variable, you need to set one or more of the USERADD_PARAM, GROUPADD_PARAM, or GROUPMEMS_PARAM variables.- USERADD_PARAM
When inheriting the useradd class, this variable specifies for a package what parameters should pass to the
useradd
command if you add a user to the system when the package is installed.Here is an example from the
dbus
recipe:USERADD_PARAM_${PN} = "--system --home ${localstatedir}/lib/dbus \ --no-create-home --shell /bin/false \ --user-group messagebus"
For information on the standard Linux shell command
useradd
, see https://linux.die.net/man/8/useradd.- USERADD_UID_TABLES
Specifies a password file to use for obtaining static user identification (
uid
) values when the OpenEmbedded build system adds a user to the system during package installation.When applying static user identification (
uid
) values, the OpenEmbedded build system looks in BBPATH for afiles/passwd
file and then applies thoseuid
values. Set the variable as follows in yourlocal.conf
file:USERADD_UID_TABLES = "files/passwd"
Note
Setting the USERADDEXTENSION variable to “useradd-staticids” causes the build system to use static
uid
values.- USERADDEXTENSION
When set to “useradd-staticids”, causes the OpenEmbedded build system to base all user and group additions on a static
passwd
andgroup
files found in BBPATH.To use static user identification (
uid
) and group identification (gid
) values, set the variable as follows in yourlocal.conf
file: USERADDEXTENSION = “useradd-staticids”Note
Setting this variable to use static
uid
andgid
values causes the OpenEmbedded build system to employ the useradd*.bbclass class.If you use static
uid
andgid
information, you must also specify thefiles/passwd
andfiles/group
files by setting the USERADD_UID_TABLES and USERADD_GID_TABLES variables. Additionally, you should also set the USERADD_ERROR_DYNAMIC variable.- VOLATILE_LOG_DIR
Specifies the persistence of the target’s
/var/log
directory, which is used to house postinstall target log files.By default,
VOLATILE_LOG_DIR
is set to “yes”, which means the file is not persistent. You can override this setting by setting the variable to “no” to make the log directory persistent.- WARN_QA
Specifies the quality assurance checks whose failures are reported as warnings by the OpenEmbedded build system. You set this variable in your distribution configuration file. For a list of the checks you can control with this variable, see the “insane.bbclass” section.
- WKS_FILE
Specifies the location of the Wic kickstart file that is used by the OpenEmbedded build system to create a partitioned image (image
.wic
). For information on how to create a partitioned image, see the “Creating Partitioned Images Using Wic” section in the Yocto Project Development Tasks Manual. For details on the kickstart file format, see the “OpenEmbedded Kickstart (.wks) Reference” Chapter.- WKS_FILE_DEPENDS
When placed in the recipe that builds your image, this variable lists build-time dependencies. The
WKS_FILE_DEPENDS
variable is only applicable when Wic images are active (i.e. when IMAGE_FSTYPES contains entries related to Wic). If your recipe does not create Wic images, the variable has no effect.The
WKS_FILE_DEPENDS
variable is similar to the DEPENDS variable. When you use the variable in your recipe that builds the Wic image, dependencies you list in theWKS_FILE_DEPENDS
variable are added to theDEPENDS
variable.With the
WKS_FILE_DEPENDS
variable, you have the possibility to specify a list of additional dependencies (e.g. native tools, bootloaders, and so forth), that are required to build Wic images. Following is an example:WKS_FILE_DEPENDS = "some-native-tool"
In the previous example, some-native-tool would be replaced with an actual native tool on which the build would depend.
- WORKDIR
The pathname of the work directory in which the OpenEmbedded build system builds a recipe. This directory is located within the TMPDIR directory structure and is specific to the recipe being built and the system for which it is being built.
The
WORKDIR
directory is defined as follows:${TMPDIR}/work/${MULTIMACH_TARGET_SYS}/${PN}/${EXTENDPE}${PV}-${PR}
The actual directory depends on several things:
TMPDIR: The top-level build output directory
MULTIMACH_TARGET_SYS: The target system identifier
PN: The recipe name
EXTENDPE: The epoch - (if PE is not specified, which is usually the case for most recipes, then EXTENDPE is blank)
PV: The recipe version
PR: The recipe revision
As an example, assume a Source Directory top-level folder name
poky
, a default Build Directory atpoky/build
, and aqemux86-poky-linux
machine target system. Furthermore, suppose your recipe is namedfoo_1.3.0-r0.bb
. In this case, the work directory the build system uses to build the package would be as follows:poky/build/tmp/work/qemux86-poky-linux/foo/1.3.0-r0
- XSERVER
Specifies the packages that should be installed to provide an X server and drivers for the current machine, assuming your image directly includes
packagegroup-core-x11-xserver
or, perhaps indirectly, includes “x11-base” in IMAGE_FEATURES.The default value of
XSERVER
, if not specified in the machine configuration, is “xserver-xorg xf86-video-fbdev xf86-input-evdev”.
14 Variable Context
While you can use most variables in almost any context such as
.conf
, .bbclass
, .inc
, and .bb
files, some variables are
often associated with a particular locality or context. This chapter
describes some common associations.
14.1 Configuration
The following subsections provide lists of variables whose context is configuration: distribution, machine, and local.
14.1.1 Distribution (Distro)
This section lists variables whose configuration context is the distribution, or distro.
14.1.2 Machine
This section lists variables whose configuration context is the machine.
14.1.3 Local
This section lists variables whose configuration context is the local
configuration through the local.conf
file.
14.2 Recipes
The following subsections provide lists of variables whose context is recipes: required, dependencies, path, and extra build information.
14.2.1 Required
This section lists variables that are required for recipes.
SRC_URI - used in recipes that fetch local or remote files.
14.2.2 Dependencies
This section lists variables that define recipe dependencies.
14.2.3 Paths
This section lists variables that define recipe paths.
14.2.4 Extra Build Information
This section lists variables that define extra build information for recipes.
15 FAQ
Q: How does Poky differ from OpenEmbedded?
A: The term Poky
refers to the specific reference build
system that the Yocto Project provides. Poky is based on
OpenEmbedded-Core (OE-Core) and BitBake. Thus, the
generic term used here for the build system is the “OpenEmbedded build
system.” Development in the Yocto Project using Poky is closely tied to
OpenEmbedded, with changes always being merged to OE-Core or BitBake
first before being pulled back into Poky. This practice benefits both
projects immediately.
Q: My development system does not meet the required Git, tar, and Python versions. In particular, I do not have Python 3.6.0 or greater. Can I still use the Yocto Project?
A: You can get the required tools on your host development system a couple different ways (i.e. building a tarball or downloading a tarball). See the “Required Git, tar, Python and gcc Versions” section for steps on how to update your build tools.
Q: How can you claim Poky / OpenEmbedded-Core is stable?
A: There are three areas that help with stability;
The Yocto Project team keeps OpenEmbedded-Core (OE-Core) small and focused, containing around 830 recipes as opposed to the thousands available in other OpenEmbedded community layers. Keeping it small makes it easy to test and maintain.
The Yocto Project team runs manual and automated tests using a small, fixed set of reference hardware as well as emulated targets.
The Yocto Project uses an autobuilder, which provides continuous build and integration tests.
Q: How do I get support for my board added to the Yocto Project?
A: Support for an additional board is added by creating a Board Support Package (BSP) layer for it. For more information on how to create a BSP layer, see the “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual and the Yocto Project Board Support Package Developer’s Guide.
Usually, if the board is not completely exotic, adding support in the Yocto Project is fairly straightforward.
Q: Are there any products built using the OpenEmbedded build system?
A: The software running on the Vernier LabQuest is built using the OpenEmbedded build system. See the Vernier LabQuest website for more information. There are a number of pre-production devices using the OpenEmbedded build system and the Yocto Project team announces them as soon as they are released.
Q: What does the OpenEmbedded build system produce as output?
A: Because you can use the same set of recipes to create output of various formats, the output of an OpenEmbedded build depends on how you start it. Usually, the output is a flashable image ready for the target device.
Q: How do I add my package to the Yocto Project?
A: To add a package, you need to create a BitBake recipe. For information on how to create a BitBake recipe, see the “Writing a New Recipe” section in the Yocto Project Development Tasks Manual.
Q: Do I have to reflash my entire board with a new Yocto Project image when recompiling a package?
A: The OpenEmbedded build system can build packages in various
formats such as IPK for OPKG, Debian package (.deb
), or RPM. You can
then upgrade the packages using the package tools on the device, much
like on a desktop distribution such as Ubuntu or Fedora. However,
package management on the target is entirely optional.
Q: I see the error
‘chmod: XXXXX new permissions are r-xrwxrwx, not r-xr-xr-x
’. What is
wrong?
A: You are probably running the build on an NTFS filesystem. Use
ext2
, ext3
, or ext4
instead.
Q: I see lots of 404 responses for files when the OpenEmbedded build system is trying to download sources. Is something wrong?
A: Nothing is wrong. The OpenEmbedded build system checks any configured source mirrors before downloading from the upstream sources. The build system does this searching for both source archives and pre-checked out versions of SCM-managed software. These checks help in large installations because it can reduce load on the SCM servers themselves. The address above is one of the default mirrors configured into the build system. Consequently, if an upstream source disappears, the team can place sources there so builds continue to work.
Q: I have machine-specific data in a package for one machine only but the package is being marked as machine-specific in all cases, how do I prevent this?
A: Set SRC_URI_OVERRIDES_PACKAGE_ARCH
= “0” in the .bb
file
but make sure the package is manually marked as machine-specific for the
case that needs it. The code that handles
SRC_URI_OVERRIDES_PACKAGE_ARCH
is in the
meta/classes/base.bbclass
file.
Q: I’m behind a firewall and need to use a proxy server. How do I do that?
A: Most source fetching by the OpenEmbedded build system is done by
wget
and you therefore need to specify the proxy settings in a
.wgetrc
file, which can be in your home directory if you are a
single user or can be in /usr/local/etc/wgetrc
as a global user
file.
Following is the applicable code for setting various proxy types in the
.wgetrc
file. By default, these settings are disabled with comments.
To use them, remove the comments:
# You can set the default proxies for Wget to use for http, https, and ftp.
# They will override the value in the environment.
#https_proxy = http://proxy.yoyodyne.com:18023/
#http_proxy = http://proxy.yoyodyne.com:18023/
#ftp_proxy = http://proxy.yoyodyne.com:18023/
# If you do not want to use proxy at all, set this to off.
#use_proxy = on
The Yocto Project also includes a
meta-poky/conf/site.conf.sample
file that shows how to configure CVS
and Git proxy servers if needed. For more information on setting up
various proxy types and configuring proxy servers, see the
“Working Behind a Network Proxy”
Wiki page.
Q: What’s the difference between target and target-native
?
A: The *-native
targets are designed to run on the system being
used for the build. These are usually tools that are needed to assist
the build in some way such as quilt-native
, which is used to apply
patches. The non-native version is the one that runs on the target
device.
Q: I’m seeing random build failures. Help?!
A: If the same build is failing in totally different and random ways, the most likely explanation is:
The hardware you are running the build on has some problem.
You are running the build under virtualization, in which case the virtualization probably has bugs.
The OpenEmbedded build system processes a massive amount of data that causes lots of network, disk and CPU activity and is sensitive to even single-bit failures in any of these areas. True random failures have always been traced back to hardware or virtualization issues.
Q: When I try to build a native recipe, the build fails with
iconv.h
problems.
A: If you get an error message that indicates GNU libiconv
is
not in use but iconv.h
has been included from libiconv
, you need
to check to see if you have a previously installed version of the header
file in /usr/local/include
.
#error GNU libiconv not in use but included iconv.h is from libiconv
If you find a previously installed file, you should either uninstall it or temporarily rename it and try the build again.
This issue is just a single manifestation of “system leakage” issues
caused when the OpenEmbedded build system finds and uses previously
installed files during a native build. This type of issue might not be
limited to iconv.h
. Be sure that leakage cannot occur from
/usr/local/include
and /opt
locations.
Q: What do we need to ship for license compliance?
A: This is a difficult question and you need to consult your lawyer for the answer for your specific case. It is worth bearing in mind that for GPL compliance, there needs to be enough information shipped to allow someone else to rebuild and produce the same end result you are shipping. This means sharing the source code, any patches applied to it, and also any configuration information about how that package was configured and built.
You can find more information on licensing in the “Licensing” section in the Yocto Project Overview and Concepts Manual and also in the “Maintaining Open Source License Compliance During Your Product’s Lifecycle” section in the Yocto Project Development Tasks Manual.
Q: How do I disable the cursor on my touchscreen device?
A: You need to create a form factor file as described in the
“Miscellaneous BSP-Specific Recipe Files” section in
the Yocto Project Board Support Packages (BSP) Developer’s Guide. Set
the HAVE_TOUCHSCREEN
variable equal to one as follows:
HAVE_TOUCHSCREEN=1
Q: How do I make sure connected network interfaces are brought up by default?
A: The default interfaces file provided by the netbase recipe does not automatically bring up network interfaces. Therefore, you will need to add a BSP-specific netbase that includes an interfaces file. See the “Miscellaneous BSP-Specific Recipe Files” section in the Yocto Project Board Support Packages (BSP) Developer’s Guide for information on creating these types of miscellaneous recipe files.
For example, add the following files to your layer:
meta-MACHINE/recipes-bsp/netbase/netbase/MACHINE/interfaces
meta-MACHINE/recipes-bsp/netbase/netbase_5.0.bbappend
Q: How do I create images with more free space?
A: By default, the OpenEmbedded build system creates images that are 1.3 times the size of the populated root filesystem. To affect the image size, you need to set various configurations:
Image Size: The OpenEmbedded build system uses the IMAGE_ROOTFS_SIZE variable to define the size of the image in Kbytes. The build system determines the size by taking into account the initial root filesystem size before any modifications such as requested size for the image and any requested additional free disk space to be added to the image.
Overhead: Use the IMAGE_OVERHEAD_FACTOR variable to define the multiplier that the build system applies to the initial image size, which is 1.3 by default.
Additional Free Space: Use the IMAGE_ROOTFS_EXTRA_SPACE variable to add additional free space to the image. The build system adds this space to the image after it determines its
IMAGE_ROOTFS_SIZE
.
Q: Why don’t you support directories with spaces in the pathnames?
A: The Yocto Project team has tried to do this before but too many
of the tools the OpenEmbedded build system depends on, such as
autoconf
, break when they find spaces in pathnames. Until that
situation changes, the team will not support spaces in pathnames.
Q: How do I use an external toolchain?
A: The toolchain configuration is very flexible and customizable. It
is primarily controlled with the TCMODE
variable. This variable
controls which tcmode-*.inc
file to include from the
meta/conf/distro/include
directory within the Source Directory.
The default value of TCMODE
is “default”, which tells the
OpenEmbedded build system to use its internally built toolchain (i.e.
tcmode-default.inc
). However, other patterns are accepted. In
particular, “external-*” refers to external toolchains. One example is
the Sourcery G++ Toolchain. The support for this toolchain resides in
the separate meta-sourcery
layer at
https://github.com/MentorEmbedded/meta-sourcery/.
In addition to the toolchain configuration, you also need a
corresponding toolchain recipe file. This recipe file needs to package
up any pre-built objects in the toolchain such as libgcc
,
libstdcc++
, any locales, and libc
.
Q: How does the OpenEmbedded build system obtain source code and will it work behind my firewall or proxy server?
A: The way the build system obtains source code is highly configurable. You can setup the build system to get source code in most environments if HTTP transport is available.
When the build system searches for source code, it first tries the local download directory. If that location fails, Poky tries PREMIRRORS, the upstream source, and then MIRRORS in that order.
Assuming your distribution is “poky”, the OpenEmbedded build system uses
the Yocto Project source PREMIRRORS
by default for SCM-based
sources, upstreams for normal tarballs, and then falls back to a number
of other mirrors including the Yocto Project source mirror if those
fail.
As an example, you could add a specific server for the build system to
attempt before any others by adding something like the following to the
local.conf
configuration file:
PREMIRRORS_prepend = "\
git://.*/.* http://www.yoctoproject.org/sources/ \n \
ftp://.*/.* http://www.yoctoproject.org/sources/ \n \
http://.*/.* http://www.yoctoproject.org/sources/ \n \
https://.*/.* http://www.yoctoproject.org/sources/ \n"
These changes cause the build system to intercept Git, FTP, HTTP, and
HTTPS requests and direct them to the http://
sources mirror. You
can use file://
URLs to point to local directories or network shares
as well.
Aside from the previous technique, these options also exist:
BB_NO_NETWORK = "1"
This statement tells BitBake to issue an error instead of trying to access the Internet. This technique is useful if you want to ensure code builds only from local sources.
Here is another technique:
BB_FETCH_PREMIRRORONLY = "1"
This statement
limits the build system to pulling source from the PREMIRRORS
only.
Again, this technique is useful for reproducing builds.
Here is another technique:
BB_GENERATE_MIRROR_TARBALLS = "1"
This statement tells the build system to generate mirror tarballs. This technique is useful if you want to create a mirror server. If not, however, the technique can simply waste time during the build.
Finally, consider an example where you are behind an HTTP-only firewall.
You could make the following changes to the local.conf
configuration
file as long as the PREMIRRORS
server is current:
PREMIRRORS_prepend = "\
ftp://.*/.* http://www.yoctoproject.org/sources/ \n \
http://.*/.* http://www.yoctoproject.org/sources/ \n \
https://.*/.* http://www.yoctoproject.org/sources/ \n"
BB_FETCH_PREMIRRORONLY = "1"
These changes would cause the build system to successfully fetch source
over HTTP and any network accesses to anything other than the
PREMIRRORS
would fail.
The build system also honors the standard shell environment variables
http_proxy
, ftp_proxy
, https_proxy
, and all_proxy
to
redirect requests through proxy servers.
Note
You can find more information on the “Working Behind a Network Proxy” Wiki page.
Q: Can I get rid of build output so I can start over?
A: Yes - you can easily do this. When you use BitBake to build an
image, all the build output goes into the directory created when you run
the build environment setup script (i.e.
oe-init-build-env). By default, this Build Directory
is named build
but can be named
anything you want.
Within the Build Directory, is the tmp
directory. To remove all the
build output yet preserve any source code or downloaded files from
previous builds, simply remove the tmp
directory.
Q: Why do ${bindir}
and ${libdir}
have strange values for
-native
recipes?
A: Executables and libraries might need to be used from a directory other than the directory into which they were initially installed. Complicating this situation is the fact that sometimes these executables and libraries are compiled with the expectation of being run from that initial installation target directory. If this is the case, moving them causes problems.
This scenario is a fundamental problem for package maintainers of
mainstream Linux distributions as well as for the OpenEmbedded build
system. As such, a well-established solution exists. Makefiles,
Autotools configuration scripts, and other build systems are expected to
respect environment variables such as bindir
, libdir
, and
sysconfdir
that indicate where executables, libraries, and data
reside when a program is actually run. They are also expected to respect
a DESTDIR
environment variable, which is prepended to all the other
variables when the build system actually installs the files. It is
understood that the program does not actually run from within
DESTDIR
.
When the OpenEmbedded build system uses a recipe to build a
target-architecture program (i.e. one that is intended for inclusion on
the image being built), that program eventually runs from the root file
system of that image. Thus, the build system provides a value of
“/usr/bin” for bindir
, a value of “/usr/lib” for libdir
, and so
forth.
Meanwhile, DESTDIR
is a path within the Build Directory.
However, when the recipe builds a
native program (i.e. one that is intended to run on the build machine),
that program is never installed directly to the build machine’s root
file system. Consequently, the build system uses paths within the Build
Directory for DESTDIR
, bindir
and related variables. To better
understand this, consider the following two paths where the first is
relatively normal and the second is not:
Note
Due to these lengthy examples, the paths are artificially broken across lines for readability.
/home/maxtothemax/poky-bootchart2/build/tmp/work/i586-poky-linux/zlib/
1.2.8-r0/sysroot-destdir/usr/bin
/home/maxtothemax/poky-bootchart2/build/tmp/work/x86_64-linux/
zlib-native/1.2.8-r0/sysroot-destdir/home/maxtothemax/poky-bootchart2/
build/tmp/sysroots/x86_64-linux/usr/bin
Even if the paths look unusual,
they both are correct - the first for a target and the second for a
native recipe. These paths are a consequence of the DESTDIR
mechanism and while they appear strange, they are correct and in
practice very effective.
Q: The files provided by my *-native
recipe do not appear to be
available to other recipes. Files are missing from the native sysroot,
my recipe is installing to the wrong place, or I am getting permissions
errors during the do_install task in my recipe! What is wrong?
A: This situation results when a build system does not recognize the
environment variables supplied to it by BitBake. The
incident that prompted this FAQ entry involved a Makefile that used an
environment variable named BINDIR
instead of the more standard
variable bindir
. The makefile’s hardcoded default value of
“/usr/bin” worked most of the time, but not for the recipe’s -native
variant. For another example, permissions errors might be caused by a
Makefile that ignores DESTDIR
or uses a different name for that
environment variable. Check the build system to see if these kinds
of issues exist.
Q: I’m adding a binary in a recipe but it’s different in the image, what is changing it?
A: The first most obvious change is the system stripping debug symbols from it. Setting INHIBIT_PACKAGE_STRIP to stop debug symbols being stripped and/or INHIBIT_PACKAGE_DEBUG_SPLIT to stop debug symbols being split into a separate file will ensure the binary is unchanged. The other less obvious thing that can happen is prelinking of the image. This is set by default in local.conf via USER_CLASSES which can contain ‘image-prelink’. If you remove that, the image will not be prelinked meaning the binaries would be unchanged.
16 Contributions and Additional Information
16.1 Introduction
The Yocto Project team is happy for people to experiment with the Yocto Project. A number of places exist to find help if you run into difficulties or find bugs. This presents information about contributing and participating in the Yocto Project.
16.2 Contributions
The Yocto Project gladly accepts contributions. You can submit changes to the project either by creating and sending pull requests, or by submitting patches through email. For information on how to do both as well as information on how to identify the maintainer for each area of code, see the “Submitting a Change to the Yocto Project” section in the Yocto Project Development Tasks Manual.
16.3 Yocto Project Bugzilla
The Yocto Project uses its own implementation of Bugzilla to track defects (bugs). Implementations of Bugzilla work well for group development because they track bugs and code changes, can be used to communicate changes and problems with developers, can be used to submit and review patches, and can be used to manage quality assurance.
Sometimes it is helpful to submit, investigate, or track a bug against the Yocto Project itself (e.g. when discovering an issue with some component of the build system that acts contrary to the documentation or your expectations).
A general procedure and guidelines exist for when you use Bugzilla to submit a bug. For information on how to use Bugzilla to submit a bug against the Yocto Project, see the following:
The “Submitting a Defect Against the Yocto Project” section in the Yocto Project Development Tasks Manual.
The Yocto Project Bugzilla wiki page
For information on Bugzilla in general, see https://www.bugzilla.org/about/.
16.4 Mailing lists
A number of mailing lists maintained by the Yocto Project exist as well as related OpenEmbedded mailing lists for discussion, patch submission and announcements. To subscribe to one of the following mailing lists, click on the appropriate URL in the following list and follow the instructions:
https://lists.yoctoproject.org/g/yocto - General Yocto Project discussion mailing list.
https://lists.openembedded.org/g/openembedded-core - Discussion mailing list about OpenEmbedded-Core (the core metadata).
https://lists.openembedded.org/g/openembedded-devel - Discussion mailing list about OpenEmbedded.
https://lists.openembedded.org/g/bitbake-devel - Discussion mailing list about the BitBake build tool.
https://lists.yoctoproject.org/g/poky - Discussion mailing list about Poky.
https://lists.yoctoproject.org/g/yocto-announce - Mailing list to receive official Yocto Project release and milestone announcements.
For more Yocto Project-related mailing lists, see the Yocto Project Website.
16.5 Internet Relay Chat (IRC)
Two IRC channels on Freenode are available for the Yocto Project and Poky discussions:
#yocto
#poky
17 Manual Revision History
Revision |
Date |
Note |
---|---|---|
0.9 |
November 2010 |
The initial document released with the Yocto Project 0.9 Release |
1.0 |
April 2011 |
Released with the Yocto Project 1.0 Release. |
1.1 |
October 2011 |
Released with the Yocto Project 1.1 Release. |
1.2 |
April 2012 |
Released with the Yocto Project 1.2 Release. |
1.3 |
October 2012 |
Released with the Yocto Project 1.3 Release. |
1.4 |
April 2013 |
Released with the Yocto Project 1.4 Release. |
1.5 |
October 2013 |
Released with the Yocto Project 1.5 Release. |
1.6 |
April 2014 |
Released with the Yocto Project 1.6 Release. |
1.7 |
October 2014 |
Released with the Yocto Project 1.7 Release. |
1.8 |
April 2015 |
Released with the Yocto Project 1.8 Release. |
2.0 |
October 2015 |
Released with the Yocto Project 2.0 Release. |
2.1 |
April 2016 |
Released with the Yocto Project 2.1 Release. |
2.2 |
October 2016 |
Released with the Yocto Project 2.2 Release. |
2.3 |
May 2017 |
Released with the Yocto Project 2.3 Release. |
2.4 |
October 2017 |
Released with the Yocto Project 2.4 Release. |
2.5 |
May 2018 |
Released with the Yocto Project 2.5 Release. |
2.6 |
November 2018 |
Released with the Yocto Project 2.6 Release. |
2.7 |
May 2019 |
Released with the Yocto Project 2.7 Release. |
3.0 |
October 2019 |
Released with the Yocto Project 3.0 Release. |
3.1 |
April 2020 |
Released with the Yocto Project 3.1 Release. |
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Yocto Project Board Support Package Developer’s Guide
1 Board Support Packages (BSP) - Developer’s Guide
A Board Support Package (BSP) is a collection of information that defines how to support a particular hardware device, set of devices, or hardware platform. The BSP includes information about the hardware features present on the device and kernel configuration information along with any additional hardware drivers required. The BSP also lists any additional software components required in addition to a generic Linux software stack for both essential and optional platform features.
This guide presents information about BSP layers, defines a structure for components so that BSPs follow a commonly understood layout, discusses how to customize a recipe for a BSP, addresses BSP licensing, and provides information that shows you how to create a BSP Layer using the bitbake-layers tool.
1.1 BSP Layers
A BSP consists of a file structure inside a base directory. Collectively, you can think of the base directory, its file structure, and the contents as a BSP layer. Although not a strict requirement, BSP layers in the Yocto Project use the following well-established naming convention:
meta-bsp_root_name
The string “meta-” is prepended to the machine or platform name, which is “bsp_root_name” in the above form.
Note
Because the BSP layer naming convention is well-established, it is
advisable to follow it when creating layers. Technically speaking, a
BSP layer name does not need to start with meta-
.
However, various scripts and tools in the Yocto Project development
environment assume this convention.
To help understand the BSP layer concept, consider the BSPs that the Yocto Project supports and provides with each release. You can see the layers in the Yocto Project Source Repositories through a web interface at https://git.yoctoproject.org/cgit/cgit.cgi/. If you go to that interface, you will find a list of repositories under “Yocto Metadata Layers”.
Note
Layers that are no longer actively supported as part of the Yocto Project appear under the heading “Yocto Metadata Layer Archive.”
Each repository is a BSP layer supported by the Yocto Project (e.g.
meta-raspberrypi
and meta-intel
). Each of these layers is a
repository unto itself and clicking on the layer name displays two URLs
from which you can clone the layer’s repository to your local system.
Here is an example that clones the Raspberry Pi BSP layer:
$ git clone git://git.yoctoproject.org/meta-raspberrypi
In addition to BSP layers, the meta-yocto-bsp
layer is part of the
shipped poky
repository. The meta-yocto-bsp
layer maintains
several “reference” BSPs including the ARM-based Beaglebone, MIPS-based
EdgeRouter, and generic versions of both 32-bit and 64-bit IA machines.
For information on typical BSP development workflow, see the Developing a Board Support Package (BSP) section. For more information on how to set up a local copy of source files from a Git repository, see the Locating Yocto Project Source Files section in the Yocto Project Development Tasks Manual.
The BSP layer’s base directory (meta-bsp_root_name
) is the root
directory of that Layer. This directory is what you add to the
BBLAYERS variable in the
conf/bblayers.conf
file found in your
Build Directory, which is
established after you run the OpenEmbedded build environment setup
script (i.e. oe-init-build-env).
Adding the root directory allows the OpenEmbedded Build System
to recognize the BSP
layer and from it build an image. Here is an example:
BBLAYERS ?= " \
/usr/local/src/yocto/meta \
/usr/local/src/yocto/meta-poky \
/usr/local/src/yocto/meta-yocto-bsp \
/usr/local/src/yocto/meta-mylayer \
"
Note
Ordering and BBFILE_PRIORITY for the layers listed in BBLAYERS
matter. For example, if multiple layers define a machine configuration, the
OpenEmbedded build system uses the last layer searched given similar layer
priorities. The build system works from the top-down through the layers
listed in BBLAYERS
.
Some BSPs require or depend on additional layers beyond the BSP’s root
layer in order to be functional. In this case, you need to specify these
layers in the README
“Dependencies” section of the BSP’s root layer.
Additionally, if any build instructions exist for the BSP, you must add
them to the “Dependencies” section.
Some layers function as a layer to hold other BSP layers. These layers
are known as “container layers”. An example of
this type of layer is OpenEmbedded’s
meta-openembedded
layer. The meta-openembedded
layer contains many meta-*
layers.
In cases like this, you need to include the names of the actual layers
you want to work with, such as:
BBLAYERS ?= " \
/usr/local/src/yocto/meta \
/usr/local/src/yocto/meta-poky \
/usr/local/src/yocto/meta-yocto-bsp \
/usr/local/src/yocto/meta-mylayer \
.../meta-openembedded/meta-oe \
.../meta-openembedded/meta-perl \
.../meta-openembedded/meta-networking \
"
and so on.
For more information on layers, see the “Understanding and Creating Layers” section of the Yocto Project Development Tasks Manual.
1.2 Preparing Your Build Host to Work With BSP Layers
This section describes how to get your build host ready to work with BSP layers. Once you have the host set up, you can create the layer as described in the “Creating a new BSP Layer Using the bitbake-layers Script” section.
Note
For structural information on BSPs, see the Example Filesystem Layout section.
Set Up the Build Environment: Be sure you are set up to use BitBake in a shell. See the “Preparing the Build Host” section in the Yocto Project Development Tasks Manual for information on how to get a build host ready that is either a native Linux machine or a machine that uses CROPS.
Clone the poky Repository: You need to have a local copy of the Yocto Project Source Directory (i.e. a local
poky
repository). See the “Cloning the poky Repository” and possibly the “Checking Out by Branch in Poky” or “Checking Out by Tag in Poky” sections all in the Yocto Project Development Tasks Manual for information on how to clone thepoky
repository and check out the appropriate branch for your work.Determine the BSP Layer You Want: The Yocto Project supports many BSPs, which are maintained in their own layers or in layers designed to contain several BSPs. To get an idea of machine support through BSP layers, you can look at the index of machines for the release.
Optionally Clone the meta-intel BSP Layer: If your hardware is based on current Intel CPUs and devices, you can leverage this BSP layer. For details on the
meta-intel
BSP layer, see the layer’s README file.Navigate to Your Source Directory: Typically, you set up the
meta-intel
Git repository inside the Source Directory (e.g.poky
).$ cd /home/you/poky
Clone the Layer:
$ git clone git://git.yoctoproject.org/meta-intel.git Cloning into 'meta-intel'... remote: Counting objects: 15585, done. remote: Compressing objects: 100% (5056/5056), done. remote: Total 15585 (delta 9123), reused 15329 (delta 8867) Receiving objects: 100% (15585/15585), 4.51 MiB | 3.19 MiB/s, done. Resolving deltas: 100% (9123/9123), done. Checking connectivity... done.
Check Out the Proper Branch: The branch you check out for
meta-intel
must match the same branch you are using for the Yocto Project release (e.g.hardknott
):$ cd meta-intel $ git checkout -b hardknott remotes/origin/hardknott Branch hardknott set up to track remote branch hardknott from origin. Switched to a new branch 'hardknott'
Note
To see the available branch names in a cloned repository, use the
git branch -al
command. See the “Checking Out by Branch in Poky” section in the Yocto Project Development Tasks Manual for more information.
Optionally Set Up an Alternative BSP Layer: If your hardware can be more closely leveraged to an existing BSP not within the
meta-intel
BSP layer, you can clone that BSP layer.The process is identical to the process used for the
meta-intel
layer except for the layer’s name. For example, if you determine that your hardware most closely matches themeta-raspberrypi
, clone that layer:$ git clone git://git.yoctoproject.org/meta-raspberrypi Cloning into 'meta-raspberrypi'... remote: Counting objects: 4743, done. remote: Compressing objects: 100% (2185/2185), done. remote: Total 4743 (delta 2447), reused 4496 (delta 2258) Receiving objects: 100% (4743/4743), 1.18 MiB | 0 bytes/s, done. Resolving deltas: 100% (2447/2447), done. Checking connectivity... done.
Initialize the Build Environment: While in the root directory of the Source Directory (i.e.
poky
), run the oe-init-build-env environment setup script to define the OpenEmbedded build environment on your build host.$ source oe-init-build-env
Among other things, the script creates the Build Directory, which is
build
in this case and is located in the Source Directory. After the script runs, your current working directory is set to thebuild
directory.
1.3 Example Filesystem Layout
Defining a common BSP directory structure allows end-users to understand and become familiar with that standard. A common format also encourages standardization of software support for hardware.
The proposed form described in this section does have elements that are specific to the OpenEmbedded build system. It is intended that developers can use this structure with other build systems besides the OpenEmbedded build system. It is also intended that it will be simple to extract information and convert it to other formats if required. The OpenEmbedded build system, through its standard layers mechanism, can directly accept the format described as a layer. The BSP layer captures all the hardware-specific details in one place using a standard format, which is useful for any person wishing to use the hardware platform regardless of the build system they are using.
The BSP specification does not include a build system or other tools - the specification is concerned with the hardware-specific components only. At the end-distribution point, you can ship the BSP layer combined with a build system and other tools. Realize that it is important to maintain the distinction that the BSP layer, a build system, and tools are separate components that could be combined in certain end products.
Before looking at the recommended form for the directory structure inside a BSP layer, you should be aware that some requirements do exist in order for a BSP layer to be considered compliant with the Yocto Project. For that list of requirements, see the “Released BSP Requirements” section.
Below is the typical directory structure for a BSP layer. While this basic form represents the standard, realize that the actual layout for individual BSPs could differ.
meta-bsp_root_name/
meta-bsp_root_name/bsp_license_file
meta-bsp_root_name/README
meta-bsp_root_name/README.sources
meta-bsp_root_name/binary/bootable_images
meta-bsp_root_name/conf/layer.conf
meta-bsp_root_name/conf/machine/*.conf
meta-bsp_root_name/recipes-bsp/*
meta-bsp_root_name/recipes-core/*
meta-bsp_root_name/recipes-graphics/*
meta-bsp_root_name/recipes-kernel/linux/linux-yocto_kernel_rev.bbappend
Below is an example of the Raspberry Pi BSP layer that is available from the Source Repositories:
meta-raspberrypi/COPYING.MIT
meta-raspberrypi/README.md
meta-raspberrypi/classes
meta-raspberrypi/classes/sdcard_image-rpi.bbclass
meta-raspberrypi/conf/
meta-raspberrypi/conf/layer.conf
meta-raspberrypi/conf/machine/
meta-raspberrypi/conf/machine/raspberrypi-cm.conf
meta-raspberrypi/conf/machine/raspberrypi-cm3.conf
meta-raspberrypi/conf/machine/raspberrypi.conf
meta-raspberrypi/conf/machine/raspberrypi0-wifi.conf
meta-raspberrypi/conf/machine/raspberrypi0.conf
meta-raspberrypi/conf/machine/raspberrypi2.conf
meta-raspberrypi/conf/machine/raspberrypi3-64.conf
meta-raspberrypi/conf/machine/raspberrypi3.conf
meta-raspberrypi/conf/machine/include
meta-raspberrypi/conf/machine/include/rpi-base.inc
meta-raspberrypi/conf/machine/include/rpi-default-providers.inc
meta-raspberrypi/conf/machine/include/rpi-default-settings.inc
meta-raspberrypi/conf/machine/include/rpi-default-versions.inc
meta-raspberrypi/conf/machine/include/tune-arm1176jzf-s.inc
meta-raspberrypi/docs
meta-raspberrypi/docs/Makefile
meta-raspberrypi/docs/conf.py
meta-raspberrypi/docs/contributing.md
meta-raspberrypi/docs/extra-apps.md
meta-raspberrypi/docs/extra-build-config.md
meta-raspberrypi/docs/index.rst
meta-raspberrypi/docs/layer-contents.md
meta-raspberrypi/docs/readme.md
meta-raspberrypi/files
meta-raspberrypi/files/custom-licenses
meta-raspberrypi/files/custom-licenses/Broadcom
meta-raspberrypi/recipes-bsp
meta-raspberrypi/recipes-bsp/bootfiles
meta-raspberrypi/recipes-bsp/bootfiles/bcm2835-bootfiles.bb
meta-raspberrypi/recipes-bsp/bootfiles/rpi-config_git.bb
meta-raspberrypi/recipes-bsp/common
meta-raspberrypi/recipes-bsp/common/firmware.inc
meta-raspberrypi/recipes-bsp/formfactor
meta-raspberrypi/recipes-bsp/formfactor/formfactor
meta-raspberrypi/recipes-bsp/formfactor/formfactor/raspberrypi
meta-raspberrypi/recipes-bsp/formfactor/formfactor/raspberrypi/machconfig
meta-raspberrypi/recipes-bsp/formfactor/formfactor_0.0.bbappend
meta-raspberrypi/recipes-bsp/rpi-u-boot-src
meta-raspberrypi/recipes-bsp/rpi-u-boot-src/files
meta-raspberrypi/recipes-bsp/rpi-u-boot-src/files/boot.cmd.in
meta-raspberrypi/recipes-bsp/rpi-u-boot-src/rpi-u-boot-scr.bb
meta-raspberrypi/recipes-bsp/u-boot
meta-raspberrypi/recipes-bsp/u-boot/u-boot
meta-raspberrypi/recipes-bsp/u-boot/u-boot/*.patch
meta-raspberrypi/recipes-bsp/u-boot/u-boot_%.bbappend
meta-raspberrypi/recipes-connectivity
meta-raspberrypi/recipes-connectivity/bluez5
meta-raspberrypi/recipes-connectivity/bluez5/bluez5
meta-raspberrypi/recipes-connectivity/bluez5/bluez5/*.patch
meta-raspberrypi/recipes-connectivity/bluez5/bluez5/BCM43430A1.hcd
meta-raspberrypi/recipes-connectivity/bluez5/bluez5brcm43438.service
meta-raspberrypi/recipes-connectivity/bluez5/bluez5_%.bbappend
meta-raspberrypi/recipes-core
meta-raspberrypi/recipes-core/images
meta-raspberrypi/recipes-core/images/rpi-basic-image.bb
meta-raspberrypi/recipes-core/images/rpi-hwup-image.bb
meta-raspberrypi/recipes-core/images/rpi-test-image.bb
meta-raspberrypi/recipes-core/packagegroups
meta-raspberrypi/recipes-core/packagegroups/packagegroup-rpi-test.bb
meta-raspberrypi/recipes-core/psplash
meta-raspberrypi/recipes-core/psplash/files
meta-raspberrypi/recipes-core/psplash/files/psplash-raspberrypi-img.h
meta-raspberrypi/recipes-core/psplash/psplash_git.bbappend
meta-raspberrypi/recipes-core/udev
meta-raspberrypi/recipes-core/udev/udev-rules-rpi
meta-raspberrypi/recipes-core/udev/udev-rules-rpi/99-com.rules
meta-raspberrypi/recipes-core/udev/udev-rules-rpi.bb
meta-raspberrypi/recipes-devtools
meta-raspberrypi/recipes-devtools/bcm2835
meta-raspberrypi/recipes-devtools/bcm2835/bcm2835_1.52.bb
meta-raspberrypi/recipes-devtools/pi-blaster
meta-raspberrypi/recipes-devtools/pi-blaster/files
meta-raspberrypi/recipes-devtools/pi-blaster/files/*.patch
meta-raspberrypi/recipes-devtools/pi-blaster/pi-blaster_git.bb
meta-raspberrypi/recipes-devtools/python
meta-raspberrypi/recipes-devtools/python/python-rtimu
meta-raspberrypi/recipes-devtools/python/python-rtimu/*.patch
meta-raspberrypi/recipes-devtools/python/python-rtimu_git.bb
meta-raspberrypi/recipes-devtools/python/python-sense-hat_2.2.0.bb
meta-raspberrypi/recipes-devtools/python/rpi-gpio
meta-raspberrypi/recipes-devtools/python/rpi-gpio/*.patch
meta-raspberrypi/recipes-devtools/python/rpi-gpio_0.6.3.bb
meta-raspberrypi/recipes-devtools/python/rpio
meta-raspberrypi/recipes-devtools/python/rpio/*.patch
meta-raspberrypi/recipes-devtools/python/rpio_0.10.0.bb
meta-raspberrypi/recipes-devtools/wiringPi
meta-raspberrypi/recipes-devtools/wiringPi/files
meta-raspberrypi/recipes-devtools/wiringPi/files/*.patch
meta-raspberrypi/recipes-devtools/wiringPi/wiringpi_git.bb
meta-raspberrypi/recipes-graphics
meta-raspberrypi/recipes-graphics/eglinfo
meta-raspberrypi/recipes-graphics/eglinfo/eglinfo-fb_%.bbappend
meta-raspberrypi/recipes-graphics/eglinfo/eglinfo-x11_%.bbappend
meta-raspberrypi/recipes-graphics/mesa
meta-raspberrypi/recipes-graphics/mesa/mesa-gl_%.bbappend
meta-raspberrypi/recipes-graphics/mesa/mesa_%.bbappend
meta-raspberrypi/recipes-graphics/userland
meta-raspberrypi/recipes-graphics/userland/userland
meta-raspberrypi/recipes-graphics/userland/userland/*.patch
meta-raspberrypi/recipes-graphics/userland/userland_git.bb
meta-raspberrypi/recipes-graphics/vc-graphics
meta-raspberrypi/recipes-graphics/vc-graphics/files
meta-raspberrypi/recipes-graphics/vc-graphics/files/egl.pc
meta-raspberrypi/recipes-graphics/vc-graphics/files/vchiq.sh
meta-raspberrypi/recipes-graphics/vc-graphics/vc-graphics-hardfp.bb
meta-raspberrypi/recipes-graphics/vc-graphics/vc-graphics.bb
meta-raspberrypi/recipes-graphics/vc-graphics/vc-graphics.inc
meta-raspberrypi/recipes-graphics/wayland
meta-raspberrypi/recipes-graphics/wayland/weston_%.bbappend
meta-raspberrypi/recipes-graphics/xorg-xserver
meta-raspberrypi/recipes-graphics/xorg-xserver/xserver-xf86-config
meta-raspberrypi/recipes-graphics/xorg-xserver/xserver-xf86-config/rpi
meta-raspberrypi/recipes-graphics/xorg-xserver/xserver-xf86-config/rpi/xorg.conf
meta-raspberrypi/recipes-graphics/xorg-xserver/xserver-xf86-config/rpi/xorg.conf.d
meta-raspberrypi/recipes-graphics/xorg-xserver/xserver-xf86-config/rpi/xorg.conf.d/10-evdev.conf
meta-raspberrypi/recipes-graphics/xorg-xserver/xserver-xf86-config/rpi/xorg.conf.d/98-pitft.conf
meta-raspberrypi/recipes-graphics/xorg-xserver/xserver-xf86-config/rpi/xorg.conf.d/99-calibration.conf
meta-raspberrypi/recipes-graphics/xorg-xserver/xserver-xf86-config_0.1.bbappend
meta-raspberrypi/recipes-graphics/xorg-xserver/xserver-xorg_%.bbappend
meta-raspberrypi/recipes-kernel
meta-raspberrypi/recipes-kernel/linux-firmware
meta-raspberrypi/recipes-kernel/linux-firmware/files
meta-raspberrypi/recipes-kernel/linux-firmware/files/brcmfmac43430-sdio.bin
meta-raspberrypi/recipes-kernel/linux-firmware/files/brcfmac43430-sdio.txt
meta-raspberrypi/recipes-kernel/linux-firmware/linux-firmware_%.bbappend
meta-raspberrypi/recipes-kernel/linux
meta-raspberrypi/recipes-kernel/linux/linux-raspberrypi-dev.bb
meta-raspberrypi/recipes-kernel/linux/linux-raspberrypi.inc
meta-raspberrypi/recipes-kernel/linux/linux-raspberrypi_4.14.bb
meta-raspberrypi/recipes-kernel/linux/linux-raspberrypi_4.9.bb
meta-raspberrypi/recipes-multimedia
meta-raspberrypi/recipes-multimedia/gstreamer
meta-raspberrypi/recipes-multimedia/gstreamer/gstreamer1.0-omx
meta-raspberrypi/recipes-multimedia/gstreamer/gstreamer1.0-omx/*.patch
meta-raspberrypi/recipes-multimedia/gstreamer/gstreamer1.0-omx_%.bbappend
meta-raspberrypi/recipes-multimedia/gstreamer/gstreamer1.0-plugins-bad_%.bbappend
meta-raspberrypi/recipes-multimedia/gstreamer/gstreamer1.0-omx-1.12
meta-raspberrypi/recipes-multimedia/gstreamer/gstreamer1.0-omx-1.12/*.patch
meta-raspberrypi/recipes-multimedia/omxplayer
meta-raspberrypi/recipes-multimedia/omxplayer/omxplayer
meta-raspberrypi/recipes-multimedia/omxplayer/omxplayer/*.patch
meta-raspberrypi/recipes-multimedia/omxplayer/omxplayer_git.bb
meta-raspberrypi/recipes-multimedia/x264
meta-raspberrypi/recipes-multimedia/x264/x264_git.bbappend
meta-raspberrypi/wic meta-raspberrypi/wic/sdimage-raspberrypi.wks
The following sections describe each part of the proposed BSP format.
1.3.1 License Files
You can find these files in the BSP Layer at:
meta-bsp_root_name/bsp_license_file
These optional files satisfy licensing requirements for the BSP. The
type or types of files here can vary depending on the licensing
requirements. For example, in the Raspberry Pi BSP, all licensing
requirements are handled with the COPYING.MIT
file.
Licensing files can be MIT, BSD, GPLv*, and so forth. These files are recommended for the BSP but are optional and totally up to the BSP developer. For information on how to maintain license compliance, see the “Maintaining Open Source License Compliance During Your Product’s Lifecycle” section in the Yocto Project Development Tasks Manual.
1.3.2 README File
You can find this file in the BSP Layer at:
meta-bsp_root_name/README
This file provides information on how to boot the live images that are
optionally included in the binary/
directory. The README
file
also provides information needed for building the image.
At a minimum, the README
file must contain a list of dependencies,
such as the names of any other layers on which the BSP depends and the
name of the BSP maintainer with his or her contact information.
1.3.3 README.sources File
You can find this file in the BSP Layer at:
meta-bsp_root_name/README.sources
This file provides information on where to locate the BSP source files
used to build the images (if any) that reside in
meta-bsp_root_name/binary
. Images in the binary
would be images
released with the BSP. The information in the README.sources
file
also helps you find the Metadata
used to generate the images that ship with the BSP.
Note
If the BSP’s binary
directory is missing or the directory has no images, an
existing README.sources
file is meaningless and usually does not exist.
1.3.4 Pre-built User Binaries
You can find these files in the BSP Layer at:
meta-bsp_root_name/binary/bootable_images
This optional area contains useful pre-built kernels and user-space filesystem images released with the BSP that are appropriate to the target system. This directory typically contains graphical (e.g. Sato) and minimal live images when the BSP tarball has been created and made available in the Yocto Project website. You can use these kernels and images to get a system running and quickly get started on development tasks.
The exact types of binaries present are highly hardware-dependent. The README file should be present in the BSP Layer and it explains how to use the images with the target hardware. Additionally, the README.sources file should be present to locate the sources used to build the images and provide information on the Metadata.
1.3.5 Layer Configuration File
You can find this file in the BSP Layer at:
meta-bsp_root_name/conf/layer.conf
The conf/layer.conf
file identifies the file structure as a layer,
identifies the contents of the layer, and contains information about how
the build system should use it. Generally, a standard boilerplate file
such as the following works. In the following example, you would replace
“bsp” with the actual name of the BSP (i.e. “bsp_root_name” from the example
template).
# We have a conf and classes directory, add to BBPATH
BBPATH .= ":${LAYERDIR}"
# We have a recipes directory containing .bb and .bbappend files, add to BBFILES
BBFILES += "${LAYERDIR}/recipes-*/*/*.bb \
${LAYERDIR}/recipes-*/*/*.bbappend"
BBFILE_COLLECTIONS += "bsp"
BBFILE_PATTERN_bsp = "^${LAYERDIR}/"
BBFILE_PRIORITY_bsp = "6"
LAYERDEPENDS_bsp = "intel"
To illustrate the string substitutions, here are the corresponding
statements from the Raspberry Pi conf/layer.conf
file:
# We have a conf and classes directory, append to BBPATH
BBPATH .= ":${LAYERDIR}"
# We have a recipes directory containing .bb and .bbappend files, add to BBFILES
BBFILES += "${LAYERDIR}/recipes*/*/*.bb \
${LAYERDIR}/recipes*/*/*.bbappend"
BBFILE_COLLECTIONS += "raspberrypi"
BBFILE_PATTERN_raspberrypi := "^${LAYERDIR}/"
BBFILE_PRIORITY_raspberrypi = "9"
# Additional license directories.
LICENSE_PATH += "${LAYERDIR}/files/custom-licenses"
.
.
.
This file simply makes BitBake aware of the recipes and configuration directories. The file must exist so that the OpenEmbedded build system can recognize the BSP.
1.3.6 Hardware Configuration Options
You can find these files in the BSP Layer at:
meta-bsp_root_name/conf/machine/*.conf
The machine files bind together all the information contained elsewhere in the BSP into a format that the build system can understand. Each BSP Layer requires at least one machine file. If the BSP supports multiple machines, multiple machine configuration files can exist. These filenames correspond to the values to which users have set the MACHINE variable.
These files define things such as the kernel package to use (PREFERRED_PROVIDER of virtual/kernel), the hardware drivers to include in different types of images, any special software components that are needed, any bootloader information, and also any special image format requirements.
This configuration file could also include a hardware “tuning” file that is commonly used to define the package architecture and specify optimization flags, which are carefully chosen to give best performance on a given processor.
Tuning files are found in the meta/conf/machine/include
directory
within the Source Directory.
For example, many tune-*
files (e.g. tune-arm1136jf-s.inc
,
tune-1586-nlp.inc
, and so forth) reside in the
poky/meta/conf/machine/include
directory.
To use an include file, you simply include them in the machine
configuration file. For example, the Raspberry Pi BSP
raspberrypi3.conf
contains the following statement:
include conf/machine/include/rpi-base.inc
1.3.7 Miscellaneous BSP-Specific Recipe Files
You can find these files in the BSP Layer at:
meta-bsp_root_name/recipes-bsp/*
This optional directory contains miscellaneous recipe files for the BSP.
Most notably would be the formfactor files. For example, in the
Raspberry Pi BSP, there is the formfactor_0.0.bbappend
file, which
is an append file used to augment the recipe that starts the build.
Furthermore, there are machine-specific settings used during the build
that are defined by the machconfig
file further down in the
directory. Here is the machconfig
file for the Raspberry Pi BSP:
HAVE_TOUCHSCREEN=0
HAVE_KEYBOARD=1
DISPLAY_CAN_ROTATE=0
DISPLAY_ORIENTATION=0
DISPLAY_DPI=133
Note
If a BSP does not have a formfactor entry, defaults are established
according to the formfactor configuration file that is installed by
the main formfactor recipe
meta/recipes-bsp/formfactor/formfactor_0.0.bb
, which is found in
the Source Directory.
1.3.8 Display Support Files
You can find these files in the BSP Layer at:
meta-bsp_root_name/recipes-graphics/*
This optional directory contains recipes for the BSP if it has special requirements for graphics support. All files that are needed for the BSP to support a display are kept here.
1.3.9 Linux Kernel Configuration
You can find these files in the BSP Layer at:
meta-bsp_root_name/recipes-kernel/linux/linux*.bbappend
meta-bsp_root_name/recipes-kernel/linux/*.bb
Append files (*.bbappend
) modify the main kernel recipe being used
to build the image. The *.bb
files would be a developer-supplied
kernel recipe. This area of the BSP hierarchy can contain both these
types of files although, in practice, it is likely that you would have
one or the other.
For your BSP, you typically want to use an existing Yocto Project kernel
recipe found in the Source Directory
at
meta/recipes-kernel/linux
. You can append machine-specific changes
to the kernel recipe by using a similarly named append file, which is
located in the BSP Layer for your target device (e.g. the
meta-bsp_root_name/recipes-kernel/linux
directory).
Suppose you are using the linux-yocto_4.4.bb
recipe to build the
kernel. In other words, you have selected the kernel in your
"bsp_root_name".conf
file by adding
PREFERRED_PROVIDER and PREFERRED_VERSION
statements as follows:
PREFERRED_PROVIDER_virtual/kernel ?= "linux-yocto"
PREFERRED_VERSION_linux-yocto ?= "4.4%"
Note
When the preferred provider is assumed by default, the PREFERRED_PROVIDER
statement does not appear in the "bsp_root_name".conf
file.
You would use the linux-yocto_4.4.bbappend
file to append specific
BSP settings to the kernel, thus configuring the kernel for your
particular BSP.
You can find more information on what your append file should contain in the “Creating the Append File” section in the Yocto Project Linux Kernel Development Manual.
An alternate scenario is when you create your own kernel recipe for the
BSP. A good example of this is the Raspberry Pi BSP. If you examine the
recipes-kernel/linux
directory you see the following:
linux-raspberrypi-dev.bb
linux-raspberrypi.inc
linux-raspberrypi_4.14.bb
linux-raspberrypi_4.9.bb
The directory contains three kernel recipes and a common include file.
1.4 Developing a Board Support Package (BSP)
This section describes the high-level procedure you can follow to create
a BSP. Although not required for BSP creation, the meta-intel
repository, which contains many BSPs supported by the Yocto Project, is
part of the example.
For an example that shows how to create a new layer using the tools, see the “Creating a new BSP Layer Using the bitbake-layers Script” section.
The following illustration and list summarize the BSP creation general workflow.
Set up Your Host Development System to Support Development Using the Yocto Project: See the “Preparing the Build Host” section in the Yocto Project Development Tasks Manual for options on how to get a system ready to use the Yocto Project.
Establish the meta-intel Repository on Your System: Having local copies of these supported BSP layers on your system gives you access to layers you might be able to leverage when creating your BSP. For information on how to get these files, see the “Preparing Your Build Host to Work With BSP Layers” section.
Create Your Own BSP Layer Using the bitbake-layers Script: Layers are ideal for isolating and storing work for a given piece of hardware. A layer is really just a location or area in which you place the recipes and configurations for your BSP. In fact, a BSP is, in itself, a special type of layer. The simplest way to create a new BSP layer that is compliant with the Yocto Project is to use the
bitbake-layers
script. For information about that script, see the “Creating a new BSP Layer Using the bitbake-layers Script” section.Another example that illustrates a layer is an application. Suppose you are creating an application that has library or other dependencies in order for it to compile and run. The layer, in this case, would be where all the recipes that define those dependencies are kept. The key point for a layer is that it is an isolated area that contains all the relevant information for the project that the OpenEmbedded build system knows about. For more information on layers, see the “The Yocto Project Layer Model” section in the Yocto Project Overview and Concepts Manual. You can also reference the “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual. For more information on BSP layers, see the “BSP Layers” section.
Note
Four hardware reference BSPs exist that are part of the Yocto Project release and are located in the
poky/meta-yocto-bsp
BSP layer:Texas Instruments Beaglebone (
beaglebone-yocto
)Ubiquiti Networks EdgeRouter Lite (
edgerouter
)Two general IA platforms (
genericx86
andgenericx86-64
)
Three core Intel BSPs exist as part of the Yocto Project release in the
meta-intel
layer:intel-core2-32
, which is a BSP optimized for the Core2 family of CPUs as well as all CPUs prior to the Silvermont core.intel-corei7-64
, which is a BSP optimized for Nehalem and later Core and Xeon CPUs as well as Silvermont and later Atom CPUs, such as the Baytrail SoCs.intel-quark
, which is a BSP optimized for the Intel Galileo gen1 & gen2 development boards.
When you set up a layer for a new BSP, you should follow a standard layout. This layout is described in the “Example Filesystem Layout” section. In the standard layout, notice the suggested structure for recipes and configuration information. You can see the standard layout for a BSP by examining any supported BSP found in the
meta-intel
layer inside the Source Directory.Make Configuration Changes to Your New BSP Layer: The standard BSP layer structure organizes the files you need to edit in
conf
and severalrecipes-*
directories within the BSP layer. Configuration changes identify where your new layer is on the local system and identifies the kernel you are going to use. When you run thebitbake-layers
script, you are able to interactively configure many things for the BSP (e.g. keyboard, touchscreen, and so forth).Make Recipe Changes to Your New BSP Layer: Recipe changes include altering recipes (
*.bb
files), removing recipes you do not use, and adding new recipes or append files (.bbappend
) that support your hardware.Prepare for the Build: Once you have made all the changes to your BSP layer, there remains a few things you need to do for the OpenEmbedded build system in order for it to create your image. You need to get the build environment ready by sourcing an environment setup script (i.e.
oe-init-build-env
) and you need to be sure two key configuration files are configured appropriately: theconf/local.conf
and theconf/bblayers.conf
file. You must make the OpenEmbedded build system aware of your new layer. See the “Enabling Your Layer” section in the Yocto Project Development Tasks Manual for information on how to let the build system know about your new layer.Build the Image: The OpenEmbedded build system uses the BitBake tool to build images based on the type of image you want to create. You can find more information about BitBake in the BitBake User Manual.
The build process supports several types of images to satisfy different needs. See the “Images” chapter in the Yocto Project Reference Manual for information on supported images.
1.5 Requirements and Recommendations for Released BSPs
Certain requirements exist for a released BSP to be considered compliant with the Yocto Project. Additionally, recommendations also exist. This section describes the requirements and recommendations for released BSPs.
1.5.1 Released BSP Requirements
Before looking at BSP requirements, you should consider the following:
The requirements here assume the BSP layer is a well-formed, “legal” layer that can be added to the Yocto Project. For guidelines on creating a layer that meets these base requirements, see the “BSP Layers” section in this manual and the “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual.
The requirements in this section apply regardless of how you package a BSP. You should consult the packaging and distribution guidelines for your specific release process. For an example of packaging and distribution requirements, see the “Third Party BSP Release Process” wiki page.
The requirements for the BSP as it is made available to a developer are completely independent of the released form of the BSP. For example, the BSP Metadata can be contained within a Git repository and could have a directory structure completely different from what appears in the officially released BSP layer.
It is not required that specific packages or package modifications exist in the BSP layer, beyond the requirements for general compliance with the Yocto Project. For example, no requirement exists dictating that a specific kernel or kernel version be used in a given BSP.
Following are the requirements for a released BSP that conform to the Yocto Project:
Layer Name: The BSP must have a layer name that follows the Yocto Project standards. For information on BSP layer names, see the “BSP Layers” section.
File System Layout: When possible, use the same directory names in your BSP layer as listed in the
recipes.txt
file, which is found inpoky/meta
directory of the Source Directory or in the OpenEmbedded-Core Layer (openembedded-core
) at https://git.openembedded.org/openembedded-core/tree/meta.You should place recipes (
*.bb
files) and recipe modifications (*.bbappend
files) intorecipes-*
subdirectories by functional area as outlined inrecipes.txt
. If you cannot find a category inrecipes.txt
to fit a particular recipe, you can make up your ownrecipes-*
subdirectory.Within any particular
recipes-*
category, the layout should match what is found in the OpenEmbedded-Core Git repository (openembedded-core
) or the Source Directory (poky
). In other words, make sure you place related files in appropriately-relatedrecipes-*
subdirectories specific to the recipe’s function, or within a subdirectory containing a set of closely-related recipes. The recipes themselves should follow the general guidelines for recipes used in the Yocto Project found in the “OpenEmbedded Style Guide”.License File: You must include a license file in the
meta-bsp_root_name
directory. This license covers the BSP Metadata as a whole. You must specify which license to use since no default license exists when one is not specified. See the COPYING.MIT file for the Raspberry Pi BSP in themeta-raspberrypi
BSP layer as an example.README File: You must include a
README
file in themeta-bsp_root_name
directory. See the README.md file for the Raspberry Pi BSP in themeta-raspberrypi
BSP layer as an example.At a minimum, the
README
file should contain the following:A brief description of the target hardware.
A list of all the dependencies of the BSP. These dependencies are typically a list of required layers needed to build the BSP. However, the dependencies should also contain information regarding any other dependencies the BSP might have.
Any required special licensing information. For example, this information includes information on special variables needed to satisfy a EULA, or instructions on information needed to build or distribute binaries built from the BSP Metadata.
The name and contact information for the BSP layer maintainer. This is the person to whom patches and questions should be sent. For information on how to find the right person, see the “Submitting a Change to the Yocto Project” section in the Yocto Project Development Tasks Manual.
Instructions on how to build the BSP using the BSP layer.
Instructions on how to boot the BSP build from the BSP layer.
Instructions on how to boot the binary images contained in the
binary
directory, if present.Information on any known bugs or issues that users should know about when either building or booting the BSP binaries.
README.sources File: If your BSP contains binary images in the
binary
directory, you must include aREADME.sources
file in themeta-bsp_root_name
directory. This file specifies exactly where you can find the sources used to generate the binary images.Layer Configuration File: You must include a
conf/layer.conf
file in themeta-bsp_root_name
directory. This file identifies themeta-bsp_root_name
BSP layer as a layer to the build system.Machine Configuration File: You must include one or more
conf/machine/bsp_root_name.conf
files in themeta-bsp_root_name
directory. These configuration files define machine targets that can be built using the BSP layer. Multiple machine configuration files define variations of machine configurations that the BSP supports. If a BSP supports multiple machine variations, you need to adequately describe each variation in the BSPREADME
file. Do not use multiple machine configuration files to describe disparate hardware. If you do have very different targets, you should create separate BSP layers for each target.Note
It is completely possible for a developer to structure the working repository as a conglomeration of unrelated BSP files, and to possibly generate BSPs targeted for release from that directory using scripts or some other mechanism (e.g.
meta-yocto-bsp
layer). Such considerations are outside the scope of this document.
1.5.2 Released BSP Recommendations
Following are recommendations for released BSPs that conform to the Yocto Project:
Bootable Images: Released BSPs can contain one or more bootable images. Including bootable images allows users to easily try out the BSP using their own hardware.
In some cases, it might not be convenient to include a bootable image. If so, you might want to make two versions of the BSP available: one that contains binary images, and one that does not. The version that does not contain bootable images avoids unnecessary download times for users not interested in the images.
If you need to distribute a BSP and include bootable images or build kernel and filesystems meant to allow users to boot the BSP for evaluation purposes, you should put the images and artifacts within a
binary/
subdirectory located in themeta-bsp_root_name
directory.Note
If you do include a bootable image as part of the BSP and the image was built by software covered by the GPL or other open source licenses, it is your responsibility to understand and meet all licensing requirements, which could include distribution of source files.
Use a Yocto Linux Kernel: Kernel recipes in the BSP should be based on a Yocto Linux kernel. Basing your recipes on these kernels reduces the costs for maintaining the BSP and increases its scalability. See the
Yocto Linux Kernel
category in the Source Repositories for these kernels.
1.6 Customizing a Recipe for a BSP
If you plan on customizing a recipe for a particular BSP, you need to do the following:
Create a
*.bbappend
file for the modified recipe. For information on using append files, see the “Using .bbappend Files in Your Layer” section in the Yocto Project Development Tasks Manual.Ensure your directory structure in the BSP layer that supports your machine is such that the OpenEmbedded build system can find it. See the example later in this section for more information.
Put the append file in a directory whose name matches the machine’s name and is located in an appropriate sub-directory inside the BSP layer (i.e.
recipes-bsp
,recipes-graphics
,recipes-core
, and so forth).Place the BSP-specific files in the proper directory inside the BSP layer. How expansive the layer is affects where you must place these files. For example, if your layer supports several different machine types, you need to be sure your layer’s directory structure includes hierarchy that separates the files according to machine. If your layer does not support multiple machines, the layer would not have that additional hierarchy and the files would obviously not be able to reside in a machine-specific directory.
Following is a specific example to help you better understand the
process. This example customizes a recipe by adding a
BSP-specific configuration file named interfaces
to the
init-ifupdown_1.0.bb
recipe for machine “xyz” where the BSP layer
also supports several other machines:
Edit the
init-ifupdown_1.0.bbappend
file so that it contains the following:FILESEXTRAPATHS_prepend := "${THISDIR}/files:"
The append file needs to be in the
meta-xyz/recipes-core/init-ifupdown
directory.Create and place the new
interfaces
configuration file in the BSP’s layer here:meta-xyz/recipes-core/init-ifupdown/files/xyz-machine-one/interfaces
Note
If the
meta-xyz
layer did not support multiple machines, you would place the interfaces configuration file in the layer here:meta-xyz/recipes-core/init-ifupdown/files/interfaces
The FILESEXTRAPATHS variable in the append files extends the search path the build system uses to find files during the build. Consequently, for this example you need to have the
files
directory in the same location as your append file.
1.7 BSP Licensing Considerations
In some cases, a BSP contains separately-licensed Intellectual Property (IP) for a component or components. For these cases, you are required to accept the terms of a commercial or other type of license that requires some kind of explicit End User License Agreement (EULA). Once you accept the license, the OpenEmbedded build system can then build and include the corresponding component in the final BSP image. If the BSP is available as a pre-built image, you can download the image after agreeing to the license or EULA.
You could find that some separately-licensed components that are essential for normal operation of the system might not have an unencumbered (or free) substitute. Without these essential components, the system would be non-functional. Then again, you might find that other licensed components that are simply ‘good-to-have’ or purely elective do have an unencumbered, free replacement component that you can use rather than agreeing to the separately-licensed component. Even for components essential to the system, you might find an unencumbered component that is not identical but will work as a less-capable version of the licensed version in the BSP recipe.
For cases where you can substitute a free component and still maintain the system’s functionality, the “DOWNLOADS” selection from the “SOFTWARE” tab on the Yocto Project Website makes available de-featured BSPs that are completely free of any IP encumbrances. For these cases, you can use the substitution directly and without any further licensing requirements. If present, these fully de-featured BSPs are named appropriately different as compared to the names of their respective encumbered BSPs. If available, these substitutions are your simplest and most preferred options. Obviously, use of these substitutions assumes the resulting functionality meets system requirements.
Note
If however, a non-encumbered version is unavailable or it provides unsuitable functionality or quality, you can use an encumbered version.
A couple different methods exist within the OpenEmbedded build system to satisfy the licensing requirements for an encumbered BSP. The following list describes them in order of preference:
Use the LICENSE_FLAGS Variable to Define the Recipes that Have Commercial or Other Types of Specially-Licensed Packages: For each of those recipes, you can specify a matching license string in a
local.conf
variable named LICENSE_FLAGS_WHITELIST. Specifying the matching license string signifies that you agree to the license. Thus, the build system can build the corresponding recipe and include the component in the image. See the “Enabling Commercially Licensed Recipes” section in the Yocto Project Development Tasks Manual for details on how to use these variables.If you build as you normally would, without specifying any recipes in the
LICENSE_FLAGS_WHITELIST
, the build stops and provides you with the list of recipes that you have tried to include in the image that need entries in theLICENSE_FLAGS_WHITELIST
. Once you enter the appropriate license flags into the whitelist, restart the build to continue where it left off. During the build, the prompt will not appear again since you have satisfied the requirement.Once the appropriate license flags are on the white list in the
LICENSE_FLAGS_WHITELIST
variable, you can build the encumbered image with no change at all to the normal build process.Get a Pre-Built Version of the BSP: You can get this type of BSP by selecting the “DOWNLOADS” item from the “SOFTWARE” tab on the Yocto Project website. You can download BSP tarballs that contain proprietary components after agreeing to the licensing requirements of each of the individually encumbered packages as part of the download process. Obtaining the BSP this way allows you to access an encumbered image immediately after agreeing to the click-through license agreements presented by the website. If you want to build the image yourself using the recipes contained within the BSP tarball, you will still need to create an appropriate
LICENSE_FLAGS_WHITELIST
to match the encumbered recipes in the BSP.
Note
Pre-compiled images are bundled with a time-limited kernel that runs for a predetermined amount of time (10 days) before it forces the system to reboot. This limitation is meant to discourage direct redistribution of the image. You must eventually rebuild the image if you want to remove this restriction.
1.8 Creating a new BSP Layer Using the bitbake-layers
Script
The bitbake-layers create-layer
script automates creating a BSP
layer. What makes a layer a “BSP layer” is the presence of at least one
machine configuration file. Additionally, a BSP layer usually has a
kernel recipe or an append file that leverages off an existing kernel
recipe. The primary requirement, however, is the machine configuration.
Use these steps to create a BSP layer:
Create a General Layer: Use the
bitbake-layers
script with thecreate-layer
subcommand to create a new general layer. For instructions on how to create a general layer using thebitbake-layers
script, see the “Creating a General Layer Using the bitbake-layers Script” section in the Yocto Project Development Tasks Manual.Create a Layer Configuration File: Every layer needs a layer configuration file. This configuration file establishes locations for the layer’s recipes, priorities for the layer, and so forth. You can find examples of
layer.conf
files in the Yocto Project Source Repositories. To get examples of what you need in your configuration file, locate a layer (e.g. “meta-ti”) and examine the local.conf file.Create a Machine Configuration File: Create a
conf/machine/bsp_root_name.conf
file. See meta-yocto-bsp/conf/machine for samplebsp_root_name.conf
files. Other samples such as meta-ti and meta-freescale exist from other vendors that have more specific machine and tuning examples.Create a Kernel Recipe: Create a kernel recipe in
recipes-kernel/linux
by either using a kernel append file or a new custom kernel recipe file (e.g.yocto-linux_4.12.bb
). The BSP layers mentioned in the previous step also contain different kernel examples. See the “Modifying an Existing Recipe” section in the Yocto Project Linux Kernel Development Manual for information on how to create a custom kernel.
The remainder of this section provides a description of the Yocto Project reference BSP for Beaglebone, which resides in the meta-yocto-bsp layer.
1.8.1 BSP Layer Configuration Example
The layer’s conf
directory contains the layer.conf
configuration
file. In this example, the conf/layer.conf
is the following:
# We have a conf and classes directory, add to BBPATH
BBPATH .= ":${LAYERDIR}"
# We have a recipes directory containing .bb and .bbappend files, add to BBFILES
BBFILES += "${LAYERDIR}/recipes-*/*/*.bb \
${LAYERDIR}/recipes-*/*/*.bbappend"
BBFILE_COLLECTIONS += "yoctobsp"
BBFILE_PATTERN_yoctobsp = "^${LAYERDIR}/"
BBFILE_PRIORITY_yoctobsp = "5"
LAYERVERSION_yoctobsp = "4"
LAYERSERIES_COMPAT_yoctobsp = "hardknott"
The variables used in this file configure the layer. A good way to learn about layer configuration files is to examine various files for BSP from the Source Repositories.
For a detailed description of this particular layer configuration file, see “step 3” in the discussion that describes how to create layers in the Yocto Project Development Tasks Manual.
1.8.2 BSP Machine Configuration Example
As mentioned earlier in this section, the existence of a machine configuration file is what makes a layer a BSP layer as compared to a general or kernel layer.
One or more machine configuration files exist in the
bsp_layer/conf/machine/
directory of the layer:
bsp_layer/conf/machine/machine1\.conf
bsp_layer/conf/machine/machine2\.conf
bsp_layer/conf/machine/machine3\.conf
... more ...
For example, the machine configuration file for the BeagleBone and
BeagleBone Black development boards is
located in the layer poky/meta-yocto-bsp/conf/machine
and is named
beaglebone-yocto.conf
:
#@TYPE: Machine
#@NAME: Beaglebone-yocto machine
#@DESCRIPTION: Reference machine configuration for http://beagleboard.org/bone and http://beagleboard.org/black boards
PREFERRED_PROVIDER_virtual/xserver ?= "xserver-xorg"
XSERVER ?= "xserver-xorg \
xf86-video-modesetting \
"
MACHINE_EXTRA_RRECOMMENDS = "kernel-modules kernel-devicetree"
EXTRA_IMAGEDEPENDS += "u-boot"
DEFAULTTUNE ?= "cortexa8hf-neon"
include conf/machine/include/tune-cortexa8.inc
IMAGE_FSTYPES += "tar.bz2 jffs2 wic wic.bmap"
EXTRA_IMAGECMD_jffs2 = "-lnp "
WKS_FILE ?= "beaglebone-yocto.wks"
IMAGE_INSTALL_append = " kernel-devicetree kernel-image-zimage"
do_image_wic[depends] += "mtools-native:do_populate_sysroot dosfstools-native:do_populate_sysroot"
SERIAL_CONSOLES ?= "115200;ttyS0 115200;ttyO0"
SERIAL_CONSOLES_CHECK = "${SERIAL_CONSOLES}"
PREFERRED_PROVIDER_virtual/kernel ?= "linux-yocto"
PREFERRED_VERSION_linux-yocto ?= "5.0%"
KERNEL_IMAGETYPE = "zImage"
KERNEL_DEVICETREE = "am335x-bone.dtb am335x-boneblack.dtb am335x-bonegreen.dtb"
KERNEL_EXTRA_ARGS += "LOADADDR=${UBOOT_ENTRYPOINT}"
SPL_BINARY = "MLO"
UBOOT_SUFFIX = "img"
UBOOT_MACHINE = "am335x_evm_defconfig"
UBOOT_ENTRYPOINT = "0x80008000"
UBOOT_LOADADDRESS = "0x80008000"
MACHINE_FEATURES = "usbgadget usbhost vfat alsa"
IMAGE_BOOT_FILES ?= "u-boot.${UBOOT_SUFFIX} MLO zImage am335x-bone.dtb am335x-boneblack.dtb am335x-bonegreen.dtb"
The variables used to configure the machine define machine-specific properties; for example, machine-dependent packages, machine tunings, the type of kernel to build, and U-Boot configurations.
The following list provides some explanation for the statements found in the example reference machine configuration file for the BeagleBone development boards. Realize that much more can be defined as part of a machine’s configuration file. In general, you can learn about related variables that this example does not have by locating the variables in the “Variables Glossary” in the Yocto Project Reference Manual.
PREFERRED_PROVIDER_virtual/xserver: The recipe that provides “virtual/xserver” when more than one provider is found. In this case, the recipe that provides “virtual/xserver” is “xserver-xorg”, which exists in
poky/meta/recipes-graphics/xorg-xserver
.XSERVER: The packages that should be installed to provide an X server and drivers for the machine. In this example, the “xserver-xorg” and “xf86-video-modesetting” are installed.
MACHINE_EXTRA_RRECOMMENDS: A list of machine-dependent packages not essential for booting the image. Thus, the build does not fail if the packages do not exist. However, the packages are required for a fully-featured image.
Tip
Many
MACHINE*
variables exist that help you configure a particular piece of hardware.EXTRA_IMAGEDEPENDS: Recipes to build that do not provide packages for installing into the root filesystem but building the image depends on the recipes. Sometimes a recipe is required to build the final image but is not needed in the root filesystem. In this case, the U-Boot recipe must be built for the image.
DEFAULTTUNE: Machines use tunings to optimize machine, CPU, and application performance. These features, which are collectively known as “tuning features”, exist in the OpenEmbedded-Core (OE-Core) layer (e.g.
poky/meta/conf/machine/include
). In this example, the default tuning file is “cortexa8hf-neon”.Note
The include statement that pulls in the
conf/machine/include/tune-cortexa8.inc
file provides many tuning possibilities.IMAGE_FSTYPES: The formats the OpenEmbedded build system uses during the build when creating the root filesystem. In this example, four types of images are supported.
EXTRA_IMAGECMD: Specifies additional options for image creation commands. In this example, the “-lnp “ option is used when creating the JFFS2 image.
WKS_FILE: The location of the Wic kickstart file used by the OpenEmbedded build system to create a partitioned image (image.wic).
IMAGE_INSTALL: Specifies packages to install into an image through the image class. Recipes use the
IMAGE_INSTALL
variable.do_image_wic[depends]
: A task that is constructed during the build. In this example, the task depends on specific tools in order to create the sysroot when building a Wic image.SERIAL_CONSOLES: Defines a serial console (TTY) to enable using getty. In this case, the baud rate is “115200” and the device name is “ttyO0”.
PREFERRED_PROVIDER_virtual/kernel: Specifies the recipe that provides “virtual/kernel” when more than one provider is found. In this case, the recipe that provides “virtual/kernel” is “linux-yocto”, which exists in the layer’s
recipes-kernel/linux
directory.PREFERRED_VERSION_linux-yocto: Defines the version of the recipe used to build the kernel, which is “5.0” in this case.
KERNEL_IMAGETYPE: The type of kernel to build for the device. In this case, the OpenEmbedded build system creates a “zImage” image type.
KERNEL_DEVICETREE: The names of the generated Linux kernel device trees (i.e. the
*.dtb
) files. All the device trees for the various BeagleBone devices are included.KERNEL_EXTRA_ARGS: Additional
make
command-line arguments the OpenEmbedded build system passes on when compiling the kernel. In this example,LOADADDR=${UBOOT_ENTRYPOINT}
is passed as a command-line argument.SPL_BINARY: Defines the Secondary Program Loader (SPL) binary type. In this case, the SPL binary is set to “MLO”, which stands for Multimedia card LOader.
The BeagleBone development board requires an SPL to boot and that SPL file type must be MLO. Consequently, the machine configuration needs to define
SPL_BINARY
asMLO
.Note
For more information on how the SPL variables are used, see the u-boot.inc include file.
UBOOT_*: Defines various U-Boot configurations needed to build a U-Boot image. In this example, a U-Boot image is required to boot the BeagleBone device. See the following variables for more information:
UBOOT_SUFFIX: Points to the generated U-Boot extension.
UBOOT_MACHINE: Specifies the value passed on the make command line when building a U-Boot image.
UBOOT_ENTRYPOINT: Specifies the entry point for the U-Boot image.
UBOOT_LOADADDRESS: Specifies the load address for the U-Boot image.
MACHINE_FEATURES: Specifies the list of hardware features the BeagleBone device is capable of supporting. In this case, the device supports “usbgadget usbhost vfat alsa”.
IMAGE_BOOT_FILES: Files installed into the device’s boot partition when preparing the image using the Wic tool with the
bootimg-partition
orbootimg-efi
source plugin.
1.8.3 BSP Kernel Recipe Example
The kernel recipe used to build the kernel image for the BeagleBone device was established in the machine configuration:
PREFERRED_PROVIDER_virtual/kernel ?= "linux-yocto"
PREFERRED_VERSION_linux-yocto ?= "5.0%"
The meta-yocto-bsp/recipes-kernel/linux
directory in the layer contains
metadata used to build the kernel. In this case, a kernel append file
(i.e. linux-yocto_5.0.bbappend
) is used to override an established
kernel recipe (i.e. linux-yocto_5.0.bb
), which is located in
https://git.yoctoproject.org/cgit/cgit.cgi/poky/tree/meta/recipes-kernel/linux.
Following is the contents of the append file:
KBRANCH_genericx86 = "v5.0/standard/base"
KBRANCH_genericx86-64 = "v5.0/standard/base"
KBRANCH_edgerouter = "v5.0/standard/edgerouter"
KBRANCH_beaglebone-yocto = "v5.0/standard/beaglebone"
KMACHINE_genericx86 ?= "common-pc"
KMACHINE_genericx86-64 ?= "common-pc-64"
KMACHINE_beaglebone-yocto ?= "beaglebone"
SRCREV_machine_genericx86 ?= "3df4aae6074e94e794e27fe7f17451d9353cdf3d"
SRCREV_machine_genericx86-64 ?= "3df4aae6074e94e794e27fe7f17451d9353cdf3d"
SRCREV_machine_edgerouter ?= "3df4aae6074e94e794e27fe7f17451d9353cdf3d"
SRCREV_machine_beaglebone-yocto ?= "3df4aae6074e94e794e27fe7f17451d9353cdf3d"
COMPATIBLE_MACHINE_genericx86 = "genericx86"
COMPATIBLE_MACHINE_genericx86-64 = "genericx86-64"
COMPATIBLE_MACHINE_edgerouter = "edgerouter"
COMPATIBLE_MACHINE_beaglebone-yocto = "beaglebone-yocto"
LINUX_VERSION_genericx86 = "5.0.3"
LINUX_VERSION_genericx86-64 = "5.0.3"
LINUX_VERSION_edgerouter = "5.0.3"
LINUX_VERSION_beaglebone-yocto = "5.0.3"
This particular append file works for all the machines that are
part of the meta-yocto-bsp
layer. The relevant statements are
appended with the “beaglebone-yocto” string. The OpenEmbedded build
system uses these statements to override similar statements in the
kernel recipe:
KBRANCH: Identifies the kernel branch that is validated, patched, and configured during the build.
KMACHINE: Identifies the machine name as known by the kernel, which is sometimes a different name than what is known by the OpenEmbedded build system.
SRCREV: Identifies the revision of the source code used to build the image.
COMPATIBLE_MACHINE: A regular expression that resolves to one or more target machines with which the recipe is compatible.
LINUX_VERSION: The Linux version from kernel.org used by the OpenEmbedded build system to build the kernel image.
2 Manual Revision History
Revision |
Date |
Note |
---|---|---|
0.9 |
November 2010 |
The initial document released with the Yocto Project 0.9 Release |
1.0 |
April 2011 |
Released with the Yocto Project 1.0 Release. |
1.1 |
October 2011 |
Released with the Yocto Project 1.1 Release. |
1.2 |
April 2012 |
Released with the Yocto Project 1.2 Release. |
1.3 |
October 2012 |
Released with the Yocto Project 1.3 Release. |
1.4 |
April 2013 |
Released with the Yocto Project 1.4 Release. |
1.5 |
October 2013 |
Released with the Yocto Project 1.5 Release. |
1.6 |
April 2014 |
Released with the Yocto Project 1.6 Release. |
1.7 |
October 2014 |
Released with the Yocto Project 1.7 Release. |
1.8 |
April 2015 |
Released with the Yocto Project 1.8 Release. |
2.0 |
October 2015 |
Released with the Yocto Project 2.0 Release. |
2.1 |
April 2016 |
Released with the Yocto Project 2.1 Release. |
2.2 |
October 2016 |
Released with the Yocto Project 2.2 Release. |
2.3 |
May 2017 |
Released with the Yocto Project 2.3 Release. |
2.4 |
October 2017 |
Released with the Yocto Project 2.4 Release. |
2.5 |
May 2018 |
Released with the Yocto Project 2.5 Release. |
2.6 |
November 2018 |
Released with the Yocto Project 2.6 Release. |
2.7 |
May 2019 |
Released with the Yocto Project 2.7 Release. |
3.0 |
October 2019 |
Released with the Yocto Project 3.0 Release. |
3.1 |
April 2020 |
Released with the Yocto Project 3.1 Release. |
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Yocto Project Development Tasks Manual
1 The Yocto Project Development Tasks Manual
1.1 Welcome
Welcome to the Yocto Project Development Tasks Manual! This manual provides relevant procedures necessary for developing in the Yocto Project environment (i.e. developing embedded Linux images and user-space applications that run on targeted devices). The manual groups related procedures into higher-level sections. Procedures can consist of high-level steps or low-level steps depending on the topic.
This manual provides the following:
Procedures that help you get going with the Yocto Project. For example, procedures that show you how to set up a build host and work with the Yocto Project source repositories.
Procedures that show you how to submit changes to the Yocto Project. Changes can be improvements, new features, or bug fixes.
Procedures related to “everyday” tasks you perform while developing images and applications using the Yocto Project. For example, procedures to create a layer, customize an image, write a new recipe, and so forth.
This manual does not provide the following:
Redundant Step-by-step Instructions: For example, the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual contains detailed instructions on how to install an SDK, which is used to develop applications for target hardware.
Reference or Conceptual Material: This type of material resides in an appropriate reference manual. For example, system variables are documented in the Yocto Project Reference Manual.
Detailed Public Information Not Specific to the Yocto Project: For example, exhaustive information on how to use the Source Control Manager Git is better covered with Internet searches and official Git Documentation than through the Yocto Project documentation.
1.2 Other Information
Because this manual presents information for many different topics, supplemental information is recommended for full comprehension. For introductory information on the Yocto Project, see the Yocto Project Website. If you want to build an image with no knowledge of Yocto Project as a way of quickly testing it out, see the Yocto Project Quick Build document.
For a comprehensive list of links and other documentation, see the “Links and Related Documentation” section in the Yocto Project Reference Manual.
2 Setting Up to Use the Yocto Project
This chapter provides guidance on how to prepare to use the Yocto Project. You can learn about creating a team environment to develop using the Yocto Project, how to set up a build host, how to locate Yocto Project source repositories, and how to create local Git repositories.
2.1 Creating a Team Development Environment
It might not be immediately clear how you can use the Yocto Project in a team development environment, or how to scale it for a large team of developers. You can adapt the Yocto Project to many different use cases and scenarios; however, this flexibility could cause difficulties if you are trying to create a working setup that scales effectively.
To help you understand how to set up this type of environment, this section presents a procedure that gives you information that can help you get the results you want. The procedure is high-level and presents some of the project’s most successful experiences, practices, solutions, and available technologies that have proved to work well in the past; however, keep in mind, the procedure here is simply a starting point. You can build off these steps and customize the procedure to fit any particular working environment and set of practices.
Determine Who is Going to be Developing: You first need to understand who is going to be doing anything related to the Yocto Project and determine their roles. Making this determination is essential to completing subsequent steps, which are to get your equipment together and set up your development environment’s hardware topology.
The following roles exist:
Application Developer: This type of developer does application level work on top of an existing software stack.
Core System Developer: This type of developer works on the contents of the operating system image itself.
Build Engineer: This type of developer manages Autobuilders and releases. Depending on the specifics of the environment, not all situations might need a Build Engineer.
Test Engineer: This type of developer creates and manages automated tests that are used to ensure all application and core system development meets desired quality standards.
Gather the Hardware: Based on the size and make-up of the team, get the hardware together. Ideally, any development, build, or test engineer uses a system that runs a supported Linux distribution. These systems, in general, should be high performance (e.g. dual, six-core Xeons with 24 Gbytes of RAM and plenty of disk space). You can help ensure efficiency by having any machines used for testing or that run Autobuilders be as high performance as possible.
Note
Given sufficient processing power, you might also consider building Yocto Project development containers to be run under Docker, which is described later.
Understand the Hardware Topology of the Environment: Once you understand the hardware involved and the make-up of the team, you can understand the hardware topology of the development environment. You can get a visual idea of the machines and their roles across the development environment.
Use Git as Your Source Control Manager (SCM): Keeping your Metadata (i.e. recipes, configuration files, classes, and so forth) and any software you are developing under the control of an SCM system that is compatible with the OpenEmbedded build system is advisable. Of all of the SCMs supported by BitBake, the Yocto Project team strongly recommends using Git. Git is a distributed system that is easy to back up, allows you to work remotely, and then connects back to the infrastructure.
Note
For information about BitBake, see the BitBake User Manual.
It is relatively easy to set up Git services and create infrastructure like https://git.yoctoproject.org/cgit/cgit.cgi/, which is based on server software called
gitolite
withcgit
being used to generate the web interface that lets you view the repositories. Thegitolite
software identifies users using SSH keys and allows branch-based access controls to repositories that you can control as little or as much as necessary.Note
The setup of these services is beyond the scope of this manual. However, sites such as the following exist that describe how to perform setup:
Gitolite: Information for
gitolite
.Interfaces, frontends, and tools: Documentation on how to create interfaces and frontends for Git.
Set up the Application Development Machines: As mentioned earlier, application developers are creating applications on top of existing software stacks. Following are some best practices for setting up machines used for application development:
Use a pre-built toolchain that contains the software stack itself. Then, develop the application code on top of the stack. This method works well for small numbers of relatively isolated applications.
Keep your cross-development toolchains updated. You can do this through provisioning either as new toolchain downloads or as updates through a package update mechanism using
opkg
to provide updates to an existing toolchain. The exact mechanics of how and when to do this depend on local policy.Use multiple toolchains installed locally into different locations to allow development across versions.
Set up the Core Development Machines: As mentioned earlier, core developers work on the contents of the operating system itself. Following are some best practices for setting up machines used for developing images:
Have the OpenEmbedded Build System available on the developer workstations so developers can run their own builds and directly rebuild the software stack.
Keep the core system unchanged as much as possible and do your work in layers on top of the core system. Doing so gives you a greater level of portability when upgrading to new versions of the core system or Board Support Packages (BSPs).
Share layers amongst the developers of a particular project and contain the policy configuration that defines the project.
Set up an Autobuilder: Autobuilders are often the core of the development environment. It is here that changes from individual developers are brought together and centrally tested. Based on this automated build and test environment, subsequent decisions about releases can be made. Autobuilders also allow for “continuous integration” style testing of software components and regression identification and tracking.
See “Yocto Project Autobuilder” for more information and links to buildbot. The Yocto Project team has found this implementation works well in this role. A public example of this is the Yocto Project Autobuilders, which the Yocto Project team uses to test the overall health of the project.
The features of this system are:
Highlights when commits break the build.
Populates an sstate cache from which developers can pull rather than requiring local builds.
Allows commit hook triggers, which trigger builds when commits are made.
Allows triggering of automated image booting and testing under the QuickEMUlator (QEMU).
Supports incremental build testing and from-scratch builds.
Shares output that allows developer testing and historical regression investigation.
Creates output that can be used for releases.
Allows scheduling of builds so that resources can be used efficiently.
Set up Test Machines: Use a small number of shared, high performance systems for testing purposes. Developers can use these systems for wider, more extensive testing while they continue to develop locally using their primary development system.
Document Policies and Change Flow: The Yocto Project uses a hierarchical structure and a pull model. Scripts exist to create and send pull requests (i.e.
create-pull-request
andsend-pull-request
). This model is in line with other open source projects where maintainers are responsible for specific areas of the project and a single maintainer handles the final “top-of-tree” merges.Note
You can also use a more collective push model. The
gitolite
software supports both the push and pull models quite easily.As with any development environment, it is important to document the policy used as well as any main project guidelines so they are understood by everyone. It is also a good idea to have well-structured commit messages, which are usually a part of a project’s guidelines. Good commit messages are essential when looking back in time and trying to understand why changes were made.
If you discover that changes are needed to the core layer of the project, it is worth sharing those with the community as soon as possible. Chances are if you have discovered the need for changes, someone else in the community needs them also.
Development Environment Summary: Aside from the previous steps, some best practices exist within the Yocto Project development environment. Consider the following:
Use Git as the source control system.
Maintain your Metadata in layers that make sense for your situation. See the “The Yocto Project Layer Model” section in the Yocto Project Overview and Concepts Manual and the “Understanding and Creating Layers” section for more information on layers.
Separate the project’s Metadata and code by using separate Git repositories. See the “Yocto Project Source Repositories” section in the Yocto Project Overview and Concepts Manual for information on these repositories. See the “Locating Yocto Project Source Files” section for information on how to set up local Git repositories for related upstream Yocto Project Git repositories.
Set up the directory for the shared state cache (SSTATE_DIR) where it makes sense. For example, set up the sstate cache on a system used by developers in the same organization and share the same source directories on their machines.
Set up an Autobuilder and have it populate the sstate cache and source directories.
The Yocto Project community encourages you to send patches to the project to fix bugs or add features. If you do submit patches, follow the project commit guidelines for writing good commit messages. See the “Submitting a Change to the Yocto Project” section.
Send changes to the core sooner than later as others are likely to run into the same issues. For some guidance on mailing lists to use, see the list in the “Submitting a Change to the Yocto Project” section. For a description of the available mailing lists, see the “Mailing lists” section in the Yocto Project Reference Manual.
2.2 Preparing the Build Host
This section provides procedures to set up a system to be used as your Build Host for development using the Yocto Project. Your build host can be a native Linux machine (recommended), it can be a machine (Linux, Mac, or Windows) that uses CROPS, which leverages Docker Containers or it can be a Windows machine capable of running Windows Subsystem For Linux v2 (WSL).
Note
The Yocto Project is not compatible with Windows Subsystem for Linux v1. It is compatible but not officially supported nor validated with WSLv2. If you still decide to use WSL please upgrade to WSLv2.
Once your build host is set up to use the Yocto Project, further steps are necessary depending on what you want to accomplish. See the following references for information on how to prepare for Board Support Package (BSP) development and kernel development:
BSP Development: See the “Preparing Your Build Host to Work With BSP Layers” section in the Yocto Project Board Support Package (BSP) Developer’s Guide.
Kernel Development: See the “Preparing the Build Host to Work on the Kernel” section in the Yocto Project Linux Kernel Development Manual.
2.2.1 Setting Up a Native Linux Host
Follow these steps to prepare a native Linux machine as your Yocto Project Build Host:
Use a Supported Linux Distribution: You should have a reasonably current Linux-based host system. You will have the best results with a recent release of Fedora, openSUSE, Debian, Ubuntu, RHEL or CentOS as these releases are frequently tested against the Yocto Project and officially supported. For a list of the distributions under validation and their status, see the “Supported Linux Distributions” section in the Yocto Project Reference Manual and the wiki page at Distribution Support.
Have Enough Free Memory: Your system should have at least 50 Gbytes of free disk space for building images.
Meet Minimal Version Requirements: The OpenEmbedded build system should be able to run on any modern distribution that has the following versions for Git, tar, Python and gcc.
Git 1.8.3.1 or greater
tar 1.28 or greater
Python 3.6.0 or greater.
gcc 5.0 or greater.
If your build host does not meet any of these three listed version requirements, you can take steps to prepare the system so that you can still use the Yocto Project. See the “Required Git, tar, Python and gcc Versions” section in the Yocto Project Reference Manual for information.
Install Development Host Packages: Required development host packages vary depending on your build host and what you want to do with the Yocto Project. Collectively, the number of required packages is large if you want to be able to cover all cases.
For lists of required packages for all scenarios, see the “Required Packages for the Build Host” section in the Yocto Project Reference Manual.
Once you have completed the previous steps, you are ready to continue using a given development path on your native Linux machine. If you are going to use BitBake, see the “Cloning the poky Repository” section. If you are going to use the Extensible SDK, see the “Using the Extensible SDK” Chapter in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual. If you want to work on the kernel, see the Yocto Project Linux Kernel Development Manual. If you are going to use Toaster, see the “Setting Up and Using Toaster” section in the Toaster User Manual.
2.2.2 Setting Up to Use CROss PlatformS (CROPS)
With CROPS, which leverages Docker Containers, you can create a Yocto Project development environment that is operating system agnostic. You can set up a container in which you can develop using the Yocto Project on a Windows, Mac, or Linux machine.
Follow these general steps to prepare a Windows, Mac, or Linux machine as your Yocto Project build host:
Determine What Your Build Host Needs: Docker is a software container platform that you need to install on the build host. Depending on your build host, you might have to install different software to support Docker containers. Go to the Docker installation page and read about the platform requirements in “Supported Platforms” your build host needs to run containers.
Choose What To Install: Depending on whether or not your build host meets system requirements, you need to install “Docker CE Stable” or the “Docker Toolbox”. Most situations call for Docker CE. However, if you have a build host that does not meet requirements (e.g. Pre-Windows 10 or Windows 10 “Home” version), you must install Docker Toolbox instead.
Go to the Install Site for Your Platform: Click the link for the Docker edition associated with your build host’s native software. For example, if your build host is running Microsoft Windows Version 10 and you want the Docker CE Stable edition, click that link under “Supported Platforms”.
Install the Software: Once you have understood all the pre-requisites, you can download and install the appropriate software. Follow the instructions for your specific machine and the type of the software you need to install:
Install Docker CE for Windows for Windows build hosts that meet requirements.
Install Docker CE for MacOs for Mac build hosts that meet requirements.
Install Docker Toolbox for Windows for Windows build hosts that do not meet Docker requirements.
Install Docker Toolbox for MacOS for Mac build hosts that do not meet Docker requirements.
Install Docker CE for CentOS for Linux build hosts running the CentOS distribution.
Install Docker CE for Debian for Linux build hosts running the Debian distribution.
Install Docker CE for Fedora for Linux build hosts running the Fedora distribution.
Install Docker CE for Ubuntu for Linux build hosts running the Ubuntu distribution.
Optionally Orient Yourself With Docker: If you are unfamiliar with Docker and the container concept, you can learn more here - https://docs.docker.com/get-started/.
Launch Docker or Docker Toolbox: You should be able to launch Docker or the Docker Toolbox and have a terminal shell on your development host.
Set Up the Containers to Use the Yocto Project: Go to https://github.com/crops/docker-win-mac-docs/wiki and follow the directions for your particular build host (i.e. Linux, Mac, or Windows).
Once you complete the setup instructions for your machine, you have the Poky, Extensible SDK, and Toaster containers available. You can click those links from the page and learn more about using each of those containers.
Once you have a container set up, everything is in place to develop just as if you were running on a native Linux machine. If you are going to use the Poky container, see the “Cloning the poky Repository” section. If you are going to use the Extensible SDK container, see the “Using the Extensible SDK” Chapter in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual. If you are going to use the Toaster container, see the “Setting Up and Using Toaster” section in the Toaster User Manual.
2.2.3 Setting Up to Use Windows Subsystem For Linux (WSLv2)
With Windows Subsystem for Linux (WSLv2), you can create a Yocto Project development environment that allows you to build on Windows. You can set up a Linux distribution inside Windows in which you can develop using the Yocto Project.
Follow these general steps to prepare a Windows machine using WSLv2 as your Yocto Project build host:
Make sure your Windows 10 machine is capable of running WSLv2: WSLv2 is only available for Windows 10 builds > 18917. To check which build version you are running, you may open a command prompt on Windows and execute the command “ver”.
C:\Users\myuser> ver Microsoft Windows [Version 10.0.19041.153]
If your build is capable of running WSLv2 you may continue, for more information on this subject or instructions on how to upgrade to WSLv2 visit Windows 10 WSLv2
Install the Linux distribution of your choice inside Windows 10: Once you know your version of Windows 10 supports WSLv2, you can install the distribution of your choice from the Microsoft Store. Open the Microsoft Store and search for Linux. While there are several Linux distributions available, the assumption is that your pick will be one of the distributions supported by the Yocto Project as stated on the instructions for using a native Linux host. After making your selection, simply click “Get” to download and install the distribution.
Check your Linux distribution is using WSLv2: Open a Windows PowerShell and run:
C:\WINDOWS\system32> wsl -l -v NAME STATE VERSION *Ubuntu Running 2
Note the version column which says the WSL version being used by your distribution, on compatible systems, this can be changed back at any point in time.
Optionally Orient Yourself on WSL: If you are unfamiliar with WSL, you can learn more here - https://docs.microsoft.com/en-us/windows/wsl/wsl2-about.
Launch your WSL Distibution: From the Windows start menu simply launch your WSL distribution just like any other application.
Optimize your WSLv2 storage often: Due to the way storage is handled on WSLv2, the storage space used by the undelying Linux distribution is not reflected immedately, and since bitbake heavily uses storage, after several builds, you may be unaware you are running out of space. WSLv2 uses a VHDX file for storage, this issue can be easily avoided by manually optimizing this file often, this can be done in the following way:
Find the location of your VHDX file: First you need to find the distro app package directory, to achieve this open a Windows Powershell as Administrator and run:
C:\WINDOWS\system32> Get-AppxPackage -Name "*Ubuntu*" | Select PackageFamilyName PackageFamilyName ----------------- CanonicalGroupLimited.UbuntuonWindows_79abcdefgh
You should now replace the PackageFamilyName and your user on the following path to find your VHDX file:
ls C:\Users\myuser\AppData\Local\Packages\CanonicalGroupLimited.UbuntuonWindows_79abcdefgh\LocalState\ Mode LastWriteTime Length Name -a---- 3/14/2020 9:52 PM 57418973184 ext4.vhdx
Your VHDX file path is:
C:\Users\myuser\AppData\Local\Packages\CanonicalGroupLimited.UbuntuonWindows_79abcdefgh\LocalState\ext4.vhdx
Optimize your VHDX file: Open a Windows Powershell as Administrator to optimize your VHDX file, shutting down WSL first:
C:\WINDOWS\system32> wsl --shutdown C:\WINDOWS\system32> optimize-vhd -Path C:\Users\myuser\AppData\Local\Packages\CanonicalGroupLimited.UbuntuonWindows_79abcdefgh\LocalState\ext4.vhdx -Mode full
A progress bar should be shown while optimizing the VHDX file, and storage should now be reflected correctly on the Windows Explorer.
Note
The current implementation of WSLv2 does not have out-of-the-box
access to external devices such as those connected through a USB
port, but it automatically mounts your C:
drive on /mnt/c/
(and others), which you can use to share deploy artifacts to be later
flashed on hardware through Windows, but your build directory should
not reside inside this mountpoint.
Once you have WSLv2 set up, everything is in place to develop just as if you were running on a native Linux machine. If you are going to use the Extensible SDK container, see the “Using the Extensible SDK” Chapter in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual. If you are going to use the Toaster container, see the “Setting Up and Using Toaster” section in the Toaster User Manual.
2.3 Locating Yocto Project Source Files
This section shows you how to locate, fetch and configure the source files you’ll need to work with the Yocto Project.
Note
For concepts and introductory information about Git as it is used in the Yocto Project, see the “Git” section in the Yocto Project Overview and Concepts Manual.
For concepts on Yocto Project source repositories, see the “Yocto Project Source Repositories” section in the Yocto Project Overview and Concepts Manual.”
2.3.1 Accessing Source Repositories
Working from a copy of the upstream Accessing Source Repositories is the
preferred method for obtaining and using a Yocto Project release. You
can view the Yocto Project Source Repositories at
https://git.yoctoproject.org/cgit/cgit.cgi/. In particular, you can find the poky
repository at https://git.yoctoproject.org/cgit/cgit.cgi/poky.
Use the following procedure to locate the latest upstream copy of the
poky
Git repository:
Access Repositories: Open a browser and go to https://git.yoctoproject.org/cgit/cgit.cgi/ to access the GUI-based interface into the Yocto Project source repositories.
Select the Repository: Click on the repository in which you are interested (e.g.
poky
).Find the URL Used to Clone the Repository: At the bottom of the page, note the URL used to clone that repository (e.g. https://git.yoctoproject.org/cgit/cgit.cgi/poky).
Note
For information on cloning a repository, see the “Cloning the poky Repository” section.
2.3.2 Accessing Index of Releases
Yocto Project maintains an Index of Releases area that contains related files that contribute to the Yocto Project. Rather than Git repositories, these files are tarballs that represent snapshots in time of a given component.
Note
The recommended method for accessing Yocto Project components is to use Git to clone the upstream repository and work from within that locally cloned repository. The procedure in this section exists should you desire a tarball snapshot of any given component.
Follow these steps to locate and download a particular tarball:
Access the Index of Releases: Open a browser and go to Index of Releases. The list represents released components (e.g.
bitbake
,sato
, and so on).Note
The
yocto
directory contains the full array of released Poky tarballs. Thepoky
directory in the Index of Releases was historically used for very early releases and exists now only for retroactive completeness.Select a Component: Click on any released component in which you are interested (e.g.
yocto
).Find the Tarball: Drill down to find the associated tarball. For example, click on
yocto-3.3.1
to view files associated with the Yocto Project 3.3.1 release (e.g.poky-hardknott-25.0.1.tar.bz2
, which is the released Poky tarball).Download the Tarball: Click the tarball to download and save a snapshot of the given component.
2.3.3 Using the Downloads Page
The Yocto Project Website uses a “DOWNLOADS” page from which you can locate and download tarballs of any Yocto Project release. Rather than Git repositories, these files represent snapshot tarballs similar to the tarballs located in the Index of Releases described in the “Accessing Index of Releases” section.
Note
The recommended method for accessing Yocto Project components is to use Git to clone a repository and work from within that local repository. The procedure in this section exists should you desire a tarball snapshot of any given component.
Go to the Yocto Project Website: Open The Yocto Project Website in your browser.
Get to the Downloads Area: Select the “DOWNLOADS” item from the pull-down “SOFTWARE” tab menu near the top of the page.
Select a Yocto Project Release: Use the menu next to “RELEASE” to display and choose a recent or past supported Yocto Project release (e.g. hardknott, gatesgarth, and so forth).
Note
For a “map” of Yocto Project releases to version numbers, see the Releases wiki page.
You can use the “RELEASE ARCHIVE” link to reveal a menu of all Yocto Project releases.
Download Tools or Board Support Packages (BSPs): From the “DOWNLOADS” page, you can download tools or BSPs as well. Just scroll down the page and look for what you need.
2.3.4 Accessing Nightly Builds
Yocto Project maintains an area for nightly builds that contains tarball releases at https://autobuilder.yocto.io//pub/nightly/. These builds include Yocto Project releases (“poky”), toolchains, and builds for supported machines.
Should you ever want to access a nightly build of a particular Yocto Project component, use the following procedure:
Locate the Index of Nightly Builds: Open a browser and go to https://autobuilder.yocto.io//pub/nightly/ to access the Nightly Builds.
Select a Date: Click on the date in which you are interested. If you want the latest builds, use “CURRENT”.
Select a Build: Choose the area in which you are interested. For example, if you are looking for the most recent toolchains, select the “toolchain” link.
Find the Tarball: Drill down to find the associated tarball.
Download the Tarball: Click the tarball to download and save a snapshot of the given component.
2.4 Cloning and Checking Out Branches
To use the Yocto Project for development, you need a release locally installed on your development system. This locally installed set of files is referred to as the Source Directory in the Yocto Project documentation.
The preferred method of creating your Source Directory is by using
Git to clone a local copy of the upstream
poky
repository. Working from a cloned copy of the upstream
repository allows you to contribute back into the Yocto Project or to
simply work with the latest software on a development branch. Because
Git maintains and creates an upstream repository with a complete history
of changes and you are working with a local clone of that repository,
you have access to all the Yocto Project development branches and tag
names used in the upstream repository.
2.4.1 Cloning the poky
Repository
Follow these steps to create a local version of the upstream Poky Git repository.
Set Your Directory: Change your working directory to where you want to create your local copy of
poky
.Clone the Repository: The following example command clones the
poky
repository and uses the default name “poky” for your local repository:$ git clone git://git.yoctoproject.org/poky Cloning into 'poky'... remote: Counting objects: 432160, done. remote: Compressing objects: 100% (102056/102056), done. remote: Total 432160 (delta 323116), reused 432037 (delta 323000) Receiving objects: 100% (432160/432160), 153.81 MiB | 8.54 MiB/s, done. Resolving deltas: 100% (323116/323116), done. Checking connectivity... done.
Unless you specify a specific development branch or tag name, Git clones the “master” branch, which results in a snapshot of the latest development changes for “master”. For information on how to check out a specific development branch or on how to check out a local branch based on a tag name, see the “Checking Out by Branch in Poky” and “Checking Out by Tag in Poky” sections, respectively.
Once the local repository is created, you can change to that directory and check its status. Here, the single “master” branch exists on your system and by default, it is checked out:
$ cd poky $ git status On branch master Your branch is up-to-date with 'origin/master'. nothing to commit, working directory clean $ git branch * master
Your local repository of poky is identical to the upstream poky repository at the time from which it was cloned. As you work with the local branch, you can periodically use the
git pull --rebase
command to be sure you are up-to-date with the upstream branch.
2.4.2 Checking Out by Branch in Poky
When you clone the upstream poky repository, you have access to all its development branches. Each development branch in a repository is unique as it forks off the “master” branch. To see and use the files of a particular development branch locally, you need to know the branch name and then specifically check out that development branch.
Note
Checking out an active development branch by branch name gives you a snapshot of that particular branch at the time you check it out. Further development on top of the branch that occurs after check it out can occur.
Switch to the Poky Directory: If you have a local poky Git repository, switch to that directory. If you do not have the local copy of poky, see the “Cloning the poky Repository” section.
Determine Existing Branch Names:
$ git branch -a * master remotes/origin/1.1_M1 remotes/origin/1.1_M2 remotes/origin/1.1_M3 remotes/origin/1.1_M4 remotes/origin/1.2_M1 remotes/origin/1.2_M2 remotes/origin/1.2_M3 . . . remotes/origin/thud remotes/origin/thud-next remotes/origin/warrior remotes/origin/warrior-next remotes/origin/zeus remotes/origin/zeus-next ... and so on ...
Check out the Branch: Check out the development branch in which you want to work. For example, to access the files for the Yocto Project 3.3.1 Release (Hardknott), use the following command:
$ git checkout -b hardknott origin/hardknott Branch hardknott set up to track remote branch hardknott from origin. Switched to a new branch 'hardknott'
The previous command checks out the “hardknott” development branch and reports that the branch is tracking the upstream “origin/hardknott” branch.
The following command displays the branches that are now part of your local poky repository. The asterisk character indicates the branch that is currently checked out for work:
$ git branch master * hardknott
2.4.3 Checking Out by Tag in Poky
Similar to branches, the upstream repository uses tags to mark specific commits associated with significant points in a development branch (i.e. a release point or stage of a release). You might want to set up a local branch based on one of those points in the repository. The process is similar to checking out by branch name except you use tag names.
Note
Checking out a branch based on a tag gives you a stable set of files not affected by development on the branch above the tag.
Switch to the Poky Directory: If you have a local poky Git repository, switch to that directory. If you do not have the local copy of poky, see the “Cloning the poky Repository” section.
Fetch the Tag Names: To checkout the branch based on a tag name, you need to fetch the upstream tags into your local repository:
$ git fetch --tags $
List the Tag Names: You can list the tag names now:
$ git tag 1.1_M1.final 1.1_M1.rc1 1.1_M1.rc2 1.1_M2.final 1.1_M2.rc1 . . . yocto-2.5 yocto-2.5.1 yocto-2.5.2 yocto-2.5.3 yocto-2.6 yocto-2.6.1 yocto-2.6.2 yocto-2.7 yocto_1.5_M5.rc8
Check out the Branch:
$ git checkout tags/yocto-3.3.1 -b my_yocto_3.3.1 Switched to a new branch 'my_yocto_3.3.1' $ git branch master * my_yocto_3.3.1
The previous command creates and checks out a local branch named “my_yocto_3.3.1”, which is based on the commit in the upstream poky repository that has the same tag. In this example, the files you have available locally as a result of the
checkout
command are a snapshot of the “hardknott” development branch at the point where Yocto Project 3.3.1 was released.
3 Common Tasks
This chapter describes fundamental procedures such as creating layers, adding new software packages, extending or customizing images, porting work to new hardware (adding a new machine), and so forth. You will find that the procedures documented here occur often in the development cycle using the Yocto Project.
3.1 Understanding and Creating Layers
The OpenEmbedded build system supports organizing Metadata into multiple layers. Layers allow you to isolate different types of customizations from each other. For introductory information on the Yocto Project Layer Model, see the “The Yocto Project Layer Model” section in the Yocto Project Overview and Concepts Manual.
3.1.1 Creating Your Own Layer
It is very easy to create your own layers to use with the OpenEmbedded build system. The Yocto Project ships with tools that speed up creating layers. This section describes the steps you perform by hand to create layers so that you can better understand them. For information about the layer-creation tools, see the “Creating a new BSP Layer Using the bitbake-layers Script” section in the Yocto Project Board Support Package (BSP) Developer’s Guide and the “Creating a General Layer Using the bitbake-layers Script” section further down in this manual.
Follow these general steps to create your layer without using tools:
Check Existing Layers: Before creating a new layer, you should be sure someone has not already created a layer containing the Metadata you need. You can see the OpenEmbedded Metadata Index for a list of layers from the OpenEmbedded community that can be used in the Yocto Project. You could find a layer that is identical or close to what you need.
Create a Directory: Create the directory for your layer. When you create the layer, be sure to create the directory in an area not associated with the Yocto Project Source Directory (e.g. the cloned
poky
repository).While not strictly required, prepend the name of the directory with the string “meta-”. For example:
meta-mylayer meta-GUI_xyz meta-mymachine
With rare exceptions, a layer’s name follows this form:
meta-root_name
Following this layer naming convention can save you trouble later when tools, components, or variables “assume” your layer name begins with “meta-”. A notable example is in configuration files as shown in the following step where layer names without the “meta-” string are appended to several variables used in the configuration.
Create a Layer Configuration File: Inside your new layer folder, you need to create a
conf/layer.conf
file. It is easiest to take an existing layer configuration file and copy that to your layer’sconf
directory and then modify the file as needed.The
meta-yocto-bsp/conf/layer.conf
file in the Yocto Project Source Repositories demonstrates the required syntax. For your layer, you need to replace “yoctobsp” with a unique identifier for your layer (e.g. “machinexyz” for a layer named “meta-machinexyz”):# We have a conf and classes directory, add to BBPATH BBPATH .= ":${LAYERDIR}" # We have recipes-* directories, add to BBFILES BBFILES += "${LAYERDIR}/recipes-*/*/*.bb \ ${LAYERDIR}/recipes-*/*/*.bbappend" BBFILE_COLLECTIONS += "yoctobsp" BBFILE_PATTERN_yoctobsp = "^${LAYERDIR}/" BBFILE_PRIORITY_yoctobsp = "5" LAYERVERSION_yoctobsp = "4" LAYERSERIES_COMPAT_yoctobsp = "dunfell"
Following is an explanation of the layer configuration file:
BBPATH: Adds the layer’s root directory to BitBake’s search path. Through the use of the
BBPATH
variable, BitBake locates class files (.bbclass
), configuration files, and files that are included withinclude
andrequire
statements. For these cases, BitBake uses the first file that matches the name found inBBPATH
. This is similar to the way thePATH
variable is used for binaries. It is recommended, therefore, that you use unique class and configuration filenames in your custom layer.BBFILES: Defines the location for all recipes in the layer.
BBFILE_COLLECTIONS: Establishes the current layer through a unique identifier that is used throughout the OpenEmbedded build system to refer to the layer. In this example, the identifier “yoctobsp” is the representation for the container layer named “meta-yocto-bsp”.
BBFILE_PATTERN: Expands immediately during parsing to provide the directory of the layer.
BBFILE_PRIORITY: Establishes a priority to use for recipes in the layer when the OpenEmbedded build finds recipes of the same name in different layers.
LAYERVERSION: Establishes a version number for the layer. You can use this version number to specify this exact version of the layer as a dependency when using the LAYERDEPENDS variable.
LAYERDEPENDS: Lists all layers on which this layer depends (if any).
LAYERSERIES_COMPAT: Lists the Yocto Project releases for which the current version is compatible. This variable is a good way to indicate if your particular layer is current.
Add Content: Depending on the type of layer, add the content. If the layer adds support for a machine, add the machine configuration in a
conf/machine/
file within the layer. If the layer adds distro policy, add the distro configuration in aconf/distro/
file within the layer. If the layer introduces new recipes, put the recipes you need inrecipes-*
subdirectories within the layer.Note
For an explanation of layer hierarchy that is compliant with the Yocto Project, see the “Example Filesystem Layout” section in the Yocto Project Board Support Package (BSP) Developer’s Guide.
Optionally Test for Compatibility: If you want permission to use the Yocto Project Compatibility logo with your layer or application that uses your layer, perform the steps to apply for compatibility. See the “Making Sure Your Layer is Compatible With Yocto Project” section for more information.
3.1.2 Following Best Practices When Creating Layers
To create layers that are easier to maintain and that will not impact builds for other machines, you should consider the information in the following list:
Avoid “Overlaying” Entire Recipes from Other Layers in Your Configuration: In other words, do not copy an entire recipe into your layer and then modify it. Rather, use an append file (
.bbappend
) to override only those parts of the original recipe you need to modify.Avoid Duplicating Include Files: Use append files (
.bbappend
) for each recipe that uses an include file. Or, if you are introducing a new recipe that requires the included file, use the path relative to the original layer directory to refer to the file. For example, userequire recipes-core/
package/
file.inc
instead ofrequire
file.inc
. If you’re finding you have to overlay the include file, it could indicate a deficiency in the include file in the layer to which it originally belongs. If this is the case, you should try to address that deficiency instead of overlaying the include file. For example, you could address this by getting the maintainer of the include file to add a variable or variables to make it easy to override the parts needing to be overridden.Structure Your Layers: Proper use of overrides within append files and placement of machine-specific files within your layer can ensure that a build is not using the wrong Metadata and negatively impacting a build for a different machine. Following are some examples:
Modify Variables to Support a Different Machine: Suppose you have a layer named
meta-one
that adds support for building machine “one”. To do so, you use an append file namedbase-files.bbappend
and create a dependency on “foo” by altering the DEPENDS variable:DEPENDS = "foo"
The dependency is created during any build that includes the layer
meta-one
. However, you might not want this dependency for all machines. For example, suppose you are building for machine “two” but yourbblayers.conf
file has themeta-one
layer included. During the build, thebase-files
for machine “two” will also have the dependency onfoo
.To make sure your changes apply only when building machine “one”, use a machine override with the
DEPENDS
statement:DEPENDS_one = "foo"
You should follow the same strategy when using
_append
and_prepend
operations:DEPENDS_append_one = " foo" DEPENDS_prepend_one = "foo "
As an actual example, here’s a snippet from the generic kernel include file
linux-yocto.inc
, wherein the kernel compile and link options are adjusted in the case of a subset of the supported architectures:DEPENDS_append_aarch64 = " libgcc" KERNEL_CC_append_aarch64 = " ${TOOLCHAIN_OPTIONS}" KERNEL_LD_append_aarch64 = " ${TOOLCHAIN_OPTIONS}" DEPENDS_append_nios2 = " libgcc" KERNEL_CC_append_nios2 = " ${TOOLCHAIN_OPTIONS}" KERNEL_LD_append_nios2 = " ${TOOLCHAIN_OPTIONS}" DEPENDS_append_arc = " libgcc" KERNEL_CC_append_arc = " ${TOOLCHAIN_OPTIONS}" KERNEL_LD_append_arc = " ${TOOLCHAIN_OPTIONS}" KERNEL_FEATURES_append_qemuall=" features/debug/printk.scc"
Note
Avoiding “+=” and “=+” and using machine-specific
_append
and_prepend
operations is recommended as well.Place Machine-Specific Files in Machine-Specific Locations: When you have a base recipe, such as
base-files.bb
, that contains a SRC_URI statement to a file, you can use an append file to cause the build to use your own version of the file. For example, an append file in your layer atmeta-one/recipes-core/base-files/base-files.bbappend
could extend FILESPATH using FILESEXTRAPATHS as follows:FILESEXTRAPATHS_prepend := "${THISDIR}/${BPN}:"
The build for machine “one” will pick up your machine-specific file as long as you have the file in
meta-one/recipes-core/base-files/base-files/
. However, if you are building for a different machine and thebblayers.conf
file includes themeta-one
layer and the location of your machine-specific file is the first location where that file is found according toFILESPATH
, builds for all machines will also use that machine-specific file.You can make sure that a machine-specific file is used for a particular machine by putting the file in a subdirectory specific to the machine. For example, rather than placing the file in
meta-one/recipes-core/base-files/base-files/
as shown above, put it inmeta-one/recipes-core/base-files/base-files/one/
. Not only does this make sure the file is used only when building for machine “one”, but the build process locates the file more quickly.In summary, you need to place all files referenced from
SRC_URI
in a machine-specific subdirectory within the layer in order to restrict those files to machine-specific builds.
Perform Steps to Apply for Yocto Project Compatibility: If you want permission to use the Yocto Project Compatibility logo with your layer or application that uses your layer, perform the steps to apply for compatibility. See the “Making Sure Your Layer is Compatible With Yocto Project” section for more information.
Follow the Layer Naming Convention: Store custom layers in a Git repository that use the
meta-layer_name
format.Group Your Layers Locally: Clone your repository alongside other cloned
meta
directories from the Source Directory.
3.1.3 Making Sure Your Layer is Compatible With Yocto Project
When you create a layer used with the Yocto Project, it is advantageous to make sure that the layer interacts well with existing Yocto Project layers (i.e. the layer is compatible with the Yocto Project). Ensuring compatibility makes the layer easy to be consumed by others in the Yocto Project community and could allow you permission to use the Yocto Project Compatible Logo.
Note
Only Yocto Project member organizations are permitted to use the Yocto Project Compatible Logo. The logo is not available for general use. For information on how to become a Yocto Project member organization, see the Yocto Project Website.
The Yocto Project Compatibility Program consists of a layer application process that requests permission to use the Yocto Project Compatibility Logo for your layer and application. The process consists of two parts:
Successfully passing a script (
yocto-check-layer
) that when run against your layer, tests it against constraints based on experiences of how layers have worked in the real world and where pitfalls have been found. Getting a “PASS” result from the script is required for successful compatibility registration.Completion of an application acceptance form, which you can find at https://www.yoctoproject.org/webform/yocto-project-compatible-registration.
To be granted permission to use the logo, you need to satisfy the following:
Be able to check the box indicating that you got a “PASS” when running the script against your layer.
Answer “Yes” to the questions on the form or have an acceptable explanation for any questions answered “No”.
Be a Yocto Project Member Organization.
The remainder of this section presents information on the registration
form and on the yocto-check-layer
script.
3.1.3.1 Yocto Project Compatible Program Application
Use the form to apply for your layer’s approval. Upon successful application, you can use the Yocto Project Compatibility Logo with your layer and the application that uses your layer.
To access the form, use this link: https://www.yoctoproject.org/webform/yocto-project-compatible-registration. Follow the instructions on the form to complete your application.
The application consists of the following sections:
Contact Information: Provide your contact information as the fields require. Along with your information, provide the released versions of the Yocto Project for which your layer is compatible.
Acceptance Criteria: Provide “Yes” or “No” answers for each of the items in the checklist. Space exists at the bottom of the form for any explanations for items for which you answered “No”.
Recommendations: Provide answers for the questions regarding Linux kernel use and build success.
3.1.3.2 yocto-check-layer
Script
The yocto-check-layer
script provides you a way to assess how
compatible your layer is with the Yocto Project. You should run this
script prior to using the form to apply for compatibility as described
in the previous section. You need to achieve a “PASS” result in order to
have your application form successfully processed.
The script divides tests into three areas: COMMON, BSP, and DISTRO. For example, given a distribution layer (DISTRO), the layer must pass both the COMMON and DISTRO related tests. Furthermore, if your layer is a BSP layer, the layer must pass the COMMON and BSP set of tests.
To execute the script, enter the following commands from your build directory:
$ source oe-init-build-env
$ yocto-check-layer your_layer_directory
Be sure to provide the actual directory for your layer as part of the command.
Entering the command causes the script to determine the type of layer and then to execute a set of specific tests against the layer. The following list overviews the test:
common.test_readme
: Tests if aREADME
file exists in the layer and the file is not empty.common.test_parse
: Tests to make sure that BitBake can parse the files without error (i.e.bitbake -p
).common.test_show_environment
: Tests that the global or per-recipe environment is in order without errors (i.e.bitbake -e
).common.test_world
: Verifies thatbitbake world
works.common.test_signatures
: Tests to be sure that BSP and DISTRO layers do not come with recipes that change signatures.common.test_layerseries_compat
: Verifies layer compatibility is set properly.bsp.test_bsp_defines_machines
: Tests if a BSP layer has machine configurations.bsp.test_bsp_no_set_machine
: Tests to ensure a BSP layer does not set the machine when the layer is added.bsp.test_machine_world
: Verifies thatbitbake world
works regardless of which machine is selected.bsp.test_machine_signatures
: Verifies that building for a particular machine affects only the signature of tasks specific to that machine.distro.test_distro_defines_distros
: Tests if a DISTRO layer has distro configurations.distro.test_distro_no_set_distros
: Tests to ensure a DISTRO layer does not set the distribution when the layer is added.
3.1.4 Enabling Your Layer
Before the OpenEmbedded build system can use your new layer, you need to
enable it. To enable your layer, simply add your layer’s path to the
BBLAYERS
variable in your conf/bblayers.conf
file, which is
found in the Build Directory.
The following example shows how to enable a layer named
meta-mylayer
:
# POKY_BBLAYERS_CONF_VERSION is increased each time build/conf/bblayers.conf
# changes incompatibly
POKY_BBLAYERS_CONF_VERSION = "2"
BBPATH = "${TOPDIR}"
BBFILES ?= ""
BBLAYERS ?= " \
/home/user/poky/meta \
/home/user/poky/meta-poky \
/home/user/poky/meta-yocto-bsp \
/home/user/poky/meta-mylayer \
"
BitBake parses each conf/layer.conf
file from the top down as
specified in the BBLAYERS
variable within the conf/bblayers.conf
file. During the processing of each conf/layer.conf
file, BitBake
adds the recipes, classes and configurations contained within the
particular layer to the source directory.
3.1.5 Using .bbappend Files in Your Layer
A recipe that appends Metadata to another recipe is called a BitBake
append file. A BitBake append file uses the .bbappend
file type
suffix, while the corresponding recipe to which Metadata is being
appended uses the .bb
file type suffix.
You can use a .bbappend
file in your layer to make additions or
changes to the content of another layer’s recipe without having to copy
the other layer’s recipe into your layer. Your .bbappend
file
resides in your layer, while the main .bb
recipe file to which you
are appending Metadata resides in a different layer.
Being able to append information to an existing recipe not only avoids duplication, but also automatically applies recipe changes from a different layer into your layer. If you were copying recipes, you would have to manually merge changes as they occur.
When you create an append file, you must use the same root name as the
corresponding recipe file. For example, the append file
someapp_3.1.bbappend
must apply to someapp_3.1.bb
. This
means the original recipe and append file names are version
number-specific. If the corresponding recipe is renamed to update to a
newer version, you must also rename and possibly update the
corresponding .bbappend
as well. During the build process, BitBake
displays an error on starting if it detects a .bbappend
file that
does not have a corresponding recipe with a matching name. See the
BB_DANGLINGAPPENDS_WARNONLY
variable for information on how to handle this error.
As an example, consider the main formfactor recipe and a corresponding
formfactor append file both from the Source Directory.
Here is the main
formfactor recipe, which is named formfactor_0.0.bb
and located in
the “meta” layer at meta/recipes-bsp/formfactor
:
SUMMARY = "Device formfactor information"
DESCRIPTION = "A formfactor configuration file provides information about the \
target hardware for which the image is being built and information that the \
build system cannot obtain from other sources such as the kernel."
SECTION = "base"
LICENSE = "MIT"
LIC_FILES_CHKSUM = "file://${COREBASE}/meta/COPYING.MIT;md5=3da9cfbcb788c80a0384361b4de20420"
PR = "r45"
SRC_URI = "file://config file://machconfig"
S = "${WORKDIR}"
PACKAGE_ARCH = "${MACHINE_ARCH}"
INHIBIT_DEFAULT_DEPS = "1"
do_install() {
# Install file only if it has contents
install -d ${D}${sysconfdir}/formfactor/
install -m 0644 ${S}/config ${D}${sysconfdir}/formfactor/
if [ -s "${S}/machconfig" ]; then
install -m 0644 ${S}/machconfig ${D}${sysconfdir}/formfactor/
fi
}
In the main recipe, note the SRC_URI variable, which tells the OpenEmbedded build system where to find files during the build.
Following is the append file, which is named formfactor_0.0.bbappend
and is from the Raspberry Pi BSP Layer named meta-raspberrypi
. The
file is in the layer at recipes-bsp/formfactor
:
FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:"
By default, the build system uses the
FILESPATH variable to
locate files. This append file extends the locations by setting the
FILESEXTRAPATHS
variable. Setting this variable in the .bbappend
file is the most
reliable and recommended method for adding directories to the search
path used by the build system to find files.
The statement in this example extends the directories to include
${
THISDIR}/${
PN}
,
which resolves to a directory named formfactor
in the same directory
in which the append file resides (i.e.
meta-raspberrypi/recipes-bsp/formfactor
. This implies that you must
have the supporting directory structure set up that will contain any
files or patches you will be including from the layer.
Using the immediate expansion assignment operator :=
is important
because of the reference to THISDIR
. The trailing colon character is
important as it ensures that items in the list remain colon-separated.
Note
BitBake automatically defines the THISDIR
variable. You should
never set this variable yourself. Using “_prepend” as part of the
FILESEXTRAPATHS
ensures your path will be searched prior to other
paths in the final list.
Also, not all append files add extra files. Many append files simply
exist to add build options (e.g. systemd
). For these cases, your
append file would not even use the FILESEXTRAPATHS
statement.
3.1.6 Prioritizing Your Layer
Each layer is assigned a priority value. Priority values control which
layer takes precedence if there are recipe files with the same name in
multiple layers. For these cases, the recipe file from the layer with a
higher priority number takes precedence. Priority values also affect the
order in which multiple .bbappend
files for the same recipe are
applied. You can either specify the priority manually, or allow the
build system to calculate it based on the layer’s dependencies.
To specify the layer’s priority manually, use the BBFILE_PRIORITY variable and append the layer’s root name:
BBFILE_PRIORITY_mylayer = "1"
Note
It is possible for a recipe with a lower version number PV in a layer that has a higher priority to take precedence.
Also, the layer priority does not currently affect the precedence
order of .conf
or .bbclass
files. Future versions of BitBake
might address this.
3.1.7 Managing Layers
You can use the BitBake layer management tool bitbake-layers
to
provide a view into the structure of recipes across a multi-layer
project. Being able to generate output that reports on configured layers
with their paths and priorities and on .bbappend
files and their
applicable recipes can help to reveal potential problems.
For help on the BitBake layer management tool, use the following command:
$ bitbake-layers --help
NOTE: Starting bitbake server...
usage: bitbake-layers [-d] [-q] [-F] [--color COLOR] [-h] <subcommand> ...
BitBake layers utility
optional arguments:
-d, --debug Enable debug output
-q, --quiet Print only errors
-F, --force Force add without recipe parse verification
--color COLOR Colorize output (where COLOR is auto, always, never)
-h, --help show this help message and exit
subcommands:
<subcommand>
layerindex-fetch Fetches a layer from a layer index along with its
dependent layers, and adds them to conf/bblayers.conf.
layerindex-show-depends
Find layer dependencies from layer index.
add-layer Add one or more layers to bblayers.conf.
remove-layer Remove one or more layers from bblayers.conf.
flatten flatten layer configuration into a separate output
directory.
show-layers show current configured layers.
show-overlayed list overlayed recipes (where the same recipe exists
in another layer)
show-recipes list available recipes, showing the layer they are
provided by
show-appends list bbappend files and recipe files they apply to
show-cross-depends Show dependencies between recipes that cross layer
boundaries.
create-layer Create a basic layer
Use bitbake-layers <subcommand> --help to get help on a specific command
The following list describes the available commands:
help:
Displays general help or help on a specified command.show-layers:
Shows the current configured layers.show-overlayed:
Lists overlayed recipes. A recipe is overlayed when a recipe with the same name exists in another layer that has a higher layer priority.show-recipes:
Lists available recipes and the layers that provide them.show-appends:
Lists.bbappend
files and the recipe files to which they apply.show-cross-depends:
Lists dependency relationships between recipes that cross layer boundaries.add-layer:
Adds a layer tobblayers.conf
.remove-layer:
Removes a layer frombblayers.conf
flatten:
Flattens the layer configuration into a separate output directory. Flattening your layer configuration builds a “flattened” directory that contains the contents of all layers, with any overlayed recipes removed and any.bbappend
files appended to the corresponding recipes. You might have to perform some manual cleanup of the flattened layer as follows:Non-recipe files (such as patches) are overwritten. The flatten command shows a warning for these files.
Anything beyond the normal layer setup has been added to the
layer.conf
file. Only the lowest priority layer’slayer.conf
is used.Overridden and appended items from
.bbappend
files need to be cleaned up. The contents of each.bbappend
end up in the flattened recipe. However, if there are appended or changed variable values, you need to tidy these up yourself. Consider the following example. Here, thebitbake-layers
command adds the line#### bbappended ...
so that you know where the following lines originate:... DESCRIPTION = "A useful utility" ... EXTRA_OECONF = "--enable-something" ... #### bbappended from meta-anotherlayer #### DESCRIPTION = "Customized utility" EXTRA_OECONF += "--enable-somethingelse"
Ideally, you would tidy up these utilities as follows:
... DESCRIPTION = "Customized utility" ... EXTRA_OECONF = "--enable-something --enable-somethingelse" ...
layerindex-fetch
: Fetches a layer from a layer index, along with its dependent layers, and adds the layers to theconf/bblayers.conf
file.layerindex-show-depends
: Finds layer dependencies from the layer index.create-layer
: Creates a basic layer.
3.1.8 Creating a General Layer Using the bitbake-layers
Script
The bitbake-layers
script with the create-layer
subcommand
simplifies creating a new general layer.
Note
For information on BSP layers, see the “BSP Layers” section in the Yocto Project Board Specific (BSP) Developer’s Guide.
In order to use a layer with the OpenEmbedded build system, you need to add the layer to your
bblayers.conf
configuration file. See the “Adding a Layer Using the bitbake-layers Script” section for more information.
The default mode of the script’s operation with this subcommand is to create a layer with the following:
A layer priority of 6.
A
conf
subdirectory that contains alayer.conf
file.A
recipes-example
subdirectory that contains a further subdirectory namedexample
, which contains anexample.bb
recipe file.A
COPYING.MIT
, which is the license statement for the layer. The script assumes you want to use the MIT license, which is typical for most layers, for the contents of the layer itself.A
README
file, which is a file describing the contents of your new layer.
In its simplest form, you can use the following command form to create a layer. The command creates a layer whose name corresponds to “your_layer_name” in the current directory:
$ bitbake-layers create-layer your_layer_name
As an example, the following command creates a layer named meta-scottrif
in your home directory:
$ cd /usr/home
$ bitbake-layers create-layer meta-scottrif
NOTE: Starting bitbake server...
Add your new layer with 'bitbake-layers add-layer meta-scottrif'
If you want to set the priority of the layer to other than the default
value of “6”, you can either use the --priority
option or you
can edit the
BBFILE_PRIORITY value
in the conf/layer.conf
after the script creates it. Furthermore, if
you want to give the example recipe file some name other than the
default, you can use the --example-recipe-name
option.
The easiest way to see how the bitbake-layers create-layer
command
works is to experiment with the script. You can also read the usage
information by entering the following:
$ bitbake-layers create-layer --help
NOTE: Starting bitbake server...
usage: bitbake-layers create-layer [-h] [--priority PRIORITY]
[--example-recipe-name EXAMPLERECIPE]
layerdir
Create a basic layer
positional arguments:
layerdir Layer directory to create
optional arguments:
-h, --help show this help message and exit
--priority PRIORITY, -p PRIORITY
Layer directory to create
--example-recipe-name EXAMPLERECIPE, -e EXAMPLERECIPE
Filename of the example recipe
3.1.9 Adding a Layer Using the bitbake-layers
Script
Once you create your general layer, you must add it to your
bblayers.conf
file. Adding the layer to this configuration file
makes the OpenEmbedded build system aware of your layer so that it can
search it for metadata.
Add your layer by using the bitbake-layers add-layer
command:
$ bitbake-layers add-layer your_layer_name
Here is an example that adds a
layer named meta-scottrif
to the configuration file. Following the
command that adds the layer is another bitbake-layers
command that
shows the layers that are in your bblayers.conf
file:
$ bitbake-layers add-layer meta-scottrif
NOTE: Starting bitbake server...
Parsing recipes: 100% |##########################################################| Time: 0:00:49
Parsing of 1441 .bb files complete (0 cached, 1441 parsed). 2055 targets, 56 skipped, 0 masked, 0 errors.
$ bitbake-layers show-layers
NOTE: Starting bitbake server...
layer path priority
==========================================================================
meta /home/scottrif/poky/meta 5
meta-poky /home/scottrif/poky/meta-poky 5
meta-yocto-bsp /home/scottrif/poky/meta-yocto-bsp 5
workspace /home/scottrif/poky/build/workspace 99
meta-scottrif /home/scottrif/poky/build/meta-scottrif 6
Adding the layer to this file enables the build system to locate the layer during the build.
Note
During a build, the OpenEmbedded build system looks in the layers from the top of the list down to the bottom in that order.
3.2 Customizing Images
You can customize images to satisfy particular requirements. This section describes several methods and provides guidelines for each.
3.2.1 Customizing Images Using local.conf
Probably the easiest way to customize an image is to add a package by
way of the local.conf
configuration file. Because it is limited to
local use, this method generally only allows you to add packages and is
not as flexible as creating your own customized image. When you add
packages using local variables this way, you need to realize that these
variable changes are in effect for every build and consequently affect
all images, which might not be what you require.
To add a package to your image using the local configuration file, use
the IMAGE_INSTALL
variable with the _append
operator:
IMAGE_INSTALL_append = " strace"
Use of the syntax is important -
specifically, the space between the quote and the package name, which is
strace
in this example. This space is required since the _append
operator does not add the space.
Furthermore, you must use _append
instead of the +=
operator if
you want to avoid ordering issues. The reason for this is because doing
so unconditionally appends to the variable and avoids ordering problems
due to the variable being set in image recipes and .bbclass
files
with operators like ?=
. Using _append
ensures the operation
takes effect.
As shown in its simplest use, IMAGE_INSTALL_append
affects all
images. It is possible to extend the syntax so that the variable applies
to a specific image only. Here is an example:
IMAGE_INSTALL_append_pn-core-image-minimal = " strace"
This example adds strace
to the core-image-minimal
image only.
You can add packages using a similar approach through the
CORE_IMAGE_EXTRA_INSTALL
variable. If you use this variable, only
core-image-*
images are affected.
3.2.2 Customizing Images Using Custom IMAGE_FEATURES
and EXTRA_IMAGE_FEATURES
Another method for customizing your image is to enable or disable
high-level image features by using the
IMAGE_FEATURES and
EXTRA_IMAGE_FEATURES
variables. Although the functions for both variables are nearly
equivalent, best practices dictate using IMAGE_FEATURES
from within
a recipe and using EXTRA_IMAGE_FEATURES
from within your
local.conf
file, which is found in the
Build Directory.
To understand how these features work, the best reference is
meta/classes/core-image.bbclass
. This class lists out the available
IMAGE_FEATURES
of which most map to package groups while some, such
as debug-tweaks
and read-only-rootfs
, resolve as general
configuration settings.
In summary, the file looks at the contents of the IMAGE_FEATURES
variable and then maps or configures the feature accordingly. Based on
this information, the build system automatically adds the appropriate
packages or configurations to the
IMAGE_INSTALL variable.
Effectively, you are enabling extra features by extending the class or
creating a custom class for use with specialized image .bb
files.
Use the EXTRA_IMAGE_FEATURES
variable from within your local
configuration file. Using a separate area from which to enable features
with this variable helps you avoid overwriting the features in the image
recipe that are enabled with IMAGE_FEATURES
. The value of
EXTRA_IMAGE_FEATURES
is added to IMAGE_FEATURES
within
meta/conf/bitbake.conf
.
To illustrate how you can use these variables to modify your image,
consider an example that selects the SSH server. The Yocto Project ships
with two SSH servers you can use with your images: Dropbear and OpenSSH.
Dropbear is a minimal SSH server appropriate for resource-constrained
environments, while OpenSSH is a well-known standard SSH server
implementation. By default, the core-image-sato
image is configured
to use Dropbear. The core-image-full-cmdline
and core-image-lsb
images both include OpenSSH. The core-image-minimal
image does not
contain an SSH server.
You can customize your image and change these defaults. Edit the
IMAGE_FEATURES
variable in your recipe or use the
EXTRA_IMAGE_FEATURES
in your local.conf
file so that it
configures the image you are working with to include
ssh-server-dropbear
or ssh-server-openssh
.
Note
See the “Image Features” section in the Yocto Project Reference Manual for a complete list of image features that ship with the Yocto Project.
3.2.3 Customizing Images Using Custom .bb Files
You can also customize an image by creating a custom recipe that defines additional software as part of the image. The following example shows the form for the two lines you need:
IMAGE_INSTALL = "packagegroup-core-x11-base package1 package2"
inherit core-image
Defining the software using a custom recipe gives you total control over
the contents of the image. It is important to use the correct names of
packages in the IMAGE_INSTALL
variable. You must use the
OpenEmbedded notation and not the Debian notation for the names (e.g.
glibc-dev
instead of libc6-dev
).
The other method for creating a custom image is to base it on an
existing image. For example, if you want to create an image based on
core-image-sato
but add the additional package strace
to the
image, copy the meta/recipes-sato/images/core-image-sato.bb
to a new
.bb
and add the following line to the end of the copy:
IMAGE_INSTALL += "strace"
3.2.4 Customizing Images Using Custom Package Groups
For complex custom images, the best approach for customizing an image is
to create a custom package group recipe that is used to build the image
or images. A good example of a package group recipe is
meta/recipes-core/packagegroups/packagegroup-base.bb
.
If you examine that recipe, you see that the PACKAGES
variable lists
the package group packages to produce. The inherit packagegroup
statement sets appropriate default values and automatically adds
-dev
, -dbg
, and -ptest
complementary packages for each
package specified in the PACKAGES
statement.
Note
The inherit packagegroup
line should be located near the top of the
recipe, certainly before the PACKAGES
statement.
For each package you specify in PACKAGES
, you can use RDEPENDS
and RRECOMMENDS
entries to provide a list of packages the parent
task package should contain. You can see examples of these further down
in the packagegroup-base.bb
recipe.
Here is a short, fabricated example showing the same basic pieces for a
hypothetical packagegroup defined in packagegroup-custom.bb
, where
the variable PN
is the standard way to abbreviate the reference to
the full packagegroup name packagegroup-custom
:
DESCRIPTION = "My Custom Package Groups"
inherit packagegroup
PACKAGES = "\
${PN}-apps \
${PN}-tools \
"
RDEPENDS_${PN}-apps = "\
dropbear \
portmap \
psplash"
RDEPENDS_${PN}-tools = "\
oprofile \
oprofileui-server \
lttng-tools"
RRECOMMENDS_${PN}-tools = "\
kernel-module-oprofile"
In the previous example, two package group packages are created with
their dependencies and their recommended package dependencies listed:
packagegroup-custom-apps
, and packagegroup-custom-tools
. To
build an image using these package group packages, you need to add
packagegroup-custom-apps
and/or packagegroup-custom-tools
to
IMAGE_INSTALL
. For other forms of image dependencies see the other
areas of this section.
3.2.5 Customizing an Image Hostname
By default, the configured hostname (i.e. /etc/hostname
) in an image
is the same as the machine name. For example, if
MACHINE equals “qemux86”, the
configured hostname written to /etc/hostname
is “qemux86”.
You can customize this name by altering the value of the “hostname”
variable in the base-files
recipe using either an append file or a
configuration file. Use the following in an append file:
hostname = "myhostname"
Use the following in a configuration file:
hostname_pn-base-files = "myhostname"
Changing the default value of the variable “hostname” can be useful in certain situations. For example, suppose you need to do extensive testing on an image and you would like to easily identify the image under test from existing images with typical default hostnames. In this situation, you could change the default hostname to “testme”, which results in all the images using the name “testme”. Once testing is complete and you do not need to rebuild the image for test any longer, you can easily reset the default hostname.
Another point of interest is that if you unset the variable, the image will have no default hostname in the filesystem. Here is an example that unsets the variable in a configuration file:
hostname_pn-base-files = ""
Having no default hostname in the filesystem is suitable for environments that use dynamic hostnames such as virtual machines.
3.3 Writing a New Recipe
Recipes (.bb
files) are fundamental components in the Yocto Project
environment. Each software component built by the OpenEmbedded build
system requires a recipe to define the component. This section describes
how to create, write, and test a new recipe.
Note
For information on variables that are useful for recipes and for information about recipe naming issues, see the “Recipes” section of the Yocto Project Reference Manual.
3.3.1 Overview
The following figure shows the basic process for creating a new recipe. The remainder of the section provides details for the steps.
3.3.2 Locate or Automatically Create a Base Recipe
You can always write a recipe from scratch. However, three choices exist that can help you quickly get a start on a new recipe:
devtool add
: A command that assists in creating a recipe and an environment conducive to development.recipetool create
: A command provided by the Yocto Project that automates creation of a base recipe based on the source files.Existing Recipes: Location and modification of an existing recipe that is similar in function to the recipe you need.
Note
For information on recipe syntax, see the “Recipe Syntax” section.
3.3.2.1 Creating the Base Recipe Using devtool add
The devtool add
command uses the same logic for auto-creating the
recipe as recipetool create
, which is listed below. Additionally,
however, devtool add
sets up an environment that makes it easy for
you to patch the source and to make changes to the recipe as is often
necessary when adding a recipe to build a new piece of software to be
included in a build.
You can find a complete description of the devtool add
command in
the “A Closer Look at devtool add” section
in the Yocto Project Application Development and the Extensible Software
Development Kit (eSDK) manual.
3.3.2.2 Creating the Base Recipe Using recipetool create
recipetool create
automates creation of a base recipe given a set of
source code files. As long as you can extract or point to the source
files, the tool will construct a recipe and automatically configure all
pre-build information into the recipe. For example, suppose you have an
application that builds using Autotools. Creating the base recipe using
recipetool
results in a recipe that has the pre-build dependencies,
license requirements, and checksums configured.
To run the tool, you just need to be in your Build Directory and have sourced the build environment setup script (i.e. oe-init-build-env). To get help on the tool, use the following command:
$ recipetool -h
NOTE: Starting bitbake server...
usage: recipetool [-d] [-q] [--color COLOR] [-h] <subcommand> ...
OpenEmbedded recipe tool
options:
-d, --debug Enable debug output
-q, --quiet Print only errors
--color COLOR Colorize output (where COLOR is auto, always, never)
-h, --help show this help message and exit
subcommands:
create Create a new recipe
newappend Create a bbappend for the specified target in the specified
layer
setvar Set a variable within a recipe
appendfile Create/update a bbappend to replace a target file
appendsrcfiles Create/update a bbappend to add or replace source files
appendsrcfile Create/update a bbappend to add or replace a source file
Use recipetool <subcommand> --help to get help on a specific command
Running recipetool create -o OUTFILE
creates the base recipe and
locates it properly in the layer that contains your source files.
Following are some syntax examples:
Use this syntax to generate a recipe based on source. Once generated, the recipe resides in the existing source code layer:
recipetool create -o OUTFILE sourceUse this syntax to generate a recipe using code that you extract from source. The extracted code is placed in its own layer defined by
EXTERNALSRC
.recipetool create -o OUTFILE -x EXTERNALSRC sourceUse this syntax to generate a recipe based on source. The options direct
recipetool
to generate debugging information. Once generated, the recipe resides in the existing source code layer:recipetool create -d -o OUTFILE source
3.3.2.3 Locating and Using a Similar Recipe
Before writing a recipe from scratch, it is often useful to discover whether someone else has already written one that meets (or comes close to meeting) your needs. The Yocto Project and OpenEmbedded communities maintain many recipes that might be candidates for what you are doing. You can find a good central index of these recipes in the OpenEmbedded Layer Index.
Working from an existing recipe or a skeleton recipe is the best way to get started. Here are some points on both methods:
Locate and modify a recipe that is close to what you want to do: This method works when you are familiar with the current recipe space. The method does not work so well for those new to the Yocto Project or writing recipes.
Some risks associated with this method are using a recipe that has areas totally unrelated to what you are trying to accomplish with your recipe, not recognizing areas of the recipe that you might have to add from scratch, and so forth. All these risks stem from unfamiliarity with the existing recipe space.
Use and modify the following skeleton recipe: If for some reason you do not want to use
recipetool
and you cannot find an existing recipe that is close to meeting your needs, you can use the following structure to provide the fundamental areas of a new recipe.DESCRIPTION = "" HOMEPAGE = "" LICENSE = "" SECTION = "" DEPENDS = "" LIC_FILES_CHKSUM = "" SRC_URI = ""
3.3.3 Storing and Naming the Recipe
Once you have your base recipe, you should put it in your own layer and name it appropriately. Locating it correctly ensures that the OpenEmbedded build system can find it when you use BitBake to process the recipe.
Storing Your Recipe: The OpenEmbedded build system locates your recipe through the layer’s
conf/layer.conf
file and the BBFILES variable. This variable sets up a path from which the build system can locate recipes. Here is the typical use:BBFILES += "${LAYERDIR}/recipes-*/*/*.bb \ ${LAYERDIR}/recipes-*/*/*.bbappend"
Consequently, you need to be sure you locate your new recipe inside your layer such that it can be found.
You can find more information on how layers are structured in the “Understanding and Creating Layers” section.
Naming Your Recipe: When you name your recipe, you need to follow this naming convention:
basename_version.bb
Use lower-cased characters and do not include the reserved suffixes
-native
,-cross
,-initial
, or-dev
casually (i.e. do not use them as part of your recipe name unless the string applies). Here are some examples:cups_1.7.0.bb gawk_4.0.2.bb irssi_0.8.16-rc1.bb
3.3.4 Running a Build on the Recipe
Creating a new recipe is usually an iterative process that requires using BitBake to process the recipe multiple times in order to progressively discover and add information to the recipe file.
Assuming you have sourced the build environment setup script (i.e.
oe-init-build-env) and you are in
the Build Directory, use
BitBake to process your recipe. All you need to provide is the
basename
of the recipe as described in the previous section:
$ bitbake basename
During the build, the OpenEmbedded build system creates a temporary work
directory for each recipe
(${
WORKDIR}
)
where it keeps extracted source files, log files, intermediate
compilation and packaging files, and so forth.
The path to the per-recipe temporary work directory depends on the context in which it is being built. The quickest way to find this path is to have BitBake return it by running the following:
$ bitbake -e basename | grep ^WORKDIR=
As an example, assume a Source Directory
top-level folder named poky
, a default Build Directory at
poky/build
, and a qemux86-poky-linux
machine target system.
Furthermore, suppose your recipe is named foo_1.3.0.bb
. In this
case, the work directory the build system uses to build the package
would be as follows:
poky/build/tmp/work/qemux86-poky-linux/foo/1.3.0-r0
Inside this directory you can find sub-directories such as image
,
packages-split
, and temp
. After the build, you can examine these
to determine how well the build went.
Note
You can find log files for each task in the recipe’s temp
directory (e.g. poky/build/tmp/work/qemux86-poky-linux/foo/1.3.0-r0/temp
).
Log files are named log.taskname
(e.g. log.do_configure
,
log.do_fetch
, and log.do_compile
).
You can find more information about the build process in “The Yocto Project Development Environment” chapter of the Yocto Project Overview and Concepts Manual.
3.3.5 Fetching Code
The first thing your recipe must do is specify how to fetch the source
files. Fetching is controlled mainly through the
SRC_URI variable. Your recipe
must have a SRC_URI
variable that points to where the source is
located. For a graphical representation of source locations, see the
“Sources” section in
the Yocto Project Overview and Concepts Manual.
The do_fetch task uses
the prefix of each entry in the SRC_URI
variable value to determine
which fetcher to use to get your
source files. It is the SRC_URI
variable that triggers the fetcher.
The do_patch task uses
the variable after source is fetched to apply patches. The OpenEmbedded
build system uses
FILESOVERRIDES for
scanning directory locations for local files in SRC_URI
.
The SRC_URI
variable in your recipe must define each unique location
for your source files. It is good practice to not hard-code version
numbers in a URL used in SRC_URI
. Rather than hard-code these
values, use ${
PV}
,
which causes the fetch process to use the version specified in the
recipe filename. Specifying the version in this manner means that
upgrading the recipe to a future version is as simple as renaming the
recipe to match the new version.
Here is a simple example from the
meta/recipes-devtools/strace/strace_5.5.bb
recipe where the source
comes from a single tarball. Notice the use of the
PV variable:
SRC_URI = "https://strace.io/files/${PV}/strace-${PV}.tar.xz \
Files mentioned in SRC_URI
whose names end in a typical archive
extension (e.g. .tar
, .tar.gz
, .tar.bz2
, .zip
, and so
forth), are automatically extracted during the
do_unpack task. For
another example that specifies these types of files, see the
“Autotooled Package” section.
Another way of specifying source is from an SCM. For Git repositories,
you must specify SRCREV and
you should specify PV to include
the revision with SRCPV. Here
is an example from the recipe
meta/recipes-kernel/blktrace/blktrace_git.bb
:
SRCREV = "d6918c8832793b4205ed3bfede78c2f915c23385"
PR = "r6"
PV = "1.0.5+git${SRCPV}"
SRC_URI = "git://git.kernel.dk/blktrace.git \
file://ldflags.patch"
If your SRC_URI
statement includes URLs pointing to individual files
fetched from a remote server other than a version control system,
BitBake attempts to verify the files against checksums defined in your
recipe to ensure they have not been tampered with or otherwise modified
since the recipe was written. Two checksums are used:
SRC_URI[md5sum]
and SRC_URI[sha256sum]
.
If your SRC_URI
variable points to more than a single URL (excluding
SCM URLs), you need to provide the md5
and sha256
checksums for
each URL. For these cases, you provide a name for each URL as part of
the SRC_URI
and then reference that name in the subsequent checksum
statements. Here is an example combining lines from the files
git.inc
and git_2.24.1.bb
:
SRC_URI = "${KERNELORG_MIRROR}/software/scm/git/git-${PV}.tar.gz;name=tarball \
${KERNELORG_MIRROR}/software/scm/git/git-manpages-${PV}.tar.gz;name=manpages"
SRC_URI[tarball.md5sum] = "166bde96adbbc11c8843d4f8f4f9811b"
SRC_URI[tarball.sha256sum] = "ad5334956301c86841eb1e5b1bb20884a6bad89a10a6762c958220c7cf64da02"
SRC_URI[manpages.md5sum] = "31c2272a8979022497ba3d4202df145d"
SRC_URI[manpages.sha256sum] = "9a7ae3a093bea39770eb96ca3e5b40bff7af0b9f6123f089d7821d0e5b8e1230"
Proper values for md5
and sha256
checksums might be available
with other signatures on the download page for the upstream source (e.g.
md5
, sha1
, sha256
, GPG
, and so forth). Because the
OpenEmbedded build system only deals with sha256sum
and md5sum
,
you should verify all the signatures you find by hand.
If no SRC_URI
checksums are specified when you attempt to build the
recipe, or you provide an incorrect checksum, the build will produce an
error for each missing or incorrect checksum. As part of the error
message, the build system provides the checksum string corresponding to
the fetched file. Once you have the correct checksums, you can copy and
paste them into your recipe and then run the build again to continue.
Note
As mentioned, if the upstream source provides signatures for verifying the downloaded source code, you should verify those manually before setting the checksum values in the recipe and continuing with the build.
This final example is a bit more complicated and is from the
meta/recipes-sato/rxvt-unicode/rxvt-unicode_9.20.bb
recipe. The
example’s SRC_URI
statement identifies multiple files as the source
files for the recipe: a tarball, a patch file, a desktop file, and an
icon.
SRC_URI = "http://dist.schmorp.de/rxvt-unicode/Attic/rxvt-unicode-${PV}.tar.bz2 \
file://xwc.patch \
file://rxvt.desktop \
file://rxvt.png"
When you specify local files using the file://
URI protocol, the
build system fetches files from the local machine. The path is relative
to the FILESPATH variable
and searches specific directories in a certain order:
${
BP}
,
${
BPN}
, and
files
. The directories are assumed to be subdirectories of the
directory in which the recipe or append file resides. For another
example that specifies these types of files, see the
“Single .c File Package (Hello World!)” section.
The previous example also specifies a patch file. Patch files are files
whose names usually end in .patch
or .diff
but can end with
compressed suffixes such as diff.gz
and patch.bz2
, for example.
The build system automatically applies patches as described in the
“Patching Code” section.
3.3.6 Unpacking Code
During the build, the
do_unpack task unpacks
the source with ${
S}
pointing to where it is unpacked.
If you are fetching your source files from an upstream source archived
tarball and the tarball’s internal structure matches the common
convention of a top-level subdirectory named
${
BPN}-${
PV}
,
then you do not need to set S
. However, if SRC_URI
specifies to
fetch source from an archive that does not use this convention, or from
an SCM like Git or Subversion, your recipe needs to define S
.
If processing your recipe using BitBake successfully unpacks the source
files, you need to be sure that the directory pointed to by ${S}
matches the structure of the source.
3.3.7 Patching Code
Sometimes it is necessary to patch code after it has been fetched. Any
files mentioned in SRC_URI
whose names end in .patch
or
.diff
or compressed versions of these suffixes (e.g. diff.gz
are
treated as patches. The
do_patch task
automatically applies these patches.
The build system should be able to apply patches with the “-p1” option
(i.e. one directory level in the path will be stripped off). If your
patch needs to have more directory levels stripped off, specify the
number of levels using the “striplevel” option in the SRC_URI
entry
for the patch. Alternatively, if your patch needs to be applied in a
specific subdirectory that is not specified in the patch file, use the
“patchdir” option in the entry.
As with all local files referenced in
SRC_URI using file://
,
you should place patch files in a directory next to the recipe either
named the same as the base name of the recipe
(BP and
BPN) or “files”.
3.3.8 Licensing
Your recipe needs to have both the LICENSE and LIC_FILES_CHKSUM variables:
LICENSE
: This variable specifies the license for the software. If you do not know the license under which the software you are building is distributed, you should go to the source code and look for that information. Typical files containing this information includeCOPYING
,LICENSE
, andREADME
files. You could also find the information near the top of a source file. For example, given a piece of software licensed under the GNU General Public License version 2, you would setLICENSE
as follows:LICENSE = "GPLv2"
The licenses you specify within
LICENSE
can have any name as long as you do not use spaces, since spaces are used as separators between license names. For standard licenses, use the names of the files inmeta/files/common-licenses/
or theSPDXLICENSEMAP
flag names defined inmeta/conf/licenses.conf
.LIC_FILES_CHKSUM
: The OpenEmbedded build system uses this variable to make sure the license text has not changed. If it has, the build produces an error and it affords you the chance to figure it out and correct the problem.You need to specify all applicable licensing files for the software. At the end of the configuration step, the build process will compare the checksums of the files to be sure the text has not changed. Any differences result in an error with the message containing the current checksum. For more explanation and examples of how to set the
LIC_FILES_CHKSUM
variable, see the “Tracking License Changes” section.To determine the correct checksum string, you can list the appropriate files in the
LIC_FILES_CHKSUM
variable with incorrect md5 strings, attempt to build the software, and then note the resulting error messages that will report the correct md5 strings. See the “Fetching Code” section for additional information.Here is an example that assumes the software has a
COPYING
file:LIC_FILES_CHKSUM = "file://COPYING;md5=xxx"
When you try to build the software, the build system will produce an error and give you the correct string that you can substitute into the recipe file for a subsequent build.
3.3.9 Dependencies
Most software packages have a short list of other packages that they require, which are called dependencies. These dependencies fall into two main categories: build-time dependencies, which are required when the software is built; and runtime dependencies, which are required to be installed on the target in order for the software to run.
Within a recipe, you specify build-time dependencies using the
DEPENDS variable. Although
nuances exist, items specified in DEPENDS
should be names of other
recipes. It is important that you specify all build-time dependencies
explicitly.
Another consideration is that configure scripts might automatically check for optional dependencies and enable corresponding functionality if those dependencies are found. If you wish to make a recipe that is more generally useful (e.g. publish the recipe in a layer for others to use), instead of hard-disabling the functionality, you can use the PACKAGECONFIG variable to allow functionality and the corresponding dependencies to be enabled and disabled easily by other users of the recipe.
Similar to build-time dependencies, you specify runtime dependencies
through a variable -
RDEPENDS, which is
package-specific. All variables that are package-specific need to have
the name of the package added to the end as an override. Since the main
package for a recipe has the same name as the recipe, and the recipe’s
name can be found through the
${
PN}
variable, then
you specify the dependencies for the main package by setting
RDEPENDS_${PN}
. If the package were named ${PN}-tools
, then you
would set RDEPENDS_${PN}-tools
, and so forth.
Some runtime dependencies will be set automatically at packaging time. These dependencies include any shared library dependencies (i.e. if a package “example” contains “libexample” and another package “mypackage” contains a binary that links to “libexample” then the OpenEmbedded build system will automatically add a runtime dependency to “mypackage” on “example”). See the “Automatically Added Runtime Dependencies” section in the Yocto Project Overview and Concepts Manual for further details.
3.3.10 Configuring the Recipe
Most software provides some means of setting build-time configuration options before compilation. Typically, setting these options is accomplished by running a configure script with options, or by modifying a build configuration file.
Note
As of Yocto Project Release 1.7, some of the core recipes that
package binary configuration scripts now disable the scripts due to
the scripts previously requiring error-prone path substitution. The
OpenEmbedded build system uses pkg-config
now, which is much more
robust. You can find a list of the *-config
scripts that are disabled
in the “Binary Configuration Scripts Disabled” section
in the Yocto Project Reference Manual.
A major part of build-time configuration is about checking for build-time dependencies and possibly enabling optional functionality as a result. You need to specify any build-time dependencies for the software you are building in your recipe’s DEPENDS value, in terms of other recipes that satisfy those dependencies. You can often find build-time or runtime dependencies described in the software’s documentation.
The following list provides configuration items of note based on how your software is built:
Autotools: If your source files have a
configure.ac
file, then your software is built using Autotools. If this is the case, you just need to worry about modifying the configuration.When using Autotools, your recipe needs to inherit the autotools class and your recipe does not have to contain a do_configure task. However, you might still want to make some adjustments. For example, you can set EXTRA_OECONF or PACKAGECONFIG_CONFARGS to pass any needed configure options that are specific to the recipe.
CMake: If your source files have a
CMakeLists.txt
file, then your software is built using CMake. If this is the case, you just need to worry about modifying the configuration.When you use CMake, your recipe needs to inherit the cmake class and your recipe does not have to contain a do_configure task. You can make some adjustments by setting EXTRA_OECMAKE to pass any needed configure options that are specific to the recipe.
Note
If you need to install one or more custom CMake toolchain files that are supplied by the application you are building, install the files to
${D}${datadir}/cmake/Modules
duringdo_install
.Other: If your source files do not have a
configure.ac
orCMakeLists.txt
file, then your software is built using some method other than Autotools or CMake. If this is the case, you normally need to provide a do_configure task in your recipe unless, of course, there is nothing to configure.Even if your software is not being built by Autotools or CMake, you still might not need to deal with any configuration issues. You need to determine if configuration is even a required step. You might need to modify a Makefile or some configuration file used for the build to specify necessary build options. Or, perhaps you might need to run a provided, custom configure script with the appropriate options.
For the case involving a custom configure script, you would run
./configure --help
and look for the options you need to set.
Once configuration succeeds, it is always good practice to look at the
log.do_configure
file to ensure that the appropriate options have
been enabled and no additional build-time dependencies need to be added
to DEPENDS
. For example, if the configure script reports that it
found something not mentioned in DEPENDS
, or that it did not find
something that it needed for some desired optional functionality, then
you would need to add those to DEPENDS
. Looking at the log might
also reveal items being checked for, enabled, or both that you do not
want, or items not being found that are in DEPENDS
, in which case
you would need to look at passing extra options to the configure script
as needed. For reference information on configure options specific to
the software you are building, you can consult the output of the
./configure --help
command within ${S}
or consult the software’s
upstream documentation.
3.3.11 Using Headers to Interface with Devices
If your recipe builds an application that needs to communicate with some
device or needs an API into a custom kernel, you will need to provide
appropriate header files. Under no circumstances should you ever modify
the existing
meta/recipes-kernel/linux-libc-headers/linux-libc-headers.inc
file.
These headers are used to build libc
and must not be compromised
with custom or machine-specific header information. If you customize
libc
through modified headers all other applications that use
libc
thus become affected.
Note
Never copy and customize the libc
header file (i.e.
meta/recipes-kernel/linux-libc-headers/linux-libc-headers.inc
).
The correct way to interface to a device or custom kernel is to use a separate package that provides the additional headers for the driver or other unique interfaces. When doing so, your application also becomes responsible for creating a dependency on that specific provider.
Consider the following:
Never modify
linux-libc-headers.inc
. Consider that file to be part of thelibc
system, and not something you use to access the kernel directly. You should accesslibc
through specificlibc
calls.Applications that must talk directly to devices should either provide necessary headers themselves, or establish a dependency on a special headers package that is specific to that driver.
For example, suppose you want to modify an existing header that adds I/O control or network support. If the modifications are used by a small number programs, providing a unique version of a header is easy and has little impact. When doing so, bear in mind the guidelines in the previous list.
Note
If for some reason your changes need to modify the behavior of the libc
,
and subsequently all other applications on the system, use a .bbappend
to modify the linux-kernel-headers.inc
file. However, take care to not
make the changes machine specific.
Consider a case where your kernel is older and you need an older
libc
ABI. The headers installed by your recipe should still be a
standard mainline kernel, not your own custom one.
When you use custom kernel headers you need to get them from STAGING_KERNEL_DIR, which is the directory with kernel headers that are required to build out-of-tree modules. Your recipe will also need the following:
do_configure[depends] += "virtual/kernel:do_shared_workdir"
3.3.12 Compilation
During a build, the do_compile
task happens after source is fetched,
unpacked, and configured. If the recipe passes through do_compile
successfully, nothing needs to be done.
However, if the compile step fails, you need to diagnose the failure. Here are some common issues that cause failures.
Note
For cases where improper paths are detected for configuration files
or for when libraries/headers cannot be found, be sure you are using
the more robust pkg-config
. See the note in section
“Configuring the Recipe” for additional information.
Parallel build failures: These failures manifest themselves as intermittent errors, or errors reporting that a file or directory that should be created by some other part of the build process could not be found. This type of failure can occur even if, upon inspection, the file or directory does exist after the build has failed, because that part of the build process happened in the wrong order.
To fix the problem, you need to either satisfy the missing dependency in the Makefile or whatever script produced the Makefile, or (as a workaround) set PARALLEL_MAKE to an empty string:
PARALLEL_MAKE = ""
For information on parallel Makefile issues, see the “Debugging Parallel Make Races” section.
Improper host path usage: This failure applies to recipes building for the target or
nativesdk
only. The failure occurs when the compilation process uses improper headers, libraries, or other files from the host system when cross-compiling for the target.To fix the problem, examine the
log.do_compile
file to identify the host paths being used (e.g./usr/include
,/usr/lib
, and so forth) and then either add configure options, apply a patch, or do both.Failure to find required libraries/headers: If a build-time dependency is missing because it has not been declared in DEPENDS, or because the dependency exists but the path used by the build process to find the file is incorrect and the configure step did not detect it, the compilation process could fail. For either of these failures, the compilation process notes that files could not be found. In these cases, you need to go back and add additional options to the configure script as well as possibly add additional build-time dependencies to
DEPENDS
.Occasionally, it is necessary to apply a patch to the source to ensure the correct paths are used. If you need to specify paths to find files staged into the sysroot from other recipes, use the variables that the OpenEmbedded build system provides (e.g.
STAGING_BINDIR
,STAGING_INCDIR
,STAGING_DATADIR
, and so forth).
3.3.13 Installing
During do_install
, the task copies the built files along with their
hierarchy to locations that would mirror their locations on the target
device. The installation process copies files from the
${
S}
,
${
B}
, and
${
WORKDIR}
directories to the ${
D}
directory to create the structure as it should appear on the target
system.
How your software is built affects what you must do to be sure your software is installed correctly. The following list describes what you must do for installation depending on the type of build system used by the software being built:
Autotools and CMake: If the software your recipe is building uses Autotools or CMake, the OpenEmbedded build system understands how to install the software. Consequently, you do not have to have a
do_install
task as part of your recipe. You just need to make sure the install portion of the build completes with no issues. However, if you wish to install additional files not already being installed bymake install
, you should do this using ado_install_append
function using the install command as described in the “Manual” bulleted item later in this list.Other (using
make install
): You need to define ado_install
function in your recipe. The function should calloe_runmake install
and will likely need to pass in the destination directory as well. How you pass that path is dependent on how theMakefile
being run is written (e.g.DESTDIR=${D}
,PREFIX=${D}
,INSTALLROOT=${D}
, and so forth).For an example recipe using
make install
, see the “Makefile-Based Package” section.Manual: You need to define a
do_install
function in your recipe. The function must first useinstall -d
to create the directories under${
D}
. Once the directories exist, your function can useinstall
to manually install the built software into the directories.You can find more information on
install
at https://www.gnu.org/software/coreutils/manual/html_node/install-invocation.html.
For the scenarios that do not use Autotools or CMake, you need to track
the installation and diagnose and fix any issues until everything
installs correctly. You need to look in the default location of
${D}
, which is ${WORKDIR}/image
, to be sure your files have been
installed correctly.
Note
During the installation process, you might need to modify some of the installed files to suit the target layout. For example, you might need to replace hard-coded paths in an initscript with values of variables provided by the build system, such as replacing
/usr/bin/
with${bindir}
. If you do perform such modifications duringdo_install
, be sure to modify the destination file after copying rather than before copying. Modifying after copying ensures that the build system can re-executedo_install
if needed.oe_runmake install
, which can be run directly or can be run indirectly by the autotools and cmake classes, runsmake install
in parallel. Sometimes, a Makefile can have missing dependencies between targets that can result in race conditions. If you experience intermittent failures duringdo_install
, you might be able to work around them by disabling parallel Makefile installs by adding the following to the recipe:PARALLEL_MAKEINST = ""
See PARALLEL_MAKEINST for additional information.
If you need to install one or more custom CMake toolchain files that are supplied by the application you are building, install the files to
${D}${datadir}/cmake/Modules
during do_install.
3.3.14 Enabling System Services
If you want to install a service, which is a process that usually starts on boot and runs in the background, then you must include some additional definitions in your recipe.
If you are adding services and the service initialization script or the
service file itself is not installed, you must provide for that
installation in your recipe using a do_install_append
function. If
your recipe already has a do_install
function, update the function
near its end rather than adding an additional do_install_append
function.
When you create the installation for your services, you need to
accomplish what is normally done by make install
. In other words,
make sure your installation arranges the output similar to how it is
arranged on the target system.
The OpenEmbedded build system provides support for starting services two different ways:
SysVinit: SysVinit is a system and service manager that manages the init system used to control the very basic functions of your system. The init program is the first program started by the Linux kernel when the system boots. Init then controls the startup, running and shutdown of all other programs.
To enable a service using SysVinit, your recipe needs to inherit the update-rc.d class. The class helps facilitate safely installing the package on the target.
You will need to set the INITSCRIPT_PACKAGES, INITSCRIPT_NAME, and INITSCRIPT_PARAMS variables within your recipe.
systemd: System Management Daemon (systemd) was designed to replace SysVinit and to provide enhanced management of services. For more information on systemd, see the systemd homepage at https://freedesktop.org/wiki/Software/systemd/.
To enable a service using systemd, your recipe needs to inherit the systemd class. See the
systemd.bbclass
file located in your Source Directory section for more information.
3.3.15 Packaging
Successful packaging is a combination of automated processes performed by the OpenEmbedded build system and some specific steps you need to take. The following list describes the process:
Splitting Files: The
do_package
task splits the files produced by the recipe into logical components. Even software that produces a single binary might still have debug symbols, documentation, and other logical components that should be split out. Thedo_package
task ensures that files are split up and packaged correctly.Running QA Checks: The insane class adds a step to the package generation process so that output quality assurance checks are generated by the OpenEmbedded build system. This step performs a range of checks to be sure the build’s output is free of common problems that show up during runtime. For information on these checks, see the insane class and the “QA Error and Warning Messages” chapter in the Yocto Project Reference Manual.
Hand-Checking Your Packages: After you build your software, you need to be sure your packages are correct. Examine the
${
WORKDIR}/packages-split
directory and make sure files are where you expect them to be. If you discover problems, you can set PACKAGES, FILES,do_install(_append)
, and so forth as needed.Splitting an Application into Multiple Packages: If you need to split an application into several packages, see the “Splitting an Application into Multiple Packages” section for an example.
Installing a Post-Installation Script: For an example showing how to install a post-installation script, see the “Post-Installation Scripts” section.
Marking Package Architecture: Depending on what your recipe is building and how it is configured, it might be important to mark the packages produced as being specific to a particular machine, or to mark them as not being specific to a particular machine or architecture at all.
By default, packages apply to any machine with the same architecture as the target machine. When a recipe produces packages that are machine-specific (e.g. the MACHINE value is passed into the configure script or a patch is applied only for a particular machine), you should mark them as such by adding the following to the recipe:
PACKAGE_ARCH = "${MACHINE_ARCH}"
On the other hand, if the recipe produces packages that do not contain anything specific to the target machine or architecture at all (e.g. recipes that simply package script files or configuration files), you should use the allarch class to do this for you by adding this to your recipe:
inherit allarch
Ensuring that the package architecture is correct is not critical while you are doing the first few builds of your recipe. However, it is important in order to ensure that your recipe rebuilds (or does not rebuild) appropriately in response to changes in configuration, and to ensure that you get the appropriate packages installed on the target machine, particularly if you run separate builds for more than one target machine.
3.3.17 Using Virtual Providers
Prior to a build, if you know that several different recipes provide the
same functionality, you can use a virtual provider (i.e. virtual/*
)
as a placeholder for the actual provider. The actual provider is
determined at build-time.
A common scenario where a virtual provider is used would be for the
kernel recipe. Suppose you have three kernel recipes whose
PN values map to kernel-big
,
kernel-mid
, and kernel-small
. Furthermore, each of these recipes
in some way uses a PROVIDES
statement that essentially identifies itself as being able to provide
virtual/kernel
. Here is one way through the
kernel class:
PROVIDES += "${@ "virtual/kernel" if (d.getVar("KERNEL_PACKAGE_NAME") == "kernel") else "" }"
Any recipe that inherits the kernel
class is
going to utilize a PROVIDES
statement that identifies that recipe as
being able to provide the virtual/kernel
item.
Now comes the time to actually build an image and you need a kernel
recipe, but which one? You can configure your build to call out the
kernel recipe you want by using the PREFERRED_PROVIDER variable. As
an example, consider the x86-base.inc include file, which is a
machine (i.e. MACHINE) configuration file. This include file is the
reason all x86-based machines use the linux-yocto
kernel. Here are the
relevant lines from the include file:
PREFERRED_PROVIDER_virtual/kernel ??= "linux-yocto"
PREFERRED_VERSION_linux-yocto ??= "4.15%"
When you use a virtual provider, you do not have to “hard code” a recipe
name as a build dependency. You can use the
DEPENDS variable to state the
build is dependent on virtual/kernel
for example:
DEPENDS = "virtual/kernel"
During the build, the OpenEmbedded build system picks
the correct recipe needed for the virtual/kernel
dependency based on
the PREFERRED_PROVIDER
variable. If you want to use the small kernel
mentioned at the beginning of this section, configure your build as
follows:
PREFERRED_PROVIDER_virtual/kernel ??= "kernel-small"
Note
Any recipe that PROVIDES
a virtual/*
item that is ultimately not
selected through PREFERRED_PROVIDER
does not get built. Preventing these
recipes from building is usually the desired behavior since this mechanism’s
purpose is to select between mutually exclusive alternative providers.
The following lists specific examples of virtual providers:
virtual/kernel
: Provides the name of the kernel recipe to use when building a kernel image.virtual/bootloader
: Provides the name of the bootloader to use when building an image.virtual/libgbm
: Providesgbm.pc
.virtual/egl
: Providesegl.pc
and possiblywayland-egl.pc
.virtual/libgl
: Providesgl.pc
(i.e. libGL).virtual/libgles1
: Providesglesv1_cm.pc
(i.e. libGLESv1_CM).virtual/libgles2
: Providesglesv2.pc
(i.e. libGLESv2).
3.3.18 Properly Versioning Pre-Release Recipes
Sometimes the name of a recipe can lead to versioning problems when the
recipe is upgraded to a final release. For example, consider the
irssi_0.8.16-rc1.bb
recipe file in the list of example recipes in
the “Storing and Naming the Recipe” section.
This recipe is at a release candidate stage (i.e. “rc1”). When the recipe is
released, the recipe filename becomes irssi_0.8.16.bb
. The version
change from 0.8.16-rc1
to 0.8.16
is seen as a decrease by the
build system and package managers, so the resulting packages will not
correctly trigger an upgrade.
In order to ensure the versions compare properly, the recommended convention is to set PV within the recipe to “previous_version+current_version”. You can use an additional variable so that you can use the current version elsewhere. Here is an example:
REALPV = "0.8.16-rc1"
PV = "0.8.15+${REALPV}"
3.3.19 Post-Installation Scripts
Post-installation scripts run immediately after installing a package on
the target or during image creation when a package is included in an
image. To add a post-installation script to a package, add a
pkg_postinst_
PACKAGENAME()
function to the recipe file
(.bb
) and replace PACKAGENAME with the name of the package you want
to attach to the postinst
script. To apply the post-installation
script to the main package for the recipe, which is usually what is
required, specify
${
PN}
in place of
PACKAGENAME.
A post-installation function has the following structure:
pkg_postinst_PACKAGENAME() {
# Commands to carry out
}
The script defined in the post-installation function is called when the root filesystem is created. If the script succeeds, the package is marked as installed.
Note
Any RPM post-installation script that runs on the target should return a 0 exit code. RPM does not allow non-zero exit codes for these scripts, and the RPM package manager will cause the package to fail installation on the target.
Sometimes it is necessary for the execution of a post-installation
script to be delayed until the first boot. For example, the script might
need to be executed on the device itself. To delay script execution
until boot time, you must explicitly mark post installs to defer to the
target. You can use pkg_postinst_ontarget()
or call
postinst_intercept delay_to_first_boot
from pkg_postinst()
. Any
failure of a pkg_postinst()
script (including exit 1) triggers an
error during the
do_rootfs task.
If you have recipes that use pkg_postinst
function and they require
the use of non-standard native tools that have dependencies during
rootfs construction, you need to use the
PACKAGE_WRITE_DEPS
variable in your recipe to list these tools. If you do not use this
variable, the tools might be missing and execution of the
post-installation script is deferred until first boot. Deferring the
script to first boot is undesirable and for read-only rootfs impossible.
Note
Equivalent support for pre-install, pre-uninstall, and post-uninstall
scripts exist by way of pkg_preinst
, pkg_prerm
, and pkg_postrm
,
respectively. These scrips work in exactly the same way as does
pkg_postinst
with the exception that they run at different times. Also,
because of when they run, they are not applicable to being run at image
creation time like pkg_postinst
.
3.3.20 Testing
The final step for completing your recipe is to be sure that the software you built runs correctly. To accomplish runtime testing, add the build’s output packages to your image and test them on the target.
For information on how to customize your image by adding specific packages, see “Customizing Images” section.
3.3.21 Examples
To help summarize how to write a recipe, this section provides some examples given various scenarios:
Recipes that use local files
Using an Autotooled package
Using a Makefile-based package
Splitting an application into multiple packages
Adding binaries to an image
3.3.21.1 Single .c File Package (Hello World!)
Building an application from a single file that is stored locally (e.g.
under files
) requires a recipe that has the file listed in the
SRC_URI
variable. Additionally, you need to manually write the
do_compile
and do_install
tasks. The S
variable defines the
directory containing the source code, which is set to
WORKDIR in this case - the
directory BitBake uses for the build.
SUMMARY = "Simple helloworld application"
SECTION = "examples"
LICENSE = "MIT"
LIC_FILES_CHKSUM = "file://${COMMON_LICENSE_DIR}/MIT;md5=0835ade698e0bcf8506ecda2f7b4f302"
SRC_URI = "file://helloworld.c"
S = "${WORKDIR}"
do_compile() {
${CC} ${LDFLAGS} helloworld.c -o helloworld
}
do_install() {
install -d ${D}${bindir}
install -m 0755 helloworld ${D}${bindir}
}
By default, the helloworld
, helloworld-dbg
, and
helloworld-dev
packages are built. For information on how to
customize the packaging process, see the
“Splitting an Application into Multiple Packages”
section.
3.3.21.2 Autotooled Package
Applications that use Autotools such as autoconf
and automake
require a recipe that has a source archive listed in SRC_URI
and
also inherit the
autotools class,
which contains the definitions of all the steps needed to build an
Autotool-based application. The result of the build is automatically
packaged. And, if the application uses NLS for localization, packages
with local information are generated (one package per language).
Following is one example: (hello_2.3.bb
)
SUMMARY = "GNU Helloworld application"
SECTION = "examples"
LICENSE = "GPLv2+"
LIC_FILES_CHKSUM = "file://COPYING;md5=751419260aa954499f7abaabaa882bbe"
SRC_URI = "${GNU_MIRROR}/hello/hello-${PV}.tar.gz"
inherit autotools gettext
The variable LIC_FILES_CHKSUM
is used to track source license
changes as described in the
“Tracking License Changes” section in
the Yocto Project Overview and Concepts Manual. You can quickly create
Autotool-based recipes in a manner similar to the previous example.
3.3.21.3 Makefile-Based Package
Applications that use GNU make
also require a recipe that has the
source archive listed in SRC_URI
. You do not need to add a
do_compile
step since by default BitBake starts the make
command
to compile the application. If you need additional make
options, you
should store them in the
EXTRA_OEMAKE or
PACKAGECONFIG_CONFARGS
variables. BitBake passes these options into the GNU make
invocation. Note that a do_install
task is still required.
Otherwise, BitBake runs an empty do_install
task by default.
Some applications might require extra parameters to be passed to the
compiler. For example, the application might need an additional header
path. You can accomplish this by adding to the CFLAGS
variable. The
following example shows this:
CFLAGS_prepend = "-I ${S}/include "
In the following example, mtd-utils
is a makefile-based package:
SUMMARY = "Tools for managing memory technology devices"
SECTION = "base"
DEPENDS = "zlib lzo e2fsprogs util-linux"
HOMEPAGE = "http://www.linux-mtd.infradead.org/"
LICENSE = "GPLv2+"
LIC_FILES_CHKSUM = "file://COPYING;md5=0636e73ff0215e8d672dc4c32c317bb3 \
file://include/common.h;beginline=1;endline=17;md5=ba05b07912a44ea2bf81ce409380049c"
# Use the latest version at 26 Oct, 2013
SRCREV = "9f107132a6a073cce37434ca9cda6917dd8d866b"
SRC_URI = "git://git.infradead.org/mtd-utils.git \
file://add-exclusion-to-mkfs-jffs2-git-2.patch \
"
PV = "1.5.1+git${SRCPV}"
S = "${WORKDIR}/git"
EXTRA_OEMAKE = "'CC=${CC}' 'RANLIB=${RANLIB}' 'AR=${AR}' 'CFLAGS=${CFLAGS} -I${S}/include -DWITHOUT_XATTR' 'BUILDDIR=${S}'"
do_install () {
oe_runmake install DESTDIR=${D} SBINDIR=${sbindir} MANDIR=${mandir} INCLUDEDIR=${includedir}
}
PACKAGES =+ "mtd-utils-jffs2 mtd-utils-ubifs mtd-utils-misc"
FILES_mtd-utils-jffs2 = "${sbindir}/mkfs.jffs2 ${sbindir}/jffs2dump ${sbindir}/jffs2reader ${sbindir}/sumtool"
FILES_mtd-utils-ubifs = "${sbindir}/mkfs.ubifs ${sbindir}/ubi*"
FILES_mtd-utils-misc = "${sbindir}/nftl* ${sbindir}/ftl* ${sbindir}/rfd* ${sbindir}/doc* ${sbindir}/serve_image ${sbindir}/recv_image"
PARALLEL_MAKE = ""
BBCLASSEXTEND = "native"
3.3.21.4 Splitting an Application into Multiple Packages
You can use the variables PACKAGES
and FILES
to split an
application into multiple packages.
Following is an example that uses the libxpm
recipe. By default,
this recipe generates a single package that contains the library along
with a few binaries. You can modify the recipe to split the binaries
into separate packages:
require xorg-lib-common.inc
SUMMARY = "Xpm: X Pixmap extension library"
LICENSE = "BSD"
LIC_FILES_CHKSUM = "file://COPYING;md5=51f4270b012ecd4ab1a164f5f4ed6cf7"
DEPENDS += "libxext libsm libxt"
PE = "1"
XORG_PN = "libXpm"
PACKAGES =+ "sxpm cxpm"
FILES_cxpm = "${bindir}/cxpm"
FILES_sxpm = "${bindir}/sxpm"
In the previous example, we want to ship the sxpm
and cxpm
binaries in separate packages. Since bindir
would be packaged into
the main PN
package by default, we prepend the PACKAGES
variable
so additional package names are added to the start of list. This results
in the extra FILES_*
variables then containing information that
define which files and directories go into which packages. Files
included by earlier packages are skipped by latter packages. Thus, the
main PN
package does not include the above listed files.
3.3.21.5 Packaging Externally Produced Binaries
Sometimes, you need to add pre-compiled binaries to an image. For example, suppose that binaries for proprietary code exist, which are created by a particular division of a company. Your part of the company needs to use those binaries as part of an image that you are building using the OpenEmbedded build system. Since you only have the binaries and not the source code, you cannot use a typical recipe that expects to fetch the source specified in SRC_URI and then compile it.
One method is to package the binaries and then install them as part of the image. Generally, it is not a good idea to package binaries since, among other things, it can hinder the ability to reproduce builds and could lead to compatibility problems with ABI in the future. However, sometimes you have no choice.
The easiest solution is to create a recipe that uses the
bin_package class
and to be sure that you are using default locations for build artifacts.
In most cases, the bin_package
class handles “skipping” the
configure and compile steps as well as sets things up to grab packages
from the appropriate area. In particular, this class sets noexec
on
both the do_configure
and do_compile tasks,
sets FILES_${PN}
to “/” so that it picks up all files, and sets up a
do_install task, which
effectively copies all files from ${S}
to ${D}
. The
bin_package
class works well when the files extracted into ${S}
are already laid out in the way they should be laid out on the target.
For more information on these variables, see the
FILES,
PN,
S, and
D variables in the Yocto Project
Reference Manual’s variable glossary.
Note
Using DEPENDS is a good idea even for components distributed in binary form, and is often necessary for shared libraries. For a shared library, listing the library dependencies in
DEPENDS
makes sure that the libraries are available in the staging sysroot when other recipes link against the library, which might be necessary for successful linking.Using
DEPENDS
also allows runtime dependencies between packages to be added automatically. See the “Automatically Added Runtime Dependencies” section in the Yocto Project Overview and Concepts Manual for more information.
If you cannot use the bin_package
class, you need to be sure you are
doing the following:
Create a recipe where the do_configure and do_compile tasks do nothing: It is usually sufficient to just not define these tasks in the recipe, because the default implementations do nothing unless a Makefile is found in
${
S}
.If
${S}
might contain a Makefile, or if you inherit some class that replacesdo_configure
anddo_compile
with custom versions, then you can use the[
noexec]
flag to turn the tasks into no-ops, as follows:do_configure[noexec] = "1" do_compile[noexec] = "1"
Unlike Deleting a Task, using the flag preserves the dependency chain from the do_fetch, do_unpack, and do_patch tasks to the do_install task.
Make sure your
do_install
task installs the binaries appropriately.Ensure that you set up FILES (usually
FILES_${
PN}
) to point to the files you have installed, which of course depends on where you have installed them and whether those files are in different locations than the defaults.
Note
If image prelinking is enabled (e.g. “image-prelink” is in USER_CLASSES which it is by default), prelink will change the binaries in the generated images and this often catches people out. Remove that class to ensure binaries are preserved exactly if that is necessary.
3.3.22 Following Recipe Style Guidelines
When writing recipes, it is good to conform to existing style guidelines. The OpenEmbedded Styleguide wiki page provides rough guidelines for preferred recipe style.
It is common for existing recipes to deviate a bit from this style.
However, aiming for at least a consistent style is a good idea. Some
practices, such as omitting spaces around =
operators in assignments
or ordering recipe components in an erratic way, are widely seen as poor
style.
3.3.23 Recipe Syntax
Understanding recipe file syntax is important for writing recipes. The following list overviews the basic items that make up a BitBake recipe file. For more complete BitBake syntax descriptions, see the “bitbake-user-manual/bitbake-user-manual-metadata” chapter of the BitBake User Manual.
Variable Assignments and Manipulations: Variable assignments allow a value to be assigned to a variable. The assignment can be static text or might include the contents of other variables. In addition to the assignment, appending and prepending operations are also supported.
The following example shows some of the ways you can use variables in recipes:
S = "${WORKDIR}/postfix-${PV}" CFLAGS += "-DNO_ASM" SRC_URI_append = " file://fixup.patch"
Functions: Functions provide a series of actions to be performed. You usually use functions to override the default implementation of a task function or to complement a default function (i.e. append or prepend to an existing function). Standard functions use
sh
shell syntax, although access to OpenEmbedded variables and internal methods are also available.The following is an example function from the
sed
recipe:do_install () { autotools_do_install install -d ${D}${base_bindir} mv ${D}${bindir}/sed ${D}${base_bindir}/sed rmdir ${D}${bindir}/ }
It is also possible to implement new functions that are called between existing tasks as long as the new functions are not replacing or complementing the default functions. You can implement functions in Python instead of shell. Both of these options are not seen in the majority of recipes.
Keywords: BitBake recipes use only a few keywords. You use keywords to include common functions (
inherit
), load parts of a recipe from other files (include
andrequire
) and export variables to the environment (export
).The following example shows the use of some of these keywords:
export POSTCONF = "${STAGING_BINDIR}/postconf" inherit autoconf require otherfile.inc
Comments (#): Any lines that begin with the hash character (
#
) are treated as comment lines and are ignored:# This is a comment
This next list summarizes the most important and most commonly used parts of the recipe syntax. For more information on these parts of the syntax, you can reference the Syntax and Operators chapter in the BitBake User Manual.
Line Continuation (\): Use the backward slash (
\
) character to split a statement over multiple lines. Place the slash character at the end of the line that is to be continued on the next line:VAR = "A really long \ line"
Note
You cannot have any characters including spaces or tabs after the slash character.
Using Variables (${VARNAME}): Use the
${VARNAME}
syntax to access the contents of a variable:SRC_URI = "${SOURCEFORGE_MIRROR}/libpng/zlib-${PV}.tar.gz"
Note
It is important to understand that the value of a variable expressed in this form does not get substituted automatically. The expansion of these expressions happens on-demand later (e.g. usually when a function that makes reference to the variable executes). This behavior ensures that the values are most appropriate for the context in which they are finally used. On the rare occasion that you do need the variable expression to be expanded immediately, you can use the := operator instead of = when you make the assignment, but this is not generally needed.
Quote All Assignments (“value”): Use double quotes around values in all variable assignments (e.g.
"value"
). Following is an example:VAR1 = "${OTHERVAR}" VAR2 = "The version is ${PV}"
Conditional Assignment (?=): Conditional assignment is used to assign a value to a variable, but only when the variable is currently unset. Use the question mark followed by the equal sign (
?=
) to make a “soft” assignment used for conditional assignment. Typically, “soft” assignments are used in thelocal.conf
file for variables that are allowed to come through from the external environment.Here is an example where
VAR1
is set to “New value” if it is currently empty. However, ifVAR1
has already been set, it remains unchanged:VAR1 ?= "New value"
In this next example,
VAR1
is left with the value “Original value”:VAR1 = "Original value" VAR1 ?= "New value"
Appending (+=): Use the plus character followed by the equals sign (
+=
) to append values to existing variables.Note
This operator adds a space between the existing content of the variable and the new content.
Here is an example:
SRC_URI += "file://fix-makefile.patch"
Prepending (=+): Use the equals sign followed by the plus character (
=+
) to prepend values to existing variables.Note
This operator adds a space between the new content and the existing content of the variable.
Here is an example:
VAR =+ "Starts"
Appending (_append): Use the
_append
operator to append values to existing variables. This operator does not add any additional space. Also, the operator is applied after all the+=
, and=+
operators have been applied and after all=
assignments have occurred.The following example shows the space being explicitly added to the start to ensure the appended value is not merged with the existing value:
SRC_URI_append = " file://fix-makefile.patch"
You can also use the
_append
operator with overrides, which results in the actions only being performed for the specified target or machine:SRC_URI_append_sh4 = " file://fix-makefile.patch"
Prepending (_prepend): Use the
_prepend
operator to prepend values to existing variables. This operator does not add any additional space. Also, the operator is applied after all the+=
, and=+
operators have been applied and after all=
assignments have occurred.The following example shows the space being explicitly added to the end to ensure the prepended value is not merged with the existing value:
CFLAGS_prepend = "-I${S}/myincludes "
You can also use the
_prepend
operator with overrides, which results in the actions only being performed for the specified target or machine:CFLAGS_prepend_sh4 = "-I${S}/myincludes "
Overrides: You can use overrides to set a value conditionally, typically based on how the recipe is being built. For example, to set the KBRANCH variable’s value to “standard/base” for any target MACHINE, except for qemuarm where it should be set to “standard/arm-versatile-926ejs”, you would do the following:
KBRANCH = "standard/base" KBRANCH_qemuarm = "standard/arm-versatile-926ejs"
Overrides are also used to separate alternate values of a variable in other situations. For example, when setting variables such as FILES and RDEPENDS that are specific to individual packages produced by a recipe, you should always use an override that specifies the name of the package.
Indentation: Use spaces for indentation rather than tabs. For shell functions, both currently work. However, it is a policy decision of the Yocto Project to use tabs in shell functions. Realize that some layers have a policy to use spaces for all indentation.
Using Python for Complex Operations: For more advanced processing, it is possible to use Python code during variable assignments (e.g. search and replacement on a variable).
You indicate Python code using the
${@python_code}
syntax for the variable assignment:SRC_URI = "ftp://ftp.info-zip.org/pub/infozip/src/zip${@d.getVar('PV',1).replace('.', '')}.tgz
Shell Function Syntax: Write shell functions as if you were writing a shell script when you describe a list of actions to take. You should ensure that your script works with a generic
sh
and that it does not require anybash
or other shell-specific functionality. The same considerations apply to various system utilities (e.g.sed
,grep
,awk
, and so forth) that you might wish to use. If in doubt, you should check with multiple implementations - including those from BusyBox.
3.4 Adding a New Machine
Adding a new machine to the Yocto Project is a straightforward process. This section describes how to add machines that are similar to those that the Yocto Project already supports.
Note
Although well within the capabilities of the Yocto Project, adding a
totally new architecture might require changes to gcc
/glibc
and to the site information, which is beyond the scope of this
manual.
For a complete example that shows how to add a new machine, see the “Creating a new BSP Layer Using the bitbake-layers Script” section in the Yocto Project Board Support Package (BSP) Developer’s Guide.
3.4.1 Adding the Machine Configuration File
To add a new machine, you need to add a new machine configuration file
to the layer’s conf/machine
directory. This configuration file
provides details about the device you are adding.
The OpenEmbedded build system uses the root name of the machine
configuration file to reference the new machine. For example, given a
machine configuration file named crownbay.conf
, the build system
recognizes the machine as “crownbay”.
The most important variables you must set in your machine configuration file or include from a lower-level configuration file are as follows:
TARGET_ARCH
(e.g. “arm”)PREFERRED_PROVIDER_virtual/kernel
MACHINE_FEATURES
(e.g. “apm screen wifi”)
You might also need these variables:
SERIAL_CONSOLES
(e.g. “115200;ttyS0 115200;ttyS1”)KERNEL_IMAGETYPE
(e.g. “zImage”)IMAGE_FSTYPES
(e.g. “tar.gz jffs2”)
You can find full details on these variables in the reference section.
You can leverage existing machine .conf
files from
meta-yocto-bsp/conf/machine/
.
3.4.2 Adding a Kernel for the Machine
The OpenEmbedded build system needs to be able to build a kernel for the
machine. You need to either create a new kernel recipe for this machine,
or extend an existing kernel recipe. You can find several kernel recipe
examples in the Source Directory at meta/recipes-kernel/linux
that
you can use as references.
If you are creating a new kernel recipe, normal recipe-writing rules
apply for setting up a SRC_URI
. Thus, you need to specify any
necessary patches and set S
to point at the source code. You need to
create a do_configure
task that configures the unpacked kernel with
a defconfig
file. You can do this by using a make defconfig
command or, more commonly, by copying in a suitable defconfig
file
and then running make oldconfig
. By making use of inherit kernel
and potentially some of the linux-*.inc
files, most other
functionality is centralized and the defaults of the class normally work
well.
If you are extending an existing kernel recipe, it is usually a matter
of adding a suitable defconfig
file. The file needs to be added into
a location similar to defconfig
files used for other machines in a
given kernel recipe. A possible way to do this is by listing the file in
the SRC_URI
and adding the machine to the expression in
COMPATIBLE_MACHINE
:
COMPATIBLE_MACHINE = '(qemux86|qemumips)'
For more information on defconfig
files, see the
“Changing the Configuration”
section in the Yocto Project Linux Kernel Development Manual.
3.4.3 Adding a Formfactor Configuration File
A formfactor configuration file provides information about the target hardware for which the image is being built and information that the build system cannot obtain from other sources such as the kernel. Some examples of information contained in a formfactor configuration file include framebuffer orientation, whether or not the system has a keyboard, the positioning of the keyboard in relation to the screen, and the screen resolution.
The build system uses reasonable defaults in most cases. However, if
customization is necessary, you need to create a machconfig
file in
the meta/recipes-bsp/formfactor/files
directory. This directory
contains directories for specific machines such as qemuarm
and
qemux86
. For information about the settings available and the
defaults, see the meta/recipes-bsp/formfactor/files/config
file
found in the same area.
Following is an example for “qemuarm” machine:
HAVE_TOUCHSCREEN=1
HAVE_KEYBOARD=1
DISPLAY_CAN_ROTATE=0
DISPLAY_ORIENTATION=0
#DISPLAY_WIDTH_PIXELS=640
#DISPLAY_HEIGHT_PIXELS=480
#DISPLAY_BPP=16
DISPLAY_DPI=150
DISPLAY_SUBPIXEL_ORDER=vrgb
3.5 Upgrading Recipes
Over time, upstream developers publish new versions for software built by layer recipes. It is recommended to keep recipes up-to-date with upstream version releases.
While several methods exist that allow you upgrade a recipe, you might
consider checking on the upgrade status of a recipe first. You can do so
using the devtool check-upgrade-status
command. See the
“Checking on the Upgrade Status of a Recipe”
section in the Yocto Project Reference Manual for more information.
The remainder of this section describes three ways you can upgrade a
recipe. You can use the Automated Upgrade Helper (AUH) to set up
automatic version upgrades. Alternatively, you can use
devtool upgrade
to set up semi-automatic version upgrades. Finally,
you can manually upgrade a recipe by editing the recipe itself.
3.5.1 Using the Auto Upgrade Helper (AUH)
The AUH utility works in conjunction with the OpenEmbedded build system in order to automatically generate upgrades for recipes based on new versions being published upstream. Use AUH when you want to create a service that performs the upgrades automatically and optionally sends you an email with the results.
AUH allows you to update several recipes with a single use. You can also optionally perform build and integration tests using images with the results saved to your hard drive and emails of results optionally sent to recipe maintainers. Finally, AUH creates Git commits with appropriate commit messages in the layer’s tree for the changes made to recipes.
Note
Conditions do exist when you should not use AUH to upgrade recipes
and you should instead use either devtool upgrade
or upgrade your
recipes manually:
When AUH cannot complete the upgrade sequence. This situation usually results because custom patches carried by the recipe cannot be automatically rebased to the new version. In this case,
devtool upgrade
allows you to manually resolve conflicts.When for any reason you want fuller control over the upgrade process. For example, when you want special arrangements for testing.
The following steps describe how to set up the AUH utility:
Be Sure the Development Host is Set Up: You need to be sure that your development host is set up to use the Yocto Project. For information on how to set up your host, see the “Preparing the Build Host” section.
Make Sure Git is Configured: The AUH utility requires Git to be configured because AUH uses Git to save upgrades. Thus, you must have Git user and email configured. The following command shows your configurations:
$ git config --list
If you do not have the user and email configured, you can use the following commands to do so:
$ git config --global user.name some_name $ git config --global user.email username@domain.com
Clone the AUH Repository: To use AUH, you must clone the repository onto your development host. The following command uses Git to create a local copy of the repository on your system:
$ git clone git://git.yoctoproject.org/auto-upgrade-helper Cloning into 'auto-upgrade-helper'... remote: Counting objects: 768, done. remote: Compressing objects: 100% (300/300), done. remote: Total 768 (delta 499), reused 703 (delta 434) Receiving objects: 100% (768/768), 191.47 KiB | 98.00 KiB/s, done. Resolving deltas: 100% (499/499), done. Checking connectivity... done.
AUH is not part of the OpenEmbedded-Core (OE-Core) or Poky repositories.
Create a Dedicated Build Directory: Run the oe-init-build-env script to create a fresh build directory that you use exclusively for running the AUH utility:
$ cd poky $ source oe-init-build-env your_AUH_build_directory
Re-using an existing build directory and its configurations is not recommended as existing settings could cause AUH to fail or behave undesirably.
Make Configurations in Your Local Configuration File: Several settings need to exist in the
local.conf
file in the build directory you just created for AUH. Make these following configurations:If you want to enable Build History, which is optional, you need the following lines in the
conf/local.conf
file:INHERIT =+ "buildhistory" BUILDHISTORY_COMMIT = "1"
With this configuration and a successful upgrade, a build history “diff” file appears in the
upgrade-helper/work/recipe/buildhistory-diff.txt
file found in your build directory.If you want to enable testing through the testimage class, which is optional, you need to have the following set in your
conf/local.conf
file:INHERIT += "testimage"
Note
If your distro does not enable by default ptest, which Poky does, you need the following in your
local.conf
file:DISTRO_FEATURES_append = " ptest"
Optionally Start a vncserver: If you are running in a server without an X11 session, you need to start a vncserver:
$ vncserver :1 $ export DISPLAY=:1
Create and Edit an AUH Configuration File: You need to have the
upgrade-helper/upgrade-helper.conf
configuration file in your build directory. You can find a sample configuration file in the AUH source repository.Read through the sample file and make configurations as needed. For example, if you enabled build history in your
local.conf
as described earlier, you must enable it inupgrade-helper.conf
.Also, if you are using the default
maintainers.inc
file supplied with Poky and located inmeta-yocto
and you do not set a “maintainers_whitelist” or “global_maintainer_override” in theupgrade-helper.conf
configuration, and you specify “-e all” on the AUH command-line, the utility automatically sends out emails to all the default maintainers. Please avoid this.
This next set of examples describes how to use the AUH:
Upgrading a Specific Recipe: To upgrade a specific recipe, use the following form:
$ upgrade-helper.py recipe_name
For example, this command upgrades the
xmodmap
recipe:$ upgrade-helper.py xmodmap
Upgrading a Specific Recipe to a Particular Version: To upgrade a specific recipe to a particular version, use the following form:
$ upgrade-helper.py recipe_name -t version
For example, this command upgrades the
xmodmap
recipe to version 1.2.3:$ upgrade-helper.py xmodmap -t 1.2.3
Upgrading all Recipes to the Latest Versions and Suppressing Email Notifications: To upgrade all recipes to their most recent versions and suppress the email notifications, use the following command:
$ upgrade-helper.py all
Upgrading all Recipes to the Latest Versions and Send Email Notifications: To upgrade all recipes to their most recent versions and send email messages to maintainers for each attempted recipe as well as a status email, use the following command:
$ upgrade-helper.py -e all
Once you have run the AUH utility, you can find the results in the AUH build directory:
${BUILDDIR}/upgrade-helper/timestamp
The AUH utility also creates recipe update commits from successful upgrade attempts in the layer tree.
You can easily set up to run the AUH utility on a regular basis by using a cron job. See the weeklyjob.sh file distributed with the utility for an example.
3.5.2 Using devtool upgrade
As mentioned earlier, an alternative method for upgrading recipes to
newer versions is to use
devtool upgrade.
You can read about devtool upgrade
in general in the
“Use devtool upgrade to Create a Version of the Recipe that Supports a Newer Version of the Software”
section in the Yocto Project Application Development and the Extensible
Software Development Kit (eSDK) Manual.
To see all the command-line options available with devtool upgrade
,
use the following help command:
$ devtool upgrade -h
If you want to find out what version a recipe is currently at upstream without any attempt to upgrade your local version of the recipe, you can use the following command:
$ devtool latest-version recipe_name
As mentioned in the previous section describing AUH, devtool upgrade
works in a less-automated manner than AUH. Specifically,
devtool upgrade
only works on a single recipe that you name on the
command line, cannot perform build and integration testing using images,
and does not automatically generate commits for changes in the source
tree. Despite all these “limitations”, devtool upgrade
updates the
recipe file to the new upstream version and attempts to rebase custom
patches contained by the recipe as needed.
Note
AUH uses much of devtool upgrade
behind the scenes making AUH somewhat
of a “wrapper” application for devtool upgrade
.
A typical scenario involves having used Git to clone an upstream
repository that you use during build operations. Because you have built the
recipe in the past, the layer is likely added to your
configuration already. If for some reason, the layer is not added, you
could add it easily using the
“bitbake-layers”
script. For example, suppose you use the nano.bb
recipe from the
meta-oe
layer in the meta-openembedded
repository. For this
example, assume that the layer has been cloned into following area:
/home/scottrif/meta-openembedded
The following command from your
Build Directory adds the layer to
your build configuration (i.e. ${BUILDDIR}/conf/bblayers.conf
):
$ bitbake-layers add-layer /home/scottrif/meta-openembedded/meta-oe
NOTE: Starting bitbake server...
Parsing recipes: 100% |##########################################| Time: 0:00:55
Parsing of 1431 .bb files complete (0 cached, 1431 parsed). 2040 targets, 56 skipped, 0 masked, 0 errors.
Removing 12 recipes from the x86_64 sysroot: 100% |##############| Time: 0:00:00
Removing 1 recipes from the x86_64_i586 sysroot: 100% |##########| Time: 0:00:00
Removing 5 recipes from the i586 sysroot: 100% |#################| Time: 0:00:00
Removing 5 recipes from the qemux86 sysroot: 100% |##############| Time: 0:00:00
For this example, assume that the nano.bb
recipe that
is upstream has a 2.9.3 version number. However, the version in the
local repository is 2.7.4. The following command from your build
directory automatically upgrades the recipe for you:
Note
Using the -V
option is not necessary. Omitting the version number causes
devtool upgrade
to upgrade the recipe to the most recent version.
$ devtool upgrade nano -V 2.9.3
NOTE: Starting bitbake server...
NOTE: Creating workspace layer in /home/scottrif/poky/build/workspace
Parsing recipes: 100% |##########################################| Time: 0:00:46
Parsing of 1431 .bb files complete (0 cached, 1431 parsed). 2040 targets, 56 skipped, 0 masked, 0 errors.
NOTE: Extracting current version source...
NOTE: Resolving any missing task queue dependencies
.
.
.
NOTE: Executing SetScene Tasks
NOTE: Executing RunQueue Tasks
NOTE: Tasks Summary: Attempted 74 tasks of which 72 didn't need to be rerun and all succeeded.
Adding changed files: 100% |#####################################| Time: 0:00:00
NOTE: Upgraded source extracted to /home/scottrif/poky/build/workspace/sources/nano
NOTE: New recipe is /home/scottrif/poky/build/workspace/recipes/nano/nano_2.9.3.bb
Continuing with this example, you can use devtool build
to build the
newly upgraded recipe:
$ devtool build nano
NOTE: Starting bitbake server...
Loading cache: 100% |################################################################################################| Time: 0:00:01
Loaded 2040 entries from dependency cache.
Parsing recipes: 100% |##############################################################################################| Time: 0:00:00
Parsing of 1432 .bb files complete (1431 cached, 1 parsed). 2041 targets, 56 skipped, 0 masked, 0 errors.
NOTE: Resolving any missing task queue dependencies
.
.
.
NOTE: Executing SetScene Tasks
NOTE: Executing RunQueue Tasks
NOTE: nano: compiling from external source tree /home/scottrif/poky/build/workspace/sources/nano
NOTE: Tasks Summary: Attempted 520 tasks of which 304 didn't need to be rerun and all succeeded.
Within the devtool upgrade
workflow, opportunity
exists to deploy and test your rebuilt software. For this example,
however, running devtool finish
cleans up the workspace once the
source in your workspace is clean. This usually means using Git to stage
and submit commits for the changes generated by the upgrade process.
Once the tree is clean, you can clean things up in this example with the
following command from the ${BUILDDIR}/workspace/sources/nano
directory:
$ devtool finish nano meta-oe
NOTE: Starting bitbake server...
Loading cache: 100% |################################################################################################| Time: 0:00:00
Loaded 2040 entries from dependency cache.
Parsing recipes: 100% |##############################################################################################| Time: 0:00:01
Parsing of 1432 .bb files complete (1431 cached, 1 parsed). 2041 targets, 56 skipped, 0 masked, 0 errors.
NOTE: Adding new patch 0001-nano.bb-Stuff-I-changed-when-upgrading-nano.bb.patch
NOTE: Updating recipe nano_2.9.3.bb
NOTE: Removing file /home/scottrif/meta-openembedded/meta-oe/recipes-support/nano/nano_2.7.4.bb
NOTE: Moving recipe file to /home/scottrif/meta-openembedded/meta-oe/recipes-support/nano
NOTE: Leaving source tree /home/scottrif/poky/build/workspace/sources/nano as-is; if you no longer need it then please delete it manually
Using the devtool finish
command cleans up the workspace and creates a patch
file based on your commits. The tool puts all patch files back into the
source directory in a sub-directory named nano
in this case.
3.5.3 Manually Upgrading a Recipe
If for some reason you choose not to upgrade recipes using Using the Auto Upgrade Helper (AUH) or by Using devtool upgrade, you can manually edit the recipe files to upgrade the versions.
Note
Manually updating multiple recipes scales poorly and involves many
steps. The recommendation to upgrade recipe versions is through AUH
or devtool upgrade
, both of which automate some steps and provide
guidance for others needed for the manual process.
To manually upgrade recipe versions, follow these general steps:
Change the Version: Rename the recipe such that the version (i.e. the PV part of the recipe name) changes appropriately. If the version is not part of the recipe name, change the value as it is set for
PV
within the recipe itself.Update
SRCREV
if Needed: If the source code your recipe builds is fetched from Git or some other version control system, update SRCREV to point to the commit hash that matches the new version.Build the Software: Try to build the recipe using BitBake. Typical build failures include the following:
License statements were updated for the new version. For this case, you need to review any changes to the license and update the values of LICENSE and LIC_FILES_CHKSUM as needed.
Note
License changes are often inconsequential. For example, the license text’s copyright year might have changed.
Custom patches carried by the older version of the recipe might fail to apply to the new version. For these cases, you need to review the failures. Patches might not be necessary for the new version of the software if the upgraded version has fixed those issues. If a patch is necessary and failing, you need to rebase it into the new version.
Optionally Attempt to Build for Several Architectures: Once you successfully build the new software for a given architecture, you could test the build for other architectures by changing the MACHINE variable and rebuilding the software. This optional step is especially important if the recipe is to be released publicly.
Check the Upstream Change Log or Release Notes: Checking both these reveals if new features exist that could break backwards-compatibility. If so, you need to take steps to mitigate or eliminate that situation.
Optionally Create a Bootable Image and Test: If you want, you can test the new software by booting it onto actual hardware.
Create a Commit with the Change in the Layer Repository: After all builds work and any testing is successful, you can create commits for any changes in the layer holding your upgraded recipe.
3.6 Finding Temporary Source Code
You might find it helpful during development to modify the temporary source code used by recipes to build packages. For example, suppose you are developing a patch and you need to experiment a bit to figure out your solution. After you have initially built the package, you can iteratively tweak the source code, which is located in the Build Directory, and then you can force a re-compile and quickly test your altered code. Once you settle on a solution, you can then preserve your changes in the form of patches.
During a build, the unpacked temporary source code used by recipes to
build packages is available in the Build Directory as defined by the
S variable. Below is the default
value for the S
variable as defined in the
meta/conf/bitbake.conf
configuration file in the
Source Directory:
S = "${WORKDIR}/${BP}"
You should be aware that many recipes override the
S
variable. For example, recipes that fetch their source from Git
usually set S
to ${WORKDIR}/git
.
Note
The BP represents the base recipe name, which consists of the name and version:
BP = "${BPN}-${PV}"
The path to the work directory for the recipe (WORKDIR) is defined as follows:
${TMPDIR}/work/${MULTIMACH_TARGET_SYS}/${PN}/${EXTENDPE}${PV}-${PR}
The actual directory depends on several things:
TMPDIR: The top-level build output directory.
MULTIMACH_TARGET_SYS: The target system identifier.
PN: The recipe name.
EXTENDPE: The epoch - (if PE is not specified, which is usually the case for most recipes, then
EXTENDPE
is blank).PV: The recipe version.
PR: The recipe revision.
As an example, assume a Source Directory top-level folder named
poky
, a default Build Directory at poky/build
, and a
qemux86-poky-linux
machine target system. Furthermore, suppose your
recipe is named foo_1.3.0.bb
. In this case, the work directory the
build system uses to build the package would be as follows:
poky/build/tmp/work/qemux86-poky-linux/foo/1.3.0-r0
3.7 Using Quilt in Your Workflow
Quilt is a powerful tool that allows you to capture source code changes without having a clean source tree. This section outlines the typical workflow you can use to modify source code, test changes, and then preserve the changes in the form of a patch all using Quilt.
Note
With regard to preserving changes to source files, if you clean a
recipe or have rm_work
enabled, the
devtool workflow
as described in the Yocto Project Application Development and the
Extensible Software Development Kit (eSDK) manual is a safer
development flow than the flow that uses Quilt.
Follow these general steps:
Find the Source Code: Temporary source code used by the OpenEmbedded build system is kept in the Build Directory. See the “Finding Temporary Source Code” section to learn how to locate the directory that has the temporary source code for a particular package.
Change Your Working Directory: You need to be in the directory that has the temporary source code. That directory is defined by the S variable.
Create a New Patch: Before modifying source code, you need to create a new patch. To create a new patch file, use
quilt new
as below:$ quilt new my_changes.patch
Notify Quilt and Add Files: After creating the patch, you need to notify Quilt about the files you plan to edit. You notify Quilt by adding the files to the patch you just created:
$ quilt add file1.c file2.c file3.c
Edit the Files: Make your changes in the source code to the files you added to the patch.
Test Your Changes: Once you have modified the source code, the easiest way to test your changes is by calling the
do_compile
task as shown in the following example:$ bitbake -c compile -f package
The
-f
or--force
option forces the specified task to execute. If you find problems with your code, you can just keep editing and re-testing iteratively until things work as expected.Note
All the modifications you make to the temporary source code disappear once you run the
do_clean
ordo_cleanall
tasks using BitBake (i.e.bitbake -c clean package
andbitbake -c cleanall package
). Modifications will also disappear if you use therm_work
feature as described in the “Conserving Disk Space During Builds” section.Generate the Patch: Once your changes work as expected, you need to use Quilt to generate the final patch that contains all your modifications.
$ quilt refresh
At this point, the
my_changes.patch
file has all your edits made to thefile1.c
,file2.c
, andfile3.c
files.You can find the resulting patch file in the
patches/
subdirectory of the source (S
) directory.Copy the Patch File: For simplicity, copy the patch file into a directory named
files
, which you can create in the same directory that holds the recipe (.bb
) file or the append (.bbappend
) file. Placing the patch here guarantees that the OpenEmbedded build system will find the patch. Next, add the patch into theSRC_URI
of the recipe. Here is an example:SRC_URI += "file://my_changes.patch"
3.8 Using a Development Shell
When debugging certain commands or even when just editing packages,
devshell
can be a useful tool. When you invoke devshell
, all
tasks up to and including
do_patch are run for the
specified target. Then, a new terminal is opened and you are placed in
${
S}
, the source
directory. In the new terminal, all the OpenEmbedded build-related
environment variables are still defined so you can use commands such as
configure
and make
. The commands execute just as if the
OpenEmbedded build system were executing them. Consequently, working
this way can be helpful when debugging a build or preparing software to
be used with the OpenEmbedded build system.
Following is an example that uses devshell
on a target named
matchbox-desktop
:
$ bitbake matchbox-desktop -c devshell
This command spawns a terminal with a shell prompt within the OpenEmbedded build environment. The OE_TERMINAL variable controls what type of shell is opened.
For spawned terminals, the following occurs:
The
PATH
variable includes the cross-toolchain.The
pkgconfig
variables find the correct.pc
files.The
configure
command finds the Yocto Project site files as well as any other necessary files.
Within this environment, you can run configure or compile commands as if they were being run by the OpenEmbedded build system itself. As noted earlier, the working directory also automatically changes to the Source Directory (S).
To manually run a specific task using devshell
, run the
corresponding run.*
script in the
${
WORKDIR}/temp
directory (e.g., run.do_configure.
pid). If a task’s script does
not exist, which would be the case if the task was skipped by way of the
sstate cache, you can create the task by first running it outside of the
devshell
:
$ bitbake -c task
Note
Execution of a task’s
run.*
script and BitBake’s execution of a task are identical. In other words, running the script re-runs the task just as it would be run using thebitbake -c
command.Any
run.*
file that does not have a.pid
extension is a symbolic link (symlink) to the most recent version of that file.
Remember, that the devshell
is a mechanism that allows you to get
into the BitBake task execution environment. And as such, all commands
must be called just as BitBake would call them. That means you need to
provide the appropriate options for cross-compilation and so forth as
applicable.
When you are finished using devshell
, exit the shell or close the
terminal window.
Note
It is worth remembering that when using
devshell
you need to use the full compiler name such asarm-poky-linux-gnueabi-gcc
instead of just usinggcc
. The same applies to other applications such asbinutils
,libtool
and so forth. BitBake sets up environment variables such asCC
to assist applications, such asmake
to find the correct tools.It is also worth noting that
devshell
still works over X11 forwarding and similar situations.
3.9 Using a Development Python Shell
Similar to working within a development shell as described in the
previous section, you can also spawn and work within an interactive
Python development shell. When debugging certain commands or even when
just editing packages, devpyshell
can be a useful tool. When you
invoke devpyshell
, all tasks up to and including
do_patch are run for the
specified target. Then a new terminal is opened. Additionally, key
Python objects and code are available in the same way they are to
BitBake tasks, in particular, the data store ‘d’. So, commands such as
the following are useful when exploring the data store and running
functions:
pydevshell> d.getVar("STAGING_DIR")
'/media/build1/poky/build/tmp/sysroots'
pydevshell> d.getVar("STAGING_DIR")
'${TMPDIR}/sysroots'
pydevshell> d.setVar("FOO", "bar")
pydevshell> d.getVar("FOO")
'bar'
pydevshell> d.delVar("FOO")
pydevshell> d.getVar("FOO")
pydevshell> bb.build.exec_func("do_unpack", d)
pydevshell>
The commands execute just as if the OpenEmbedded build system were executing them. Consequently, working this way can be helpful when debugging a build or preparing software to be used with the OpenEmbedded build system.
Following is an example that uses devpyshell
on a target named
matchbox-desktop
:
$ bitbake matchbox-desktop -c devpyshell
This command spawns a terminal and places you in an interactive Python interpreter within the OpenEmbedded build environment. The OE_TERMINAL variable controls what type of shell is opened.
When you are finished using devpyshell
, you can exit the shell
either by using Ctrl+d or closing the terminal window.
3.10 Building
This section describes various build procedures. For example, the steps needed for a simple build, a target that uses multiple configurations, building an image for more than one machine, and so forth.
3.10.1 Building a Simple Image
In the development environment, you need to build an image whenever you change hardware support, add or change system libraries, or add or change services that have dependencies. Several methods exist that allow you to build an image within the Yocto Project. This section presents the basic steps you need to build a simple image using BitBake from a build host running Linux.
Note
For information on how to build an image using Toaster, see the Toaster User Manual.
For information on how to use
devtool
to build images, see the “Using devtool in Your SDK Workflow” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.For a quick example on how to build an image using the OpenEmbedded build system, see the Yocto Project Quick Build document.
The build process creates an entire Linux distribution from source and
places it in your Build Directory under
tmp/deploy/images
. For detailed information on the build process
using BitBake, see the “Images” section in the
Yocto Project Overview and Concepts Manual.
The following figure and list overviews the build process:
Set up Your Host Development System to Support Development Using the Yocto Project: See the “Setting Up to Use the Yocto Project” section for options on how to get a build host ready to use the Yocto Project.
Initialize the Build Environment: Initialize the build environment by sourcing the build environment script (i.e. oe-init-build-env):
$ source oe-init-build-env [build_dir]
When you use the initialization script, the OpenEmbedded build system uses
build
as the default Build Directory in your current work directory. You can use a build_dir argument with the script to specify a different build directory.Note
A common practice is to use a different Build Directory for different targets. For example,
~/build/x86
for aqemux86
target, and~/build/arm
for aqemuarm
target.Make Sure Your
local.conf
File is Correct: Ensure theconf/local.conf
configuration file, which is found in the Build Directory, is set up how you want it. This file defines many aspects of the build environment including the target machine architecture through theMACHINE
variable, the packaging format used during the build (PACKAGE_CLASSES), and a centralized tarball download directory through the DL_DIR variable.Build the Image: Build the image using the
bitbake
command:$ bitbake target
Note
For information on BitBake, see the BitBake User Manual.
The target is the name of the recipe you want to build. Common targets are the images in
meta/recipes-core/images
,meta/recipes-sato/images
, and so forth all found in the Source Directory. Or, the target can be the name of a recipe for a specific piece of software such as BusyBox. For more details about the images the OpenEmbedded build system supports, see the “Images” chapter in the Yocto Project Reference Manual.As an example, the following command builds the
core-image-minimal
image:$ bitbake core-image-minimal
Once an image has been built, it often needs to be installed. The images and kernels built by the OpenEmbedded build system are placed in the Build Directory in
tmp/deploy/images
. For information on how to run pre-built images such asqemux86
andqemuarm
, see the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual. For information about how to install these images, see the documentation for your particular board or machine.
3.10.2 Building Images for Multiple Targets Using Multiple Configurations
You can use a single bitbake
command to build multiple images or
packages for different targets where each image or package requires a
different configuration (multiple configuration builds). The builds, in
this scenario, are sometimes referred to as “multiconfigs”, and this
section uses that term throughout.
This section describes how to set up for multiple configuration builds and how to account for cross-build dependencies between the multiconfigs.
3.10.2.1 Setting Up and Running a Multiple Configuration Build
To accomplish a multiple configuration build, you must define each
target’s configuration separately using a parallel configuration file in
the Build Directory, and you
must follow a required file hierarchy. Additionally, you must enable the
multiple configuration builds in your local.conf
file.
Follow these steps to set up and execute multiple configuration builds:
Create Separate Configuration Files: You need to create a single configuration file for each build target (each multiconfig). Minimally, each configuration file must define the machine and the temporary directory BitBake uses for the build. Suggested practice dictates that you do not overlap the temporary directories used during the builds. However, it is possible that you can share the temporary directory (TMPDIR). For example, consider a scenario with two different multiconfigs for the same MACHINE: “qemux86” built for two distributions such as “poky” and “poky-lsb”. In this case, you might want to use the same
TMPDIR
.Here is an example showing the minimal statements needed in a configuration file for a “qemux86” target whose temporary build directory is
tmpmultix86
:MACHINE = "qemux86" TMPDIR = "${TOPDIR}/tmpmultix86"
The location for these multiconfig configuration files is specific. They must reside in the current build directory in a sub-directory of
conf
namedmulticonfig
. Following is an example that defines two configuration files for the “x86” and “arm” multiconfigs:The reason for this required file hierarchy is because the
BBPATH
variable is not constructed until the layers are parsed. Consequently, using the configuration file as a pre-configuration file is not possible unless it is located in the current working directory.Add the BitBake Multi-configuration Variable to the Local Configuration File: Use the BBMULTICONFIG variable in your
conf/local.conf
configuration file to specify each multiconfig. Continuing with the example from the previous figure, theBBMULTICONFIG
variable needs to enable two multiconfigs: “x86” and “arm” by specifying each configuration file:BBMULTICONFIG = "x86 arm"
Note
A “default” configuration already exists by definition. This configuration is named: “” (i.e. empty string) and is defined by the variables coming from your
local.conf
file. Consequently, the previous example actually adds two additional configurations to your build: “arm” and “x86” along with “”.Launch BitBake: Use the following BitBake command form to launch the multiple configuration build:
$ bitbake [mc:multiconfigname:]target [[[mc:multiconfigname:]target] ... ]
For the example in this section, the following command applies:
$ bitbake mc:x86:core-image-minimal mc:arm:core-image-sato mc::core-image-base
The previous BitBake command builds a
core-image-minimal
image that is configured through thex86.conf
configuration file, acore-image-sato
image that is configured through thearm.conf
configuration file and acore-image-base
that is configured through yourlocal.conf
configuration file.
Note
Support for multiple configuration builds in the Yocto Project 3.3.1
(Hardknott) Release does not include Shared State (sstate)
optimizations. Consequently, if a build uses the same object twice
in, for example, two different TMPDIR
directories, the build either loads from an existing sstate cache for
that build at the start or builds the object fresh.
3.10.2.2 Enabling Multiple Configuration Build Dependencies
Sometimes dependencies can exist between targets (multiconfigs) in a
multiple configuration build. For example, suppose that in order to
build a core-image-sato
image for an “x86” multiconfig, the root
filesystem of an “arm” multiconfig must exist. This dependency is
essentially that the
do_image task in the
core-image-sato
recipe depends on the completion of the
do_rootfs task of the
core-image-minimal
recipe.
To enable dependencies in a multiple configuration build, you must declare the dependencies in the recipe using the following statement form:
task_or_package[mcdepends] = "mc:from_multiconfig:to_multiconfig:recipe_name:task_on_which_to_depend"
To better show how to use this statement, consider the example scenario
from the first paragraph of this section. The following statement needs
to be added to the recipe that builds the core-image-sato
image:
do_image[mcdepends] = "mc:x86:arm:core-image-minimal:do_rootfs"
In this example, the from_multiconfig is “x86”. The to_multiconfig is “arm”. The
task on which the do_image
task in the recipe depends is the
do_rootfs
task from the core-image-minimal
recipe associated
with the “arm” multiconfig.
Once you set up this dependency, you can build the “x86” multiconfig using a BitBake command as follows:
$ bitbake mc:x86:core-image-sato
This command executes all the tasks needed to create the
core-image-sato
image for the “x86” multiconfig. Because of the
dependency, BitBake also executes through the do_rootfs
task for the
“arm” multiconfig build.
Having a recipe depend on the root filesystem of another build might not
seem that useful. Consider this change to the statement in the
core-image-sato
recipe:
do_image[mcdepends] = "mc:x86:arm:core-image-minimal:do_image"
In this case, BitBake must
create the core-image-minimal
image for the “arm” build since the
“x86” build depends on it.
Because “x86” and “arm” are enabled for multiple configuration builds and have separate configuration files, BitBake places the artifacts for each build in the respective temporary build directories (i.e. TMPDIR).
3.10.3 Building an Initial RAM Filesystem (initramfs) Image
An initial RAM filesystem (initramfs) image provides a temporary root filesystem used for early system initialization (e.g. loading of modules needed to locate and mount the “real” root filesystem).
Note
The initramfs image is the successor of initial RAM disk (initrd). It is a “copy in and out” (cpio) archive of the initial filesystem that gets loaded into memory during the Linux startup process. Because Linux uses the contents of the archive during initialization, the initramfs image needs to contain all of the device drivers and tools needed to mount the final root filesystem.
Follow these steps to create an initramfs image:
Create the initramfs Image Recipe: You can reference the
core-image-minimal-initramfs.bb
recipe found in themeta/recipes-core
directory of the Source Directory as an example from which to work.Decide if You Need to Bundle the initramfs Image Into the Kernel Image: If you want the initramfs image that is built to be bundled in with the kernel image, set the INITRAMFS_IMAGE_BUNDLE variable to “1” in your
local.conf
configuration file and set the INITRAMFS_IMAGE variable in the recipe that builds the kernel image.Note
It is recommended that you do bundle the initramfs image with the kernel image to avoid circular dependencies between the kernel recipe and the initramfs recipe should the initramfs image include kernel modules.
Setting the
INITRAMFS_IMAGE_BUNDLE
flag causes the initramfs image to be unpacked into the${B}/usr/
directory. The unpacked initramfs image is then passed to the kernel’sMakefile
using the CONFIG_INITRAMFS_SOURCE variable, allowing the initramfs image to be built into the kernel normally.Note
If you choose to not bundle the initramfs image with the kernel image, you are essentially using an Initial RAM Disk (initrd). Creating an initrd is handled primarily through the INITRD_IMAGE,
INITRD_LIVE
, andINITRD_IMAGE_LIVE
variables. For more information, see the image-live.bbclass file.Optionally Add Items to the initramfs Image Through the initramfs Image Recipe: If you add items to the initramfs image by way of its recipe, you should use PACKAGE_INSTALL rather than IMAGE_INSTALL.
PACKAGE_INSTALL
gives more direct control of what is added to the image as compared to the defaults you might not necessarily want that are set by the image or core-image classes.Build the Kernel Image and the initramfs Image: Build your kernel image using BitBake. Because the initramfs image recipe is a dependency of the kernel image, the initramfs image is built as well and bundled with the kernel image if you used the INITRAMFS_IMAGE_BUNDLE variable described earlier.
3.10.4 Building a Tiny System
Very small distributions have some significant advantages such as requiring less on-die or in-package memory (cheaper), better performance through efficient cache usage, lower power requirements due to less memory, faster boot times, and reduced development overhead. Some real-world examples where a very small distribution gives you distinct advantages are digital cameras, medical devices, and small headless systems.
This section presents information that shows you how you can trim your
distribution to even smaller sizes than the poky-tiny
distribution,
which is around 5 Mbytes, that can be built out-of-the-box using the
Yocto Project.
3.10.4.1 Tiny System Overview
The following list presents the overall steps you need to consider and perform to create distributions with smaller root filesystems, achieve faster boot times, maintain your critical functionality, and avoid initial RAM disks:
3.10.4.2 Goals and Guiding Principles
Before you can reach your destination, you need to know where you are going. Here is an example list that you can use as a guide when creating very small distributions:
Determine how much space you need (e.g. a kernel that is 1 Mbyte or less and a root filesystem that is 3 Mbytes or less).
Find the areas that are currently taking 90% of the space and concentrate on reducing those areas.
Do not create any difficult “hacks” to achieve your goals.
Leverage the device-specific options.
Work in a separate layer so that you keep changes isolated. For information on how to create layers, see the “Understanding and Creating Layers” section.
3.10.4.3 Understand What Contributes to Your Image Size
It is easiest to have something to start with when creating your own
distribution. You can use the Yocto Project out-of-the-box to create the
poky-tiny
distribution. Ultimately, you will want to make changes in
your own distribution that are likely modeled after poky-tiny
.
Note
To use poky-tiny
in your build, set the DISTRO
variable in your
local.conf
file to “poky-tiny” as described in the
“Creating Your Own Distribution”
section.
Understanding some memory concepts will help you reduce the system size.
Memory consists of static, dynamic, and temporary memory. Static memory
is the TEXT (code), DATA (initialized data in the code), and BSS
(uninitialized data) sections. Dynamic memory represents memory that is
allocated at runtime: stacks, hash tables, and so forth. Temporary
memory is recovered after the boot process. This memory consists of
memory used for decompressing the kernel and for the __init__
functions.
To help you see where you currently are with kernel and root filesystem
sizes, you can use two tools found in the Source Directory
in the
scripts/tiny/
directory:
ksize.py
: Reports component sizes for the kernel build objects.dirsize.py
: Reports component sizes for the root filesystem.
This next tool and command help you organize configuration fragments and view file dependencies in a human-readable form:
merge_config.sh
: Helps you manage configuration files and fragments within the kernel. With this tool, you can merge individual configuration fragments together. The tool allows you to make overrides and warns you of any missing configuration options. The tool is ideal for allowing you to iterate on configurations, create minimal configurations, and create configuration files for different machines without having to duplicate your process.The
merge_config.sh
script is part of the Linux Yocto kernel Git repositories (i.e.linux-yocto-3.14
,linux-yocto-3.10
,linux-yocto-3.8
, and so forth) in thescripts/kconfig
directory.For more information on configuration fragments, see the “Creating Configuration Fragments” section in the Yocto Project Linux Kernel Development Manual.
bitbake -u taskexp -g bitbake_target
: Using the BitBake command with these options brings up a Dependency Explorer from which you can view file dependencies. Understanding these dependencies allows you to make informed decisions when cutting out various pieces of the kernel and root filesystem.
3.10.4.4 Trim the Root Filesystem
The root filesystem is made up of packages for booting, libraries, and applications. To change things, you can configure how the packaging happens, which changes the way you build them. You can also modify the filesystem itself or select a different filesystem.
First, find out what is hogging your root filesystem by running the
dirsize.py
script from your root directory:
$ cd root-directory-of-image
$ dirsize.py 100000 > dirsize-100k.log
$ cat dirsize-100k.log
You can apply a filter to the script to ignore files under a certain size. The previous example filters out any files below 100 Kbytes. The sizes reported by the tool are uncompressed, and thus will be smaller by a relatively constant factor in a compressed root filesystem. When you examine your log file, you can focus on areas of the root filesystem that take up large amounts of memory.
You need to be sure that what you eliminate does not cripple the functionality you need. One way to see how packages relate to each other is by using the Dependency Explorer UI with the BitBake command:
$ cd image-directory
$ bitbake -u taskexp -g image
Use the interface to select potential packages you wish to eliminate and see their dependency relationships.
When deciding how to reduce the size, get rid of packages that result in
minimal impact on the feature set. For example, you might not need a VGA
display. Or, you might be able to get by with devtmpfs
and mdev
instead of udev
.
Use your local.conf
file to make changes. For example, to eliminate
udev
and glib
, set the following in the local configuration
file:
VIRTUAL-RUNTIME_dev_manager = ""
Finally, you should consider exactly the type of root filesystem you
need to meet your needs while also reducing its size. For example,
consider cramfs
, squashfs
, ubifs
, ext2
, or an
initramfs
using initramfs
. Be aware that ext3
requires a 1
Mbyte journal. If you are okay with running read-only, you do not need
this journal.
Note
After each round of elimination, you need to rebuild your system and then use the tools to see the effects of your reductions.
3.10.4.5 Trim the Kernel
The kernel is built by including policies for hardware-independent aspects. What subsystems do you enable? For what architecture are you building? Which drivers do you build by default?
Note
You can modify the kernel source if you want to help with boot time.
Run the ksize.py
script from the top-level Linux build directory to
get an idea of what is making up the kernel:
$ cd top-level-linux-build-directory
$ ksize.py > ksize.log
$ cat ksize.log
When you examine the log, you will see how much space is taken up with
the built-in .o
files for drivers, networking, core kernel files,
filesystem, sound, and so forth. The sizes reported by the tool are
uncompressed, and thus will be smaller by a relatively constant factor
in a compressed kernel image. Look to reduce the areas that are large
and taking up around the “90% rule.”
To examine, or drill down, into any particular area, use the -d
option with the script:
$ ksize.py -d > ksize.log
Using this option breaks out the individual file information for each area of the kernel (e.g. drivers, networking, and so forth).
Use your log file to see what you can eliminate from the kernel based on features you can let go. For example, if you are not going to need sound, you do not need any drivers that support sound.
After figuring out what to eliminate, you need to reconfigure the kernel
to reflect those changes during the next build. You could run
menuconfig
and make all your changes at once. However, that makes it
difficult to see the effects of your individual eliminations and also
makes it difficult to replicate the changes for perhaps another target
device. A better method is to start with no configurations using
allnoconfig
, create configuration fragments for individual changes,
and then manage the fragments into a single configuration file using
merge_config.sh
. The tool makes it easy for you to iterate using the
configuration change and build cycle.
Each time you make configuration changes, you need to rebuild the kernel and check to see what impact your changes had on the overall size.
3.10.4.6 Remove Package Management Requirements
Packaging requirements add size to the image. One way to reduce the size of the image is to remove all the packaging requirements from the image. This reduction includes both removing the package manager and its unique dependencies as well as removing the package management data itself.
To eliminate all the packaging requirements for an image, be sure that “package-management” is not part of your IMAGE_FEATURES statement for the image. When you remove this feature, you are removing the package manager as well as its dependencies from the root filesystem.
3.10.4.7 Look for Other Ways to Minimize Size
Depending on your particular circumstances, other areas that you can trim likely exist. The key to finding these areas is through tools and methods described here combined with experimentation and iteration. Here are a couple of areas to experiment with:
glibc
: In general, follow this process:Remove
glibc
features from DISTRO_FEATURES that you think you do not need.Build your distribution.
If the build fails due to missing symbols in a package, determine if you can reconfigure the package to not need those features. For example, change the configuration to not support wide character support as is done for
ncurses
. Or, if support for those characters is needed, determine whatglibc
features provide the support and restore the configuration.Rebuild and repeat the process.
busybox
: For BusyBox, use a process similar as described forglibc
. A difference is you will need to boot the resulting system to see if you are able to do everything you expect from the running system. You need to be sure to integrate configuration fragments into Busybox because BusyBox handles its own core features and then allows you to add configuration fragments on top.
3.10.4.8 Iterate on the Process
If you have not reached your goals on system size, you need to iterate on the process. The process is the same. Use the tools and see just what is taking up 90% of the root filesystem and the kernel. Decide what you can eliminate without limiting your device beyond what you need.
Depending on your system, a good place to look might be Busybox, which provides a stripped down version of Unix tools in a single, executable file. You might be able to drop virtual terminal services or perhaps ipv6.
3.10.5 Building Images for More than One Machine
A common scenario developers face is creating images for several different machines that use the same software environment. In this situation, it is tempting to set the tunings and optimization flags for each build specifically for the targeted hardware (i.e. “maxing out” the tunings). Doing so can considerably add to build times and package feed maintenance collectively for the machines. For example, selecting tunes that are extremely specific to a CPU core used in a system might enable some micro optimizations in GCC for that particular system but would otherwise not gain you much of a performance difference across the other systems as compared to using a more general tuning across all the builds (e.g. setting DEFAULTTUNE specifically for each machine’s build). Rather than “max out” each build’s tunings, you can take steps that cause the OpenEmbedded build system to reuse software across the various machines where it makes sense.
If build speed and package feed maintenance are considerations, you should consider the points in this section that can help you optimize your tunings to best consider build times and package feed maintenance.
Share the Build Directory: If at all possible, share the TMPDIR across builds. The Yocto Project supports switching between different MACHINE values in the same
TMPDIR
. This practice is well supported and regularly used by developers when building for multiple machines. When you use the sameTMPDIR
for multiple machine builds, the OpenEmbedded build system can reuse the existing native and often cross-recipes for multiple machines. Thus, build time decreases.Note
If DISTRO settings change or fundamental configuration settings such as the filesystem layout, you need to work with a clean
TMPDIR
. SharingTMPDIR
under these circumstances might work but since it is not guaranteed, you should use a cleanTMPDIR
.Enable the Appropriate Package Architecture: By default, the OpenEmbedded build system enables three levels of package architectures: “all”, “tune” or “package”, and “machine”. Any given recipe usually selects one of these package architectures (types) for its output. Depending for what a given recipe creates packages, making sure you enable the appropriate package architecture can directly impact the build time.
A recipe that just generates scripts can enable “all” architecture because there are no binaries to build. To specifically enable “all” architecture, be sure your recipe inherits the allarch class. This class is useful for “all” architectures because it configures many variables so packages can be used across multiple architectures.
If your recipe needs to generate packages that are machine-specific or when one of the build or runtime dependencies is already machine-architecture dependent, which makes your recipe also machine-architecture dependent, make sure your recipe enables the “machine” package architecture through the MACHINE_ARCH variable:
PACKAGE_ARCH = "${MACHINE_ARCH}"
When you do not specifically enable a package architecture through the PACKAGE_ARCH, The OpenEmbedded build system defaults to the TUNE_PKGARCH setting:
PACKAGE_ARCH = "${TUNE_PKGARCH}"
Choose a Generic Tuning File if Possible: Some tunes are more generic and can run on multiple targets (e.g. an
armv5
set of packages could run onarmv6
andarmv7
processors in most cases). Similarly,i486
binaries could work oni586
and higher processors. You should realize, however, that advances on newer processor versions would not be used.If you select the same tune for several different machines, the OpenEmbedded build system reuses software previously built, thus speeding up the overall build time. Realize that even though a new sysroot for each machine is generated, the software is not recompiled and only one package feed exists.
Manage Granular Level Packaging: Sometimes cases exist where injecting another level of package architecture beyond the three higher levels noted earlier can be useful. For example, consider how NXP (formerly Freescale) allows for the easy reuse of binary packages in their layer meta-freescale. In this example, the fsl-dynamic-packagearch class shares GPU packages for i.MX53 boards because all boards share the AMD GPU. The i.MX6-based boards can do the same because all boards share the Vivante GPU. This class inspects the BitBake datastore to identify if the package provides or depends on one of the sub-architecture values. If so, the class sets the PACKAGE_ARCH value based on the
MACHINE_SUBARCH
value. If the package does not provide or depend on one of the sub-architecture values but it matches a value in the machine-specific filter, it sets MACHINE_ARCH. This behavior reduces the number of packages built and saves build time by reusing binaries.Use Tools to Debug Issues: Sometimes you can run into situations where software is being rebuilt when you think it should not be. For example, the OpenEmbedded build system might not be using shared state between machines when you think it should be. These types of situations are usually due to references to machine-specific variables such as MACHINE, SERIAL_CONSOLES, XSERVER, MACHINE_FEATURES, and so forth in code that is supposed to only be tune-specific or when the recipe depends (DEPENDS, RDEPENDS, RRECOMMENDS, RSUGGESTS, and so forth) on some other recipe that already has PACKAGE_ARCH defined as “${MACHINE_ARCH}”.
Note
Patches to fix any issues identified are most welcome as these issues occasionally do occur.
For such cases, you can use some tools to help you sort out the situation:
state-diff-machines.sh``*:* You can find this tool in the ``scripts
directory of the Source Repositories. See the comments in the script for information on how to use the tool.BitBake’s “-S printdiff” Option: Using this option causes BitBake to try to establish the closest signature match it can (e.g. in the shared state cache) and then run
bitbake-diffsigs
over the matches to determine the stamps and delta where these two stamp trees diverge.
3.10.6 Building Software from an External Source
By default, the OpenEmbedded build system uses the Build Directory when building source code. The build process involves fetching the source files, unpacking them, and then patching them if necessary before the build takes place.
Situations exist where you might want to build software from source files that are external to and thus outside of the OpenEmbedded build system. For example, suppose you have a project that includes a new BSP with a heavily customized kernel. And, you want to minimize exposing the build system to the development team so that they can focus on their project and maintain everyone’s workflow as much as possible. In this case, you want a kernel source directory on the development machine where the development occurs. You want the recipe’s SRC_URI variable to point to the external directory and use it as is, not copy it.
To build from software that comes from an external source, all you need
to do is inherit the
externalsrc class
and then set the
EXTERNALSRC variable to
point to your external source code. Here are the statements to put in
your local.conf
file:
INHERIT += "externalsrc"
EXTERNALSRC_pn-myrecipe = "path-to-your-source-tree"
This next example shows how to accomplish the same thing by setting
EXTERNALSRC
in the recipe itself or in the recipe’s append file:
EXTERNALSRC = "path"
EXTERNALSRC_BUILD = "path"
Note
In order for these settings to take effect, you must globally or locally inherit the externalsrc class.
By default, externalsrc.bbclass
builds the source code in a
directory separate from the external source directory as specified by
EXTERNALSRC. If you need
to have the source built in the same directory in which it resides, or
some other nominated directory, you can set
EXTERNALSRC_BUILD
to point to that directory:
EXTERNALSRC_BUILD_pn-myrecipe = "path-to-your-source-tree"
3.10.7 Replicating a Build Offline
It can be useful to take a “snapshot” of upstream sources used in a build and then use that “snapshot” later to replicate the build offline. To do so, you need to first prepare and populate your downloads directory your “snapshot” of files. Once your downloads directory is ready, you can use it at any time and from any machine to replicate your build.
Follow these steps to populate your Downloads directory:
Create a Clean Downloads Directory: Start with an empty downloads directory (DL_DIR). You start with an empty downloads directory by either removing the files in the existing directory or by setting
DL_DIR
to point to either an empty location or one that does not yet exist.Generate Tarballs of the Source Git Repositories: Edit your
local.conf
configuration file as follows:DL_DIR = "/home/your-download-dir/" BB_GENERATE_MIRROR_TARBALLS = "1"
During the fetch process in the next step, BitBake gathers the source files and creates tarballs in the directory pointed to by
DL_DIR
. See the BB_GENERATE_MIRROR_TARBALLS variable for more information.Populate Your Downloads Directory Without Building: Use BitBake to fetch your sources but inhibit the build:
$ bitbake target --runonly=fetch
The downloads directory (i.e.
${DL_DIR}
) now has a “snapshot” of the source files in the form of tarballs, which can be used for the build.Optionally Remove Any Git or other SCM Subdirectories From the Downloads Directory: If you want, you can clean up your downloads directory by removing any Git or other Source Control Management (SCM) subdirectories such as
${DL_DIR}/git2/*
. The tarballs already contain these subdirectories.
Once your downloads directory has everything it needs regarding source files, you can create your “own-mirror” and build your target. Understand that you can use the files to build the target offline from any machine and at any time.
Follow these steps to build your target using the files in the downloads directory:
Using Local Files Only: Inside your
local.conf
file, add the SOURCE_MIRROR_URL variable, inherit the own-mirrors class, and use the BB_NO_NETWORK variable to yourlocal.conf
.SOURCE_MIRROR_URL ?= "file:///home/your-download-dir/" INHERIT += "own-mirrors" BB_NO_NETWORK = "1"
The
SOURCE_MIRROR_URL
andown-mirror
class set up the system to use the downloads directory as your “own mirror”. Using theBB_NO_NETWORK
variable makes sure that BitBake’s fetching process in step 3 stays local, which means files from your “own-mirror” are used.Start With a Clean Build: You can start with a clean build by removing the
${
TMPDIR}
directory or using a new Build Directory.Build Your Target: Use BitBake to build your target:
$ bitbake target
The build completes using the known local “snapshot” of source files from your mirror. The resulting tarballs for your “snapshot” of source files are in the downloads directory.
Note
The offline build does not work if recipes attempt to find the latest version of software by setting SRCREV to
${
AUTOREV}
:SRCREV = "${AUTOREV}"
When a recipe sets
SRCREV
to${AUTOREV}
, the build system accesses the network in an attempt to determine the latest version of software from the SCM. Typically, recipes that useAUTOREV
are custom or modified recipes. Recipes that reside in public repositories usually do not useAUTOREV
.If you do have recipes that use
AUTOREV
, you can take steps to still use the recipes in an offline build. Do the following:Use a configuration generated by enabling build history.
Use the
buildhistory-collect-srcrevs
command to collect the storedSRCREV
values from the build’s history. For more information on collecting these values, see the “Build History Package Information” section.Once you have the correct source revisions, you can modify those recipes to set
SRCREV
to specific versions of the software.
3.11 Speeding Up a Build
Build time can be an issue. By default, the build system uses simple controls to try and maximize build efficiency. In general, the default settings for all the following variables result in the most efficient build times when dealing with single socket systems (i.e. a single CPU). If you have multiple CPUs, you might try increasing the default values to gain more speed. See the descriptions in the glossary for each variable for more information:
BB_NUMBER_THREADS: The maximum number of threads BitBake simultaneously executes.
BB_NUMBER_PARSE_THREADS: The number of threads BitBake uses during parsing.
PARALLEL_MAKE: Extra options passed to the
make
command during the do_compile task in order to specify parallel compilation on the local build host.PARALLEL_MAKEINST: Extra options passed to the
make
command during the do_install task in order to specify parallel installation on the local build host.
As mentioned, these variables all scale to the number of processor cores available on the build system. For single socket systems, this auto-scaling ensures that the build system fundamentally takes advantage of potential parallel operations during the build based on the build machine’s capabilities.
Following are additional factors that can affect build speed:
File system type: The file system type that the build is being performed on can also influence performance. Using
ext4
is recommended as compared toext2
andext3
due toext4
improved features such as extents.Disabling the updating of access time using
noatime
: Thenoatime
mount option prevents the build system from updating file and directory access times.Setting a longer commit: Using the “commit=” mount option increases the interval in seconds between disk cache writes. Changing this interval from the five second default to something longer increases the risk of data loss but decreases the need to write to the disk, thus increasing the build performance.
Choosing the packaging backend: Of the available packaging backends, IPK is the fastest. Additionally, selecting a singular packaging backend also helps.
Using
tmpfs
for TMPDIR as a temporary file system: While this can help speed up the build, the benefits are limited due to the compiler using-pipe
. The build system goes to some lengths to avoidsync()
calls into the file system on the principle that if there was a significant failure, the Build Directory contents could easily be rebuilt.Inheriting the rm_work class: Inheriting this class has shown to speed up builds due to significantly lower amounts of data stored in the data cache as well as on disk. Inheriting this class also makes cleanup of TMPDIR faster, at the expense of being easily able to dive into the source code. File system maintainers have recommended that the fastest way to clean up large numbers of files is to reformat partitions rather than delete files due to the linear nature of partitions. This, of course, assumes you structure the disk partitions and file systems in a way that this is practical.
Aside from the previous list, you should keep some trade offs in mind that can help you speed up the build:
Remove items from DISTRO_FEATURES that you might not need.
Exclude debug symbols and other debug information: If you do not need these symbols and other debug information, disabling the
*-dbg
package generation can speed up the build. You can disable this generation by setting the INHIBIT_PACKAGE_DEBUG_SPLIT variable to “1”.Disable static library generation for recipes derived from
autoconf
orlibtool
: Following is an example showing how to disable static libraries and still provide an override to handle exceptions:STATICLIBCONF = "--disable-static" STATICLIBCONF_sqlite3-native = "" EXTRA_OECONF += "${STATICLIBCONF}"
Note
Some recipes need static libraries in order to work correctly (e.g.
pseudo-native
needssqlite3-native
). Overrides, as in the previous example, account for these kinds of exceptions.Some packages have packaging code that assumes the presence of the static libraries. If so, you might need to exclude them as well.
3.12 Working With Libraries
Libraries are an integral part of your system. This section describes some common practices you might find helpful when working with libraries to build your system:
How to use the Multilib feature to combine multiple versions of library files into a single image
How to install multiple versions of the same library in parallel on the same system
3.12.1 Including Static Library Files
If you are building a library and the library offers static linking, you
can control which static library files (*.a
files) get included in
the built library.
The PACKAGES and
FILES_* variables in the
meta/conf/bitbake.conf
configuration file define how files installed
by the do_install
task are packaged. By default, the PACKAGES
variable includes ${PN}-staticdev
, which represents all static
library files.
Note
Some previously released versions of the Yocto Project defined the
static library files through ${PN}-dev
.
Following is part of the BitBake configuration file, where you can see how the static library files are defined:
PACKAGE_BEFORE_PN ?= ""
PACKAGES = "${PN}-dbg ${PN}-staticdev ${PN}-dev ${PN}-doc ${PN}-locale ${PACKAGE_BEFORE_PN} ${PN}"
PACKAGES_DYNAMIC = "^${PN}-locale-.*"
FILES = ""
FILES_${PN} = "${bindir}/* ${sbindir}/* ${libexecdir}/* ${libdir}/lib*${SOLIBS} \
${sysconfdir} ${sharedstatedir} ${localstatedir} \
${base_bindir}/* ${base_sbindir}/* \
${base_libdir}/*${SOLIBS} \
${base_prefix}/lib/udev/rules.d ${prefix}/lib/udev/rules.d \
${datadir}/${BPN} ${libdir}/${BPN}/* \
${datadir}/pixmaps ${datadir}/applications \
${datadir}/idl ${datadir}/omf ${datadir}/sounds \
${libdir}/bonobo/servers"
FILES_${PN}-bin = "${bindir}/* ${sbindir}/*"
FILES_${PN}-doc = "${docdir} ${mandir} ${infodir} ${datadir}/gtk-doc \
${datadir}/gnome/help"
SECTION_${PN}-doc = "doc"
FILES_SOLIBSDEV ?= "${base_libdir}/lib*${SOLIBSDEV} ${libdir}/lib*${SOLIBSDEV}"
FILES_${PN}-dev = "${includedir} ${FILES_SOLIBSDEV} ${libdir}/*.la \
${libdir}/*.o ${libdir}/pkgconfig ${datadir}/pkgconfig \
${datadir}/aclocal ${base_libdir}/*.o \
${libdir}/${BPN}/*.la ${base_libdir}/*.la"
SECTION_${PN}-dev = "devel"
ALLOW_EMPTY_${PN}-dev = "1"
RDEPENDS_${PN}-dev = "${PN} (= ${EXTENDPKGV})"
FILES_${PN}-staticdev = "${libdir}/*.a ${base_libdir}/*.a ${libdir}/${BPN}/*.a"
SECTION_${PN}-staticdev = "devel"
RDEPENDS_${PN}-staticdev = "${PN}-dev (= ${EXTENDPKGV})"
3.12.2 Combining Multiple Versions of Library Files into One Image
The build system offers the ability to build libraries with different target optimizations or architecture formats and combine these together into one system image. You can link different binaries in the image against the different libraries as needed for specific use cases. This feature is called “Multilib”.
An example would be where you have most of a system compiled in 32-bit mode using 32-bit libraries, but you have something large, like a database engine, that needs to be a 64-bit application and uses 64-bit libraries. Multilib allows you to get the best of both 32-bit and 64-bit libraries.
While the Multilib feature is most commonly used for 32 and 64-bit differences, the approach the build system uses facilitates different target optimizations. You could compile some binaries to use one set of libraries and other binaries to use a different set of libraries. The libraries could differ in architecture, compiler options, or other optimizations.
Several examples exist in the meta-skeleton
layer found in the
Source Directory:
conf/multilib-example.conf
configuration fileconf/multilib-example2.conf
configuration filerecipes-multilib/images/core-image-multilib-example.bb
recipe
3.12.2.1 Preparing to Use Multilib
User-specific requirements drive the Multilib feature. Consequently, there is no one “out-of-the-box” configuration that likely exists to meet your needs.
In order to enable Multilib, you first need to ensure your recipe is
extended to support multiple libraries. Many standard recipes are
already extended and support multiple libraries. You can check in the
meta/conf/multilib.conf
configuration file in the
Source Directory to see how this is
done using the
BBCLASSEXTEND variable.
Eventually, all recipes will be covered and this list will not be
needed.
For the most part, the Multilib class extension works automatically to
extend the package name from ${PN}
to ${MLPREFIX}${PN}
, where
MLPREFIX
is the particular multilib (e.g. “lib32-” or “lib64-“).
Standard variables such as
DEPENDS,
RDEPENDS,
RPROVIDES,
RRECOMMENDS,
PACKAGES, and
PACKAGES_DYNAMIC are
automatically extended by the system. If you are extending any manual
code in the recipe, you can use the ${MLPREFIX}
variable to ensure
those names are extended correctly. This automatic extension code
resides in multilib.bbclass
.
3.12.2.2 Using Multilib
After you have set up the recipes, you need to define the actual
combination of multiple libraries you want to build. You accomplish this
through your local.conf
configuration file in the
Build Directory. An example
configuration would be as follows:
MACHINE = "qemux86-64"
require conf/multilib.conf
MULTILIBS = "multilib:lib32"
DEFAULTTUNE_virtclass-multilib-lib32 = "x86"
IMAGE_INSTALL_append = "lib32-glib-2.0"
This example enables an additional library named
lib32
alongside the normal target packages. When combining these
“lib32” alternatives, the example uses “x86” for tuning. For information
on this particular tuning, see
meta/conf/machine/include/ia32/arch-ia32.inc
.
The example then includes lib32-glib-2.0
in all the images, which
illustrates one method of including a multiple library dependency. You
can use a normal image build to include this dependency, for example:
$ bitbake core-image-sato
You can also build Multilib packages specifically with a command like this:
$ bitbake lib32-glib-2.0
3.12.2.3 Additional Implementation Details
Generic implementation details as well as details that are specific to package management systems exist. Following are implementation details that exist regardless of the package management system:
The typical convention used for the class extension code as used by Multilib assumes that all package names specified in PACKAGES that contain
${PN}
have${PN}
at the start of the name. When that convention is not followed and${PN}
appears at the middle or the end of a name, problems occur.The TARGET_VENDOR value under Multilib will be extended to “-vendormlmultilib” (e.g. “-pokymllib32” for a “lib32” Multilib with Poky). The reason for this slightly unwieldy contraction is that any “-” characters in the vendor string presently break Autoconf’s
config.sub
, and other separators are problematic for different reasons.
For the RPM Package Management System, the following implementation details exist:
A unique architecture is defined for the Multilib packages, along with creating a unique deploy folder under
tmp/deploy/rpm
in the Build Directory. For example, considerlib32
in aqemux86-64
image. The possible architectures in the system are “all”, “qemux86_64”, “lib32_qemux86_64”, and “lib32_x86”.The
${MLPREFIX}
variable is stripped from${PN}
during RPM packaging. The naming for a normal RPM package and a Multilib RPM package in aqemux86-64
system resolves to something similar tobash-4.1-r2.x86_64.rpm
andbash-4.1.r2.lib32_x86.rpm
, respectively.When installing a Multilib image, the RPM backend first installs the base image and then installs the Multilib libraries.
The build system relies on RPM to resolve the identical files in the two (or more) Multilib packages.
For the IPK Package Management System, the following implementation details exist:
The
${MLPREFIX}
is not stripped from${PN}
during IPK packaging. The naming for a normal RPM package and a Multilib IPK package in aqemux86-64
system resolves to something likebash_4.1-r2.x86_64.ipk
andlib32-bash_4.1-rw_x86.ipk
, respectively.The IPK deploy folder is not modified with
${MLPREFIX}
because packages with and without the Multilib feature can exist in the same folder due to the${PN}
differences.IPK defines a sanity check for Multilib installation using certain rules for file comparison, overridden, etc.
3.12.3 Installing Multiple Versions of the Same Library
Situations can exist where you need to install and use multiple versions of the same library on the same system at the same time. These situations almost always exist when a library API changes and you have multiple pieces of software that depend on the separate versions of the library. To accommodate these situations, you can install multiple versions of the same library in parallel on the same system.
The process is straightforward as long as the libraries use proper
versioning. With properly versioned libraries, all you need to do to
individually specify the libraries is create separate, appropriately
named recipes where the PN part of
the name includes a portion that differentiates each library version
(e.g. the major part of the version number). Thus, instead of having a
single recipe that loads one version of a library (e.g. clutter
),
you provide multiple recipes that result in different versions of the
libraries you want. As an example, the following two recipes would allow
the two separate versions of the clutter
library to co-exist on the
same system:
clutter-1.6_1.6.20.bb
clutter-1.8_1.8.4.bb
Additionally, if
you have other recipes that depend on a given library, you need to use
the DEPENDS variable to
create the dependency. Continuing with the same example, if you want to
have a recipe depend on the 1.8 version of the clutter
library, use
the following in your recipe:
DEPENDS = "clutter-1.8"
3.13 Using x32 psABI
x32 processor-specific Application Binary Interface (x32 psABI) is a native 32-bit processor-specific ABI for Intel 64 (x86-64) architectures. An ABI defines the calling conventions between functions in a processing environment. The interface determines what registers are used and what the sizes are for various C data types.
Some processing environments prefer using 32-bit applications even when running on Intel 64-bit platforms. Consider the i386 psABI, which is a very old 32-bit ABI for Intel 64-bit platforms. The i386 psABI does not provide efficient use and access of the Intel 64-bit processor resources, leaving the system underutilized. Now consider the x86_64 psABI. This ABI is newer and uses 64-bits for data sizes and program pointers. The extra bits increase the footprint size of the programs, libraries, and also increases the memory and file system size requirements. Executing under the x32 psABI enables user programs to utilize CPU and system resources more efficiently while keeping the memory footprint of the applications low. Extra bits are used for registers but not for addressing mechanisms.
The Yocto Project supports the final specifications of x32 psABI as follows:
You can create packages and images in x32 psABI format on x86_64 architecture targets.
You can successfully build recipes with the x32 toolchain.
You can create and boot
core-image-minimal
andcore-image-sato
images.RPM Package Manager (RPM) support exists for x32 binaries.
Support for large images exists.
To use the x32 psABI, you need to edit your conf/local.conf
configuration file as follows:
MACHINE = "qemux86-64"
DEFAULTTUNE = "x86-64-x32"
baselib = "${@d.getVar('BASE_LIB_tune-' + (d.getVar('DEFAULTTUNE') \
or 'INVALID')) or 'lib'}"
Once you have set up your configuration file, use BitBake to build an image that supports the x32 psABI. Here is an example:
$ bitbake core-image-sato
3.14 Enabling GObject Introspection Support
GObject introspection is the standard mechanism for accessing GObject-based software from runtime environments. GObject is a feature of the GLib library that provides an object framework for the GNOME desktop and related software. GObject Introspection adds information to GObject that allows objects created within it to be represented across different programming languages. If you want to construct GStreamer pipelines using Python, or control UPnP infrastructure using Javascript and GUPnP, GObject introspection is the only way to do it.
This section describes the Yocto Project support for generating and
packaging GObject introspection data. GObject introspection data is a
description of the API provided by libraries built on top of GLib
framework, and, in particular, that framework’s GObject mechanism.
GObject Introspection Repository (GIR) files go to -dev
packages,
typelib
files go to main packages as they are packaged together with
libraries that are introspected.
The data is generated when building such a library, by linking the library with a small executable binary that asks the library to describe itself, and then executing the binary and processing its output.
Generating this data in a cross-compilation environment is difficult because the library is produced for the target architecture, but its code needs to be executed on the build host. This problem is solved with the OpenEmbedded build system by running the code through QEMU, which allows precisely that. Unfortunately, QEMU does not always work perfectly as mentioned in the “Known Issues” section.
3.14.1 Enabling the Generation of Introspection Data
Enabling the generation of introspection data (GIR files) in your library package involves the following:
Inherit the gobject-introspection class.
Make sure introspection is not disabled anywhere in the recipe or from anything the recipe includes. Also, make sure that “gobject-introspection-data” is not in DISTRO_FEATURES_BACKFILL_CONSIDERED and that “qemu-usermode” is not in MACHINE_FEATURES_BACKFILL_CONSIDERED. If either of these conditions exist, nothing will happen.
Try to build the recipe. If you encounter build errors that look like something is unable to find
.so
libraries, check where these libraries are located in the source tree and add the following to the recipe:GIR_EXTRA_LIBS_PATH = "${B}/something/.libs"
Note
See recipes in the
oe-core
repository that use thatGIR_EXTRA_LIBS_PATH
variable as an example.Look for any other errors, which probably mean that introspection support in a package is not entirely standard, and thus breaks down in a cross-compilation environment. For such cases, custom-made fixes are needed. A good place to ask and receive help in these cases is the Yocto Project mailing lists.
Note
Using a library that no longer builds against the latest Yocto Project release and prints introspection related errors is a good candidate for the previous procedure.
3.14.2 Disabling the Generation of Introspection Data
You might find that you do not want to generate introspection data. Or, perhaps QEMU does not work on your build host and target architecture combination. If so, you can use either of the following methods to disable GIR file generations:
Add the following to your distro configuration:
DISTRO_FEATURES_BACKFILL_CONSIDERED = "gobject-introspection-data"
Adding this statement disables generating introspection data using QEMU but will still enable building introspection tools and libraries (i.e. building them does not require the use of QEMU).
Add the following to your machine configuration:
MACHINE_FEATURES_BACKFILL_CONSIDERED = "qemu-usermode"
Adding this statement disables the use of QEMU when building packages for your machine. Currently, this feature is used only by introspection recipes and has the same effect as the previously described option.
Note
Future releases of the Yocto Project might have other features affected by this option.
If you disable introspection data, you can still obtain it through other means such as copying the data from a suitable sysroot, or by generating it on the target hardware. The OpenEmbedded build system does not currently provide specific support for these techniques.
3.14.3 Testing that Introspection Works in an Image
Use the following procedure to test if generating introspection data is working in an image:
Make sure that “gobject-introspection-data” is not in DISTRO_FEATURES_BACKFILL_CONSIDERED and that “qemu-usermode” is not in MACHINE_FEATURES_BACKFILL_CONSIDERED.
Build
core-image-sato
.Launch a Terminal and then start Python in the terminal.
Enter the following in the terminal:
>>> from gi.repository import GLib >>> GLib.get_host_name()
For something a little more advanced, enter the following see: https://python-gtk-3-tutorial.readthedocs.io/en/latest/introduction.html
3.14.4 Known Issues
The following know issues exist for GObject Introspection Support:
qemu-ppc64
immediately crashes. Consequently, you cannot build introspection data on that architecture.x32 is not supported by QEMU. Consequently, introspection data is disabled.
musl causes transient GLib binaries to crash on assertion failures. Consequently, generating introspection data is disabled.
Because QEMU is not able to run the binaries correctly, introspection is disabled for some specific packages under specific architectures (e.g.
gcr
,libsecret
, andwebkit
).QEMU usermode might not work properly when running 64-bit binaries under 32-bit host machines. In particular, “qemumips64” is known to not work under i686.
3.15 Optionally Using an External Toolchain
You might want to use an external toolchain as part of your development. If this is the case, the fundamental steps you need to accomplish are as follows:
Understand where the installed toolchain resides. For cases where you need to build the external toolchain, you would need to take separate steps to build and install the toolchain.
Make sure you add the layer that contains the toolchain to your
bblayers.conf
file through the BBLAYERS variable.Set the
EXTERNAL_TOOLCHAIN
variable in yourlocal.conf
file to the location in which you installed the toolchain.
A good example of an external toolchain used with the Yocto Project is
Mentor Graphics Sourcery G++ Toolchain. You can see information on how
to use that particular layer in the README
file at
https://github.com/MentorEmbedded/meta-sourcery/. You can find
further information by reading about the
TCMODE variable in the Yocto
Project Reference Manual’s variable glossary.
3.16 Creating Partitioned Images Using Wic
Creating an image for a particular hardware target using the OpenEmbedded build system does not necessarily mean you can boot that image as is on your device. Physical devices accept and boot images in various ways depending on the specifics of the device. Usually, information about the hardware can tell you what image format the device requires. Should your device require multiple partitions on an SD card, flash, or an HDD, you can use the OpenEmbedded Image Creator, Wic, to create the properly partitioned image.
The wic
command generates partitioned images from existing
OpenEmbedded build artifacts. Image generation is driven by partitioning
commands contained in an Openembedded kickstart file (.wks
)
specified either directly on the command line or as one of a selection
of canned kickstart files as shown with the wic list images
command
in the
“Generate an Image using an Existing Kickstart File”
section. When you apply the command to a given set of build artifacts, the
result is an image or set of images that can be directly written onto media and
used on a particular system.
Note
For a kickstart file reference, see the “OpenEmbedded Kickstart (.wks) Reference” Chapter in the Yocto Project Reference Manual.
The wic
command and the infrastructure it is based on is by
definition incomplete. The purpose of the command is to allow the
generation of customized images, and as such, was designed to be
completely extensible through a plugin interface. See the
“Using the Wic Plugin Interface” section
for information on these plugins.
This section provides some background information on Wic, describes what you need to have in place to run the tool, provides instruction on how to use the Wic utility, provides information on using the Wic plugins interface, and provides several examples that show how to use Wic.
3.16.1 Background
This section provides some background on the Wic utility. While none of this information is required to use Wic, you might find it interesting.
The name “Wic” is derived from OpenEmbedded Image Creator (oeic). The “oe” diphthong in “oeic” was promoted to the letter “w”, because “oeic” is both difficult to remember and to pronounce.
Wic is loosely based on the Meego Image Creator (
mic
) framework. The Wic implementation has been heavily modified to make direct use of OpenEmbedded build artifacts instead of package installation and configuration, which are already incorporated within the OpenEmbedded artifacts.Wic is a completely independent standalone utility that initially provides easier-to-use and more flexible replacements for an existing functionality in OE-Core’s image-live class. The difference between Wic and those examples is that with Wic the functionality of those scripts is implemented by a general-purpose partitioning language, which is based on Redhat kickstart syntax.
3.16.2 Requirements
In order to use the Wic utility with the OpenEmbedded Build system, your system needs to meet the following requirements:
The Linux distribution on your development host must support the Yocto Project. See the “Supported Linux Distributions” section in the Yocto Project Reference Manual for the list of distributions that support the Yocto Project.
The standard system utilities, such as
cp
, must be installed on your development host system.You must have sourced the build environment setup script (i.e. oe-init-build-env) found in the Build Directory.
You need to have the build artifacts already available, which typically means that you must have already created an image using the Openembedded build system (e.g.
core-image-minimal
). While it might seem redundant to generate an image in order to create an image using Wic, the current version of Wic requires the artifacts in the form generated by the OpenEmbedded build system.You must build several native tools, which are built to run on the build system:
$ bitbake parted-native dosfstools-native mtools-native
Include “wic” as part of the IMAGE_FSTYPES variable.
Include the name of the wic kickstart file as part of the WKS_FILE variable
3.16.3 Getting Help
You can get general help for the wic
command by entering the wic
command by itself or by entering the command with a help argument as
follows:
$ wic -h
$ wic --help
$ wic help
Currently, Wic supports seven commands: cp
, create
, help
,
list
, ls
, rm
, and write
. You can get help for all these
commands except “help” by using the following form:
$ wic help command
For example, the following command returns help for the write
command:
$ wic help write
Wic supports help for three topics: overview
, plugins
, and
kickstart
. You can get help for any topic using the following form:
$ wic help topic
For example, the following returns overview help for Wic:
$ wic help overview
One additional level of help exists for Wic. You can get help on
individual images through the list
command. You can use the list
command to return the available Wic images as follows:
$ wic list images
genericx86 Create an EFI disk image for genericx86*
beaglebone-yocto Create SD card image for Beaglebone
edgerouter Create SD card image for Edgerouter
qemux86-directdisk Create a QEMU machine 'pcbios' direct disk image
directdisk-gpt Create a 'pcbios' direct disk image
mkefidisk Create an EFI disk image
directdisk Create a 'pcbios' direct disk image
systemd-bootdisk Create an EFI disk image with systemd-boot
mkhybridiso Create a hybrid ISO image
sdimage-bootpart Create SD card image with a boot partition
directdisk-multi-rootfs Create multi rootfs image using rootfs plugin
directdisk-bootloader-config Create a 'pcbios' direct disk image with custom bootloader config
Once you know the list of available
Wic images, you can use help
with the command to get help on a
particular image. For example, the following command returns help on the
“beaglebone-yocto” image:
$ wic list beaglebone-yocto help
Creates a partitioned SD card image for Beaglebone.
Boot files are located in the first vfat partition.
3.16.4 Operational Modes
You can use Wic in two different modes, depending on how much control you need for specifying the Openembedded build artifacts that are used for creating the image: Raw and Cooked:
Raw Mode: You explicitly specify build artifacts through Wic command-line arguments.
Cooked Mode: The current MACHINE setting and image name are used to automatically locate and provide the build artifacts. You just supply a kickstart file and the name of the image from which to use artifacts.
Regardless of the mode you use, you need to have the build artifacts ready and available.
3.16.4.1 Raw Mode
Running Wic in raw mode allows you to specify all the partitions through
the wic
command line. The primary use for raw mode is if you have
built your kernel outside of the Yocto Project
Build Directory. In other words, you
can point to arbitrary kernel, root filesystem locations, and so forth.
Contrast this behavior with cooked mode where Wic looks in the Build
Directory (e.g. tmp/deploy/images/
machine).
The general form of the wic
command in raw mode is:
$ wic create wks_file options ...
Where:
wks_file:
An OpenEmbedded kickstart file. You can provide
your own custom file or use a file from a set of
existing files as described by further options.
optional arguments:
-h, --help show this help message and exit
-o OUTDIR, --outdir OUTDIR
name of directory to create image in
-e IMAGE_NAME, --image-name IMAGE_NAME
name of the image to use the artifacts from e.g. core-
image-sato
-r ROOTFS_DIR, --rootfs-dir ROOTFS_DIR
path to the /rootfs dir to use as the .wks rootfs
source
-b BOOTIMG_DIR, --bootimg-dir BOOTIMG_DIR
path to the dir containing the boot artifacts (e.g.
/EFI or /syslinux dirs) to use as the .wks bootimg
source
-k KERNEL_DIR, --kernel-dir KERNEL_DIR
path to the dir containing the kernel to use in the
.wks bootimg
-n NATIVE_SYSROOT, --native-sysroot NATIVE_SYSROOT
path to the native sysroot containing the tools to use
to build the image
-s, --skip-build-check
skip the build check
-f, --build-rootfs build rootfs
-c {gzip,bzip2,xz}, --compress-with {gzip,bzip2,xz}
compress image with specified compressor
-m, --bmap generate .bmap
--no-fstab-update Do not change fstab file.
-v VARS_DIR, --vars VARS_DIR
directory with <image>.env files that store bitbake
variables
-D, --debug output debug information
Note
You do not need root privileges to run Wic. In fact, you should not run as root when using the utility.
3.16.4.2 Cooked Mode
Running Wic in cooked mode leverages off artifacts in the Build
Directory. In other words, you do not have to specify kernel or root
filesystem locations as part of the command. All you need to provide is
a kickstart file and the name of the image from which to use artifacts
by using the “-e” option. Wic looks in the Build Directory (e.g.
tmp/deploy/images/
machine) for artifacts.
The general form of the wic
command using Cooked Mode is as follows:
$ wic create wks_file -e IMAGE_NAME
Where:
wks_file:
An OpenEmbedded kickstart file. You can provide
your own custom file or use a file from a set of
existing files provided with the Yocto Project
release.
required argument:
-e IMAGE_NAME, --image-name IMAGE_NAME
name of the image to use the artifacts from e.g. core-
image-sato
3.16.5 Using an Existing Kickstart File
If you do not want to create your own kickstart file, you can use an existing file provided by the Wic installation. As shipped, kickstart files can be found in the Yocto Project Source Repositories in the following two locations:
poky/meta-yocto-bsp/wic
poky/scripts/lib/wic/canned-wks
Use the following command to list the available kickstart files:
$ wic list images
genericx86 Create an EFI disk image for genericx86*
beaglebone-yocto Create SD card image for Beaglebone
edgerouter Create SD card image for Edgerouter
qemux86-directdisk Create a QEMU machine 'pcbios' direct disk image
directdisk-gpt Create a 'pcbios' direct disk image
mkefidisk Create an EFI disk image
directdisk Create a 'pcbios' direct disk image
systemd-bootdisk Create an EFI disk image with systemd-boot
mkhybridiso Create a hybrid ISO image
sdimage-bootpart Create SD card image with a boot partition
directdisk-multi-rootfs Create multi rootfs image using rootfs plugin
directdisk-bootloader-config Create a 'pcbios' direct disk image with custom bootloader config
When you use an existing file, you
do not have to use the .wks
extension. Here is an example in Raw
Mode that uses the directdisk
file:
$ wic create directdisk -r rootfs_dir -b bootimg_dir \
-k kernel_dir -n native_sysroot
Here are the actual partition language commands used in the
genericx86.wks
file to generate an image:
# short-description: Create an EFI disk image for genericx86*
# long-description: Creates a partitioned EFI disk image for genericx86* machines
part /boot --source bootimg-efi --sourceparams="loader=grub-efi" --ondisk sda --label msdos --active --align 1024
part / --source rootfs --ondisk sda --fstype=ext4 --label platform --align 1024 --use-uuid
part swap --ondisk sda --size 44 --label swap1 --fstype=swap
bootloader --ptable gpt --timeout=5 --append="rootfstype=ext4 console=ttyS0,115200 console=tty0"
3.16.6 Using the Wic Plugin Interface
You can extend and specialize Wic functionality by using Wic plugins. This section explains the Wic plugin interface.
Note
Wic plugins consist of “source” and “imager” plugins. Imager plugins are beyond the scope of this section.
Source plugins provide a mechanism to customize partition content during
the Wic image generation process. You can use source plugins to map
values that you specify using --source
commands in kickstart files
(i.e. *.wks
) to a plugin implementation used to populate a given
partition.
Note
If you use plugins that have build-time dependencies (e.g. native tools, bootloaders, and so forth) when building a Wic image, you need to specify those dependencies using the WKS_FILE_DEPENDS variable.
Source plugins are subclasses defined in plugin files. As shipped, the Yocto Project provides several plugin files. You can see the source plugin files that ship with the Yocto Project here. Each of these plugin files contains source plugins that are designed to populate a specific Wic image partition.
Source plugins are subclasses of the SourcePlugin
class, which is
defined in the poky/scripts/lib/wic/pluginbase.py
file. For example,
the BootimgEFIPlugin
source plugin found in the bootimg-efi.py
file is a subclass of the SourcePlugin
class, which is found in the
pluginbase.py
file.
You can also implement source plugins in a layer outside of the Source
Repositories (external layer). To do so, be sure that your plugin files
are located in a directory whose path is
scripts/lib/wic/plugins/source/
within your external layer. When the
plugin files are located there, the source plugins they contain are made
available to Wic.
When the Wic implementation needs to invoke a partition-specific
implementation, it looks for the plugin with the same name as the
--source
parameter used in the kickstart file given to that
partition. For example, if the partition is set up using the following
command in a kickstart file:
part /boot --source bootimg-pcbios --ondisk sda --label boot --active --align 1024
The methods defined as class
members of the matching source plugin (i.e. bootimg-pcbios
) in the
bootimg-pcbios.py
plugin file are used.
To be more concrete, here is the corresponding plugin definition from
the bootimg-pcbios.py
file for the previous command along with an
example method called by the Wic implementation when it needs to prepare
a partition using an implementation-specific function:
.
.
.
class BootimgPcbiosPlugin(SourcePlugin):
"""
Create MBR boot partition and install syslinux on it.
"""
name = 'bootimg-pcbios'
.
.
.
@classmethod
def do_prepare_partition(cls, part, source_params, creator, cr_workdir,
oe_builddir, bootimg_dir, kernel_dir,
rootfs_dir, native_sysroot):
"""
Called to do the actual content population for a partition i.e. it
'prepares' the partition to be incorporated into the image.
In this case, prepare content for legacy bios boot partition.
"""
.
.
.
If a
subclass (plugin) itself does not implement a particular function, Wic
locates and uses the default version in the superclass. It is for this
reason that all source plugins are derived from the SourcePlugin
class.
The SourcePlugin
class defined in the pluginbase.py
file defines
a set of methods that source plugins can implement or override. Any
plugins (subclass of SourcePlugin
) that do not implement a
particular method inherit the implementation of the method from the
SourcePlugin
class. For more information, see the SourcePlugin
class in the pluginbase.py
file for details:
The following list describes the methods implemented in the
SourcePlugin
class:
do_prepare_partition()
: Called to populate a partition with actual content. In other words, the method prepares the final partition image that is incorporated into the disk image.do_configure_partition()
: Called beforedo_prepare_partition()
to create custom configuration files for a partition (e.g. syslinux or grub configuration files).do_install_disk()
: Called after all partitions have been prepared and assembled into a disk image. This method provides a hook to allow finalization of a disk image (e.g. writing an MBR).do_stage_partition()
: Special content-staging hook called beforedo_prepare_partition()
. This method is normally empty.Typically, a partition just uses the passed-in parameters (e.g. the unmodified value of
bootimg_dir
). However, in some cases, things might need to be more tailored. As an example, certain files might additionally need to be taken frombootimg_dir + /boot
. This hook allows those files to be staged in a customized fashion.Note
get_bitbake_var()
allows you to access non-standard variables that you might want to use for this behavior.
You can extend the source plugin mechanism. To add more hooks, create
more source plugin methods within SourcePlugin
and the corresponding
derived subclasses. The code that calls the plugin methods uses the
plugin.get_source_plugin_methods()
function to find the method or
methods needed by the call. Retrieval of those methods is accomplished
by filling up a dict with keys that contain the method names of
interest. On success, these will be filled in with the actual methods.
See the Wic implementation for examples and details.
3.16.7 Wic Examples
This section provides several examples that show how to use the Wic
utility. All the examples assume the list of requirements in the
“Requirements” section have been met. The
examples assume the previously generated image is
core-image-minimal
.
3.16.7.1 Generate an Image using an Existing Kickstart File
This example runs in Cooked Mode and uses the mkefidisk
kickstart
file:
$ wic create mkefidisk -e core-image-minimal
INFO: Building wic-tools...
.
.
.
INFO: The new image(s) can be found here:
./mkefidisk-201804191017-sda.direct
The following build artifacts were used to create the image(s):
ROOTFS_DIR: /home/stephano/build/master/build/tmp-glibc/work/qemux86-oe-linux/core-image-minimal/1.0-r0/rootfs
BOOTIMG_DIR: /home/stephano/build/master/build/tmp-glibc/work/qemux86-oe-linux/core-image-minimal/1.0-r0/recipe-sysroot/usr/share
KERNEL_DIR: /home/stephano/build/master/build/tmp-glibc/deploy/images/qemux86
NATIVE_SYSROOT: /home/stephano/build/master/build/tmp-glibc/work/i586-oe-linux/wic-tools/1.0-r0/recipe-sysroot-native
INFO: The image(s) were created using OE kickstart file:
/home/stephano/build/master/openembedded-core/scripts/lib/wic/canned-wks/mkefidisk.wks
The previous example shows the easiest way to create an image by running
in cooked mode and supplying a kickstart file and the “-e” option to
point to the existing build artifacts. Your local.conf
file needs to
have the MACHINE variable set
to the machine you are using, which is “qemux86” in this example.
Once the image builds, the output provides image location, artifact use, and kickstart file information.
Note
You should always verify the details provided in the output to make sure that the image was indeed created exactly as expected.
Continuing with the example, you can now write the image from the Build
Directory onto a USB stick, or whatever media for which you built your
image, and boot from the media. You can write the image by using
bmaptool
or dd
:
$ oe-run-native bmaptool copy mkefidisk-201804191017-sda.direct /dev/sdX
or
$ sudo dd if=mkefidisk-201804191017-sda.direct of=/dev/sdX
Note
For more information on how to use the bmaptool
to flash a device with an image, see the
“Flashing Images Using bmaptool”
section.
3.16.7.2 Using a Modified Kickstart File
Because partitioned image creation is driven by the kickstart file, it
is easy to affect image creation by changing the parameters in the file.
This next example demonstrates that through modification of the
directdisk-gpt
kickstart file.
As mentioned earlier, you can use the command wic list images
to
show the list of existing kickstart files. The directory in which the
directdisk-gpt.wks
file resides is
scripts/lib/image/canned-wks/
, which is located in the
Source Directory (e.g. poky
).
Because available files reside in this directory, you can create and add
your own custom files to the directory. Subsequent use of the
wic list images
command would then include your kickstart files.
In this example, the existing directdisk-gpt
file already does most
of what is needed. However, for the hardware in this example, the image
will need to boot from sdb
instead of sda
, which is what the
directdisk-gpt
kickstart file uses.
The example begins by making a copy of the directdisk-gpt.wks
file
in the scripts/lib/image/canned-wks
directory and then by changing
the lines that specify the target disk from which to boot.
$ cp /home/stephano/poky/scripts/lib/wic/canned-wks/directdisk-gpt.wks \
/home/stephano/poky/scripts/lib/wic/canned-wks/directdisksdb-gpt.wks
Next, the example modifies the directdisksdb-gpt.wks
file and
changes all instances of “--ondisk sda
” to “--ondisk sdb
”. The
example changes the following two lines and leaves the remaining lines
untouched:
part /boot --source bootimg-pcbios --ondisk sdb --label boot --active --align 1024
part / --source rootfs --ondisk sdb --fstype=ext4 --label platform --align 1024 --use-uuid
Once the lines are changed, the
example generates the directdisksdb-gpt
image. The command points
the process at the core-image-minimal
artifacts for the Next Unit of
Computing (nuc) MACHINE the
local.conf
.
$ wic create directdisksdb-gpt -e core-image-minimal
INFO: Building wic-tools...
.
.
.
Initialising tasks: 100% |#######################################| Time: 0:00:01
NOTE: Executing SetScene Tasks
NOTE: Executing RunQueue Tasks
NOTE: Tasks Summary: Attempted 1161 tasks of which 1157 didn't need to be rerun and all succeeded.
INFO: Creating image(s)...
INFO: The new image(s) can be found here:
./directdisksdb-gpt-201710090938-sdb.direct
The following build artifacts were used to create the image(s):
ROOTFS_DIR: /home/stephano/build/master/build/tmp-glibc/work/qemux86-oe-linux/core-image-minimal/1.0-r0/rootfs
BOOTIMG_DIR: /home/stephano/build/master/build/tmp-glibc/work/qemux86-oe-linux/core-image-minimal/1.0-r0/recipe-sysroot/usr/share
KERNEL_DIR: /home/stephano/build/master/build/tmp-glibc/deploy/images/qemux86
NATIVE_SYSROOT: /home/stephano/build/master/build/tmp-glibc/work/i586-oe-linux/wic-tools/1.0-r0/recipe-sysroot-native
INFO: The image(s) were created using OE kickstart file:
/home/stephano/poky/scripts/lib/wic/canned-wks/directdisksdb-gpt.wks
Continuing with the example, you can now directly dd
the image to a
USB stick, or whatever media for which you built your image, and boot
the resulting media:
$ sudo dd if=directdisksdb-gpt-201710090938-sdb.direct of=/dev/sdb
140966+0 records in
140966+0 records out
72174592 bytes (72 MB, 69 MiB) copied, 78.0282 s, 925 kB/s
$ sudo eject /dev/sdb
3.16.7.3 Using a Modified Kickstart File and Running in Raw Mode
This next example manually specifies each build artifact (runs in Raw
Mode) and uses a modified kickstart file. The example also uses the
-o
option to cause Wic to create the output somewhere other than the
default output directory, which is the current directory:
$ wic create /home/stephano/my_yocto/test.wks -o /home/stephano/testwic \
--rootfs-dir /home/stephano/build/master/build/tmp/work/qemux86-poky-linux/core-image-minimal/1.0-r0/rootfs \
--bootimg-dir /home/stephano/build/master/build/tmp/work/qemux86-poky-linux/core-image-minimal/1.0-r0/recipe-sysroot/usr/share \
--kernel-dir /home/stephano/build/master/build/tmp/deploy/images/qemux86 \
--native-sysroot /home/stephano/build/master/build/tmp/work/i586-poky-linux/wic-tools/1.0-r0/recipe-sysroot-native
INFO: Creating image(s)...
INFO: The new image(s) can be found here:
/home/stephano/testwic/test-201710091445-sdb.direct
The following build artifacts were used to create the image(s):
ROOTFS_DIR: /home/stephano/build/master/build/tmp-glibc/work/qemux86-oe-linux/core-image-minimal/1.0-r0/rootfs
BOOTIMG_DIR: /home/stephano/build/master/build/tmp-glibc/work/qemux86-oe-linux/core-image-minimal/1.0-r0/recipe-sysroot/usr/share
KERNEL_DIR: /home/stephano/build/master/build/tmp-glibc/deploy/images/qemux86
NATIVE_SYSROOT: /home/stephano/build/master/build/tmp-glibc/work/i586-oe-linux/wic-tools/1.0-r0/recipe-sysroot-native
INFO: The image(s) were created using OE kickstart file:
/home/stephano/my_yocto/test.wks
For this example,
MACHINE did not have to be
specified in the local.conf
file since the artifact is manually
specified.
3.16.7.4 Using Wic to Manipulate an Image
Wic image manipulation allows you to shorten turnaround time during image development. For example, you can use Wic to delete the kernel partition of a Wic image and then insert a newly built kernel. This saves you time from having to rebuild the entire image each time you modify the kernel.
Note
In order to use Wic to manipulate a Wic image as in this example,
your development machine must have the mtools
package installed.
The following example examines the contents of the Wic image, deletes the existing kernel, and then inserts a new kernel:
List the Partitions: Use the
wic ls
command to list all the partitions in the Wic image:$ wic ls tmp/deploy/images/qemux86/core-image-minimal-qemux86.wic Num Start End Size Fstype 1 1048576 25041919 23993344 fat16 2 25165824 72157183 46991360 ext4
The previous output shows two partitions in the
core-image-minimal-qemux86.wic
image.Examine a Particular Partition: Use the
wic ls
command again but in a different form to examine a particular partition.Note
You can get command usage on any Wic command using the following form:
$ wic help command
For example, the following command shows you the various ways to use the wic ls command:
$ wic help ls
The following command shows what is in partition one:
$ wic ls tmp/deploy/images/qemux86/core-image-minimal-qemux86.wic:1 Volume in drive : is boot Volume Serial Number is E894-1809 Directory for ::/ libcom32 c32 186500 2017-10-09 16:06 libutil c32 24148 2017-10-09 16:06 syslinux cfg 220 2017-10-09 16:06 vesamenu c32 27104 2017-10-09 16:06 vmlinuz 6904608 2017-10-09 16:06 5 files 7 142 580 bytes 16 582 656 bytes free
The previous output shows five files, with the
vmlinuz
being the kernel.Note
If you see the following error, you need to update or create a
~/.mtoolsrc
file and be sure to have the line “mtools_skip_check=1” in the file. Then, run the Wic command again:ERROR: _exec_cmd: /usr/bin/mdir -i /tmp/wic-parttfokuwra ::/ returned '1' instead of 0 output: Total number of sectors (47824) not a multiple of sectors per track (32)! Add mtools_skip_check=1 to your .mtoolsrc file to skip this test
Remove the Old Kernel: Use the
wic rm
command to remove thevmlinuz
file (kernel):$ wic rm tmp/deploy/images/qemux86/core-image-minimal-qemux86.wic:1/vmlinuz
Add In the New Kernel: Use the
wic cp
command to add the updated kernel to the Wic image. Depending on how you built your kernel, it could be in different places. If you useddevtool
and an SDK to build your kernel, it resides in thetmp/work
directory of the extensible SDK. If you usedmake
to build the kernel, the kernel will be in theworkspace/sources
area.The following example assumes
devtool
was used to build the kernel:$ wic cp poky_sdk/tmp/work/qemux86-poky-linux/linux-yocto/4.12.12+git999-r0/linux-yocto-4.12.12+git999/arch/x86/boot/bzImage \ poky/build/tmp/deploy/images/qemux86/core-image-minimal-qemux86.wic:1/vmlinuz
Once the new kernel is added back into the image, you can use the
dd
command or bmaptool to flash your wic image onto an SD card or USB stick and test your target.Note
Using
bmaptool
is generally 10 to 20 times faster than usingdd
.
3.17 Flashing Images Using bmaptool
A fast and easy way to flash an image to a bootable device is to use Bmaptool, which is integrated into the OpenEmbedded build system. Bmaptool is a generic tool that creates a file’s block map (bmap) and then uses that map to copy the file. As compared to traditional tools such as dd or cp, Bmaptool can copy (or flash) large files like raw system image files much faster.
Note
If you are using Ubuntu or Debian distributions, you can install the
bmap-tools
package using the following command and then use the tool without specifyingPATH
even from the root account:$ sudo apt-get install bmap-tools
If you are unable to install the
bmap-tools
package, you will need to build Bmaptool before using it. Use the following command:$ bitbake bmap-tools-native
Following, is an example that shows how to flash a Wic image. Realize that while this example uses a Wic image, you can use Bmaptool to flash any type of image. Use these steps to flash an image using Bmaptool:
Update your local.conf File: You need to have the following set in your
local.conf
file before building your image:IMAGE_FSTYPES += "wic wic.bmap"
Get Your Image: Either have your image ready (pre-built with the IMAGE_FSTYPES setting previously mentioned) or take the step to build the image:
$ bitbake image
Flash the Device: Flash the device with the image by using Bmaptool depending on your particular setup. The following commands assume the image resides in the Build Directory’s
deploy/images/
area:If you have write access to the media, use this command form:
$ oe-run-native bmap-tools-native bmaptool copy build-directory/tmp/deploy/images/machine/image.wic /dev/sdX
If you do not have write access to the media, set your permissions first and then use the same command form:
$ sudo chmod 666 /dev/sdX $ oe-run-native bmap-tools-native bmaptool copy build-directory/tmp/deploy/images/machine/image.wic /dev/sdX
For help on the bmaptool
command, use the following command:
$ bmaptool --help
3.18 Making Images More Secure
Security is of increasing concern for embedded devices. Consider the issues and problems discussed in just this sampling of work found across the Internet:
“Security Risks of Embedded Systems“ by Bruce Schneier
“Internet Census 2012“ by Carna Botnet
“Security Issues for Embedded Devices“ by Jake Edge
When securing your image is of concern, there are steps, tools, and variables that you can consider to help you reach the security goals you need for your particular device. Not all situations are identical when it comes to making an image secure. Consequently, this section provides some guidance and suggestions for consideration when you want to make your image more secure.
Note
Because the security requirements and risks are different for every type of device, this section cannot provide a complete reference on securing your custom OS. It is strongly recommended that you also consult other sources of information on embedded Linux system hardening and on security.
3.18.1 General Considerations
General considerations exist that help you create more secure images. You should consider the following suggestions to help make your device more secure:
Scan additional code you are adding to the system (e.g. application code) by using static analysis tools. Look for buffer overflows and other potential security problems.
Pay particular attention to the security for any web-based administration interface.
Web interfaces typically need to perform administrative functions and tend to need to run with elevated privileges. Thus, the consequences resulting from the interface’s security becoming compromised can be serious. Look for common web vulnerabilities such as cross-site-scripting (XSS), unvalidated inputs, and so forth.
As with system passwords, the default credentials for accessing a web-based interface should not be the same across all devices. This is particularly true if the interface is enabled by default as it can be assumed that many end-users will not change the credentials.
Ensure you can update the software on the device to mitigate vulnerabilities discovered in the future. This consideration especially applies when your device is network-enabled.
Ensure you remove or disable debugging functionality before producing the final image. For information on how to do this, see the “Considerations Specific to the OpenEmbedded Build System” section.
Ensure you have no network services listening that are not needed.
Remove any software from the image that is not needed.
Enable hardware support for secure boot functionality when your device supports this functionality.
3.18.2 Security Flags
The Yocto Project has security flags that you can enable that help make
your build output more secure. The security flags are in the
meta/conf/distro/include/security_flags.inc
file in your
Source Directory (e.g. poky
).
Note
Depending on the recipe, certain security flags are enabled and disabled by default.
Use the following line in your local.conf
file or in your custom
distribution configuration file to enable the security compiler and
linker flags for your build:
require conf/distro/include/security_flags.inc
3.18.3 Considerations Specific to the OpenEmbedded Build System
You can take some steps that are specific to the OpenEmbedded build system to make your images more secure:
Ensure “debug-tweaks” is not one of your selected IMAGE_FEATURES. When creating a new project, the default is to provide you with an initial
local.conf
file that enables this feature using the EXTRA_IMAGE_FEATURES variable with the line:EXTRA_IMAGE_FEATURES = "debug-tweaks"
To disable that feature, simply comment out that line in your
local.conf
file, or make sureIMAGE_FEATURES
does not contain “debug-tweaks” before producing your final image. Among other things, leaving this in place sets the root password as blank, which makes logging in for debugging or inspection easy during development but also means anyone can easily log in during production.It is possible to set a root password for the image and also to set passwords for any extra users you might add (e.g. administrative or service type users). When you set up passwords for multiple images or users, you should not duplicate passwords.
To set up passwords, use the extrausers class, which is the preferred method. For an example on how to set up both root and user passwords, see the “extrausers.bbclass” section.
Note
When adding extra user accounts or setting a root password, be cautious about setting the same password on every device. If you do this, and the password you have set is exposed, then every device is now potentially compromised. If you need this access but want to ensure security, consider setting a different, random password for each device. Typically, you do this as a separate step after you deploy the image onto the device.
Consider enabling a Mandatory Access Control (MAC) framework such as SMACK or SELinux and tuning it appropriately for your device’s usage. You can find more information in the meta-selinux layer.
3.18.4 Tools for Hardening Your Image
The Yocto Project provides tools for making your image more secure. You
can find these tools in the meta-security
layer of the
Yocto Project Source Repositories.
3.19 Creating Your Own Distribution
When you build an image using the Yocto Project and do not alter any distribution Metadata, you are creating a Poky distribution. If you wish to gain more control over package alternative selections, compile-time options, and other low-level configurations, you can create your own distribution.
To create your own distribution, the basic steps consist of creating your own distribution layer, creating your own distribution configuration file, and then adding any needed code and Metadata to the layer. The following steps provide some more detail:
Create a layer for your new distro: Create your distribution layer so that you can keep your Metadata and code for the distribution separate. It is strongly recommended that you create and use your own layer for configuration and code. Using your own layer as compared to just placing configurations in a
local.conf
configuration file makes it easier to reproduce the same build configuration when using multiple build machines. See the “Creating a General Layer Using the bitbake-layers Script” section for information on how to quickly set up a layer.Create the distribution configuration file: The distribution configuration file needs to be created in the
conf/distro
directory of your layer. You need to name it using your distribution name (e.g.mydistro.conf
).Note
The DISTRO variable in your
local.conf
file determines the name of your distribution.You can split out parts of your configuration file into include files and then “require” them from within your distribution configuration file. Be sure to place the include files in the
conf/distro/include
directory of your layer. A common example usage of include files would be to separate out the selection of desired version and revisions for individual recipes.Your configuration file needs to set the following required variables:
These following variables are optional and you typically set them from the distribution configuration file:
Tip
If you want to base your distribution configuration file on the very basic configuration from OE-Core, you can use
conf/distro/defaultsetup.conf
as a reference and just include variables that differ as compared todefaultsetup.conf
. Alternatively, you can create a distribution configuration file from scratch using thedefaultsetup.conf
file or configuration files from other distributions such as Poky or Angstrom as references.Provide miscellaneous variables: Be sure to define any other variables for which you want to create a default or enforce as part of the distribution configuration. You can include nearly any variable from the
local.conf
file. The variables you use are not limited to the list in the previous bulleted item.Point to Your distribution configuration file: In your
local.conf
file in the Build Directory, set your DISTRO variable to point to your distribution’s configuration file. For example, if your distribution’s configuration file is namedmydistro.conf
, then you point to it as follows:DISTRO = "mydistro"
Add more to the layer if necessary: Use your layer to hold other information needed for the distribution:
Add recipes for installing distro-specific configuration files that are not already installed by another recipe. If you have distro-specific configuration files that are included by an existing recipe, you should add an append file (
.bbappend
) for those. For general information and recommendations on how to add recipes to your layer, see the “Creating Your Own Layer” and “Following Best Practices When Creating Layers” sections.Add any image recipes that are specific to your distribution.
Add a
psplash
append file for a branded splash screen. For information on append files, see the “Using .bbappend Files in Your Layer” section.Add any other append files to make custom changes that are specific to individual recipes.
3.20 Creating a Custom Template Configuration Directory
If you are producing your own customized version of the build system for
use by other users, you might want to customize the message shown by the
setup script or you might want to change the template configuration
files (i.e. local.conf
and bblayers.conf
) that are created in a
new build directory.
The OpenEmbedded build system uses the environment variable
TEMPLATECONF
to locate the directory from which it gathers
configuration information that ultimately ends up in the
Build Directory conf
directory.
By default, TEMPLATECONF
is set as follows in the poky
repository:
TEMPLATECONF=${TEMPLATECONF:-meta-poky/conf}
This is the
directory used by the build system to find templates from which to build
some key configuration files. If you look at this directory, you will
see the bblayers.conf.sample
, local.conf.sample
, and
conf-notes.txt
files. The build system uses these files to form the
respective bblayers.conf
file, local.conf
file, and display the
list of BitBake targets when running the setup script.
To override these default configuration files with configurations you
want used within every new Build Directory, simply set the
TEMPLATECONF
variable to your directory. The TEMPLATECONF
variable is set in the .templateconf
file, which is in the top-level
Source Directory folder
(e.g. poky
). Edit the .templateconf
so that it can locate your
directory.
Best practices dictate that you should keep your template configuration
directory in your custom distribution layer. For example, suppose you
have a layer named meta-mylayer
located in your home directory and
you want your template configuration directory named myconf
.
Changing the .templateconf
as follows causes the OpenEmbedded build
system to look in your directory and base its configuration files on the
*.sample
configuration files it finds. The final configuration files
(i.e. local.conf
and bblayers.conf
ultimately still end up in
your Build Directory, but they are based on your *.sample
files.
TEMPLATECONF=${TEMPLATECONF:-meta-mylayer/myconf}
Aside from the *.sample
configuration files, the conf-notes.txt
also resides in the default meta-poky/conf
directory. The script
that sets up the build environment (i.e.
oe-init-build-env) uses this file to
display BitBake targets as part of the script output. Customizing this
conf-notes.txt
file is a good way to make sure your list of custom
targets appears as part of the script’s output.
Here is the default list of targets displayed as a result of running either of the setup scripts:
You can now run 'bitbake <target>'
Common targets are:
core-image-minimal
core-image-sato
meta-toolchain
meta-ide-support
Changing the listed common targets is as easy as editing your version of
conf-notes.txt
in your custom template configuration directory and
making sure you have TEMPLATECONF
set to your directory.
3.21 Conserving Disk Space During Builds
To help conserve disk space during builds, you can add the following
statement to your project’s local.conf
configuration file found in
the Build Directory:
INHERIT += "rm_work"
Adding this statement deletes the work directory used for building a recipe once the recipe is built. For more information on “rm_work”, see the rm_work class in the Yocto Project Reference Manual.
3.22 Working with Packages
This section describes a few tasks that involve packages:
3.22.1 Excluding Packages from an Image
You might find it necessary to prevent specific packages from being installed into an image. If so, you can use several variables to direct the build system to essentially ignore installing recommended packages or to not install a package at all.
The following list introduces variables you can use to prevent packages
from being installed into your image. Each of these variables only works
with IPK and RPM package types. Support for Debian packages does not
exist. Also, you can use these variables from your local.conf
file
or attach them to a specific image recipe by using a recipe name
override. For more detail on the variables, see the descriptions in the
Yocto Project Reference Manual’s glossary chapter.
BAD_RECOMMENDATIONS: Use this variable to specify “recommended-only” packages that you do not want installed.
NO_RECOMMENDATIONS: Use this variable to prevent all “recommended-only” packages from being installed.
PACKAGE_EXCLUDE: Use this variable to prevent specific packages from being installed regardless of whether they are “recommended-only” or not. You need to realize that the build process could fail with an error when you prevent the installation of a package whose presence is required by an installed package.
3.22.2 Incrementing a Package Version
This section provides some background on how binary package versioning is accomplished and presents some of the services, variables, and terminology involved.
In order to understand binary package versioning, you need to consider the following:
Binary Package: The binary package that is eventually built and installed into an image.
Binary Package Version: The binary package version is composed of two components - a version and a revision.
Note
Technically, a third component, the “epoch” (i.e. PE) is involved but this discussion for the most part ignores
PE
.The version and revision are taken from the PV and PR variables, respectively.
PV
: The recipe version.PV
represents the version of the software being packaged. Do not confusePV
with the binary package version.PR
: The recipe revision.SRCPV: The OpenEmbedded build system uses this string to help define the value of
PV
when the source code revision needs to be included in it.PR Service: A network-based service that helps automate keeping package feeds compatible with existing package manager applications such as RPM, APT, and OPKG.
Whenever the binary package content changes, the binary package version
must change. Changing the binary package version is accomplished by
changing or “bumping” the PR
and/or PV
values. Increasing these
values occurs one of two ways:
Automatically using a Package Revision Service (PR Service).
Manually incrementing the
PR
and/orPV
variables.
Given a primary challenge of any build system and its users is how to maintain a package feed that is compatible with existing package manager applications such as RPM, APT, and OPKG, using an automated system is much preferred over a manual system. In either system, the main requirement is that binary package version numbering increases in a linear fashion and that a number of version components exist that support that linear progression. For information on how to ensure package revisioning remains linear, see the “Automatically Incrementing a Package Version Number” section.
The following three sections provide related information on the PR
Service, the manual method for “bumping” PR
and/or PV
, and on
how to ensure binary package revisioning remains linear.
3.22.2.1 Working With a PR Service
As mentioned, attempting to maintain revision numbers in the Metadata is error prone, inaccurate, and causes problems for people submitting recipes. Conversely, the PR Service automatically generates increasing numbers, particularly the revision field, which removes the human element.
Note
For additional information on using a PR Service, you can see the PR Service wiki page.
The Yocto Project uses variables in order of decreasing priority to facilitate revision numbering (i.e. PE, PV, and PR for epoch, version, and revision, respectively). The values are highly dependent on the policies and procedures of a given distribution and package feed.
Because the OpenEmbedded build system uses
“signatures”, which are
unique to a given build, the build system knows when to rebuild
packages. All the inputs into a given task are represented by a
signature, which can trigger a rebuild when different. Thus, the build
system itself does not rely on the PR
, PV
, and PE
numbers to
trigger a rebuild. The signatures, however, can be used to generate
these values.
The PR Service works with both OEBasic
and OEBasicHash
generators. The value of PR
bumps when the checksum changes and the
different generator mechanisms change signatures under different
circumstances.
As implemented, the build system includes values from the PR Service
into the PR
field as an addition using the form “.x
” so r0
becomes r0.1
, r0.2
and so forth. This scheme allows existing
PR
values to be used for whatever reasons, which include manual
PR
bumps, should it be necessary.
By default, the PR Service is not enabled or running. Thus, the packages generated are just “self consistent”. The build system adds and removes packages and there are no guarantees about upgrade paths but images will be consistent and correct with the latest changes.
The simplest form for a PR Service is for it to exist for a single host
development system that builds the package feed (building system). For
this scenario, you can enable a local PR Service by setting
PRSERV_HOST in your
local.conf
file in the Build Directory:
PRSERV_HOST = "localhost:0"
Once the service is started, packages will automatically
get increasing PR
values and BitBake takes care of starting and
stopping the server.
If you have a more complex setup where multiple host development systems
work against a common, shared package feed, you have a single PR Service
running and it is connected to each building system. For this scenario,
you need to start the PR Service using the bitbake-prserv
command:
bitbake-prserv --host ip --port port --start
In addition to
hand-starting the service, you need to update the local.conf
file of
each building system as described earlier so each system points to the
server and port.
It is also recommended you use build history, which adds some sanity
checks to binary package versions, in conjunction with the server that
is running the PR Service. To enable build history, add the following to
each building system’s local.conf
file:
# It is recommended to activate "buildhistory" for testing the PR service
INHERIT += "buildhistory"
BUILDHISTORY_COMMIT = "1"
For information on build history, see the “Maintaining Build Output Quality” section.
Note
The OpenEmbedded build system does not maintain PR
information as
part of the shared state (sstate) packages. If you maintain an sstate
feed, it’s expected that either all your building systems that
contribute to the sstate feed use a shared PR Service, or you do not
run a PR Service on any of your building systems. Having some systems
use a PR Service while others do not leads to obvious problems.
For more information on shared state, see the “Shared State Cache” section in the Yocto Project Overview and Concepts Manual.
3.22.2.2 Manually Bumping PR
The alternative to setting up a PR Service is to manually “bump” the PR variable.
If a committed change results in changing the package output, then the
value of the PR variable needs to be increased (or “bumped”) as part of
that commit. For new recipes you should add the PR
variable and set
its initial value equal to “r0”, which is the default. Even though the
default value is “r0”, the practice of adding it to a new recipe makes
it harder to forget to bump the variable when you make changes to the
recipe in future.
If you are sharing a common .inc
file with multiple recipes, you can
also use the INC_PR
variable to ensure that the recipes sharing the
.inc
file are rebuilt when the .inc
file itself is changed. The
.inc
file must set INC_PR
(initially to “r0”), and all recipes
referring to it should set PR
to “${INC_PR}.0” initially,
incrementing the last number when the recipe is changed. If the .inc
file is changed then its INC_PR
should be incremented.
When upgrading the version of a binary package, assuming the PV
changes, the PR
variable should be reset to “r0” (or “${INC_PR}.0”
if you are using INC_PR
).
Usually, version increases occur only to binary packages. However, if
for some reason PV
changes but does not increase, you can increase
the PE
variable (Package Epoch). The PE
variable defaults to
“0”.
Binary package version numbering strives to follow the Debian Version Field Policy Guidelines. These guidelines define how versions are compared and what “increasing” a version means.
3.22.2.3 Automatically Incrementing a Package Version Number
When fetching a repository, BitBake uses the
SRCREV variable to determine
the specific source code revision from which to build. You set the
SRCREV
variable to
AUTOREV to cause the
OpenEmbedded build system to automatically use the latest revision of
the software:
SRCREV = "${AUTOREV}"
Furthermore, you need to reference SRCPV
in PV
in order to
automatically update the version whenever the revision of the source
code changes. Here is an example:
PV = "1.0+git${SRCPV}"
The OpenEmbedded build system substitutes SRCPV
with the following:
AUTOINC+source_code_revision
The build system replaces the AUTOINC
with a number. The number used depends on the state of the PR Service:
If PR Service is enabled, the build system increments the number, which is similar to the behavior of PR. This behavior results in linearly increasing package versions, which is desirable. Here is an example:
hello-world-git_0.0+git0+b6558dd387-r0.0_armv7a-neon.ipk hello-world-git_0.0+git1+dd2f5c3565-r0.0_armv7a-neon.ipk
If PR Service is not enabled, the build system replaces the
AUTOINC
placeholder with zero (i.e. “0”). This results in changing the package version since the source revision is included. However, package versions are not increased linearly. Here is an example:hello-world-git_0.0+git0+b6558dd387-r0.0_armv7a-neon.ipk hello-world-git_0.0+git0+dd2f5c3565-r0.0_armv7a-neon.ipk
In summary, the OpenEmbedded build system does not track the history of
binary package versions for this purpose. AUTOINC
, in this case, is
comparable to PR
. If PR server is not enabled, AUTOINC
in the
package version is simply replaced by “0”. If PR server is enabled, the
build system keeps track of the package versions and bumps the number
when the package revision changes.
3.22.3 Handling Optional Module Packaging
Many pieces of software split functionality into optional modules (or plugins) and the plugins that are built might depend on configuration options. To avoid having to duplicate the logic that determines what modules are available in your recipe or to avoid having to package each module by hand, the OpenEmbedded build system provides functionality to handle module packaging dynamically.
To handle optional module packaging, you need to do two things:
Ensure the module packaging is actually done.
Ensure that any dependencies on optional modules from other recipes are satisfied by your recipe.
3.22.3.1 Making Sure the Packaging is Done
To ensure the module packaging actually gets done, you use the
do_split_packages
function within the populate_packages
Python
function in your recipe. The do_split_packages
function searches for
a pattern of files or directories under a specified path and creates a
package for each one it finds by appending to the
PACKAGES variable and
setting the appropriate values for FILES_packagename
,
RDEPENDS_packagename
, DESCRIPTION_packagename
, and so forth.
Here is an example from the lighttpd
recipe:
python populate_packages_prepend () {
lighttpd_libdir = d.expand('${libdir}')
do_split_packages(d, lighttpd_libdir, '^mod_(.*).so$',
'lighttpd-module-%s', 'Lighttpd module for %s',
extra_depends='')
}
The previous example specifies a number of things in the call to
do_split_packages
.
A directory within the files installed by your recipe through
do_install
in which to search.A regular expression used to match module files in that directory. In the example, note the parentheses () that mark the part of the expression from which the module name should be derived.
A pattern to use for the package names.
A description for each package.
An empty string for
extra_depends
, which disables the default dependency on the mainlighttpd
package. Thus, if a file in${libdir}
calledmod_alias.so
is found, a package calledlighttpd-module-alias
is created for it and the DESCRIPTION is set to “Lighttpd module for alias”.
Often, packaging modules is as simple as the previous example. However,
more advanced options exist that you can use within
do_split_packages
to modify its behavior. And, if you need to, you
can add more logic by specifying a hook function that is called for each
package. It is also perfectly acceptable to call do_split_packages
multiple times if you have more than one set of modules to package.
For more examples that show how to use do_split_packages
, see the
connman.inc
file in the meta/recipes-connectivity/connman/
directory of the poky
source repository. You can
also find examples in meta/classes/kernel.bbclass
.
Following is a reference that shows do_split_packages
mandatory and
optional arguments:
Mandatory arguments
root
The path in which to search
file_regex
Regular expression to match searched files.
Use parentheses () to mark the part of this
expression that should be used to derive the
module name (to be substituted where %s is
used in other function arguments as noted below)
output_pattern
Pattern to use for the package names. Must
include %s.
description
Description to set for each package. Must
include %s.
Optional arguments
postinst
Postinstall script to use for all packages
(as a string)
recursive
True to perform a recursive search - default
False
hook
A hook function to be called for every match.
The function will be called with the following
arguments (in the order listed):
f
Full path to the file/directory match
pkg
The package name
file_regex
As above
output_pattern
As above
modulename
The module name derived using file_regex
extra_depends
Extra runtime dependencies (RDEPENDS) to be
set for all packages. The default value of None
causes a dependency on the main package
(${PN}) - if you do not want this, pass empty
string '' for this parameter.
aux_files_pattern
Extra item(s) to be added to FILES for each
package. Can be a single string item or a list
of strings for multiple items. Must include %s.
postrm
postrm script to use for all packages (as a
string)
allow_dirs
True to allow directories to be matched -
default False
prepend
If True, prepend created packages to PACKAGES
instead of the default False which appends them
match_path
match file_regex on the whole relative path to
the root rather than just the file name
aux_files_pattern_verbatim
Extra item(s) to be added to FILES for each
package, using the actual derived module name
rather than converting it to something legal
for a package name. Can be a single string item
or a list of strings for multiple items. Must
include %s.
allow_links
True to allow symlinks to be matched - default
False
summary
Summary to set for each package. Must include %s;
defaults to description if not set.
3.22.3.2 Satisfying Dependencies
The second part for handling optional module packaging is to ensure that
any dependencies on optional modules from other recipes are satisfied by
your recipe. You can be sure these dependencies are satisfied by using
the PACKAGES_DYNAMIC
variable. Here is an example that continues with the lighttpd
recipe
shown earlier:
PACKAGES_DYNAMIC = "lighttpd-module-.*"
The name
specified in the regular expression can of course be anything. In this
example, it is lighttpd-module-
and is specified as the prefix to
ensure that any RDEPENDS and
RRECOMMENDS on a package
name starting with the prefix are satisfied during build time. If you
are using do_split_packages
as described in the previous section,
the value you put in PACKAGES_DYNAMIC
should correspond to the name
pattern specified in the call to do_split_packages
.
3.22.4 Using Runtime Package Management
During a build, BitBake always transforms a recipe into one or more
packages. For example, BitBake takes the bash
recipe and produces a
number of packages (e.g. bash
, bash-bashbug
,
bash-completion
, bash-completion-dbg
, bash-completion-dev
,
bash-completion-extra
, bash-dbg
, and so forth). Not all
generated packages are included in an image.
In several situations, you might need to update, add, remove, or query the packages on a target device at runtime (i.e. without having to generate a new image). Examples of such situations include:
You want to provide in-the-field updates to deployed devices (e.g. security updates).
You want to have a fast turn-around development cycle for one or more applications that run on your device.
You want to temporarily install the “debug” packages of various applications on your device so that debugging can be greatly improved by allowing access to symbols and source debugging.
You want to deploy a more minimal package selection of your device but allow in-the-field updates to add a larger selection for customization.
In all these situations, you have something similar to a more traditional Linux distribution in that in-field devices are able to receive pre-compiled packages from a server for installation or update. Being able to install these packages on a running, in-field device is what is termed “runtime package management”.
In order to use runtime package management, you need a host or server machine that serves up the pre-compiled packages plus the required metadata. You also need package manipulation tools on the target. The build machine is a likely candidate to act as the server. However, that machine does not necessarily have to be the package server. The build machine could push its artifacts to another machine that acts as the server (e.g. Internet-facing). In fact, doing so is advantageous for a production environment as getting the packages away from the development system’s build directory prevents accidental overwrites.
A simple build that targets just one device produces more than one
package database. In other words, the packages produced by a build are
separated out into a couple of different package groupings based on
criteria such as the target’s CPU architecture, the target board, or the
C library used on the target. For example, a build targeting the
qemux86
device produces the following three package databases:
noarch
, i586
, and qemux86
. If you wanted your qemux86
device to be aware of all the packages that were available to it, you
would need to point it to each of these databases individually. In a
similar way, a traditional Linux distribution usually is configured to
be aware of a number of software repositories from which it retrieves
packages.
Using runtime package management is completely optional and not required for a successful build or deployment in any way. But if you want to make use of runtime package management, you need to do a couple things above and beyond the basics. The remainder of this section describes what you need to do.
3.22.4.1 Build Considerations
This section describes build considerations of which you need to be aware in order to provide support for runtime package management.
When BitBake generates packages, it needs to know what format or formats to use. In your configuration, you use the PACKAGE_CLASSES variable to specify the format:
Open the
local.conf
file inside your Build Directory (e.g.poky/build/conf/local.conf
).Select the desired package format as follows:
PACKAGE_CLASSES ?= "package_packageformat"
where packageformat can be “ipk”, “rpm”, “deb”, or “tar” which are the supported package formats.
Note
Because the Yocto Project supports four different package formats, you can set the variable with more than one argument. However, the OpenEmbedded build system only uses the first argument when creating an image or Software Development Kit (SDK).
If you would like your image to start off with a basic package database containing the packages in your current build as well as to have the relevant tools available on the target for runtime package management, you can include “package-management” in the IMAGE_FEATURES variable. Including “package-management” in this configuration variable ensures that when the image is assembled for your target, the image includes the currently-known package databases as well as the target-specific tools required for runtime package management to be performed on the target. However, this is not strictly necessary. You could start your image off without any databases but only include the required on-target package tool(s). As an example, you could include “opkg” in your IMAGE_INSTALL variable if you are using the IPK package format. You can then initialize your target’s package database(s) later once your image is up and running.
Whenever you perform any sort of build step that can potentially generate a package or modify existing package, it is always a good idea to re-generate the package index after the build by using the following command:
$ bitbake package-index
It might be tempting to build the package and the package index at the same time with a command such as the following:
$ bitbake some-package package-index
Do not do this as BitBake does not schedule the package index for after the completion of the package you are building. Consequently, you cannot be sure of the package index including information for the package you just built. Thus, be sure to run the package update step separately after building any packages.
You can use the PACKAGE_FEED_ARCHS, PACKAGE_FEED_BASE_PATHS, and PACKAGE_FEED_URIS variables to pre-configure target images to use a package feed. If you do not define these variables, then manual steps as described in the subsequent sections are necessary to configure the target. You should set these variables before building the image in order to produce a correctly configured image.
When your build is complete, your packages reside in the
${TMPDIR}/deploy/packageformat
directory. For example, if
${
TMPDIR}
is
tmp
and your selected package type is RPM, then your RPM packages
are available in tmp/deploy/rpm
.
3.22.4.2 Host or Server Machine Setup
Although other protocols are possible, a server using HTTP typically serves packages. If you want to use HTTP, then set up and configure a web server such as Apache 2, lighttpd, or Python web server on the machine serving the packages.
To keep things simple, this section describes how to set up a Python web server to share package feeds from the developer’s machine. Although this server might not be the best for a production environment, the setup is simple and straight forward. Should you want to use a different server more suited for production (e.g. Apache 2, Lighttpd, or Nginx), take the appropriate steps to do so.
From within the build directory where you have built an image based on
your packaging choice (i.e. the
PACKAGE_CLASSES
setting), simply start the server. The following example assumes a build
directory of poky/build/tmp/deploy/rpm
and a PACKAGE_CLASSES
setting of “package_rpm”:
$ cd poky/build/tmp/deploy/rpm
$ python3 -m http.server
3.22.4.3 Target Setup
Setting up the target differs depending on the package management system. This section provides information for RPM, IPK, and DEB.
3.22.4.3.1 Using RPM
The Dandified Packaging
Tool (DNF) performs
runtime package management of RPM packages. In order to use DNF for
runtime package management, you must perform an initial setup on the
target machine for cases where the PACKAGE_FEED_*
variables were not
set as part of the image that is running on the target. This means if
you built your image and did not use these variables as part of the
build and your image is now running on the target, you need to perform
the steps in this section if you want to use runtime package management.
Note
For information on the PACKAGE_FEED_*
variables, see
PACKAGE_FEED_ARCHS, PACKAGE_FEED_BASE_PATHS, and
PACKAGE_FEED_URIS in the Yocto Project Reference Manual variables
glossary.
On the target, you must inform DNF that package databases are available.
You do this by creating a file named
/etc/yum.repos.d/oe-packages.repo
and defining the oe-packages
.
As an example, assume the target is able to use the following package
databases: all
, i586
, and qemux86
from a server named
my.server
. The specifics for setting up the web server are up to
you. The critical requirement is that the URIs in the target repository
configuration point to the correct remote location for the feeds.
Note
For development purposes, you can point the web server to the build
system’s deploy
directory. However, for production use, it is better to
copy the package directories to a location outside of the build area and use
that location. Doing so avoids situations where the build system
overwrites or changes the deploy
directory.
When telling DNF where to look for the package databases, you must declare individual locations per architecture or a single location used for all architectures. You cannot do both:
Create an Explicit List of Architectures: Define individual base URLs to identify where each package database is located:
[oe-packages] baseurl=http://my.server/rpm/i586 http://my.server/rpm/qemux86 http://my.server/rpm/all
This example informs DNF about individual package databases for all three architectures.
Create a Single (Full) Package Index: Define a single base URL that identifies where a full package database is located:
[oe-packages] baseurl=http://my.server/rpm
This example informs DNF about a single package database that contains all the package index information for all supported architectures.
Once you have informed DNF where to find the package databases, you need to fetch them:
# dnf makecache
DNF is now able to find, install, and upgrade packages from the specified repository or repositories.
Note
See the DNF documentation for additional information.
3.22.4.3.2 Using IPK
The opkg
application performs runtime package management of IPK
packages. You must perform an initial setup for opkg
on the target
machine if the
PACKAGE_FEED_ARCHS,
PACKAGE_FEED_BASE_PATHS,
and
PACKAGE_FEED_URIS
variables have not been set or the target image was built before the
variables were set.
The opkg
application uses configuration files to find available
package databases. Thus, you need to create a configuration file inside
the /etc/opkg/
direction, which informs opkg
of any repository
you want to use.
As an example, suppose you are serving packages from a ipk/
directory containing the i586
, all
, and qemux86
databases
through an HTTP server named my.server
. On the target, create a
configuration file (e.g. my_repo.conf
) inside the /etc/opkg/
directory containing the following:
src/gz all http://my.server/ipk/all
src/gz i586 http://my.server/ipk/i586
src/gz qemux86 http://my.server/ipk/qemux86
Next, instruct opkg
to fetch the
repository information:
# opkg update
The opkg
application is now able to find, install, and upgrade packages
from the specified repository.
3.22.4.3.3 Using DEB
The apt
application performs runtime package management of DEB
packages. This application uses a source list file to find available
package databases. You must perform an initial setup for apt
on the
target machine if the
PACKAGE_FEED_ARCHS,
PACKAGE_FEED_BASE_PATHS,
and
PACKAGE_FEED_URIS
variables have not been set or the target image was built before the
variables were set.
To inform apt
of the repository you want to use, you might create a
list file (e.g. my_repo.list
) inside the
/etc/apt/sources.list.d/
directory. As an example, suppose you are
serving packages from a deb/
directory containing the i586
,
all
, and qemux86
databases through an HTTP server named
my.server
. The list file should contain:
deb http://my.server/deb/all ./
deb http://my.server/deb/i586 ./
deb http://my.server/deb/qemux86 ./
Next, instruct the apt
application
to fetch the repository information:
# apt-get update
After this step,
apt
is able to find, install, and upgrade packages from the
specified repository.
3.22.5 Generating and Using Signed Packages
In order to add security to RPM packages used during a build, you can take steps to securely sign them. Once a signature is verified, the OpenEmbedded build system can use the package in the build. If security fails for a signed package, the build system aborts the build.
This section describes how to sign RPM packages during a build and how to use signed package feeds (repositories) when doing a build.
3.22.5.1 Signing RPM Packages
To enable signing RPM packages, you must set up the following
configurations in either your local.config
or distro.config
file:
# Inherit sign_rpm.bbclass to enable signing functionality
INHERIT += " sign_rpm"
# Define the GPG key that will be used for signing.
RPM_GPG_NAME = "key_name"
# Provide passphrase for the key
RPM_GPG_PASSPHRASE = "passphrase"
Note
Be sure to supply appropriate values for both key_name and passphrase.
Aside from the RPM_GPG_NAME
and RPM_GPG_PASSPHRASE
variables in
the previous example, two optional variables related to signing exist:
GPG_BIN: Specifies a
gpg
binary/wrapper that is executed when the package is signed.GPG_PATH: Specifies the
gpg
home directory used when the package is signed.
3.22.5.2 Processing Package Feeds
In addition to being able to sign RPM packages, you can also enable signed package feeds for IPK and RPM packages.
The steps you need to take to enable signed package feed use are similar
to the steps used to sign RPM packages. You must define the following in
your local.config
or distro.config
file:
INHERIT += "sign_package_feed"
PACKAGE_FEED_GPG_NAME = "key_name"
PACKAGE_FEED_GPG_PASSPHRASE_FILE = "path_to_file_containing_passphrase"
For signed package feeds, the passphrase must exist in a separate file,
which is pointed to by the PACKAGE_FEED_GPG_PASSPHRASE_FILE
variable. Regarding security, keeping a plain text passphrase out of the
configuration is more secure.
Aside from the PACKAGE_FEED_GPG_NAME
and
PACKAGE_FEED_GPG_PASSPHRASE_FILE
variables, three optional variables
related to signed package feeds exist:
GPG_BIN Specifies a
gpg
binary/wrapper that is executed when the package is signed.GPG_PATH: Specifies the
gpg
home directory used when the package is signed.PACKAGE_FEED_GPG_SIGNATURE_TYPE: Specifies the type of
gpg
signature. This variable applies only to RPM and IPK package feeds. Allowable values for thePACKAGE_FEED_GPG_SIGNATURE_TYPE
are “ASC”, which is the default and specifies ascii armored, and “BIN”, which specifies binary.
3.22.6 Testing Packages With ptest
A Package Test (ptest) runs tests against packages built by the
OpenEmbedded build system on the target machine. A ptest contains at
least two items: the actual test, and a shell script (run-ptest
)
that starts the test. The shell script that starts the test must not
contain the actual test - the script only starts the test. On the other
hand, the test can be anything from a simple shell script that runs a
binary and checks the output to an elaborate system of test binaries and
data files.
The test generates output in the format used by Automake:
result: testname
where the result can be PASS
, FAIL
, or SKIP
, and
the testname can be any identifying string.
For a list of Yocto Project recipes that are already enabled with ptest, see the Ptest wiki page.
Note
A recipe is “ptest-enabled” if it inherits the ptest class.
3.22.6.1 Adding ptest to Your Build
To add package testing to your build, add the
DISTRO_FEATURES and
EXTRA_IMAGE_FEATURES
variables to your local.conf
file, which is found in the
Build Directory:
DISTRO_FEATURES_append = " ptest"
EXTRA_IMAGE_FEATURES += "ptest-pkgs"
Once your build is complete, the ptest files are installed into the
/usr/lib/package/ptest
directory within the image, where package
is the name of the package.
3.22.6.2 Running ptest
The ptest-runner
package installs a shell script that loops through
all installed ptest test suites and runs them in sequence. Consequently,
you might want to add this package to your image.
3.22.6.3 Getting Your Package Ready
In order to enable a recipe to run installed ptests on target hardware, you need to prepare the recipes that build the packages you want to test. Here is what you have to do for each recipe:
Be sure the recipe inherits the ptest class: Include the following line in each recipe:
inherit ptest
Create run-ptest: This script starts your test. Locate the script where you will refer to it using SRC_URI. Here is an example that starts a test for
dbus
:#!/bin/sh cd test make -k runtest-TESTS
Ensure dependencies are met: If the test adds build or runtime dependencies that normally do not exist for the package (such as requiring “make” to run the test suite), use the DEPENDS and RDEPENDS variables in your recipe in order for the package to meet the dependencies. Here is an example where the package has a runtime dependency on “make”:
RDEPENDS_${PN}-ptest += "make"
Add a function to build the test suite: Not many packages support cross-compilation of their test suites. Consequently, you usually need to add a cross-compilation function to the package.
Many packages based on Automake compile and run the test suite by using a single command such as
make check
. However, the hostmake check
builds and runs on the same computer, while cross-compiling requires that the package is built on the host but executed for the target architecture (though often, as in the case for ptest, the execution occurs on the host). The built version of Automake that ships with the Yocto Project includes a patch that separates building and execution. Consequently, packages that use the unaltered, patched version ofmake check
automatically cross-compiles.Regardless, you still must add a
do_compile_ptest
function to build the test suite. Add a function similar to the following to your recipe:do_compile_ptest() { oe_runmake buildtest-TESTS }
Ensure special configurations are set: If the package requires special configurations prior to compiling the test code, you must insert a
do_configure_ptest
function into the recipe.Install the test suite: The
ptest
class automatically copies the filerun-ptest
to the target and then runs makeinstall-ptest
to run the tests. If this is not enough, you need to create ado_install_ptest
function and make sure it gets called after the “make install-ptest” completes.
3.22.7 Creating Node Package Manager (NPM) Packages
NPM is a package manager for the JavaScript programming language. The Yocto Project supports the NPM fetcher. You can use this fetcher in combination with devtool to create recipes that produce NPM packages.
Two workflows exist that allow you to create NPM packages using
devtool
: the NPM registry modules method and the NPM project code
method.
Note
While it is possible to create NPM recipes manually, using
devtool
is far simpler.
Additionally, some requirements and caveats exist.
3.22.7.1 Requirements and Caveats
You need to be aware of the following before using devtool
to create
NPM packages:
Of the two methods that you can use
devtool
to create NPM packages, the registry approach is slightly simpler. However, you might consider the project approach because you do not have to publish your module in the NPM registry (npm-registry), which is NPM’s public registry.Be familiar with devtool.
The NPM host tools need the native
nodejs-npm
package, which is part of the OpenEmbedded environment. You need to get the package by cloning the https://github.com/openembedded/meta-openembedded repository out of GitHub. Be sure to add the path to your local copy to yourbblayers.conf
file.devtool
cannot detect native libraries in module dependencies. Consequently, you must manually add packages to your recipe.While deploying NPM packages,
devtool
cannot determine which dependent packages are missing on the target (e.g. the node runtimenodejs
). Consequently, you need to find out what files are missing and be sure they are on the target.Although you might not need NPM to run your node package, it is useful to have NPM on your target. The NPM package name is
nodejs-npm
.
3.22.7.2 Using the Registry Modules Method
This section presents an example that uses the cute-files
module,
which is a file browser web application.
Note
You must know the cute-files
module version.
The first thing you need to do is use devtool
and the NPM fetcher to
create the recipe:
$ devtool add "npm://registry.npmjs.org;package=cute-files;version=1.0.2"
The
devtool add
command runs recipetool create
and uses the same
fetch URI to download each dependency and capture license details where
possible. The result is a generated recipe.
The recipe file is fairly simple and contains every license that
recipetool
finds and includes the licenses in the recipe’s
LIC_FILES_CHKSUM
variables. You need to examine the variables and look for those with
“unknown” in the LICENSE
field. You need to track down the license information for “unknown”
modules and manually add the information to the recipe.
recipetool
creates a “shrinkwrap” file for your recipe. Shrinkwrap
files capture the version of all dependent modules. Many packages do not
provide shrinkwrap files. recipetool
create a shrinkwrap file as it
runs.
Note
A package is created for each sub-module. This policy is the only practical way to have the licenses for all of the dependencies represented in the license manifest of the image.
The devtool edit-recipe
command lets you take a look at the recipe:
$ devtool edit-recipe cute-files
SUMMARY = "Turn any folder on your computer into a cute file browser, available on the local network."
LICENSE = "MIT & ISC & Unknown"
LIC_FILES_CHKSUM = "file://LICENSE;md5=71d98c0a1db42956787b1909c74a86ca \
file://node_modules/toidentifier/LICENSE;md5=1a261071a044d02eb6f2bb47f51a3502 \
file://node_modules/debug/LICENSE;md5=ddd815a475e7338b0be7a14d8ee35a99 \
...
SRC_URI = " \
npm://registry.npmjs.org/;package=cute-files;version=${PV} \
npmsw://${THISDIR}/${BPN}/npm-shrinkwrap.json \
"
S = "${WORKDIR}/npm"
inherit npm LICENSE_${PN} = "MIT"
LICENSE_${PN}-accepts = "MIT"
LICENSE_${PN}-array-flatten = "MIT"
...
LICENSE_${PN}-vary = "MIT"
Three key points exist in the previous example:
SRC_URI uses the NPM scheme so that the NPM fetcher is used.
recipetool
collects all the license information. If a sub-module’s license is unavailable, the sub-module’s name appears in the comments.The
inherit npm
statement causes the npm class to package up all the modules.
You can run the following command to build the cute-files
package:
$ devtool build cute-files
Remember that nodejs
must be installed on
the target before your package.
Assuming 192.168.7.2 for the target’s IP address, use the following command to deploy your package:
$ devtool deploy-target -s cute-files root@192.168.7.2
Once the package is installed on the target, you can test the application:
Note
Because of a known issue, you cannot simply run cute-files
as you would
if you had run npm install
.
$ cd /usr/lib/node_modules/cute-files
$ node cute-files.js
On a browser,
go to http://192.168.7.2:3000
and you see the following:
You can find the recipe in workspace/recipes/cute-files
. You can use
the recipe in any layer you choose.
3.22.7.3 Using the NPM Projects Code Method
Although it is useful to package modules already in the NPM registry,
adding node.js
projects under development is a more common developer
use case.
This section covers the NPM projects code method, which is very similar
to the “registry” approach described in the previous section. In the NPM
projects method, you provide devtool
with an URL that points to the
source files.
Replicating the same example, (i.e. cute-files
) use the following
command:
$ devtool add https://github.com/martinaglv/cute-files.git
The
recipe this command generates is very similar to the recipe created in
the previous section. However, the SRC_URI
looks like the following:
SRC_URI = " \
git://github.com/martinaglv/cute-files.git;protocol=https \
npmsw://${THISDIR}/${BPN}/npm-shrinkwrap.json \
"
In this example, the main module is taken from the Git repository and dependencies are taken from the NPM registry. Other than those differences, the recipe is basically the same between the two methods. You can build and deploy the package exactly as described in the previous section that uses the registry modules method.
3.22.8 Adding custom metadata to packages
The variable PACKAGE_ADD_METADATA can be used to add additional metadata to packages. This is reflected in the package control/spec file. To take the ipk format for example, the CONTROL file stored inside would contain the additional metadata as additional lines.
The variable can be used in multiple ways, including using suffixes to set it for a specific package type and/or package. Note that the order of precedence is the same as this list:
PACKAGE_ADD_METADATA_<PKGTYPE>_<PN>
PACKAGE_ADD_METADATA_<PKGTYPE>
PACKAGE_ADD_METADATA_<PN>
PACKAGE_ADD_METADATA
<PKGTYPE> is a parameter and expected to be a distinct name of specific package type:
IPK for .ipk packages
DEB for .deb packages
RPM for .rpm packages
<PN> is a parameter and expected to be a package name.
The variable can contain multiple [one-line] metadata fields separated
by the literal sequence ‘\n’. The separator can be redefined using the
variable flag separator
.
The following is an example that adds two custom fields for ipk packages:
PACKAGE_ADD_METADATA_IPK = "Vendor: CustomIpk\nGroup:Applications/Spreadsheets"
3.23 Efficiently Fetching Source Files During a Build
The OpenEmbedded build system works with source files located through the SRC_URI variable. When you build something using BitBake, a big part of the operation is locating and downloading all the source tarballs. For images, downloading all the source for various packages can take a significant amount of time.
This section shows you how you can use mirrors to speed up fetching source files and how you can pre-fetch files all of which leads to more efficient use of resources and time.
3.23.1 Setting up Effective Mirrors
A good deal that goes into a Yocto Project build is simply downloading all of the source tarballs. Maybe you have been working with another build system (OpenEmbedded or Angstrom) for which you have built up a sizable directory of source tarballs. Or, perhaps someone else has such a directory for which you have read access. If so, you can save time by adding statements to your configuration file so that the build process checks local directories first for existing tarballs before checking the Internet.
Here is an efficient way to set it up in your local.conf
file:
SOURCE_MIRROR_URL ?= "file:///home/you/your-download-dir/"
INHERIT += "own-mirrors"
BB_GENERATE_MIRROR_TARBALLS = "1"
# BB_NO_NETWORK = "1"
In the previous example, the BB_GENERATE_MIRROR_TARBALLS variable causes the OpenEmbedded build system to generate tarballs of the Git repositories and store them in the DL_DIR directory. Due to performance reasons, generating and storing these tarballs is not the build system’s default behavior.
You can also use the PREMIRRORS variable. For an example, see the variable’s glossary entry in the Yocto Project Reference Manual.
3.23.2 Getting Source Files and Suppressing the Build
Another technique you can use to ready yourself for a successive string of build operations, is to pre-fetch all the source files without actually starting a build. This technique lets you work through any download issues and ultimately gathers all the source files into your download directory build/downloads/, which is located with DL_DIR.
Use the following BitBake command form to fetch all the necessary sources without starting the build:
$ bitbake target --runall=fetch
This variation of the BitBake command guarantees that you have all the sources for that BitBake target should you disconnect from the Internet and want to do the build later offline.
3.24 Selecting an Initialization Manager
By default, the Yocto Project uses SysVinit as the initialization manager. However, support also exists for systemd, which is a full replacement for init with parallel starting of services, reduced shell overhead and other features that are used by many distributions.
Within the system, SysVinit treats system components as services. These
services are maintained as shell scripts stored in the /etc/init.d/
directory. Services organize into different run levels. This
organization is maintained by putting links to the services in the
/etc/rcN.d/
directories, where N/ is one of the following options:
“S”, “0”, “1”, “2”, “3”, “4”, “5”, or “6”.
Note
Each runlevel has a dependency on the previous runlevel. This dependency allows the services to work properly.
In comparison, systemd treats components as units. Using units is a broader concept as compared to using a service. A unit includes several different types of entities. Service is one of the types of entities. The runlevel concept in SysVinit corresponds to the concept of a target in systemd, where target is also a type of supported unit.
In a SysVinit-based system, services load sequentially (i.e. one by one) during init and parallelization is not supported. With systemd, services start in parallel. Needless to say, the method can have an impact on system startup performance.
If you want to use SysVinit, you do not have to do anything. But, if you want to use systemd, you must take some steps as described in the following sections.
3.24.1 Using systemd Exclusively
Set these variables in your distribution configuration file as follows:
DISTRO_FEATURES_append = " systemd"
VIRTUAL-RUNTIME_init_manager = "systemd"
You can also prevent the SysVinit distribution feature from being automatically enabled as follows:
DISTRO_FEATURES_BACKFILL_CONSIDERED = "sysvinit"
Doing so removes any redundant SysVinit scripts.
To remove initscripts from your image altogether, set this variable also:
VIRTUAL-RUNTIME_initscripts = ""
For information on the backfill variable, see DISTRO_FEATURES_BACKFILL_CONSIDERED.
3.24.2 Using systemd for the Main Image and Using SysVinit for the Rescue Image
Set these variables in your distribution configuration file as follows:
DISTRO_FEATURES_append = " systemd"
VIRTUAL-RUNTIME_init_manager = "systemd"
Doing so causes your main image to use the
packagegroup-core-boot.bb
recipe and systemd. The rescue/minimal
image cannot use this package group. However, it can install SysVinit
and the appropriate packages will have support for both systemd and
SysVinit.
3.25 Selecting a Device Manager
The Yocto Project provides multiple ways to manage the device manager
(/dev
):
Persistent and Pre-Populated
/dev
: For this case, the/dev
directory is persistent and the required device nodes are created during the build.Use
devtmpfs
with a Device Manager: For this case, the/dev
directory is provided by the kernel as an in-memory file system and is automatically populated by the kernel at runtime. Additional configuration of device nodes is done in user space by a device manager likeudev
orbusybox-mdev
.
3.25.1 Using Persistent and Pre-Populated/dev
To use the static method for device population, you need to set the USE_DEVFS variable to “0” as follows:
USE_DEVFS = "0"
The content of the resulting /dev
directory is defined in a Device
Table file. The
IMAGE_DEVICE_TABLES
variable defines the Device Table to use and should be set in the
machine or distro configuration file. Alternatively, you can set this
variable in your local.conf
configuration file.
If you do not define the IMAGE_DEVICE_TABLES
variable, the default
device_table-minimal.txt
is used:
IMAGE_DEVICE_TABLES = "device_table-mymachine.txt"
The population is handled by the makedevs
utility during image
creation:
3.25.2 Using devtmpfs
and a Device Manager
To use the dynamic method for device population, you need to use (or be sure to set) the USE_DEVFS variable to “1”, which is the default:
USE_DEVFS = "1"
With this
setting, the resulting /dev
directory is populated by the kernel
using devtmpfs
. Make sure the corresponding kernel configuration
variable CONFIG_DEVTMPFS
is set when building you build a Linux
kernel.
All devices created by devtmpfs
will be owned by root
and have
permissions 0600
.
To have more control over the device nodes, you can use a device manager
like udev
or busybox-mdev
. You choose the device manager by
defining the VIRTUAL-RUNTIME_dev_manager
variable in your machine or
distro configuration file. Alternatively, you can set this variable in
your local.conf
configuration file:
VIRTUAL-RUNTIME_dev_manager = "udev"
# Some alternative values
# VIRTUAL-RUNTIME_dev_manager = "busybox-mdev"
# VIRTUAL-RUNTIME_dev_manager = "systemd"
3.26 Using an External SCM
If you’re working on a recipe that pulls from an external Source Code Manager (SCM), it is possible to have the OpenEmbedded build system notice new recipe changes added to the SCM and then build the resulting packages that depend on the new recipes by using the latest versions. This only works for SCMs from which it is possible to get a sensible revision number for changes. Currently, you can do this with Apache Subversion (SVN), Git, and Bazaar (BZR) repositories.
To enable this behavior, the PV of the recipe needs to reference SRCPV. Here is an example:
PV = "1.2.3+git${SRCPV}"
Then, you can add the following to your
local.conf
:
SRCREV_pn-PN = "${AUTOREV}"
PN is the name of the recipe for which you want to enable automatic source revision updating.
If you do not want to update your local configuration file, you can add the following directly to the recipe to finish enabling the feature:
SRCREV = "${AUTOREV}"
The Yocto Project provides a distribution named poky-bleeding
, whose
configuration file contains the line:
require conf/distro/include/poky-floating-revisions.inc
This line pulls in the listed include file that contains numerous lines of exactly that form:
#SRCREV_pn-opkg-native ?= "${AUTOREV}"
#SRCREV_pn-opkg-sdk ?= "${AUTOREV}"
#SRCREV_pn-opkg ?= "${AUTOREV}"
#SRCREV_pn-opkg-utils-native ?= "${AUTOREV}"
#SRCREV_pn-opkg-utils ?= "${AUTOREV}"
SRCREV_pn-gconf-dbus ?= "${AUTOREV}"
SRCREV_pn-matchbox-common ?= "${AUTOREV}"
SRCREV_pn-matchbox-config-gtk ?= "${AUTOREV}"
SRCREV_pn-matchbox-desktop ?= "${AUTOREV}"
SRCREV_pn-matchbox-keyboard ?= "${AUTOREV}"
SRCREV_pn-matchbox-panel-2 ?= "${AUTOREV}"
SRCREV_pn-matchbox-themes-extra ?= "${AUTOREV}"
SRCREV_pn-matchbox-terminal ?= "${AUTOREV}"
SRCREV_pn-matchbox-wm ?= "${AUTOREV}"
SRCREV_pn-settings-daemon ?= "${AUTOREV}"
SRCREV_pn-screenshot ?= "${AUTOREV}"
. . .
These lines allow you to experiment with building a distribution that tracks the latest development source for numerous packages.
Note
The poky-bleeding
distribution is not tested on a regular basis. Keep
this in mind if you use it.
3.27 Creating a Read-Only Root Filesystem
Suppose, for security reasons, you need to disable your target device’s root filesystem’s write permissions (i.e. you need a read-only root filesystem). Or, perhaps you are running the device’s operating system from a read-only storage device. For either case, you can customize your image for that behavior.
Note
Supporting a read-only root filesystem requires that the system and applications do not try to write to the root filesystem. You must configure all parts of the target system to write elsewhere, or to gracefully fail in the event of attempting to write to the root filesystem.
3.27.1 Creating the Root Filesystem
To create the read-only root filesystem, simply add the
“read-only-rootfs” feature to your image, normally in one of two ways.
The first way is to add the “read-only-rootfs” image feature in the
image’s recipe file via the IMAGE_FEATURES
variable:
IMAGE_FEATURES += "read-only-rootfs"
As an alternative, you can add the same feature
from within your build directory’s local.conf
file with the
associated EXTRA_IMAGE_FEATURES
variable, as in:
EXTRA_IMAGE_FEATURES = "read-only-rootfs"
For more information on how to use these variables, see the “Customizing Images Using Custom IMAGE_FEATURES and EXTRA_IMAGE_FEATURES” section. For information on the variables, see IMAGE_FEATURES and EXTRA_IMAGE_FEATURES.
3.27.2 Post-Installation Scripts and Read-Only Root Filesystem
It is very important that you make sure all post-Installation
(pkg_postinst
) scripts for packages that are installed into the
image can be run at the time when the root filesystem is created during
the build on the host system. These scripts cannot attempt to run during
first-boot on the target device. With the “read-only-rootfs” feature
enabled, the build system checks during root filesystem creation to make
sure all post-installation scripts succeed. If any of these scripts
still need to be run after the root filesystem is created, the build
immediately fails. These build-time checks ensure that the build fails
rather than the target device fails later during its initial boot
operation.
Most of the common post-installation scripts generated by the build system for the out-of-the-box Yocto Project are engineered so that they can run during root filesystem creation (e.g. post-installation scripts for caching fonts). However, if you create and add custom scripts, you need to be sure they can be run during this file system creation.
Here are some common problems that prevent post-installation scripts from running during root filesystem creation:
Not using $D in front of absolute paths: The build system defines
$
D when the root filesystem is created. Furthermore,$D
is blank when the script is run on the target device. This implies two purposes for$D
: ensuring paths are valid in both the host and target environments, and checking to determine which environment is being used as a method for taking appropriate actions.Attempting to run processes that are specific to or dependent on the target architecture: You can work around these attempts by using native tools, which run on the host system, to accomplish the same tasks, or by alternatively running the processes under QEMU, which has the
qemu_run_binary
function. For more information, see the qemu class.
3.27.3 Areas With Write Access
With the “read-only-rootfs” feature enabled, any attempt by the target
to write to the root filesystem at runtime fails. Consequently, you must
make sure that you configure processes and applications that attempt
these types of writes do so to directories with write access (e.g.
/tmp
or /var/run
).
3.28 Maintaining Build Output Quality
Many factors can influence the quality of a build. For example, if you upgrade a recipe to use a new version of an upstream software package or you experiment with some new configuration options, subtle changes can occur that you might not detect until later. Consider the case where your recipe is using a newer version of an upstream package. In this case, a new version of a piece of software might introduce an optional dependency on another library, which is auto-detected. If that library has already been built when the software is building, the software will link to the built library and that library will be pulled into your image along with the new software even if you did not want the library.
The buildhistory class exists to help you maintain the quality of your build output. You can use the class to highlight unexpected and possibly unwanted changes in the build output. When you enable build history, it records information about the contents of each package and image and then commits that information to a local Git repository where you can examine the information.
The remainder of this section describes the following:
3.28.1 Enabling and Disabling Build History
Build history is disabled by default. To enable it, add the following
INHERIT
statement and set the
BUILDHISTORY_COMMIT
variable to “1” at the end of your conf/local.conf
file found in the
Build Directory:
INHERIT += "buildhistory"
BUILDHISTORY_COMMIT = "1"
Enabling build history as previously described causes the OpenEmbedded build system to collect build output information and commit it as a single commit to a local Git repository.
Note
Enabling build history increases your build times slightly, particularly for images, and increases the amount of disk space used during the build.
You can disable build history by removing the previous statements from
your conf/local.conf
file.
3.28.2 Understanding What the Build History Contains
Build history information is kept in
${
TOPDIR}/buildhistory
in the Build Directory as defined by the
BUILDHISTORY_DIR
variable. The following is an example abbreviated listing:
At the top level, a metadata-revs
file exists that lists the
revisions of the repositories for the enabled layers when the build was
produced. The rest of the data splits into separate packages
,
images
and sdk
directories, the contents of which are described
as follows.
3.28.2.1 Build History Package Information
The history for each package contains a text file that has name-value
pairs with information about the package. For example,
buildhistory/packages/i586-poky-linux/busybox/busybox/latest
contains the following:
PV = 1.22.1
PR = r32
RPROVIDES =
RDEPENDS = glibc (>= 2.20) update-alternatives-opkg
RRECOMMENDS = busybox-syslog busybox-udhcpc update-rc.d
PKGSIZE = 540168
FILES = /usr/bin/* /usr/sbin/* /usr/lib/busybox/* /usr/lib/lib*.so.* \
/etc /com /var /bin/* /sbin/* /lib/*.so.* /lib/udev/rules.d \
/usr/lib/udev/rules.d /usr/share/busybox /usr/lib/busybox/* \
/usr/share/pixmaps /usr/share/applications /usr/share/idl \
/usr/share/omf /usr/share/sounds /usr/lib/bonobo/servers
FILELIST = /bin/busybox /bin/busybox.nosuid /bin/busybox.suid /bin/sh \
/etc/busybox.links.nosuid /etc/busybox.links.suid
Most of these
name-value pairs correspond to variables used to produce the package.
The exceptions are FILELIST
, which is the actual list of files in
the package, and PKGSIZE
, which is the total size of files in the
package in bytes.
A file also exists that corresponds to the recipe from which the package
came (e.g. buildhistory/packages/i586-poky-linux/busybox/latest
):
PV = 1.22.1
PR = r32
DEPENDS = initscripts kern-tools-native update-rc.d-native \
virtual/i586-poky-linux-compilerlibs virtual/i586-poky-linux-gcc \
virtual/libc virtual/update-alternatives
PACKAGES = busybox-ptest busybox-httpd busybox-udhcpd busybox-udhcpc \
busybox-syslog busybox-mdev busybox-hwclock busybox-dbg \
busybox-staticdev busybox-dev busybox-doc busybox-locale busybox
Finally, for those recipes fetched from a version control system (e.g.,
Git), a file exists that lists source revisions that are specified in
the recipe and lists the actual revisions used during the build. Listed
and actual revisions might differ when
SRCREV is set to
${AUTOREV}. Here is an
example assuming
buildhistory/packages/qemux86-poky-linux/linux-yocto/latest_srcrev
):
# SRCREV_machine = "38cd560d5022ed2dbd1ab0dca9642e47c98a0aa1"
SRCREV_machine = "38cd560d5022ed2dbd1ab0dca9642e47c98a0aa1"
# SRCREV_meta = "a227f20eff056e511d504b2e490f3774ab260d6f"
SRCREV_meta ="a227f20eff056e511d504b2e490f3774ab260d6f"
You can use the
buildhistory-collect-srcrevs
command with the -a
option to
collect the stored SRCREV
values from build history and report them
in a format suitable for use in global configuration (e.g.,
local.conf
or a distro include file) to override floating
AUTOREV
values to a fixed set of revisions. Here is some example
output from this command:
$ buildhistory-collect-srcrevs -a
# i586-poky-linux
SRCREV_pn-glibc = "b8079dd0d360648e4e8de48656c5c38972621072"
SRCREV_pn-glibc-initial = "b8079dd0d360648e4e8de48656c5c38972621072"
SRCREV_pn-opkg-utils = "53274f087565fd45d8452c5367997ba6a682a37a"
SRCREV_pn-kmod = "fd56638aed3fe147015bfa10ed4a5f7491303cb4"
# x86_64-linux
SRCREV_pn-gtk-doc-stub-native = "1dea266593edb766d6d898c79451ef193eb17cfa"
SRCREV_pn-dtc-native = "65cc4d2748a2c2e6f27f1cf39e07a5dbabd80ebf"
SRCREV_pn-update-rc.d-native = "eca680ddf28d024954895f59a241a622dd575c11"
SRCREV_glibc_pn-cross-localedef-native = "b8079dd0d360648e4e8de48656c5c38972621072"
SRCREV_localedef_pn-cross-localedef-native = "c833367348d39dad7ba018990bfdaffaec8e9ed3"
SRCREV_pn-prelink-native = "faa069deec99bf61418d0bab831c83d7c1b797ca"
SRCREV_pn-opkg-utils-native = "53274f087565fd45d8452c5367997ba6a682a37a"
SRCREV_pn-kern-tools-native = "23345b8846fe4bd167efdf1bd8a1224b2ba9a5ff"
SRCREV_pn-kmod-native = "fd56638aed3fe147015bfa10ed4a5f7491303cb4"
# qemux86-poky-linux
SRCREV_machine_pn-linux-yocto = "38cd560d5022ed2dbd1ab0dca9642e47c98a0aa1"
SRCREV_meta_pn-linux-yocto = "a227f20eff056e511d504b2e490f3774ab260d6f"
# all-poky-linux
SRCREV_pn-update-rc.d = "eca680ddf28d024954895f59a241a622dd575c11"
Note
Here are some notes on using the buildhistory-collect-srcrevs
command:
By default, only values where the
SRCREV
was not hardcoded (usually whenAUTOREV
is used) are reported. Use the-a
option to see allSRCREV
values.The output statements might not have any effect if overrides are applied elsewhere in the build system configuration. Use the
-f
option to add theforcevariable
override to each output line if you need to work around this restriction.The script does apply special handling when building for multiple machines. However, the script does place a comment before each set of values that specifies which triplet to which they belong as previously shown (e.g.,
i586-poky-linux
).
3.28.2.2 Build History Image Information
The files produced for each image are as follows:
image-files:
A directory containing selected files from the root filesystem. The files are defined by BUILDHISTORY_IMAGE_FILES.build-id.txt:
Human-readable information about the build configuration and metadata source revisions. This file contains the full build header as printed by BitBake.*.dot:
Dependency graphs for the image that are compatible withgraphviz
.files-in-image.txt:
A list of files in the image with permissions, owner, group, size, and symlink information.image-info.txt:
A text file containing name-value pairs with information about the image. See the following listing example for more information.installed-package-names.txt:
A list of installed packages by name only.installed-package-sizes.txt:
A list of installed packages ordered by size.installed-packages.txt:
A list of installed packages with full package filenames.
Note
Installed package information is able to be gathered and produced even if package management is disabled for the final image.
Here is an example of image-info.txt
:
DISTRO = poky
DISTRO_VERSION = 1.7
USER_CLASSES = buildstats image-mklibs image-prelink
IMAGE_CLASSES = image_types
IMAGE_FEATURES = debug-tweaks
IMAGE_LINGUAS =
IMAGE_INSTALL = packagegroup-core-boot run-postinsts
BAD_RECOMMENDATIONS =
NO_RECOMMENDATIONS =
PACKAGE_EXCLUDE =
ROOTFS_POSTPROCESS_COMMAND = write_package_manifest; license_create_manifest; \
write_image_manifest ; buildhistory_list_installed_image ; \
buildhistory_get_image_installed ; ssh_allow_empty_password; \
postinst_enable_logging; rootfs_update_timestamp ; ssh_disable_dns_lookup ;
IMAGE_POSTPROCESS_COMMAND = buildhistory_get_imageinfo ;
IMAGESIZE = 6900
Other than IMAGESIZE
,
which is the total size of the files in the image in Kbytes, the
name-value pairs are variables that may have influenced the content of
the image. This information is often useful when you are trying to
determine why a change in the package or file listings has occurred.
3.28.2.3 Using Build History to Gather Image Information Only
As you can see, build history produces image information, including
dependency graphs, so you can see why something was pulled into the
image. If you are just interested in this information and not interested
in collecting specific package or SDK information, you can enable
writing only image information without any history by adding the
following to your conf/local.conf
file found in the
Build Directory:
INHERIT += "buildhistory"
BUILDHISTORY_COMMIT = "0"
BUILDHISTORY_FEATURES = "image"
Here, you set the BUILDHISTORY_FEATURES variable to use the image feature only.
3.28.2.4 Build History SDK Information
Build history collects similar information on the contents of SDKs (e.g.
bitbake -c populate_sdk imagename
) as compared to information it
collects for images. Furthermore, this information differs depending on
whether an extensible or standard SDK is being produced.
The following list shows the files produced for SDKs:
files-in-sdk.txt:
A list of files in the SDK with permissions, owner, group, size, and symlink information. This list includes both the host and target parts of the SDK.sdk-info.txt:
A text file containing name-value pairs with information about the SDK. See the following listing example for more information.sstate-task-sizes.txt:
A text file containing name-value pairs with information about task group sizes (e.g.do_populate_sysroot
tasks have a total size). Thesstate-task-sizes.txt
file exists only when an extensible SDK is created.sstate-package-sizes.txt:
A text file containing name-value pairs with information for the shared-state packages and sizes in the SDK. Thesstate-package-sizes.txt
file exists only when an extensible SDK is created.sdk-files:
A folder that contains copies of the files mentioned inBUILDHISTORY_SDK_FILES
if the files are present in the output. Additionally, the default value ofBUILDHISTORY_SDK_FILES
is specific to the extensible SDK although you can set it differently if you would like to pull in specific files from the standard SDK.The default files are
conf/local.conf
,conf/bblayers.conf
,conf/auto.conf
,conf/locked-sigs.inc
, andconf/devtool.conf
. Thus, for an extensible SDK, these files get copied into thesdk-files
directory.The following information appears under each of the
host
andtarget
directories for the portions of the SDK that run on the host and on the target, respectively:Note
The following files for the most part are empty when producing an extensible SDK because this type of SDK is not constructed from packages as is the standard SDK.
depends.dot:
Dependency graph for the SDK that is compatible withgraphviz
.installed-package-names.txt:
A list of installed packages by name only.installed-package-sizes.txt:
A list of installed packages ordered by size.installed-packages.txt:
A list of installed packages with full package filenames.
Here is an example of sdk-info.txt
:
DISTRO = poky
DISTRO_VERSION = 1.3+snapshot-20130327
SDK_NAME = poky-glibc-i686-arm
SDK_VERSION = 1.3+snapshot
SDKMACHINE =
SDKIMAGE_FEATURES = dev-pkgs dbg-pkgs
BAD_RECOMMENDATIONS =
SDKSIZE = 352712
Other than SDKSIZE
, which is
the total size of the files in the SDK in Kbytes, the name-value pairs
are variables that might have influenced the content of the SDK. This
information is often useful when you are trying to determine why a
change in the package or file listings has occurred.
3.28.2.5 Examining Build History Information
You can examine build history output from the command line or from a web interface.
To see any changes that have occurred (assuming you have BUILDHISTORY_COMMIT = “1”), you can simply use any Git command that allows you to view the history of a repository. Here is one method:
$ git log -p
You need to realize, however, that this method does show changes that are not significant (e.g. a package’s size changing by a few bytes).
A command-line tool called buildhistory-diff
does exist, though,
that queries the Git repository and prints just the differences that
might be significant in human-readable form. Here is an example:
$ poky/poky/scripts/buildhistory-diff . HEAD^
Changes to images/qemux86_64/glibc/core-image-minimal (files-in-image.txt):
/etc/anotherpkg.conf was added
/sbin/anotherpkg was added
* (installed-package-names.txt):
* anotherpkg was added
Changes to images/qemux86_64/glibc/core-image-minimal (installed-package-names.txt):
anotherpkg was added
packages/qemux86_64-poky-linux/v86d: PACKAGES: added "v86d-extras"
* PR changed from "r0" to "r1"
* PV changed from "0.1.10" to "0.1.12"
packages/qemux86_64-poky-linux/v86d/v86d: PKGSIZE changed from 110579 to 144381 (+30%)
* PR changed from "r0" to "r1"
* PV changed from "0.1.10" to "0.1.12"
Note
The buildhistory-diff
tool requires the GitPython
package. Be sure to install it using Pip3 as follows:
$ pip3 install GitPython --user
Alternatively, you can install python3-git
using the appropriate
distribution package manager (e.g. apt-get
, dnf
, or zipper
).
To see changes to the build history using a web interface, follow the
instruction in the README
file
here.
Here is a sample screenshot of the interface:
3.29 Performing Automated Runtime Testing
The OpenEmbedded build system makes available a series of automated
tests for images to verify runtime functionality. You can run these
tests on either QEMU or actual target hardware. Tests are written in
Python making use of the unittest
module, and the majority of them
run commands on the target system over SSH. This section describes how
you set up the environment to use these tests, run available tests, and
write and add your own tests.
For information on the test and QA infrastructure available within the Yocto Project, see the “Testing and Quality Assurance” section in the Yocto Project Reference Manual.
3.29.1 Enabling Tests
Depending on whether you are planning to run tests using QEMU or on the hardware, you have to take different steps to enable the tests. See the following subsections for information on how to enable both types of tests.
3.29.1.1 Enabling Runtime Tests on QEMU
In order to run tests, you need to do the following:
Set up to avoid interaction with sudo for networking: To accomplish this, you must do one of the following:
Add
NOPASSWD
for your user in/etc/sudoers
either for all commands or just forrunqemu-ifup
. You must provide the full path as that can change if you are using multiple clones of the source repository.Note
On some distributions, you also need to comment out “Defaults requiretty” in
/etc/sudoers
.Manually configure a tap interface for your system.
Run as root the script in
scripts/runqemu-gen-tapdevs
, which should generate a list of tap devices. This is the option typically chosen for Autobuilder-type environments.Note
Be sure to use an absolute path when calling this script with sudo.
The package recipe
qemu-helper-native
is required to run this script. Build the package using the following command:$ bitbake qemu-helper-native
Set the DISPLAY variable: You need to set this variable so that you have an X server available (e.g. start
vncserver
for a headless machine).Be sure your host’s firewall accepts incoming connections from 192.168.7.0/24: Some of the tests (in particular DNF tests) start an HTTP server on a random high number port, which is used to serve files to the target. The DNF module serves
${WORKDIR}/oe-rootfs-repo
so it can run DNF channel commands. That means your host’s firewall must accept incoming connections from 192.168.7.0/24, which is the default IP range used for tap devices byrunqemu
.Be sure your host has the correct packages installed: Depending your host’s distribution, you need to have the following packages installed:
Ubuntu and Debian:
sysstat
andiproute2
openSUSE:
sysstat
andiproute2
Fedora:
sysstat
andiproute
CentOS:
sysstat
andiproute
Once you start running the tests, the following happens:
A copy of the root filesystem is written to
${WORKDIR}/testimage
.The image is booted under QEMU using the standard
runqemu
script.A default timeout of 500 seconds occurs to allow for the boot process to reach the login prompt. You can change the timeout period by setting TEST_QEMUBOOT_TIMEOUT in the
local.conf
file.Once the boot process is reached and the login prompt appears, the tests run. The full boot log is written to
${WORKDIR}/testimage/qemu_boot_log
.Each test module loads in the order found in
TEST_SUITES
. You can find the full output of the commands run over SSH in${WORKDIR}/testimgage/ssh_target_log
.If no failures occur, the task running the tests ends successfully. You can find the output from the
unittest
in the task log at${WORKDIR}/temp/log.do_testimage
.
3.29.1.2 Enabling Runtime Tests on Hardware
The OpenEmbedded build system can run tests on real hardware, and for certain devices it can also deploy the image to be tested onto the device beforehand.
For automated deployment, a “master image” is installed onto the hardware once as part of setup. Then, each time tests are to be run, the following occurs:
The master image is booted into and used to write the image to be tested to a second partition.
The device is then rebooted using an external script that you need to provide.
The device boots into the image to be tested.
When running tests (independent of whether the image has been deployed automatically or not), the device is expected to be connected to a network on a pre-determined IP address. You can either use static IP addresses written into the image, or set the image to use DHCP and have your DHCP server on the test network assign a known IP address based on the MAC address of the device.
In order to run tests on hardware, you need to set TEST_TARGET
to an
appropriate value. For QEMU, you do not have to change anything, the
default value is “qemu”. For running tests on hardware, the following
options exist:
“simpleremote”: Choose “simpleremote” if you are going to run tests on a target system that is already running the image to be tested and is available on the network. You can use “simpleremote” in conjunction with either real hardware or an image running within a separately started QEMU or any other virtual machine manager.
“SystemdbootTarget”: Choose “SystemdbootTarget” if your hardware is an EFI-based machine with
systemd-boot
as bootloader andcore-image-testmaster
(or something similar) is installed. Also, your hardware under test must be in a DHCP-enabled network that gives it the same IP address for each reboot.If you choose “SystemdbootTarget”, there are additional requirements and considerations. See the “Selecting SystemdbootTarget” section, which follows, for more information.
“BeagleBoneTarget”: Choose “BeagleBoneTarget” if you are deploying images and running tests on the BeagleBone “Black” or original “White” hardware. For information on how to use these tests, see the comments at the top of the BeagleBoneTarget
meta-yocto-bsp/lib/oeqa/controllers/beaglebonetarget.py
file.“EdgeRouterTarget”: Choose “EdgeRouterTarget” if you are deploying images and running tests on the Ubiquiti Networks EdgeRouter Lite. For information on how to use these tests, see the comments at the top of the EdgeRouterTarget
meta-yocto-bsp/lib/oeqa/controllers/edgeroutertarget.py
file.“GrubTarget”: Choose “GrubTarget” if you are deploying images and running tests on any generic PC that boots using GRUB. For information on how to use these tests, see the comments at the top of the GrubTarget
meta-yocto-bsp/lib/oeqa/controllers/grubtarget.py
file.“your-target”: Create your own custom target if you want to run tests when you are deploying images and running tests on a custom machine within your BSP layer. To do this, you need to add a Python unit that defines the target class under
lib/oeqa/controllers/
within your layer. You must also provide an empty__init__.py
. For examples, see files inmeta-yocto-bsp/lib/oeqa/controllers/
.
3.29.1.3 Selecting SystemdbootTarget
If you did not set TEST_TARGET
to “SystemdbootTarget”, then you do
not need any information in this section. You can skip down to the
“Running Tests” section.
If you did set TEST_TARGET
to “SystemdbootTarget”, you also need to
perform a one-time setup of your master image by doing the following:
Set EFI_PROVIDER: Be sure that
EFI_PROVIDER
is as follows:EFI_PROVIDER = "systemd-boot"
Build the master image: Build the
core-image-testmaster
image. Thecore-image-testmaster
recipe is provided as an example for a “master” image and you can customize the image recipe as you would any other recipe.Here are the image recipe requirements:
Inherits
core-image
so that kernel modules are installed.Installs normal linux utilities not BusyBox ones (e.g.
bash
,coreutils
,tar
,gzip
, andkmod
).Uses a custom Initial RAM Disk (initramfs) image with a custom installer. A normal image that you can install usually creates a single rootfs partition. This image uses another installer that creates a specific partition layout. Not all Board Support Packages (BSPs) can use an installer. For such cases, you need to manually create the following partition layout on the target:
First partition mounted under
/boot
, labeled “boot”.The main rootfs partition where this image gets installed, which is mounted under
/
.Another partition labeled “testrootfs” where test images get deployed.
Install image: Install the image that you just built on the target system.
The final thing you need to do when setting TEST_TARGET
to
“SystemdbootTarget” is to set up the test image:
Set up your local.conf file: Make sure you have the following statements in your
local.conf
file:IMAGE_FSTYPES += "tar.gz" INHERIT += "testimage" TEST_TARGET = "SystemdbootTarget" TEST_TARGET_IP = "192.168.2.3"
Build your test image: Use BitBake to build the image:
$ bitbake core-image-sato
3.29.1.4 Power Control
For most hardware targets other than “simpleremote”, you can control power:
You can use
TEST_POWERCONTROL_CMD
together withTEST_POWERCONTROL_EXTRA_ARGS
as a command that runs on the host and does power cycling. The test code passes one argument to that command: off, on or cycle (off then on). Here is an example that could appear in yourlocal.conf
file:TEST_POWERCONTROL_CMD = "powercontrol.exp test 10.11.12.1 nuc1"
In this example, the expect script does the following:
ssh test@10.11.12.1 "pyctl nuc1 arg"
It then runs a Python script that controls power for a label called
nuc1
.Note
You need to customize
TEST_POWERCONTROL_CMD
andTEST_POWERCONTROL_EXTRA_ARGS
for your own setup. The one requirement is that it accepts “on”, “off”, and “cycle” as the last argument.When no command is defined, it connects to the device over SSH and uses the classic reboot command to reboot the device. Classic reboot is fine as long as the machine actually reboots (i.e. the SSH test has not failed). It is useful for scenarios where you have a simple setup, typically with a single board, and where some manual interaction is okay from time to time.
If you have no hardware to automatically perform power control but still
wish to experiment with automated hardware testing, you can use the
dialog-power-control
script that shows a dialog prompting you to perform
the required power action. This script requires either KDialog or Zenity
to be installed. To use this script, set the
TEST_POWERCONTROL_CMD
variable as follows:
TEST_POWERCONTROL_CMD = "${COREBASE}/scripts/contrib/dialog-power-control"
3.29.1.5 Serial Console Connection
For test target classes requiring a serial console to interact with the bootloader (e.g. BeagleBoneTarget, EdgeRouterTarget, and GrubTarget), you need to specify a command to use to connect to the serial console of the target machine by using the TEST_SERIALCONTROL_CMD variable and optionally the TEST_SERIALCONTROL_EXTRA_ARGS variable.
These cases could be a serial terminal program if the machine is
connected to a local serial port, or a telnet
or ssh
command
connecting to a remote console server. Regardless of the case, the
command simply needs to connect to the serial console and forward that
connection to standard input and output as any normal terminal program
does. For example, to use the picocom terminal program on serial device
/dev/ttyUSB0
at 115200bps, you would set the variable as follows:
TEST_SERIALCONTROL_CMD = "picocom /dev/ttyUSB0 -b 115200"
For local
devices where the serial port device disappears when the device reboots,
an additional “serdevtry” wrapper script is provided. To use this
wrapper, simply prefix the terminal command with
${COREBASE}/scripts/contrib/serdevtry
:
TEST_SERIALCONTROL_CMD = "${COREBASE}/scripts/contrib/serdevtry picocom -b 115200 /dev/ttyUSB0"
3.29.2 Running Tests
You can start the tests automatically or manually:
Automatically running tests: To run the tests automatically after the OpenEmbedded build system successfully creates an image, first set the TESTIMAGE_AUTO variable to “1” in your
local.conf
file in the Build Directory:TESTIMAGE_AUTO = "1"
Next, build your image. If the image successfully builds, the tests run:
bitbake core-image-sato
Manually running tests: To manually run the tests, first globally inherit the testimage class by editing your
local.conf
file:INHERIT += "testimage"
Next, use BitBake to run the tests:
bitbake -c testimage image
All test files reside in meta/lib/oeqa/runtime
in the
Source Directory. A test name maps
directly to a Python module. Each test module may contain a number of
individual tests. Tests are usually grouped together by the area tested
(e.g tests for systemd reside in meta/lib/oeqa/runtime/systemd.py
).
You can add tests to any layer provided you place them in the proper
area and you extend BBPATH in
the local.conf
file as normal. Be sure that tests reside in
layer/lib/oeqa/runtime
.
Note
Be sure that module names do not collide with module names used in
the default set of test modules in meta/lib/oeqa/runtime
.
You can change the set of tests run by appending or overriding
TEST_SUITES variable in
local.conf
. Each name in TEST_SUITES
represents a required test
for the image. Test modules named within TEST_SUITES
cannot be
skipped even if a test is not suitable for an image (e.g. running the
RPM tests on an image without rpm
). Appending “auto” to
TEST_SUITES
causes the build system to try to run all tests that are
suitable for the image (i.e. each test module may elect to skip itself).
The order you list tests in TEST_SUITES
is important and influences
test dependencies. Consequently, tests that depend on other tests should
be added after the test on which they depend. For example, since the
ssh
test depends on the ping
test, “ssh” needs to come after
“ping” in the list. The test class provides no re-ordering or dependency
handling.
Note
Each module can have multiple classes with multiple test methods.
And, Python unittest
rules apply.
Here are some things to keep in mind when running tests:
The default tests for the image are defined as:
DEFAULT_TEST_SUITES_pn-image = "ping ssh df connman syslog xorg scp vnc date rpm dnf dmesg"
Add your own test to the list of the by using the following:
TEST_SUITES_append = " mytest"
Run a specific list of tests as follows:
TEST_SUITES = "test1 test2 test3"
Remember, order is important. Be sure to place a test that is dependent on another test later in the order.
3.29.3 Exporting Tests
You can export tests so that they can run independently of the build system. Exporting tests is required if you want to be able to hand the test execution off to a scheduler. You can only export tests that are defined in TEST_SUITES.
If your image is already built, make sure the following are set in your
local.conf
file:
INHERIT += "testexport"
TEST_TARGET_IP = "IP-address-for-the-test-target"
TEST_SERVER_IP = "IP-address-for-the-test-server"
You can then export the tests with the following BitBake command form:
$ bitbake image -c testexport
Exporting the tests places them in the
Build Directory in
tmp/testexport/
image, which is controlled by the
TEST_EXPORT_DIR
variable.
You can now run the tests outside of the build environment:
$ cd tmp/testexport/image
$ ./runexported.py testdata.json
Here is a complete example that shows IP addresses and uses the
core-image-sato
image:
INHERIT += "testexport"
TEST_TARGET_IP = "192.168.7.2"
TEST_SERVER_IP = "192.168.7.1"
Use BitBake to export the tests:
$ bitbake core-image-sato -c testexport
Run the tests outside of the build environment using the following:
$ cd tmp/testexport/core-image-sato
$ ./runexported.py testdata.json
3.29.4 Writing New Tests
As mentioned previously, all new test files need to be in the proper
place for the build system to find them. New tests for additional
functionality outside of the core should be added to the layer that adds
the functionality, in layer/lib/oeqa/runtime
(as long as
BBPATH is extended in the
layer’s layer.conf
file as normal). Just remember the following:
Filenames need to map directly to test (module) names.
Do not use module names that collide with existing core tests.
Minimally, an empty
__init__.py
file must exist in the runtime directory.
To create a new test, start by copying an existing module (e.g.
syslog.py
or gcc.py
are good ones to use). Test modules can use
code from meta/lib/oeqa/utils
, which are helper classes.
Note
Structure shell commands such that you rely on them and they return a
single code for success. Be aware that sometimes you will need to
parse the output. See the df.py
and date.py
modules for examples.
You will notice that all test classes inherit oeRuntimeTest
, which
is found in meta/lib/oetest.py
. This base class offers some helper
attributes, which are described in the following sections:
3.29.4.1 Class Methods
Class methods are as follows:
hasPackage(pkg): Returns “True” if
pkg
is in the installed package list of the image, which is based on the manifest file that is generated during thedo_rootfs
task.hasFeature(feature): Returns “True” if the feature is in IMAGE_FEATURES or DISTRO_FEATURES.
3.29.4.2 Class Attributes
Class attributes are as follows:
pscmd: Equals “ps -ef” if
procps
is installed in the image. Otherwise,pscmd
equals “ps” (busybox).tc: The called test context, which gives access to the following attributes:
d: The BitBake datastore, which allows you to use stuff such as
oeRuntimeTest.tc.d.getVar("VIRTUAL-RUNTIME_init_manager")
.testslist and testsrequired: Used internally. The tests do not need these.
filesdir: The absolute path to
meta/lib/oeqa/runtime/files
, which contains helper files for tests meant for copying on the target such as small files written in C for compilation.target: The target controller object used to deploy and start an image on a particular target (e.g. Qemu, SimpleRemote, and SystemdbootTarget). Tests usually use the following:
ip: The target’s IP address.
server_ip: The host’s IP address, which is usually used by the DNF test suite.
run(cmd, timeout=None): The single, most used method. This command is a wrapper for:
ssh root@host "cmd"
. The command returns a tuple: (status, output), which are what their names imply - the return code of “cmd” and whatever output it produces. The optional timeout argument represents the number of seconds the test should wait for “cmd” to return. If the argument is “None”, the test uses the default instance’s timeout period, which is 300 seconds. If the argument is “0”, the test runs until the command returns.copy_to(localpath, remotepath):
scp localpath root@ip:remotepath
.copy_from(remotepath, localpath):
scp root@host:remotepath localpath
.
3.29.4.3 Instance Attributes
A single instance attribute exists, which is target
. The target
instance attribute is identical to the class attribute of the same name,
which is described in the previous section. This attribute exists as
both an instance and class attribute so tests can use
self.target.run(cmd)
in instance methods instead of
oeRuntimeTest.tc.target.run(cmd)
.
3.29.5 Installing Packages in the DUT Without the Package Manager
When a test requires a package built by BitBake, it is possible to
install that package. Installing the package does not require a package
manager be installed in the device under test (DUT). It does, however,
require an SSH connection and the target must be using the
sshcontrol
class.
Note
This method uses scp
to copy files from the host to the target, which
causes permissions and special attributes to be lost.
A JSON file is used to define the packages needed by a test. This file
must be in the same path as the file used to define the tests.
Furthermore, the filename must map directly to the test module name with
a .json
extension.
The JSON file must include an object with the test name as keys of an object or an array. This object (or array of objects) uses the following data:
“pkg” - A mandatory string that is the name of the package to be installed.
“rm” - An optional boolean, which defaults to “false”, that specifies to remove the package after the test.
“extract” - An optional boolean, which defaults to “false”, that specifies if the package must be extracted from the package format. When set to “true”, the package is not automatically installed into the DUT.
Following is an example JSON file that handles test “foo” installing package “bar” and test “foobar” installing packages “foo” and “bar”. Once the test is complete, the packages are removed from the DUT.
{
"foo": {
"pkg": "bar"
},
"foobar": [
{
"pkg": "foo",
"rm": true
},
{
"pkg": "bar",
"rm": true
}
]
}
3.30 Debugging Tools and Techniques
The exact method for debugging build failures depends on the nature of the problem and on the system’s area from which the bug originates. Standard debugging practices such as comparison against the last known working version with examination of the changes and the re-application of steps to identify the one causing the problem are valid for the Yocto Project just as they are for any other system. Even though it is impossible to detail every possible potential failure, this section provides some general tips to aid in debugging given a variety of situations.
Note
A useful feature for debugging is the error reporting tool. Configuring the Yocto Project to use this tool causes the OpenEmbedded build system to produce error reporting commands as part of the console output. You can enter the commands after the build completes to log error information into a common database, that can help you figure out what might be going wrong. For information on how to enable and use this feature, see the “Using the Error Reporting Tool” section.
The following list shows the debugging topics in the remainder of this section:
“Viewing Logs from Failed Tasks” describes how to find and view logs from tasks that failed during the build process.
“Viewing Variable Values” describes how to use the BitBake
-e
option to examine variable values after a recipe has been parsed.“Viewing Package Information with oe-pkgdata-util” describes how to use the
oe-pkgdata-util
utility to query PKGDATA_DIR and display package-related information for built packages.“Viewing Dependencies Between Recipes and Tasks” describes how to use the BitBake
-g
option to display recipe dependency information used during the build.“Viewing Task Variable Dependencies” describes how to use the
bitbake-dumpsig
command in conjunction with key subdirectories in the Build Directory to determine variable dependencies.“Running Specific Tasks” describes how to use several BitBake options (e.g.
-c
,-C
, and-f
) to run specific tasks in the build chain. It can be useful to run tasks “out-of-order” when trying isolate build issues.“General BitBake Problems” describes how to use BitBake’s
-D
debug output option to reveal more about what BitBake is doing during the build.“Building with No Dependencies” describes how to use the BitBake
-b
option to build a recipe while ignoring dependencies.“Recipe Logging Mechanisms” describes how to use the many recipe logging functions to produce debugging output and report errors and warnings.
“Debugging Parallel Make Races” describes how to debug situations where the build consists of several parts that are run simultaneously and when the output or result of one part is not ready for use with a different part of the build that depends on that output.
“Debugging With the GNU Project Debugger (GDB) Remotely” describes how to use GDB to allow you to examine running programs, which can help you fix problems.
“Debugging with the GNU Project Debugger (GDB) on the Target” describes how to use GDB directly on target hardware for debugging.
“Other Debugging Tips” describes miscellaneous debugging tips that can be useful.
3.30.1 Viewing Logs from Failed Tasks
You can find the log for a task in the file
${
WORKDIR}/temp/log.do_
taskname.
For example, the log for the
do_compile task of the
QEMU minimal image for the x86 machine (qemux86
) might be in
tmp/work/qemux86-poky-linux/core-image-minimal/1.0-r0/temp/log.do_compile
.
To see the commands BitBake ran
to generate a log, look at the corresponding run.do_
taskname file
in the same directory.
log.do_
taskname and run.do_
taskname are actually symbolic
links to log.do_
taskname.
pid and
log.run_
taskname.
pid, where pid is the PID the task had
when it ran. The symlinks always point to the files corresponding to the
most recent run.
3.30.2 Viewing Variable Values
Sometimes you need to know the value of a variable as a result of BitBake’s parsing step. This could be because some unexpected behavior occurred in your project. Perhaps an attempt to modify a variable did not work out as expected.
BitBake’s -e
option is used to display variable values after
parsing. The following command displays the variable values after the
configuration files (i.e. local.conf
, bblayers.conf
,
bitbake.conf
and so forth) have been parsed:
$ bitbake -e
The following command displays variable values after a specific recipe has been parsed. The variables include those from the configuration as well:
$ bitbake -e recipename
Note
Each recipe has its own private set of variables (datastore). Internally, after parsing the configuration, a copy of the resulting datastore is made prior to parsing each recipe. This copying implies that variables set in one recipe will not be visible to other recipes.
Likewise, each task within a recipe gets a private datastore based on the recipe datastore, which means that variables set within one task will not be visible to other tasks.
In the output of bitbake -e
, each variable is preceded by a
description of how the variable got its value, including temporary
values that were later overridden. This description also includes
variable flags (varflags) set on the variable. The output can be very
helpful during debugging.
Variables that are exported to the environment are preceded by
export
in the output of bitbake -e
. See the following example:
export CC="i586-poky-linux-gcc -m32 -march=i586 --sysroot=/home/ulf/poky/build/tmp/sysroots/qemux86"
In addition to variable values, the output of the bitbake -e
and
bitbake -e
recipe commands includes the following information:
The output starts with a tree listing all configuration files and classes included globally, recursively listing the files they include or inherit in turn. Much of the behavior of the OpenEmbedded build system (including the behavior of the Normal Recipe Build Tasks) is implemented in the base class and the classes it inherits, rather than being built into BitBake itself.
After the variable values, all functions appear in the output. For shell functions, variables referenced within the function body are expanded. If a function has been modified using overrides or using override-style operators like
_append
and_prepend
, then the final assembled function body appears in the output.
3.30.3 Viewing Package Information with oe-pkgdata-util
You can use the oe-pkgdata-util
command-line utility to query
PKGDATA_DIR and display
various package-related information. When you use the utility, you must
use it to view information on packages that have already been built.
Following are a few of the available oe-pkgdata-util
subcommands.
Note
You can use the standard * and ? globbing wildcards as part of package names and paths.
oe-pkgdata-util list-pkgs [pattern]
: Lists all packages that have been built, optionally limiting the match to packages that match pattern.oe-pkgdata-util list-pkg-files package ...
: Lists the files and directories contained in the given packages.Note
A different way to view the contents of a package is to look at the
${
WORKDIR}/packages-split
directory of the recipe that generates the package. This directory is created by the do_package task and has one subdirectory for each package the recipe generates, which contains the files stored in that package.If you want to inspect the
${WORKDIR}/packages-split
directory, make sure that rm_work is not enabled when you build the recipe.oe-pkgdata-util find-path path ...
: Lists the names of the packages that contain the given paths. For example, the following tells us that/usr/share/man/man1/make.1
is contained in themake-doc
package:$ oe-pkgdata-util find-path /usr/share/man/man1/make.1 make-doc: /usr/share/man/man1/make.1
oe-pkgdata-util lookup-recipe package ...
: Lists the name of the recipes that produce the given packages.
For more information on the oe-pkgdata-util
command, use the help
facility:
$ oe-pkgdata-util --help
$ oe-pkgdata-util subcommand --help
3.30.4 Viewing Dependencies Between Recipes and Tasks
Sometimes it can be hard to see why BitBake wants to build other recipes before the one you have specified. Dependency information can help you understand why a recipe is built.
To generate dependency information for a recipe, run the following command:
$ bitbake -g recipename
This command writes the following files in the current directory:
pn-buildlist
: A list of recipes/targets involved in building recipename. “Involved” here means that at least one task from the recipe needs to run when building recipename from scratch. Targets that are in ASSUME_PROVIDED are not listed.task-depends.dot
: A graph showing dependencies between tasks.
The graphs are in
DOT
format and can be converted to images (e.g. using the dot
tool from
Graphviz).
Note
DOT files use a plain text format. The graphs generated using the
bitbake -g
command are often so large as to be difficult to read without special pruning (e.g. with Bitbake’s-I
option) and processing. Despite the form and size of the graphs, the corresponding.dot
files can still be possible to read and provide useful information.As an example, the
task-depends.dot
file contains lines such as the following:"libxslt.do_configure" -> "libxml2.do_populate_sysroot"
The above example line reveals that the do_configure task in
libxslt
depends on the do_populate_sysroot task inlibxml2
, which is a normal DEPENDS dependency between the two recipes.For an example of how
.dot
files can be processed, see thescripts/contrib/graph-tool
Python script, which finds and displays paths between graph nodes.
You can use a different method to view dependency information by using the following command:
$ bitbake -g -u taskexp recipename
This command displays a GUI window from which you can view build-time and runtime dependencies for the recipes involved in building recipename.
3.30.5 Viewing Task Variable Dependencies
As mentioned in the
“Checksums (Signatures)” section of the BitBake
User Manual, BitBake tries to automatically determine what variables a
task depends on so that it can rerun the task if any values of the
variables change. This determination is usually reliable. However, if
you do things like construct variable names at runtime, then you might
have to manually declare dependencies on those variables using
vardeps
as described in the
“Variable Flags” section of the BitBake
User Manual.
If you are unsure whether a variable dependency is being picked up automatically for a given task, you can list the variable dependencies BitBake has determined by doing the following:
1. Build the recipe containing the task:
$ bitbake recipename
Inside the STAMPS_DIR directory, find the signature data (
sigdata
) file that corresponds to the task. Thesigdata
files contain a pickled Python database of all the metadata that went into creating the input checksum for the task. As an example, for the do_fetch task of thedb
recipe, thesigdata
file might be found in the following location:${BUILDDIR}/tmp/stamps/i586-poky-linux/db/6.0.30-r1.do_fetch.sigdata.7c048c18222b16ff0bcee2000ef648b1
For tasks that are accelerated through the shared state (sstate) cache, an additional
siginfo
file is written into SSTATE_DIR along with the cached task output. Thesiginfo
files contain exactly the same information assigdata
files.Run
bitbake-dumpsig
on thesigdata
orsiginfo
file. Here is an example:$ bitbake-dumpsig ${BUILDDIR}/tmp/stamps/i586-poky-linux/db/6.0.30-r1.do_fetch.sigdata.7c048c18222b16ff0bcee2000ef648b1
In the output of the above command, you will find a line like the following, which lists all the (inferred) variable dependencies for the task. This list also includes indirect dependencies from variables depending on other variables, recursively.
Task dependencies: ['PV', 'SRCREV', 'SRC_URI', 'SRC_URI[md5sum]', 'SRC_URI[sha256sum]', 'base_do_fetch']
Note
Functions (e.g.
base_do_fetch
) also count as variable dependencies. These functions in turn depend on the variables they reference.The output of
bitbake-dumpsig
also includes the value each variable had, a list of dependencies for each variable, and BB_HASHBASE_WHITELIST information.
There is also a bitbake-diffsigs
command for comparing two
siginfo
or sigdata
files. This command can be helpful when
trying to figure out what changed between two versions of a task. If you
call bitbake-diffsigs
with just one file, the command behaves like
bitbake-dumpsig
.
You can also use BitBake to dump out the signature construction information without executing tasks by using either of the following BitBake command-line options:
‐‐dump-signatures=SIGNATURE_HANDLER
-S SIGNATURE_HANDLER
Note
Two common values for SIGNATURE_HANDLER are “none” and “printdiff”, which dump only the signature or compare the dumped signature with the cached one, respectively.
Using BitBake with either of these options causes BitBake to dump out
sigdata
files in the stamps
directory for every task it would
have executed instead of building the specified target package.
3.30.8 Running Specific Tasks
Any given recipe consists of a set of tasks. The standard BitBake
behavior in most cases is: do_fetch
, do_unpack
, do_patch
,
do_configure
, do_compile
, do_install
, do_package
,
do_package_write_*
, and do_build
. The default task is
do_build
and any tasks on which it depends build first. Some tasks,
such as do_devshell
, are not part of the default build chain. If you
wish to run a task that is not part of the default build chain, you can
use the -c
option in BitBake. Here is an example:
$ bitbake matchbox-desktop -c devshell
The -c
option respects task dependencies, which means that all other
tasks (including tasks from other recipes) that the specified task
depends on will be run before the task. Even when you manually specify a
task to run with -c
, BitBake will only run the task if it considers
it “out of date”. See the
“Stamp Files and the Rerunning of Tasks”
section in the Yocto Project Overview and Concepts Manual for how
BitBake determines whether a task is “out of date”.
If you want to force an up-to-date task to be rerun (e.g. because you
made manual modifications to the recipe’s
WORKDIR that you want to try
out), then you can use the -f
option.
Note
The reason -f
is never required when running the
do_devshell task is because the
[nostamp]
variable flag is already set for the task.
The following example shows one way you can use the -f
option:
$ bitbake matchbox-desktop
.
.
make some changes to the source code in the work directory
.
.
$ bitbake matchbox-desktop -c compile -f
$ bitbake matchbox-desktop
This sequence first builds and then recompiles matchbox-desktop
. The
last command reruns all tasks (basically the packaging tasks) after the
compile. BitBake recognizes that the do_compile
task was rerun and
therefore understands that the other tasks also need to be run again.
Another, shorter way to rerun a task and all
Normal Recipe Build Tasks
that depend on it is to use the -C
option.
Note
This option is upper-cased and is separate from the -c
option, which is lower-cased.
Using this option invalidates the given task and then runs the do_build task, which is the default task if no task is given, and the tasks on which it depends. You could replace the final two commands in the previous example with the following single command:
$ bitbake matchbox-desktop -C compile
Internally, the -f
and -C
options work by tainting (modifying)
the input checksum of the specified task. This tainting indirectly
causes the task and its dependent tasks to be rerun through the normal
task dependency mechanisms.
Note
BitBake explicitly keeps track of which tasks have been tainted in this fashion, and will print warnings such as the following for builds involving such tasks:
WARNING: /home/ulf/poky/meta/recipes-sato/matchbox-desktop/matchbox-desktop_2.1.bb.do_compile is tainted from a forced run
The purpose of the warning is to let you know that the work directory and build output might not be in the clean state they would be in for a “normal” build, depending on what actions you took. To get rid of such warnings, you can remove the work directory and rebuild the recipe, as follows:
$ bitbake matchbox-desktop -c clean
$ bitbake matchbox-desktop
You can view a list of tasks in a given package by running the
do_listtasks
task as follows:
$ bitbake matchbox-desktop -c listtasks
The results appear as output to the console and are also in
the file ${WORKDIR}/temp/log.do_listtasks
.
3.30.9 General BitBake Problems
You can see debug output from BitBake by using the -D
option. The
debug output gives more information about what BitBake is doing and the
reason behind it. Each -D
option you use increases the logging
level. The most common usage is -DDD
.
The output from bitbake -DDD -v targetname
can reveal why BitBake
chose a certain version of a package or why BitBake picked a certain
provider. This command could also help you in a situation where you
think BitBake did something unexpected.
3.30.10 Building with No Dependencies
To build a specific recipe (.bb
file), you can use the following
command form:
$ bitbake -b somepath/somerecipe.bb
This command form does not check for dependencies. Consequently, you should use it only when you know existing dependencies have been met.
Note
You can also specify fragments of the filename. In this case, BitBake checks for a unique match.
3.30.11 Recipe Logging Mechanisms
The Yocto Project provides several logging functions for producing
debugging output and reporting errors and warnings. For Python
functions, the following logging functions exist. All of these functions
log to ${T}/log.do_
task, and can also log to standard output
(stdout) with the right settings:
bb.plain(msg)
: Writes msg as is to the log while also logging to stdout.bb.note(msg)
: Writes “NOTE: msg” to the log. Also logs to stdout if BitBake is called with “-v”.bb.debug(level, msg)
: Writes “DEBUG: msg” to the log. Also logs to stdout if the log level is greater than or equal to level. See the “-D” option in the BitBake User Manual for more information.bb.warn(msg)
: Writes “WARNING: msg” to the log while also logging to stdout.bb.error(msg)
: Writes “ERROR: msg” to the log while also logging to standard out (stdout).Note
Calling this function does not cause the task to fail.
bb.fatal(
msg)
: This logging function is similar tobb.error(
msg)
but also causes the calling task to fail.Note
bb.fatal()
raises an exception, which means you do not need to put a “return” statement after the function.
The same logging functions are also available in shell functions, under
the names bbplain
, bbnote
, bbdebug
, bbwarn
, bberror
,
and bbfatal
. The
logging class
implements these functions. See that class in the meta/classes
folder of the Source Directory for information.
3.30.11.1 Logging With Python
When creating recipes using Python and inserting code that handles build logs, keep in mind the goal is to have informative logs while keeping the console as “silent” as possible. Also, if you want status messages in the log, use the “debug” loglevel.
Following is an example written in Python. The code handles logging for a function that determines the number of tasks needed to be run. See the “do_listtasks” section for additional information:
python do_listtasks() {
bb.debug(2, "Starting to figure out the task list")
if noteworthy_condition:
bb.note("There are 47 tasks to run")
bb.debug(2, "Got to point xyz")
if warning_trigger:
bb.warn("Detected warning_trigger, this might be a problem later.")
if recoverable_error:
bb.error("Hit recoverable_error, you really need to fix this!")
if fatal_error:
bb.fatal("fatal_error detected, unable to print the task list")
bb.plain("The tasks present are abc")
bb.debug(2, "Finished figuring out the tasklist")
}
3.30.11.2 Logging With Bash
When creating recipes using Bash and inserting code that handles build logs, you have the same goals - informative with minimal console output. The syntax you use for recipes written in Bash is similar to that of recipes written in Python described in the previous section.
Following is an example written in Bash. The code logs the progress of
the do_my_function
function.
do_my_function() {
bbdebug 2 "Running do_my_function"
if [ exceptional_condition ]; then
bbnote "Hit exceptional_condition"
fi
bbdebug 2 "Got to point xyz"
if [ warning_trigger ]; then
bbwarn "Detected warning_trigger, this might cause a problem later."
fi
if [ recoverable_error ]; then
bberror "Hit recoverable_error, correcting"
fi
if [ fatal_error ]; then
bbfatal "fatal_error detected"
fi
bbdebug 2 "Completed do_my_function"
}
3.30.12 Debugging Parallel Make Races
A parallel make
race occurs when the build consists of several parts
that are run simultaneously and a situation occurs when the output or
result of one part is not ready for use with a different part of the
build that depends on that output. Parallel make races are annoying and
can sometimes be difficult to reproduce and fix. However, some simple
tips and tricks exist that can help you debug and fix them. This section
presents a real-world example of an error encountered on the Yocto
Project autobuilder and the process used to fix it.
Note
If you cannot properly fix a make
race condition, you can work around it
by clearing either the PARALLEL_MAKE or PARALLEL_MAKEINST
variables.
3.30.12.1 The Failure
For this example, assume that you are building an image that depends on the “neard” package. And, during the build, BitBake runs into problems and creates the following output.
Note
This example log file has longer lines artificially broken to make the listing easier to read.
If you examine the output or the log file, you see the failure during
make
:
| DEBUG: SITE files ['endian-little', 'bit-32', 'ix86-common', 'common-linux', 'common-glibc', 'i586-linux', 'common']
| DEBUG: Executing shell function do_compile
| NOTE: make -j 16
| make --no-print-directory all-am
| /bin/mkdir -p include/near
| /bin/mkdir -p include/near
| /bin/mkdir -p include/near
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/types.h include/near/types.h
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/log.h include/near/log.h
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/plugin.h include/near/plugin.h
| /bin/mkdir -p include/near
| /bin/mkdir -p include/near
| /bin/mkdir -p include/near
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/tag.h include/near/tag.h
| /bin/mkdir -p include/near
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/adapter.h include/near/adapter.h
| /bin/mkdir -p include/near
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/ndef.h include/near/ndef.h
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/tlv.h include/near/tlv.h
| /bin/mkdir -p include/near
| /bin/mkdir -p include/near
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/setting.h include/near/setting.h
| /bin/mkdir -p include/near
| /bin/mkdir -p include/near
| /bin/mkdir -p include/near
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/device.h include/near/device.h
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/nfc_copy.h include/near/nfc_copy.h
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/snep.h include/near/snep.h
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/version.h include/near/version.h
| ln -s /home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/work/i586-poky-linux/neard/
0.14-r0/neard-0.14/include/dbus.h include/near/dbus.h
| ./src/genbuiltin nfctype1 nfctype2 nfctype3 nfctype4 p2p > src/builtin.h
| i586-poky-linux-gcc -m32 -march=i586 --sysroot=/home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/
build/build/tmp/sysroots/qemux86 -DHAVE_CONFIG_H -I. -I./include -I./src -I./gdbus -I/home/pokybuild/
yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/sysroots/qemux86/usr/include/glib-2.0
-I/home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/tmp/sysroots/qemux86/usr/
lib/glib-2.0/include -I/home/pokybuild/yocto-autobuilder/yocto-slave/nightly-x86/build/build/
tmp/sysroots/qemux86/usr/include/dbus-1.0 -I/home/pokybuild/yocto-autobuilder/yocto-slave/
nightly-x86/build/build/tmp/sysroots/qemux86/usr/lib/dbus-1.0/include -I/home/pokybuild/yocto-autobuilder/
yocto-slave/nightly-x86/build/build/tmp/sysroots/qemux86/usr/include/libnl3
-DNEAR_PLUGIN_BUILTIN -DPLUGINDIR=\""/usr/lib/near/plugins"\"
-DCONFIGDIR=\""/etc/neard\"" -O2 -pipe -g -feliminate-unused-debug-types -c
-o tools/snep-send.o tools/snep-send.c
| In file included from tools/snep-send.c:16:0:
| tools/../src/near.h:41:23: fatal error: near/dbus.h: No such file or directory
| #include <near/dbus.h>
| ^
| compilation terminated.
| make[1]: *** [tools/snep-send.o] Error 1
| make[1]: *** Waiting for unfinished jobs....
| make: *** [all] Error 2
| ERROR: oe_runmake failed
3.30.12.2 Reproducing the Error
Because race conditions are intermittent, they do not manifest themselves every time you do the build. In fact, most times the build will complete without problems even though the potential race condition exists. Thus, once the error surfaces, you need a way to reproduce it.
In this example, compiling the “neard” package is causing the problem.
So the first thing to do is build “neard” locally. Before you start the
build, set the
PARALLEL_MAKE variable
in your local.conf
file to a high number (e.g. “-j 20”). Using a
high value for PARALLEL_MAKE
increases the chances of the race
condition showing up:
$ bitbake neard
Once the local build for “neard” completes, start a devshell
build:
$ bitbake neard -c devshell
For information on how to use a devshell
, see the
“Using a Development Shell” section.
In the devshell
, do the following:
$ make clean
$ make tools/snep-send.o
The devshell
commands cause the failure to clearly
be visible. In this case, a missing dependency exists for the “neard”
Makefile target. Here is some abbreviated, sample output with the
missing dependency clearly visible at the end:
i586-poky-linux-gcc -m32 -march=i586 --sysroot=/home/scott-lenovo/......
.
.
.
tools/snep-send.c
In file included from tools/snep-send.c:16:0:
tools/../src/near.h:41:23: fatal error: near/dbus.h: No such file or directory
#include <near/dbus.h>
^
compilation terminated.
make: *** [tools/snep-send.o] Error 1
$
3.30.12.3 Creating a Patch for the Fix
Because there is a missing dependency for the Makefile target, you need
to patch the Makefile.am
file, which is generated from
Makefile.in
. You can use Quilt to create the patch:
$ quilt new parallelmake.patch
Patch patches/parallelmake.patch is now on top
$ quilt add Makefile.am
File Makefile.am added to patch patches/parallelmake.patch
For more information on using Quilt, see the “Using Quilt in Your Workflow” section.
At this point you need to make the edits to Makefile.am
to add the
missing dependency. For our example, you have to add the following line
to the file:
tools/snep-send.$(OBJEXT): include/near/dbus.h
Once you have edited the file, use the refresh
command to create the
patch:
$ quilt refresh
Refreshed patch patches/parallelmake.patch
Once
the patch file exists, you need to add it back to the originating recipe
folder. Here is an example assuming a top-level
Source Directory named poky
:
$ cp patches/parallelmake.patch poky/meta/recipes-connectivity/neard/neard
The final thing you need to do to implement the fix in the build is to
update the “neard” recipe (i.e. neard-0.14.bb
) so that the
SRC_URI statement includes
the patch file. The recipe file is in the folder above the patch. Here
is what the edited SRC_URI
statement would look like:
SRC_URI = "${KERNELORG_MIRROR}/linux/network/nfc/${BPN}-${PV}.tar.xz \
file://neard.in \
file://neard.service.in \
file://parallelmake.patch \
"
With the patch complete and moved to the correct folder and the
SRC_URI
statement updated, you can exit the devshell
:
$ exit
3.30.12.4 Testing the Build
With everything in place, you can get back to trying the build again locally:
$ bitbake neard
This build should succeed.
Now you can open up a devshell
again and repeat the clean and make
operations as follows:
$ bitbake neard -c devshell
$ make clean
$ make tools/snep-send.o
The build should work without issue.
As with all solved problems, if they originated upstream, you need to submit the fix for the recipe in OE-Core and upstream so that the problem is taken care of at its source. See the “Submitting a Change to the Yocto Project” section for more information.
3.30.13 Debugging With the GNU Project Debugger (GDB) Remotely
GDB allows you to examine running programs, which in turn helps you to understand and fix problems. It also allows you to perform post-mortem style analysis of program crashes. GDB is available as a package within the Yocto Project and is installed in SDK images by default. See the “Images” chapter in the Yocto Project Reference Manual for a description of these images. You can find information on GDB at https://sourceware.org/gdb/.
Note
For best results, install debug (-dbg
) packages for the applications you
are going to debug. Doing so makes extra debug symbols available that give
you more meaningful output.
Sometimes, due to memory or disk space constraints, it is not possible to use GDB directly on the remote target to debug applications. These constraints arise because GDB needs to load the debugging information and the binaries of the process being debugged. Additionally, GDB needs to perform many computations to locate information such as function names, variable names and values, stack traces and so forth - even before starting the debugging process. These extra computations place more load on the target system and can alter the characteristics of the program being debugged.
To help get past the previously mentioned constraints, there are two methods you can use: running a debuginfod server and using gdbserver.
3.30.13.1 Using the debuginfod server method
“debuginfod” from “elfutils” is a way to distribute “debuginfo” files. Running a “debuginfod” server makes debug symbols readily available, which means you don’t need to download debugging information and the binaries of the process being debugged. You can just fetch debug symbols from the server.
To run a debuginfod server, you need to do the following:
Ensure that this variable is set in your
local.conf
file:PACKAGECONFIG_pn-elfutils-native = "debuginfod libdebuginfod"
This PACKAGECONFIG option enables debuginfod and libdebuginfod for “elfutils-native”.
Run the following commands to set up the “debuginfod” server:
$ oe-debuginfod
To use debuginfod on the target, you need the following:
Ensure that this variable is set in your
local.conf
file:DEBUGINFOD_URLS = "http://localhost:8002/"
This DEBUGINFOD_URLS option does the client configuration.
PACKAGECONFIG_pn-gdb = "debuginfod"
This PACKAGECONFIG option enables “debuginfod” for “gdb”.
3.30.13.2 Using the gdbserver method
gdbserver, which runs on the remote target and does not load any debugging information from the debugged process. Instead, a GDB instance processes the debugging information that is run on a remote computer - the host GDB. The host GDB then sends control commands to gdbserver to make it stop or start the debugged program, as well as read or write memory regions of that debugged program. All the debugging information loaded and processed as well as all the heavy debugging is done by the host GDB. Offloading these processes gives the gdbserver running on the target a chance to remain small and fast.
Because the host GDB is responsible for loading the debugging information and for doing the necessary processing to make actual debugging happen, you have to make sure the host can access the unstripped binaries complete with their debugging information and also be sure the target is compiled with no optimizations. The host GDB must also have local access to all the libraries used by the debugged program. Because gdbserver does not need any local debugging information, the binaries on the remote target can remain stripped. However, the binaries must also be compiled without optimization so they match the host’s binaries.
To remain consistent with GDB documentation and terminology, the binary being debugged on the remote target machine is referred to as the “inferior” binary. For documentation on GDB see the GDB site.
The following steps show you how to debug using the GNU project debugger.
Configure your build system to construct the companion debug filesystem:
In your
local.conf
file, set the following:IMAGE_GEN_DEBUGFS = "1" IMAGE_FSTYPES_DEBUGFS = "tar.bz2"
These options cause the OpenEmbedded build system to generate a special companion filesystem fragment, which contains the matching source and debug symbols to your deployable filesystem. The build system does this by looking at what is in the deployed filesystem, and pulling the corresponding
-dbg
packages.The companion debug filesystem is not a complete filesystem, but only contains the debug fragments. This filesystem must be combined with the full filesystem for debugging. Subsequent steps in this procedure show how to combine the partial filesystem with the full filesystem.
Configure the system to include gdbserver in the target filesystem:
Make the following addition in either your
local.conf
file or in an image recipe:IMAGE_INSTALL_append = " gdbserver"
The change makes sure the
gdbserver
package is included.Build the environment:
Use the following command to construct the image and the companion Debug Filesystem:
$ bitbake image
Build the cross GDB component and make it available for debugging. Build the SDK that matches the image. Building the SDK is best for a production build that can be used later for debugging, especially during long term maintenance:
$ bitbake -c populate_sdk image
Alternatively, you can build the minimal toolchain components that match the target. Doing so creates a smaller than typical SDK and only contains a minimal set of components with which to build simple test applications, as well as run the debugger:
$ bitbake meta-toolchain
A final method is to build Gdb itself within the build system:
$ bitbake gdb-cross-<architecture>
Doing so produces a temporary copy of
cross-gdb
you can use for debugging during development. While this is the quickest approach, the two previous methods in this step are better when considering long-term maintenance strategies.Note
If you run
bitbake gdb-cross
, the OpenEmbedded build system suggests the actual image (e.g.gdb-cross-i586
). The suggestion is usually the actual name you want to use.Set up the
debugfs
:Run the following commands to set up the
debugfs
:$ mkdir debugfs $ cd debugfs $ tar xvfj build-dir/tmp-glibc/deploy/images/machine/image.rootfs.tar.bz2 $ tar xvfj build-dir/tmp-glibc/deploy/images/machine/image-dbg.rootfs.tar.bz2
Set up GDB:
Install the SDK (if you built one) and then source the correct environment file. Sourcing the environment file puts the SDK in your
PATH
environment variable.If you are using the build system, Gdb is located in build-dir
/tmp/sysroots/
host/usr/bin/
architecture/
architecture-gdb
Boot the target:
For information on how to run QEMU, see the QEMU Documentation.
Note
Be sure to verify that your host can access the target via TCP.
Debug a program:
Debugging a program involves running gdbserver on the target and then running Gdb on the host. The example in this step debugs
gzip
:root@qemux86:~# gdbserver localhost:1234 /bin/gzip —help
For additional gdbserver options, see the GDB Server Documentation.
After running gdbserver on the target, you need to run Gdb on the host and configure it and connect to the target. Use these commands:
$ cd directory-holding-the-debugfs-directory $ arch-gdb (gdb) set sysroot debugfs (gdb) set substitute-path /usr/src/debug debugfs/usr/src/debug (gdb) target remote IP-of-target:1234
At this point, everything should automatically load (i.e. matching binaries, symbols and headers).
Note
The Gdb
set
commands in the previous example can be placed into the users~/.gdbinit
file. Upon starting, Gdb automatically runs whatever commands are in that file.Deploying without a full image rebuild:
In many cases, during development you want a quick method to deploy a new binary to the target and debug it, without waiting for a full image build.
One approach to solving this situation is to just build the component you want to debug. Once you have built the component, copy the executable directly to both the target and the host
debugfs
.If the binary is processed through the debug splitting in OpenEmbedded, you should also copy the debug items (i.e.
.debug
contents and corresponding/usr/src/debug
files) from the work directory. Here is an example:$ bitbake bash $ bitbake -c devshell bash $ cd .. $ scp packages-split/bash/bin/bash target:/bin/bash $ cp -a packages-split/bash-dbg/\* path/debugfs
3.30.14 Debugging with the GNU Project Debugger (GDB) on the Target
The previous section addressed using GDB remotely for debugging purposes, which is the most usual case due to the inherent hardware limitations on many embedded devices. However, debugging in the target hardware itself is also possible with more powerful devices. This section describes what you need to do in order to support using GDB to debug on the target hardware.
To support this kind of debugging, you need do the following:
Ensure that GDB is on the target. You can do this by adding “gdb” to IMAGE_INSTALL:
IMAGE_INSTALL_append = " gdb"
Alternatively, you can add “tools-debug” to IMAGE_FEATURES:
IMAGE_FEATURES_append = " tools-debug"
Ensure that debug symbols are present. You can make sure these symbols are present by installing
-dbg
:IMAGE_INSTALL_append = "packagename-dbg"
Alternatively, you can do the following to include all the debug symbols:
IMAGE_FEATURES_append = " dbg-pkgs"
Note
To improve the debug information accuracy, you can reduce the level
of optimization used by the compiler. For example, when adding the
following line to your local.conf
file, you will reduce optimization
from FULL_OPTIMIZATION of “-O2” to DEBUG_OPTIMIZATION
of “-O -fno-omit-frame-pointer”:
DEBUG_BUILD = "1"
Consider that this will reduce the application’s performance and is recommended only for debugging purposes.
3.30.15 Other Debugging Tips
Here are some other tips that you might find useful:
When adding new packages, it is worth watching for undesirable items making their way into compiler command lines. For example, you do not want references to local system files like
/usr/lib/
or/usr/include/
.If you want to remove the
psplash
boot splashscreen, addpsplash=false
to the kernel command line. Doing so preventspsplash
from loading and thus allows you to see the console. It is also possible to switch out of the splashscreen by switching the virtual console (e.g. Fn+Left or Fn+Right on a Zaurus).Removing TMPDIR (usually
tmp/
, within the Build Directory) can often fix temporary build issues. RemovingTMPDIR
is usually a relatively cheap operation, because task output will be cached in SSTATE_DIR (usuallysstate-cache/
, which is also in the Build Directory).Note
Removing
TMPDIR
might be a workaround rather than a fix. Consequently, trying to determine the underlying cause of an issue before removing the directory is a good idea.Understanding how a feature is used in practice within existing recipes can be very helpful. It is recommended that you configure some method that allows you to quickly search through files.
Using GNU Grep, you can use the following shell function to recursively search through common recipe-related files, skipping binary files,
.git
directories, and the Build Directory (assuming its name starts with “build”):g() { grep -Ir \ --exclude-dir=.git \ --exclude-dir='build*' \ --include='*.bb*' \ --include='*.inc*' \ --include='*.conf*' \ --include='*.py*' \ "$@" }
Following are some usage examples:
$ g FOO # Search recursively for "FOO" $ g -i foo # Search recursively for "foo", ignoring case $ g -w FOO # Search recursively for "FOO" as a word, ignoring e.g. "FOOBAR"
If figuring out how some feature works requires a lot of searching, it might indicate that the documentation should be extended or improved. In such cases, consider filing a documentation bug using the Yocto Project implementation of Bugzilla. For information on how to submit a bug against the Yocto Project, see the Yocto Project Bugzilla wiki page and the “Submitting a Defect Against the Yocto Project” section.
Note
The manuals might not be the right place to document variables that are purely internal and have a limited scope (e.g. internal variables used to implement a single
.bbclass
file).
3.31 Making Changes to the Yocto Project
Because the Yocto Project is an open-source, community-based project, you can effect changes to the project. This section presents procedures that show you how to submit a defect against the project and how to submit a change.
3.31.1 Submitting a Defect Against the Yocto Project
Use the Yocto Project implementation of Bugzilla to submit a defect (bug) against the Yocto Project. For additional information on this implementation of Bugzilla see the “Yocto Project Bugzilla” section in the Yocto Project Reference Manual. For more detail on any of the following steps, see the Yocto Project Bugzilla wiki page.
Use the following general steps to submit a bug:
Open the Yocto Project implementation of Bugzilla.
Click “File a Bug” to enter a new bug.
Choose the appropriate “Classification”, “Product”, and “Component” for which the bug was found. Bugs for the Yocto Project fall into one of several classifications, which in turn break down into several products and components. For example, for a bug against the
meta-intel
layer, you would choose “Build System, Metadata & Runtime”, “BSPs”, and “bsps-meta-intel”, respectively.Choose the “Version” of the Yocto Project for which you found the bug (e.g. 3.3.1).
Determine and select the “Severity” of the bug. The severity indicates how the bug impacted your work.
Choose the “Hardware” that the bug impacts.
Choose the “Architecture” that the bug impacts.
Choose a “Documentation change” item for the bug. Fixing a bug might or might not affect the Yocto Project documentation. If you are unsure of the impact to the documentation, select “Don’t Know”.
Provide a brief “Summary” of the bug. Try to limit your summary to just a line or two and be sure to capture the essence of the bug.
Provide a detailed “Description” of the bug. You should provide as much detail as you can about the context, behavior, output, and so forth that surrounds the bug. You can even attach supporting files for output from logs by using the “Add an attachment” button.
Click the “Submit Bug” button submit the bug. A new Bugzilla number is assigned to the bug and the defect is logged in the bug tracking system.
Once you file a bug, the bug is processed by the Yocto Project Bug Triage Team and further details concerning the bug are assigned (e.g. priority and owner). You are the “Submitter” of the bug and any further categorization, progress, or comments on the bug result in Bugzilla sending you an automated email concerning the particular change or progress to the bug.
3.31.2 Submitting a Change to the Yocto Project
Contributions to the Yocto Project and OpenEmbedded are very welcome. Because the system is extremely configurable and flexible, we recognize that developers will want to extend, configure or optimize it for their specific uses.
The Yocto Project uses a mailing list and a patch-based workflow that is
similar to the Linux kernel but contains important differences. In
general, a mailing list exists through which you can submit patches. You
should send patches to the appropriate mailing list so that they can be
reviewed and merged by the appropriate maintainer. The specific mailing
list you need to use depends on the location of the code you are
changing. Each component (e.g. layer) should have a README
file that
indicates where to send the changes and which process to follow.
You can send the patch to the mailing list using whichever approach you feel comfortable with to generate the patch. Once sent, the patch is usually reviewed by the community at large. If somebody has concerns with the patch, they will usually voice their concern over the mailing list. If a patch does not receive any negative reviews, the maintainer of the affected layer typically takes the patch, tests it, and then based on successful testing, merges the patch.
The “poky” repository, which is the Yocto Project’s reference build environment, is a hybrid repository that contains several individual pieces (e.g. BitBake, Metadata, documentation, and so forth) built using the combo-layer tool. The upstream location used for submitting changes varies by component:
Core Metadata: Send your patch to the openembedded-core mailing list. For example, a change to anything under the
meta
orscripts
directories should be sent to this mailing list.BitBake: For changes to BitBake (i.e. anything under the
bitbake
directory), send your patch to the bitbake-devel mailing list.“meta-*” trees: These trees contain Metadata. Use the poky mailing list.
Documentation: For changes to the Yocto Project documentation, use the docs mailing list.
For changes to other layers hosted in the Yocto Project source
repositories (i.e. yoctoproject.org
) and tools use the
Yocto Project general mailing list.
Note
Sometimes a layer’s documentation specifies to use a particular mailing list. If so, use that list.
For additional recipes that do not fit into the core Metadata, you
should determine which layer the recipe should go into and submit the
change in the manner recommended by the documentation (e.g. the
README
file) supplied with the layer. If in doubt, please ask on the
Yocto general mailing list or on the openembedded-devel mailing list.
You can also push a change upstream and request a maintainer to pull the change into the component’s upstream repository. You do this by pushing to a contribution repository that is upstream. See the “Git Workflows and the Yocto Project” section in the Yocto Project Overview and Concepts Manual for additional concepts on working in the Yocto Project development environment.
Maintainers commonly use -next
branches to test submissions prior to
merging patches. Thus, you can get an idea of the status of a patch based on
whether the patch has been merged into one of these branches. The commonly
used testing branches for OpenEmbedded-Core are as follows:
openembedded-core “master-next” branch: This branch is part of the openembedded-core repository and contains proposed changes to the core metadata.
poky “master-next” branch: This branch is part of the poky repository and combines proposed changes to bitbake, the core metadata and the poky distro.
Similarly, stable branches maintained by the project may have corresponding
-next
branches which collect proposed changes. For example,
hardknott-next
and gatesgarth-next
branches in both the “openembdedded-core” and “poky” repositories.
Other layers may have similar testing branches but there is no formal requirement or standard for these so please check the documentation for the layers you are contributing to.
The following sections provide procedures for submitting a change.
3.31.2.1 Preparing Changes for Submission
Make Your Changes Locally: Make your changes in your local Git repository. You should make small, controlled, isolated changes. Keeping changes small and isolated aids review, makes merging/rebasing easier and keeps the change history clean should anyone need to refer to it in future.
Stage Your Changes: Stage your changes by using the
git add
command on each file you changed.Commit Your Changes: Commit the change by using the
git commit
command. Make sure your commit information follows standards by following these accepted conventions:Be sure to include a “Signed-off-by:” line in the same style as required by the Linux kernel. Adding this line signifies that you, the submitter, have agreed to the Developer’s Certificate of Origin 1.1 as follows:
Developer's Certificate of Origin 1.1 By making a contribution to this project, I certify that: (a) The contribution was created in whole or in part by me and I have the right to submit it under the open source license indicated in the file; or (b) The contribution is based upon previous work that, to the best of my knowledge, is covered under an appropriate open source license and I have the right under that license to submit that work with modifications, whether created in whole or in part by me, under the same open source license (unless I am permitted to submit under a different license), as indicated in the file; or (c) The contribution was provided directly to me by some other person who certified (a), (b) or (c) and I have not modified it. (d) I understand and agree that this project and the contribution are public and that a record of the contribution (including all personal information I submit with it, including my sign-off) is maintained indefinitely and may be redistributed consistent with this project or the open source license(s) involved.
Provide a single-line summary of the change and, if more explanation is needed, provide more detail in the body of the commit. This summary is typically viewable in the “shortlist” of changes. Thus, providing something short and descriptive that gives the reader a summary of the change is useful when viewing a list of many commits. You should prefix this short description with the recipe name (if changing a recipe), or else with the short form path to the file being changed.
For the body of the commit message, provide detailed information that describes what you changed, why you made the change, and the approach you used. It might also be helpful if you mention how you tested the change. Provide as much detail as you can in the body of the commit message.
Note
You do not need to provide a more detailed explanation of a change if the change is minor to the point of the single line summary providing all the information.
If the change addresses a specific bug or issue that is associated with a bug-tracking ID, include a reference to that ID in your detailed description. For example, the Yocto Project uses a specific convention for bug references - any commit that addresses a specific bug should use the following form for the detailed description. Be sure to use the actual bug-tracking ID from Bugzilla for bug-id:
Fixes [YOCTO #bug-id] detailed description of change
3.31.2.2 Using Email to Submit a Patch
Depending on the components changed, you need to submit the email to a specific mailing list. For some guidance on which mailing list to use, see the list at the beginning of this section. For a description of all the available mailing lists, see the “Mailing Lists” section in the Yocto Project Reference Manual.
Here is the general procedure on how to submit a patch through email without using the scripts once the steps in Preparing Changes for Submission have been followed:
Format the Commit: Format the commit into an email message. To format commits, use the
git format-patch
command. When you provide the command, you must include a revision list or a number of patches as part of the command. For example, either of these two commands takes your most recent single commit and formats it as an email message in the current directory:$ git format-patch -1
or
$ git format-patch HEAD~
After the command is run, the current directory contains a numbered
.patch
file for the commit.If you provide several commits as part of the command, the
git format-patch
command produces a series of numbered files in the current directory – one for each commit. If you have more than one patch, you should also use the--cover
option with the command, which generates a cover letter as the first “patch” in the series. You can then edit the cover letter to provide a description for the series of patches. For information on thegit format-patch
command, seeGIT_FORMAT_PATCH(1)
displayed using theman git-format-patch
command.Note
If you are or will be a frequent contributor to the Yocto Project or to OpenEmbedded, you might consider requesting a contrib area and the necessary associated rights.
Send the patches via email: Send the patches to the recipients and relevant mailing lists by using the
git send-email
command.Note
In order to use
git send-email
, you must have the proper Git packages installed on your host. For Ubuntu, Debian, and Fedora the package isgit-email
.The
git send-email
command sends email by using a local or remote Mail Transport Agent (MTA) such asmsmtp
,sendmail
, or through a directsmtp
configuration in your Git~/.gitconfig
file. If you are submitting patches through email only, it is very important that you submit them without any whitespace or HTML formatting that either you or your mailer introduces. The maintainer that receives your patches needs to be able to save and apply them directly from your emails. A good way to verify that what you are sending will be applicable by the maintainer is to do a dry run and send them to yourself and then save and apply them as the maintainer would.The
git send-email
command is the preferred method for sending your patches using email since there is no risk of compromising whitespace in the body of the message, which can occur when you use your own mail client. The command also has several options that let you specify recipients and perform further editing of the email message. For information on how to use thegit send-email
command, seeGIT-SEND-EMAIL(1)
displayed using theman git-send-email
command.
The Yocto Project uses a Patchwork instance to track the status of patches submitted to the various mailing lists and to support automated patch testing. Each submitted patch is checked for common mistakes and deviations from the expected patch format and submitters are notified by patchtest if such mistakes are found. This process helps to reduce the burden of patch review on maintainers.
Note
This system is imperfect and changes can sometimes get lost in the flow. Asking about the status of a patch or change is reasonable if the change has been idle for a while with no feedback.
3.31.2.3 Using Scripts to Push a Change Upstream and Request a Pull
For larger patch series it is preferable to send a pull request which not
only includes the patch but also a pointer to a branch that can be pulled
from. This involves making a local branch for your changes, pushing this
branch to an accessible repository and then using the create-pull-request
and send-pull-request
scripts from openembedded-core to create and send a
patch series with a link to the branch for review.
Follow this procedure to push a change to an upstream “contrib” Git repository once the steps in Preparing Changes for Submission have been followed:
Note
You can find general Git information on how to push a change upstream in the Git Community Book.
Push Your Commits to a “Contrib” Upstream: If you have arranged for permissions to push to an upstream contrib repository, push the change to that repository:
$ git push upstream_remote_repo local_branch_name
For example, suppose you have permissions to push into the upstream
meta-intel-contrib
repository and you are working in a local branch named your_name/README
. The following command pushes your local commits to themeta-intel-contrib
upstream repository and puts the commit in a branch named your_name/README
:$ git push meta-intel-contrib your_name/README
Determine Who to Notify: Determine the maintainer or the mailing list that you need to notify for the change.
Before submitting any change, you need to be sure who the maintainer is or what mailing list that you need to notify. Use either these methods to find out:
Maintenance File: Examine the
maintainers.inc
file, which is located in the Source Directory atmeta/conf/distro/include
, to see who is responsible for code.Search by File: Using Git, you can enter the following command to bring up a short list of all commits against a specific file:
git shortlog -- filename
Just provide the name of the file for which you are interested. The information returned is not ordered by history but does include a list of everyone who has committed grouped by name. From the list, you can see who is responsible for the bulk of the changes against the file.
Examine the List of Mailing Lists: For a list of the Yocto Project and related mailing lists, see the “Mailing lists” section in the Yocto Project Reference Manual.
Make a Pull Request: Notify the maintainer or the mailing list that you have pushed a change by making a pull request.
The Yocto Project provides two scripts that conveniently let you generate and send pull requests to the Yocto Project. These scripts are
create-pull-request
andsend-pull-request
. You can find these scripts in thescripts
directory within the Source Directory (e.g.poky/scripts
).Using these scripts correctly formats the requests without introducing any whitespace or HTML formatting. The maintainer that receives your patches either directly or through the mailing list needs to be able to save and apply them directly from your emails. Using these scripts is the preferred method for sending patches.
First, create the pull request. For example, the following command runs the script, specifies the upstream repository in the contrib directory into which you pushed the change, and provides a subject line in the created patch files:
$ poky/scripts/create-pull-request -u meta-intel-contrib -s "Updated Manual Section Reference in README"
Running this script forms
*.patch
files in a folder namedpull-
PID in the current directory. One of the patch files is a cover letter.Before running the
send-pull-request
script, you must edit the cover letter patch to insert information about your change. After editing the cover letter, send the pull request. For example, the following command runs the script and specifies the patch directory and email address. In this example, the email address is a mailing list:$ poky/scripts/send-pull-request -p ~/meta-intel/pull-10565 -t meta-intel@yoctoproject.org
You need to follow the prompts as the script is interactive.
Note
For help on using these scripts, simply provide the
-h
argument as follows:$ poky/scripts/create-pull-request -h $ poky/scripts/send-pull-request -h
3.31.2.4 Responding to Patch Review
You may get feedback on your submitted patches from other community members or from the automated patchtest service. If issues are identified in your patch then it is usually necessary to address these before the patch will be accepted into the project. In this case you should amend the patch according to the feedback and submit an updated version to the relevant mailing list, copying in the reviewers who provided feedback to the previous version of the patch.
The patch should be amended using git commit --amend
or perhaps git
rebase
for more expert git users. You should also modify the [PATCH]
tag in the email subject line when sending the revised patch to mark the new
iteration as [PATCH v2]
, [PATCH v3]
, etc as appropriate. This can be
done by passing the -v
argument to git format-patch
with a version
number.
Lastly please ensure that you also test your revised changes. In particular
please don’t just edit the patch file written out by git format-patch
and
resend it.
3.31.2.5 Submitting Changes to Stable Release Branches
The process for proposing changes to a Yocto Project stable branch differs from the steps described above. Changes to a stable branch must address identified bugs or CVEs and should be made carefully in order to avoid the risk of introducing new bugs or breaking backwards compatibility. Typically bug fixes must already be accepted into the master branch before they can be backported to a stable branch unless the bug in question does not affect the master branch or the fix on the master branch is unsuitable for backporting.
The list of stable branches along with the status and maintainer for each branch can be obtained from the Releases wiki page.
Note
Changes will not typically be accepted for branches which are marked as End-Of-Life (EOL).
With this in mind, the steps to submit a change for a stable branch are as follows:
Identify the bug or CVE to be fixed: This information should be collected so that it can be included in your submission.
Check if the fix is already present in the master branch: This will result in the most straightforward path into the stable branch for the fix.
If the fix is present in the master branch - Submit a backport request by email: You should send an email to the relevant stable branch maintainer and the mailing list with details of the bug or CVE to be fixed, the commit hash on the master branch that fixes the issue and the stable branches which you would like this fix to be backported to.
If the fix is not present in the master branch - Submit the fix to the master branch first: This will ensure that the fix passes through the project’s usual patch review and test processes before being accepted. It will also ensure that bugs are not left unresolved in the master branch itself. Once the fix is accepted in the master branch a backport request can be submitted as above.
If the fix is unsuitable for the master branch - Submit a patch directly for the stable branch: This method should be considered as a last resort. It is typically necessary when the master branch is using a newer version of the software which includes an upstream fix for the issue or when the issue has been fixed on the master branch in a way that introduces backwards incompatible changes. In this case follow the steps in Preparing Changes for Submission and Using Email to Submit a Patch but modify the subject header of your patch email to include the name of the stable branch which you are targetting. This can be done using the
--subject-prefix
argument togit format-patch
, for example to submit a patch to the dunfell branch usegit format-patch --subject-prefix='gatesgarth][PATCH' ...
.
3.32 Working With Licenses
As mentioned in the “Licensing” section in the Yocto Project Overview and Concepts Manual, open source projects are open to the public and they consequently have different licensing structures in place. This section describes the mechanism by which the OpenEmbedded Build System tracks changes to licensing text and covers how to maintain open source license compliance during your project’s lifecycle. The section also describes how to enable commercially licensed recipes, which by default are disabled.
3.32.1 Tracking License Changes
The license of an upstream project might change in the future. In order to prevent these changes going unnoticed, the LIC_FILES_CHKSUM variable tracks changes to the license text. The checksums are validated at the end of the configure step, and if the checksums do not match, the build will fail.
3.32.1.1 Specifying the LIC_FILES_CHKSUM
Variable
The LIC_FILES_CHKSUM
variable contains checksums of the license text
in the source code for the recipe. Following is an example of how to
specify LIC_FILES_CHKSUM
:
LIC_FILES_CHKSUM = "file://COPYING;md5=xxxx \
file://licfile1.txt;beginline=5;endline=29;md5=yyyy \
file://licfile2.txt;endline=50;md5=zzzz \
..."
Note
When using “beginline” and “endline”, realize that line numbering begins with one and not zero. Also, the included lines are inclusive (i.e. lines five through and including 29 in the previous example for
licfile1.txt
).When a license check fails, the selected license text is included as part of the QA message. Using this output, you can determine the exact start and finish for the needed license text.
The build system uses the S
variable as the default directory when searching files listed in
LIC_FILES_CHKSUM
. The previous example employs the default
directory.
Consider this next example:
LIC_FILES_CHKSUM = "file://src/ls.c;beginline=5;endline=16;\
md5=bb14ed3c4cda583abc85401304b5cd4e"
LIC_FILES_CHKSUM = "file://${WORKDIR}/license.html;md5=5c94767cedb5d6987c902ac850ded2c6"
The first line locates a file in ${S}/src/ls.c
and isolates lines
five through 16 as license text. The second line refers to a file in
WORKDIR.
Note that LIC_FILES_CHKSUM
variable is mandatory for all recipes,
unless the LICENSE
variable is set to “CLOSED”.
3.32.1.2 Explanation of Syntax
As mentioned in the previous section, the LIC_FILES_CHKSUM
variable
lists all the important files that contain the license text for the
source code. It is possible to specify a checksum for an entire file, or
a specific section of a file (specified by beginning and ending line
numbers with the “beginline” and “endline” parameters, respectively).
The latter is useful for source files with a license notice header,
README documents, and so forth. If you do not use the “beginline”
parameter, then it is assumed that the text begins on the first line of
the file. Similarly, if you do not use the “endline” parameter, it is
assumed that the license text ends with the last line of the file.
The “md5” parameter stores the md5 checksum of the license text. If the license text changes in any way as compared to this parameter then a mismatch occurs. This mismatch triggers a build failure and notifies the developer. Notification allows the developer to review and address the license text changes. Also note that if a mismatch occurs during the build, the correct md5 checksum is placed in the build log and can be easily copied to the recipe.
There is no limit to how many files you can specify using the
LIC_FILES_CHKSUM
variable. Generally, however, every project
requires a few specifications for license tracking. Many projects have a
“COPYING” file that stores the license information for all the source
code files. This practice allows you to just track the “COPYING” file as
long as it is kept up to date.
Note
If you specify an empty or invalid “md5” parameter, BitBake returns an md5 mis-match error and displays the correct “md5” parameter value during the build. The correct parameter is also captured in the build log.
If the whole file contains only license text, you do not need to use the “beginline” and “endline” parameters.
3.32.2 Enabling Commercially Licensed Recipes
By default, the OpenEmbedded build system disables components that have
commercial or other special licensing requirements. Such requirements
are defined on a recipe-by-recipe basis through the
LICENSE_FLAGS variable
definition in the affected recipe. For instance, the
poky/meta/recipes-multimedia/gstreamer/gst-plugins-ugly
recipe
contains the following statement:
LICENSE_FLAGS = "commercial"
Here is a slightly more complicated example that contains both an explicit recipe name and version (after variable expansion):
LICENSE_FLAGS = "license_${PN}_${PV}"
In order for a component restricted by a
LICENSE_FLAGS
definition to be enabled and included in an image, it
needs to have a matching entry in the global
LICENSE_FLAGS_WHITELIST
variable, which is a variable typically defined in your local.conf
file. For example, to enable the
poky/meta/recipes-multimedia/gstreamer/gst-plugins-ugly
package, you
could add either the string “commercial_gst-plugins-ugly” or the more
general string “commercial” to LICENSE_FLAGS_WHITELIST
. See the
“License Flag Matching” section for a full
explanation of how LICENSE_FLAGS
matching works. Here is the
example:
LICENSE_FLAGS_WHITELIST = "commercial_gst-plugins-ugly"
Likewise, to additionally enable the package built from the recipe
containing LICENSE_FLAGS = "license_${PN}_${PV}"
, and assuming that
the actual recipe name was emgd_1.10.bb
, the following string would
enable that package as well as the original gst-plugins-ugly
package:
LICENSE_FLAGS_WHITELIST = "commercial_gst-plugins-ugly license_emgd_1.10"
As a convenience, you do not need to specify the complete license string in the whitelist for every package. You can use an abbreviated form, which consists of just the first portion or portions of the license string before the initial underscore character or characters. A partial string will match any license that contains the given string as the first portion of its license. For example, the following whitelist string will also match both of the packages previously mentioned as well as any other packages that have licenses starting with “commercial” or “license”.
LICENSE_FLAGS_WHITELIST = "commercial license"
3.32.2.1 License Flag Matching
License flag matching allows you to control what recipes the
OpenEmbedded build system includes in the build. Fundamentally, the
build system attempts to match LICENSE_FLAGS
strings found in
recipes against LICENSE_FLAGS_WHITELIST
strings found in the
whitelist. A match causes the build system to include a recipe in the
build, while failure to find a match causes the build system to exclude
a recipe.
In general, license flag matching is simple. However, understanding some concepts will help you correctly and effectively use matching.
Before a flag defined by a particular recipe is tested against the
contents of the whitelist, the expanded string _${PN}
is appended to
the flag. This expansion makes each LICENSE_FLAGS
value
recipe-specific. After expansion, the string is then matched against the
whitelist. Thus, specifying LICENSE_FLAGS = "commercial"
in recipe
“foo”, for example, results in the string "commercial_foo"
. And, to
create a match, that string must appear in the whitelist.
Judicious use of the LICENSE_FLAGS
strings and the contents of the
LICENSE_FLAGS_WHITELIST
variable allows you a lot of flexibility for
including or excluding recipes based on licensing. For example, you can
broaden the matching capabilities by using license flags string subsets
in the whitelist.
Note
When using a string subset, be sure to use the part of the expanded
string that precedes the appended underscore character (e.g.
usethispart_1.3
, usethispart_1.4
, and so forth).
For example, simply specifying the string “commercial” in the whitelist
matches any expanded LICENSE_FLAGS
definition that starts with the
string “commercial” such as “commercial_foo” and “commercial_bar”, which
are the strings the build system automatically generates for
hypothetical recipes named “foo” and “bar” assuming those recipes simply
specify the following:
LICENSE_FLAGS = "commercial"
Thus, you can choose to exhaustively enumerate each license flag in the whitelist and allow only specific recipes into the image, or you can use a string subset that causes a broader range of matches to allow a range of recipes into the image.
This scheme works even if the LICENSE_FLAGS
string already has
_${PN}
appended. For example, the build system turns the license
flag “commercial_1.2_foo” into “commercial_1.2_foo_foo” and would match
both the general “commercial” and the specific “commercial_1.2_foo”
strings found in the whitelist, as expected.
Here are some other scenarios:
You can specify a versioned string in the recipe such as “commercial_foo_1.2” in a “foo” recipe. The build system expands this string to “commercial_foo_1.2_foo”. Combine this license flag with a whitelist that has the string “commercial” and you match the flag along with any other flag that starts with the string “commercial”.
Under the same circumstances, you can use “commercial_foo” in the whitelist and the build system not only matches “commercial_foo_1.2” but also matches any license flag with the string “commercial_foo”, regardless of the version.
You can be very specific and use both the package and version parts in the whitelist (e.g. “commercial_foo_1.2”) to specifically match a versioned recipe.
3.32.3 Maintaining Open Source License Compliance During Your Product’s Lifecycle
One of the concerns for a development organization using open source software is how to maintain compliance with various open source licensing during the lifecycle of the product. While this section does not provide legal advice or comprehensively cover all scenarios, it does present methods that you can use to assist you in meeting the compliance requirements during a software release.
With hundreds of different open source licenses that the Yocto Project tracks, it is difficult to know the requirements of each and every license. However, the requirements of the major FLOSS licenses can begin to be covered by assuming that three main areas of concern exist:
Source code must be provided.
License text for the software must be provided.
Compilation scripts and modifications to the source code must be provided.
spdx files can be provided.
There are other requirements beyond the scope of these three and the methods described in this section (e.g. the mechanism through which source code is distributed).
As different organizations have different methods of complying with open source licensing, this section is not meant to imply that there is only one single way to meet your compliance obligations, but rather to describe one method of achieving compliance. The remainder of this section describes methods supported to meet the previously mentioned three requirements. Once you take steps to meet these requirements, and prior to releasing images, sources, and the build system, you should audit all artifacts to ensure completeness.
Note
The Yocto Project generates a license manifest during image creation
that is located in ${DEPLOY_DIR}/licenses/
image_name-
datestamp
to assist with any audits.
3.32.3.1 Providing the Source Code
Compliance activities should begin before you generate the final image. The first thing you should look at is the requirement that tops the list for most compliance groups - providing the source. The Yocto Project has a few ways of meeting this requirement.
One of the easiest ways to meet this requirement is to provide the entire DL_DIR used by the build. This method, however, has a few issues. The most obvious is the size of the directory since it includes all sources used in the build and not just the source used in the released image. It will include toolchain source, and other artifacts, which you would not generally release. However, the more serious issue for most companies is accidental release of proprietary software. The Yocto Project provides an archiver class to help avoid some of these concerns.
Before you employ DL_DIR
or the archiver
class, you need to
decide how you choose to provide source. The source archiver
class
can generate tarballs and SRPMs and can create them with various levels
of compliance in mind.
One way of doing this (but certainly not the only way) is to release
just the source as a tarball. You can do this by adding the following to
the local.conf
file found in the
Build Directory:
INHERIT += "archiver"
ARCHIVER_MODE[src] = "original"
During the creation of your
image, the source from all recipes that deploy packages to the image is
placed within subdirectories of DEPLOY_DIR/sources
based on the
LICENSE for each recipe.
Releasing the entire directory enables you to comply with requirements
concerning providing the unmodified source. It is important to note that
the size of the directory can get large.
A way to help mitigate the size issue is to only release tarballs for licenses that require the release of source. Let us assume you are only concerned with GPL code as identified by running the following script:
# Script to archive a subset of packages matching specific license(s)
# Source and license files are copied into sub folders of package folder
# Must be run from build folder
#!/bin/bash
src_release_dir="source-release"
mkdir -p $src_release_dir
for a in tmp/deploy/sources/*; do
for d in $a/*; do
# Get package name from path
p=`basename $d`
p=${p%-*}
p=${p%-*}
# Only archive GPL packages (update *GPL* regex for your license check)
numfiles=`ls tmp/deploy/licenses/$p/*GPL* 2> /dev/null | wc -l`
if [ $numfiles -gt 1 ]; then
echo Archiving $p
mkdir -p $src_release_dir/$p/source
cp $d/* $src_release_dir/$p/source 2> /dev/null
mkdir -p $src_release_dir/$p/license
cp tmp/deploy/licenses/$p/* $src_release_dir/$p/license 2> /dev/null
fi
done
done
At this point, you
could create a tarball from the gpl_source_release
directory and
provide that to the end user. This method would be a step toward
achieving compliance with section 3a of GPLv2 and with section 6 of
GPLv3.
3.32.3.2 Providing License Text
One requirement that is often overlooked is inclusion of license text.
This requirement also needs to be dealt with prior to generating the
final image. Some licenses require the license text to accompany the
binary. You can achieve this by adding the following to your
local.conf
file:
COPY_LIC_MANIFEST = "1"
COPY_LIC_DIRS = "1"
LICENSE_CREATE_PACKAGE = "1"
Adding these statements to the configuration file ensures that the licenses collected during package generation are included on your image.
Note
Setting all three variables to “1” results in the image having two
copies of the same license file. One copy resides in
/usr/share/common-licenses
and the other resides in
/usr/share/license
.
The reason for this behavior is because COPY_LIC_DIRS and COPY_LIC_MANIFEST add a copy of the license when the image is built but do not offer a path for adding licenses for newly installed packages to an image. LICENSE_CREATE_PACKAGE adds a separate package and an upgrade path for adding licenses to an image.
As the source archiver
class has already archived the original
unmodified source that contains the license files, you would have
already met the requirements for inclusion of the license information
with source as defined by the GPL and other open source licenses.
3.32.3.3 Providing Compilation Scripts and Source Code Modifications
At this point, we have addressed all we need to prior to generating the image. The next two requirements are addressed during the final packaging of the release.
By releasing the version of the OpenEmbedded build system and the layers used during the build, you will be providing both compilation scripts and the source code modifications in one step.
If the deployment team has a BSP Layer and a distro layer, and those those layers are used to patch, compile, package, or modify (in any way) any open source software included in your released images, you might be required to release those layers under section 3 of GPLv2 or section 1 of GPLv3. One way of doing that is with a clean checkout of the version of the Yocto Project and layers used during your build. Here is an example:
# We built using the dunfell branch of the poky repo
$ git clone -b dunfell git://git.yoctoproject.org/poky
$ cd poky
# We built using the release_branch for our layers
$ git clone -b release_branch git://git.mycompany.com/meta-my-bsp-layer
$ git clone -b release_branch git://git.mycompany.com/meta-my-software-layer
# clean up the .git repos
$ find . -name ".git" -type d -exec rm -rf {} \;
One
thing a development organization might want to consider for end-user
convenience is to modify meta-poky/conf/bblayers.conf.sample
to
ensure that when the end user utilizes the released build system to
build an image, the development organization’s layers are included in
the bblayers.conf
file automatically:
# POKY_BBLAYERS_CONF_VERSION is increased each time build/conf/bblayers.conf
# changes incompatibly
POKY_BBLAYERS_CONF_VERSION = "2"
BBPATH = "${TOPDIR}"
BBFILES ?= ""
BBLAYERS ?= " \
##OEROOT##/meta \
##OEROOT##/meta-poky \
##OEROOT##/meta-yocto-bsp \
##OEROOT##/meta-mylayer \
"
Creating and providing an archive of the Metadata layers (recipes, configuration files, and so forth) enables you to meet your requirements to include the scripts to control compilation as well as any modifications to the original source.
3.32.3.4 Providing spdx files
The spdx module has been integrated to a layer named meta-spdxscanner. meta-spdxscanner provides several kinds of scanner. If you want to enable this function, you have to follow the following steps:
Add meta-spdxscanner layer into
bblayers.conf
.Refer to the README in meta-spdxscanner to setup the environment (e.g, setup a fossology server) needed for the scanner.
Meta-spdxscanner provides several methods within the bbclass to create spdx files. Please choose one that you want to use and enable the spdx task. You have to add some config options in
local.conf
file in your Build Directory. The following is an example showing how to generate spdx files during bitbake using the fossology-python.bbclass:# Select fossology-python.bbclass. INHERIT += "fossology-python" # For fossology-python.bbclass, TOKEN is necessary, so, after setup a # Fossology server, you have to create a token. TOKEN = "eyJ0eXAiO..." # The fossology server is necessary for fossology-python.bbclass. FOSSOLOGY_SERVER = "http://xx.xx.xx.xx:8081/repo" # If you want to upload the source code to a special folder: FOLDER_NAME = "xxxx" //Optional # If you don't want to put spdx files in tmp/deploy/spdx, you can enable: SPDX_DEPLOY_DIR = "${DEPLOY_DIR}" //Optional
For more usage information refer to the meta-spdxscanner repository.
3.32.4 Copying Licenses that Do Not Exist
Some packages, such as the linux-firmware package, have many licenses that are not in any way common. You can avoid adding a lot of these types of common license files, which are only applicable to a specific package, by using the NO_GENERIC_LICENSE variable. Using this variable also avoids QA errors when you use a non-common, non-CLOSED license in a recipe.
The following is an example that uses the LICENSE.Abilis.txt
file as
the license from the fetched source:
NO_GENERIC_LICENSE[Firmware-Abilis] = "LICENSE.Abilis.txt"
3.33 Using the Error Reporting Tool
The error reporting tool allows you to submit errors encountered during
builds to a central database. Outside of the build environment, you can
use a web interface to browse errors, view statistics, and query for
errors. The tool works using a client-server system where the client
portion is integrated with the installed Yocto Project
Source Directory (e.g. poky
).
The server receives the information collected and saves it in a
database.
A live instance of the error reporting server exists at https://errors.yoctoproject.org. This server exists so that when you want to get help with build failures, you can submit all of the information on the failure easily and then point to the URL in your bug report or send an email to the mailing list.
Note
If you send error reports to this server, the reports become publicly visible.
3.33.1 Enabling and Using the Tool
By default, the error reporting tool is disabled. You can enable it by
inheriting the
report-error
class by adding the following statement to the end of your
local.conf
file in your
Build Directory.
INHERIT += "report-error"
By default, the error reporting feature stores information in
${
LOG_DIR}/error-report
.
However, you can specify a directory to use by adding the following to
your local.conf
file:
ERR_REPORT_DIR = "path"
Enabling error reporting causes the build process to collect the errors and store them in a file as previously described. When the build system encounters an error, it includes a command as part of the console output. You can run the command to send the error file to the server. For example, the following command sends the errors to an upstream server:
$ send-error-report /home/brandusa/project/poky/build/tmp/log/error-report/error_report_201403141617.txt
In the previous example, the errors are sent to a public database available at https://errors.yoctoproject.org, which is used by the entire community. If you specify a particular server, you can send the errors to a different database. Use the following command for more information on available options:
$ send-error-report --help
When sending the error file, you are prompted to review the data being sent as well as to provide a name and optional email address. Once you satisfy these prompts, the command returns a link from the server that corresponds to your entry in the database. For example, here is a typical link: https://errors.yoctoproject.org/Errors/Details/9522/
Following the link takes you to a web interface where you can browse, query the errors, and view statistics.
3.33.2 Disabling the Tool
To disable the error reporting feature, simply remove or comment out the
following statement from the end of your local.conf
file in your
Build Directory.
INHERIT += "report-error"
3.33.3 Setting Up Your Own Error Reporting Server
If you want to set up your own error reporting server, you can obtain the code from the Git repository at https://git.yoctoproject.org/cgit/cgit.cgi/error-report-web/. Instructions on how to set it up are in the README document.
3.34 Using Wayland and Weston
Wayland is a computer display server protocol that provides a method for compositing window managers to communicate directly with applications and video hardware and expects them to communicate with input hardware using other libraries. Using Wayland with supporting targets can result in better control over graphics frame rendering than an application might otherwise achieve.
The Yocto Project provides the Wayland protocol libraries and the
reference
Weston
compositor as part of its release. You can find the integrated packages
in the meta
layer of the Source Directory.
Specifically, you
can find the recipes that build both Wayland and Weston at
meta/recipes-graphics/wayland
.
You can build both the Wayland and Weston packages for use only with targets that accept the Mesa 3D and Direct Rendering Infrastructure, which is also known as Mesa DRI. This implies that you cannot build and use the packages if your target uses, for example, the Intel Embedded Media and Graphics Driver (Intel EMGD) that overrides Mesa DRI.
Note
Due to lack of EGL support, Weston 1.0.3 will not run directly on the emulated QEMU hardware. However, this version of Weston will run under X emulation without issues.
This section describes what you need to do to implement Wayland and use the Weston compositor when building an image for a supporting target.
3.34.1 Enabling Wayland in an Image
To enable Wayland, you need to enable it to be built and enable it to be included (installed) in the image.
3.34.1.1 Building Wayland
To cause Mesa to build the wayland-egl
platform and Weston to build
Wayland with Kernel Mode Setting
(KMS)
support, include the “wayland” flag in the
DISTRO_FEATURES
statement in your local.conf
file:
DISTRO_FEATURES_append = " wayland"
Note
If X11 has been enabled elsewhere, Weston will build Wayland with X11 support
3.34.1.2 Installing Wayland and Weston
To install the Wayland feature into an image, you must include the
following
CORE_IMAGE_EXTRA_INSTALL
statement in your local.conf
file:
CORE_IMAGE_EXTRA_INSTALL += "wayland weston"
3.34.2 Running Weston
To run Weston inside X11, enabling it as described earlier and building a Sato image is sufficient. If you are running your image under Sato, a Weston Launcher appears in the “Utility” category.
Alternatively, you can run Weston through the command-line interpretor (CLI), which is better suited for development work. To run Weston under the CLI, you need to do the following after your image is built:
Run these commands to export
XDG_RUNTIME_DIR
:mkdir -p /tmp/$USER-weston chmod 0700 /tmp/$USER-weston export XDG_RUNTIME_DIR=/tmp/$USER-weston
Launch Weston in the shell:
weston
4 Using the Quick EMUlator (QEMU)
The Yocto Project uses an implementation of the Quick EMUlator (QEMU) Open Source project as part of the Yocto Project development “tool set”. This chapter provides both procedures that show you how to use the Quick EMUlator (QEMU) and other QEMU information helpful for development purposes.
4.1 Overview
Within the context of the Yocto Project, QEMU is an emulator and virtualization machine that allows you to run a complete image you have built using the Yocto Project as just another task on your build system. QEMU is useful for running and testing images and applications on supported Yocto Project architectures without having actual hardware. Among other things, the Yocto Project uses QEMU to run automated Quality Assurance (QA) tests on final images shipped with each release.
Note
This implementation is not the same as QEMU in general.
This section provides a brief reference for the Yocto Project implementation of QEMU.
For official information and documentation on QEMU in general, see the following references:
QEMU Website: The official website for the QEMU Open Source project.
Documentation: The QEMU user manual.
4.2 Running QEMU
To use QEMU, you need to have QEMU installed and initialized as well as have the proper artifacts (i.e. image files and root filesystems) available. Follow these general steps to run QEMU:
Install QEMU: QEMU is made available with the Yocto Project a number of ways. One method is to install a Software Development Kit (SDK). See “The QEMU Emulator” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual for information on how to install QEMU.
Setting Up the Environment: How you set up the QEMU environment depends on how you installed QEMU:
If you cloned the
poky
repository or you downloaded and unpacked a Yocto Project release tarball, you can source the build environment script (i.e. oe-init-build-env):$ cd poky $ source oe-init-build-env
If you installed a cross-toolchain, you can run the script that initializes the toolchain. For example, the following commands run the initialization script from the default
poky_sdk
directory:. poky_sdk/environment-setup-core2-64-poky-linux
Ensure the Artifacts are in Place: You need to be sure you have a pre-built kernel that will boot in QEMU. You also need the target root filesystem for your target machine’s architecture:
If you have previously built an image for QEMU (e.g.
qemux86
,qemuarm
, and so forth), then the artifacts are in place in your Build Directory.If you have not built an image, you can go to the machines/qemu area and download a pre-built image that matches your architecture and can be run on QEMU.
See the “Extracting the Root Filesystem” section in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual for information on how to extract a root filesystem.
Run QEMU: The basic
runqemu
command syntax is as follows:$ runqemu [option ] [...]
Based on what you provide on the command line,
runqemu
does a good job of figuring out what you are trying to do. For example, by default, QEMU looks for the most recently built image according to the timestamp when it needs to look for an image. Minimally, through the use of options, you must provide either a machine name, a virtual machine image (*wic.vmdk
), or a kernel image (*.bin
).Here are some additional examples to help illustrate further QEMU:
This example starts QEMU with MACHINE set to “qemux86-64”. Assuming a standard Build Directory,
runqemu
automatically finds thebzImage-qemux86-64.bin
image file and thecore-image-minimal-qemux86-64-20200218002850.rootfs.ext4
(assuming the current build created acore-image-minimal
image).Note
When more than one image with the same name exists, QEMU finds and uses the most recently built image according to the timestamp.
$ runqemu qemux86-64
This example produces the exact same results as the previous example. This command, however, specifically provides the image and root filesystem type.
$ runqemu qemux86-64 core-image-minimal ext4
This example specifies to boot an initial RAM disk image and to enable audio in QEMU. For this case,
runqemu
set the internal variableFSTYPE
to “cpio.gz”. Also, for audio to be enabled, an appropriate driver must be installed (see the previous description for theaudio
option for more information).$ runqemu qemux86-64 ramfs audio
This example does not provide enough information for QEMU to launch. While the command does provide a root filesystem type, it must also minimally provide a MACHINE, KERNEL, or VM option.
$ runqemu ext4
This example specifies to boot a virtual machine image (
.wic.vmdk
file). From the.wic.vmdk
,runqemu
determines the QEMU architecture (MACHINE) to be “qemux86-64” and the root filesystem type to be “vmdk”.$ runqemu /home/scott-lenovo/vm/core-image-minimal-qemux86-64.wic.vmdk
4.3 Switching Between Consoles
When booting or running QEMU, you can switch between supported consoles by using Ctrl+Alt+number. For example, Ctrl+Alt+3 switches you to the serial console as long as that console is enabled. Being able to switch consoles is helpful, for example, if the main QEMU console breaks for some reason.
Note
Usually, “2” gets you to the main console and “3” gets you to the serial console.
4.4 Removing the Splash Screen
You can remove the splash screen when QEMU is booting by using Alt+left. Removing the splash screen allows you to see what is happening in the background.
4.5 Disabling the Cursor Grab
The default QEMU integration captures the cursor within the main window. It does this since standard mouse devices only provide relative input and not absolute coordinates. You then have to break out of the grab using the “Ctrl+Alt” key combination. However, the Yocto Project’s integration of QEMU enables the wacom USB touch pad driver by default to allow input of absolute coordinates. This default means that the mouse can enter and leave the main window without the grab taking effect leading to a better user experience.
4.6 Running Under a Network File System (NFS) Server
One method for running QEMU is to run it on an NFS server. This is useful when you need to access the same file system from both the build and the emulated system at the same time. It is also worth noting that the system does not need root privileges to run. It uses a user space NFS server to avoid that. Follow these steps to set up for running QEMU using an NFS server.
Extract a Root Filesystem: Once you are able to run QEMU in your environment, you can use the
runqemu-extract-sdk
script, which is located in thescripts
directory along with therunqemu
script.The
runqemu-extract-sdk
takes a root filesystem tarball and extracts it into a location that you specify. Here is an example that takes a file system and extracts it to a directory namedtest-nfs
:runqemu-extract-sdk ./tmp/deploy/images/qemux86-64/core-image-sato-qemux86-64.tar.bz2 test-nfs
Start QEMU: Once you have extracted the file system, you can run
runqemu
normally with the additional location of the file system. You can then also make changes to the files within./test-nfs
and see those changes appear in the image in real time. Here is an example using theqemux86
image:runqemu qemux86-64 ./test-nfs
Note
Should you need to start, stop, or restart the NFS share, you can use the following commands:
The following command starts the NFS share:
runqemu-export-rootfs start file-system-location
The following command stops the NFS share:
runqemu-export-rootfs stop file-system-location
The following command restarts the NFS share:
runqemu-export-rootfs restart file-system-location
4.7 QEMU CPU Compatibility Under KVM
By default, the QEMU build compiles for and targets 64-bit and x86 Intel Core2 Duo processors and 32-bit x86 Intel Pentium II processors. QEMU builds for and targets these CPU types because they display a broad range of CPU feature compatibility with many commonly used CPUs.
Despite this broad range of compatibility, the CPUs could support a
feature that your host CPU does not support. Although this situation is
not a problem when QEMU uses software emulation of the feature, it can
be a problem when QEMU is running with KVM enabled. Specifically,
software compiled with a certain CPU feature crashes when run on a CPU
under KVM that does not support that feature. To work around this
problem, you can override QEMU’s runtime CPU setting by changing the
QB_CPU_KVM
variable in qemuboot.conf
in the
Build Directory deploy/image
directory. This setting specifies a -cpu
option passed into QEMU in
the runqemu
script. Running qemu -cpu help
returns a list of
available supported CPU types.
4.8 QEMU Performance
Using QEMU to emulate your hardware can result in speed issues depending
on the target and host architecture mix. For example, using the
qemux86
image in the emulator on an Intel-based 32-bit (x86) host
machine is fast because the target and host architectures match. On the
other hand, using the qemuarm
image on the same Intel-based host can
be slower. But, you still achieve faithful emulation of ARM-specific
issues.
To speed things up, the QEMU images support using distcc
to call a
cross-compiler outside the emulated system. If you used runqemu
to
start QEMU, and the distccd
application is present on the host
system, any BitBake cross-compiling toolchain available from the build
system is automatically used from within QEMU simply by calling
distcc
. You can accomplish this by defining the cross-compiler
variable (e.g. export CC="distcc"
). Alternatively, if you are using
a suitable SDK image or the appropriate stand-alone toolchain is
present, the toolchain is also automatically used.
Note
Several mechanisms exist that let you connect to the system running on the QEMU emulator:
QEMU provides a framebuffer interface that makes standard consoles available.
Generally, headless embedded devices have a serial port. If so, you can configure the operating system of the running image to use that port to run a console. The connection uses standard IP networking.
SSH servers exist in some QEMU images. The
core-image-sato
QEMU image has a Dropbear secure shell (SSH) server that runs with the root password disabled. Thecore-image-full-cmdline
andcore-image-lsb
QEMU images have OpenSSH instead of Dropbear. Including these SSH servers allow you to use standardssh
andscp
commands. Thecore-image-minimal
QEMU image, however, contains no SSH server.You can use a provided, user-space NFS server to boot the QEMU session using a local copy of the root filesystem on the host. In order to make this connection, you must extract a root filesystem tarball by using the
runqemu-extract-sdk
command. After running the command, you must then point therunqemu
script to the extracted directory instead of a root filesystem image file. See the “Running Under a Network File System (NFS) Server” section for more information.
4.9 QEMU Command-Line Syntax
The basic runqemu
command syntax is as follows:
$ runqemu [option ] [...]
Based on what you provide on the command line, runqemu
does a
good job of figuring out what you are trying to do. For example, by
default, QEMU looks for the most recently built image according to the
timestamp when it needs to look for an image. Minimally, through the use
of options, you must provide either a machine name, a virtual machine
image (*wic.vmdk
), or a kernel image (*.bin
).
Following is the command-line help output for the runqemu
command:
$ runqemu --help
Usage: you can run this script with any valid combination
of the following environment variables (in any order):
KERNEL - the kernel image file to use
ROOTFS - the rootfs image file or nfsroot directory to use
MACHINE - the machine name (optional, autodetected from KERNEL filename if unspecified)
Simplified QEMU command-line options can be passed with:
nographic - disable video console
serial - enable a serial console on /dev/ttyS0
slirp - enable user networking, no root privileges is required
kvm - enable KVM when running x86/x86_64 (VT-capable CPU required)
kvm-vhost - enable KVM with vhost when running x86/x86_64 (VT-capable CPU required)
publicvnc - enable a VNC server open to all hosts
audio - enable audio
[*/]ovmf* - OVMF firmware file or base name for booting with UEFI
tcpserial=<port> - specify tcp serial port number
biosdir=<dir> - specify custom bios dir
biosfilename=<filename> - specify bios filename
qemuparams=<xyz> - specify custom parameters to QEMU
bootparams=<xyz> - specify custom kernel parameters during boot
help, -h, --help: print this text
Examples:
runqemu
runqemu qemuarm
runqemu tmp/deploy/images/qemuarm
runqemu tmp/deploy/images/qemux86/<qemuboot.conf>
runqemu qemux86-64 core-image-sato ext4
runqemu qemux86-64 wic-image-minimal wic
runqemu path/to/bzImage-qemux86.bin path/to/nfsrootdir/ serial
runqemu qemux86 iso/hddimg/wic.vmdk/wic.qcow2/wic.vdi/ramfs/cpio.gz...
runqemu qemux86 qemuparams="-m 256"
runqemu qemux86 bootparams="psplash=false"
runqemu path/to/<image>-<machine>.wic
runqemu path/to/<image>-<machine>.wic.vmdk
4.10 runqemu
Command-Line Options
Following is a description of runqemu
options you can provide on the
command line:
Note
If you do provide some “illegal” option combination or perhaps you do
not provide enough in the way of options, runqemu
provides appropriate error messaging to help you correct the problem.
QEMUARCH: The QEMU machine architecture, which must be “qemuarm”, “qemuarm64”, “qemumips”, “qemumips64”, “qemuppc”, “qemux86”, or “qemux86-64”.
VM: The virtual machine image, which must be a
.wic.vmdk
file. Use this option when you want to boot a.wic.vmdk
image. The image filename you provide must contain one of the following strings: “qemux86-64”, “qemux86”, “qemuarm”, “qemumips64”, “qemumips”, “qemuppc”, or “qemush4”.ROOTFS: A root filesystem that has one of the following filetype extensions: “ext2”, “ext3”, “ext4”, “jffs2”, “nfs”, or “btrfs”. If the filename you provide for this option uses “nfs”, it must provide an explicit root filesystem path.
KERNEL: A kernel image, which is a
.bin
file. When you provide a.bin
file,runqemu
detects it and assumes the file is a kernel image.MACHINE: The architecture of the QEMU machine, which must be one of the following: “qemux86”, “qemux86-64”, “qemuarm”, “qemuarm64”, “qemumips”, “qemumips64”, or “qemuppc”. The MACHINE and QEMUARCH options are basically identical. If you do not provide a MACHINE option,
runqemu
tries to determine it based on other options.ramfs
: Indicates you are booting an initial RAM disk (initramfs) image, which means theFSTYPE
iscpio.gz
.iso
: Indicates you are booting an ISO image, which means theFSTYPE
is.iso
.nographic
: Disables the video console, which sets the console to “ttys0”. This option is useful when you have logged into a server and you do not want to disable forwarding from the X Window System (X11) to your workstation or laptop.serial
: Enables a serial console on/dev/ttyS0
.biosdir
: Establishes a custom directory for BIOS, VGA BIOS and keymaps.biosfilename
: Establishes a custom BIOS name.qemuparams=\"xyz\"
: Specifies custom QEMU parameters. Use this option to pass options other than the simple “kvm” and “serial” options.bootparams=\"xyz\"
: Specifies custom boot parameters for the kernel.audio
: Enables audio in QEMU. The MACHINE option must be either “qemux86” or “qemux86-64” in order for audio to be enabled. Additionally, thesnd_intel8x0
orsnd_ens1370
driver must be installed in linux guest.slirp
: Enables “slirp” networking, which is a different way of networking that does not need root access but also is not as easy to use or comprehensive as the default.kvm
: Enables KVM when running “qemux86” or “qemux86-64” QEMU architectures. For KVM to work, all the following conditions must be met:Your MACHINE must be either qemux86” or “qemux86-64”.
Your build host has to have the KVM modules installed, which are
/dev/kvm
.The build host
/dev/kvm
directory has to be both writable and readable.
kvm-vhost
: Enables KVM with VHOST support when running “qemux86” or “qemux86-64” QEMU architectures. For KVM with VHOST to work, the following conditions must be met:kvm
option conditions defined above must be met.Your build host has to have virtio net device, which are
/dev/vhost-net
.The build host
/dev/vhost-net
directory has to be either readable or writable and “slirp-enabled”.
publicvnc
: Enables a VNC server open to all hosts.
5 Manual Revision History
Revision |
Date |
Note |
---|---|---|
1.1 |
October 2011 |
The initial document released with the Yocto Project 1.1 Release |
1.2 |
April 2012 |
Released with the Yocto Project 1.2 Release. |
1.3 |
October 2012 |
Released with the Yocto Project 1.3 Release. |
1.4 |
April 2013 |
Released with the Yocto Project 1.4 Release. |
1.5 |
October 2013 |
Released with the Yocto Project 1.5 Release. |
1.6 |
April 2014 |
Released with the Yocto Project 1.6 Release. |
1.7 |
October 2014 |
Released with the Yocto Project 1.7 Release. |
1.8 |
April 2015 |
Released with the Yocto Project 1.8 Release. |
2.0 |
October 2015 |
Released with the Yocto Project 2.0 Release. |
2.1 |
April 2016 |
Released with the Yocto Project 2.1 Release. |
2.2 |
October 2016 |
Released with the Yocto Project 2.2 Release. |
2.3 |
May 2017 |
Released with the Yocto Project 2.3 Release. |
2.4 |
October 2017 |
Released with the Yocto Project 2.4 Release. |
2.5 |
May 2018 |
Released with the Yocto Project 2.5 Release. |
2.6 |
November 2018 |
Released with the Yocto Project 2.6 Release. |
2.7 |
May 2019 |
Released with the Yocto Project 2.7 Release. |
3.0 |
October 2019 |
Released with the Yocto Project 3.0 Release. |
3.1 |
April 2020 |
Released with the Yocto Project 3.1 Release. |
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Yocto Project Linux Kernel Development Manual
1 Introduction
1.1 Overview
Regardless of how you intend to make use of the Yocto Project, chances are you will work with the Linux kernel. This manual describes how to set up your build host to support kernel development, introduces the kernel development process, provides background information on the Yocto Linux kernel Metadata, describes common tasks you can perform using the kernel tools, shows you how to use the kernel Metadata needed to work with the kernel inside the Yocto Project, and provides insight into how the Yocto Project team develops and maintains Yocto Linux kernel Git repositories and Metadata.
Each Yocto Project release has a set of Yocto Linux kernel recipes, whose Git repositories you can view in the Yocto Source Repositories under the “Yocto Linux Kernel” heading. New recipes for the release track the latest Linux kernel upstream developments from https://www.kernel.org and introduce newly-supported platforms. Previous recipes in the release are refreshed and supported for at least one additional Yocto Project release. As they align, these previous releases are updated to include the latest from the Long Term Support Initiative (LTSI) project. You can learn more about Yocto Linux kernels and LTSI in the “Yocto Project Kernel Development and Maintenance” section.
Also included is a Yocto Linux kernel development recipe
(linux-yocto-dev.bb
) should you want to work with the very latest in
upstream Yocto Linux kernel development and kernel Metadata development.
Note
For more on Yocto Linux kernels, see the “Yocto Project Kernel Development and Maintenance” section.
The Yocto Project also provides a powerful set of kernel tools for managing Yocto Linux kernel sources and configuration data. You can use these tools to make a single configuration change, apply multiple patches, or work with your own kernel sources.
In particular, the kernel tools allow you to generate configuration
fragments that specify only what you must, and nothing more.
Configuration fragments only need to contain the highest level visible
CONFIG
options as presented by the Yocto Linux kernel menuconfig
system. Contrast this against a complete Yocto Linux kernel .config
file, which includes all the automatically selected CONFIG
options.
This efficiency reduces your maintenance effort and allows you to
further separate your configuration in ways that make sense for your
project. A common split separates policy and hardware. For example, all
your kernels might support the proc
and sys
filesystems, but
only specific boards require sound, USB, or specific drivers. Specifying
these configurations individually allows you to aggregate them together
as needed, but maintains them in only one place. Similar logic applies
to separating source changes.
If you do not maintain your own kernel sources and need to make only minimal changes to the sources, the released recipes provide a vetted base upon which to layer your changes. Doing so allows you to benefit from the continual kernel integration and testing performed during development of the Yocto Project.
If, instead, you have a very specific Linux kernel source tree and are unable to align with one of the official Yocto Linux kernel recipes, an alternative exists by which you can use the Yocto Project Linux kernel tools with your own kernel sources.
The remainder of this manual provides instructions for completing specific Linux kernel development tasks. These instructions assume you are comfortable working with BitBake recipes and basic open-source development tools. Understanding these concepts will facilitate the process of working with the kernel recipes. If you find you need some additional background, please be sure to review and understand the following documentation:
Yocto Project Quick Build document.
devtool workflow as described in the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual.
The “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual.
The “Kernel Modification Workflow” section.
1.2 Kernel Modification Workflow
Kernel modification involves changing the Yocto Project kernel, which
could involve changing configuration options as well as adding new
kernel recipes. Configuration changes can be added in the form of
configuration fragments, while recipe modification comes through the
kernel’s recipes-kernel
area in a kernel layer you create.
This section presents a high-level overview of the Yocto Project kernel modification workflow. The illustration and accompanying list provide general information and references for further information.
Set up Your Host Development System to Support Development Using the Yocto Project: See the “Setting Up to Use the Yocto Project” section in the Yocto Project Development Tasks Manual for options on how to get a build host ready to use the Yocto Project.
Set Up Your Host Development System for Kernel Development: It is recommended that you use
devtool
and an extensible SDK for kernel development. Alternatively, you can use traditional kernel development methods with the Yocto Project. Either way, there are steps you need to take to get the development environment ready.Using
devtool
and the eSDK requires that you have a clean build of the image and that you are set up with the appropriate eSDK. For more information, see the “Getting Ready to Develop Using devtool” section.Using traditional kernel development requires that you have the kernel source available in an isolated local Git repository. For more information, see the “Getting Ready for Traditional Kernel Development” section.
Make Changes to the Kernel Source Code if applicable: Modifying the kernel does not always mean directly changing source files. However, if you have to do this, you make the changes to the files in the eSDK’s Build Directory if you are using
devtool
. For more information, see the “Using devtool to Patch the Kernel” section.If you are using traditional kernel development, you edit the source files in the kernel’s local Git repository. For more information, see the “Using Traditional Kernel Development to Patch the Kernel” section.
Make Kernel Configuration Changes if Applicable: If your situation calls for changing the kernel’s configuration, you can use menuconfig, which allows you to interactively develop and test the configuration changes you are making to the kernel. Saving changes you make with
menuconfig
updates the kernel’s.config
file.Note
Try to resist the temptation to directly edit an existing
.config
file, which is found in the Build Directory among the source code used for the build. Doing so, can produce unexpected results when the OpenEmbedded build system regenerates the configuration file.Once you are satisfied with the configuration changes made using
menuconfig
and you have saved them, you can directly compare the resulting.config
file against an existing original and gather those changes into a configuration fragment file to be referenced from within the kernel’s.bbappend
file.Additionally, if you are working in a BSP layer and need to modify the BSP’s kernel’s configuration, you can use
menuconfig
.Rebuild the Kernel Image With Your Changes: Rebuilding the kernel image applies your changes. Depending on your target hardware, you can verify your changes on actual hardware or perhaps QEMU.
The remainder of this developer’s guide covers common tasks typically used during kernel development, advanced Metadata usage, and Yocto Linux kernel maintenance concepts.
2 Common Tasks
This chapter presents several common tasks you perform when you work with the Yocto Project Linux kernel. These tasks include preparing your host development system for kernel development, preparing a layer, modifying an existing recipe, patching the kernel, configuring the kernel, iterative development, working with your own sources, and incorporating out-of-tree modules.
Note
The examples presented in this chapter work with the Yocto Project 2.4 Release and forward.
2.1 Preparing the Build Host to Work on the Kernel
Before you can do any kernel development, you need to be sure your build
host is set up to use the Yocto Project. For information on how to get
set up, see the “Setting Up to Use the Yocto Project” section in
the Yocto Project Development Tasks Manual. Part of preparing the system
is creating a local Git repository of the
Source Directory (poky
) on your system. Follow the steps in the
“Cloning the poky Repository”
section in the Yocto Project Development Tasks Manual to set up your
Source Directory.
Note
Be sure you check out the appropriate development branch or you create your local branch by checking out a specific tag to get the desired version of Yocto Project. See the “Checking Out by Branch in Poky” and “Checking Out by Tag in Poky” sections in the Yocto Project Development Tasks Manual for more information.
Kernel development is best accomplished using devtool and not through traditional kernel workflow methods. The remainder of this section provides information for both scenarios.
2.1.1 Getting Ready to Develop Using devtool
Follow these steps to prepare to update the kernel image using
devtool
. Completing this procedure leaves you with a clean kernel
image and ready to make modifications as described in the
“Using devtool to Patch the Kernel”
section:
Initialize the BitBake Environment: Before building an extensible SDK, you need to initialize the BitBake build environment by sourcing the build environment script (i.e. oe-init-build-env):
$ cd poky $ source oe-init-build-env
Note
The previous commands assume the Yocto Project Source Repositories (i.e.
poky
) have been cloned using Git and the local repository is named “poky”.Prepare Your local.conf File: By default, the MACHINE variable is set to “qemux86-64”, which is fine if you are building for the QEMU emulator in 64-bit mode. However, if you are not, you need to set the
MACHINE
variable appropriately in yourconf/local.conf
file found in the Build Directory (i.e.poky/build
in this example).Also, since you are preparing to work on the kernel image, you need to set the MACHINE_ESSENTIAL_EXTRA_RRECOMMENDS variable to include kernel modules.
In this example we wish to build for qemux86 so we must set the
MACHINE
variable to “qemux86” and also add the “kernel-modules”. As described we do this by appending toconf/local.conf
:MACHINE = "qemux86" MACHINE_ESSENTIAL_EXTRA_RRECOMMENDS += "kernel-modules"
Create a Layer for Patches: You need to create a layer to hold patches created for the kernel image. You can use the
bitbake-layers create-layer
command as follows:$ cd poky/build $ bitbake-layers create-layer ../../meta-mylayer NOTE: Starting bitbake server... Add your new layer with 'bitbake-layers add-layer ../../meta-mylayer' $
Note
For background information on working with common and BSP layers, see the “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual and the “BSP Layers” section in the Yocto Project Board Support (BSP) Developer’s Guide, respectively. For information on how to use the
bitbake-layers create-layer
command to quickly set up a layer, see the “Creating a General Layer Using the bitbake-layers Script” section in the Yocto Project Development Tasks Manual.Inform the BitBake Build Environment About Your Layer: As directed when you created your layer, you need to add the layer to the BBLAYERS variable in the
bblayers.conf
file as follows:$ cd poky/build $ bitbake-layers add-layer ../../meta-mylayer NOTE: Starting bitbake server... $
Build the Extensible SDK: Use BitBake to build the extensible SDK specifically for use with images to be run using QEMU:
$ cd poky/build $ bitbake core-image-minimal -c populate_sdk_ext
Once the build finishes, you can find the SDK installer file (i.e.
*.sh
file) in the following directory:poky/build/tmp/deploy/sdk
For this example, the installer file is named
poky-glibc-x86_64-core-image-minimal-i586-toolchain-ext-3.3.1.sh
.Install the Extensible SDK: Use the following command to install the SDK. For this example, install the SDK in the default
poky_sdk
directory:$ cd poky/build/tmp/deploy/sdk $ ./poky-glibc-x86_64-core-image-minimal-i586-toolchain-ext-3.3.1.sh Poky (Yocto Project Reference Distro) Extensible SDK installer version 3.3.1 ============================================================================ Enter target directory for SDK (default: poky_sdk): You are about to install the SDK to "/home/scottrif/poky_sdk". Proceed [Y/n]? Y Extracting SDK......................................done Setting it up... Extracting buildtools... Preparing build system... Parsing recipes: 100% |#################################################################| Time: 0:00:52 Initializing tasks: 100% |############## ###############################################| Time: 0:00:04 Checking sstate mirror object availability: 100% |######################################| Time: 0:00:00 Parsing recipes: 100% |#################################################################| Time: 0:00:33 Initializing tasks: 100% |##############################################################| Time: 0:00:00 done SDK has been successfully set up and is ready to be used. Each time you wish to use the SDK in a new shell session, you need to source the environment setup script e.g. $ . /home/scottrif/poky_sdk/environment-setup-i586-poky-linux
Set Up a New Terminal to Work With the Extensible SDK: You must set up a new terminal to work with the SDK. You cannot use the same BitBake shell used to build the installer.
After opening a new shell, run the SDK environment setup script as directed by the output from installing the SDK:
$ source poky_sdk/environment-setup-i586-poky-linux "SDK environment now set up; additionally you may now run devtool to perform development tasks. Run devtool --help for further details.
Note
If you get a warning about attempting to use the extensible SDK in an environment set up to run BitBake, you did not use a new shell.
Build the Clean Image: The final step in preparing to work on the kernel is to build an initial image using
devtool
in the new terminal you just set up and initialized for SDK work:$ devtool build-image Parsing recipes: 100% |##########################################| Time: 0:00:05 Parsing of 830 .bb files complete (0 cached, 830 parsed). 1299 targets, 47 skipped, 0 masked, 0 errors. WARNING: No packages to add, building image core-image-minimal unmodified Loading cache: 100% |############################################| Time: 0:00:00 Loaded 1299 entries from dependency cache. NOTE: Resolving any missing task queue dependencies Initializing tasks: 100% |#######################################| Time: 0:00:07 Checking sstate mirror object availability: 100% |###############| Time: 0:00:00 NOTE: Executing SetScene Tasks NOTE: Executing RunQueue Tasks NOTE: Tasks Summary: Attempted 2866 tasks of which 2604 didn't need to be rerun and all succeeded. NOTE: Successfully built core-image-minimal. You can find output files in /home/scottrif/poky_sdk/tmp/deploy/images/qemux86
If you were building for actual hardware and not for emulation, you could flash the image to a USB stick on
/dev/sdd
and boot your device. For an example that uses a Minnowboard, see the TipsAndTricks/KernelDevelopmentWithEsdk Wiki page.
At this point you have set up to start making modifications to the kernel by using the extensible SDK. For a continued example, see the “Using devtool to Patch the Kernel” section.
2.1.2 Getting Ready for Traditional Kernel Development
Getting ready for traditional kernel development using the Yocto Project involves many of the same steps as described in the previous section. However, you need to establish a local copy of the kernel source since you will be editing these files.
Follow these steps to prepare to update the kernel image using traditional kernel development flow with the Yocto Project. Completing this procedure leaves you ready to make modifications to the kernel source as described in the “Using Traditional Kernel Development to Patch the Kernel” section:
Initialize the BitBake Environment: Before you can do anything using BitBake, you need to initialize the BitBake build environment by sourcing the build environment script (i.e. oe-init-build-env). Also, for this example, be sure that the local branch you have checked out for
poky
is the Yocto Project Hardknott branch. If you need to checkout out the Hardknott branch, see the “Checking Out by Branch in Poky” section in the Yocto Project Development Tasks Manual.$ cd poky $ git branch master * hardknott $ source oe-init-build-env
Note
The previous commands assume the Yocto Project Source Repositories (i.e.
poky
) have been cloned using Git and the local repository is named “poky”.Prepare Your local.conf File: By default, the MACHINE variable is set to “qemux86-64”, which is fine if you are building for the QEMU emulator in 64-bit mode. However, if you are not, you need to set the
MACHINE
variable appropriately in yourconf/local.conf
file found in the Build Directory (i.e.poky/build
in this example).Also, since you are preparing to work on the kernel image, you need to set the MACHINE_ESSENTIAL_EXTRA_RRECOMMENDS variable to include kernel modules.
In this example we wish to build for qemux86 so we must set the
MACHINE
variable to “qemux86” and also add the “kernel-modules”. As described we do this by appending toconf/local.conf
:MACHINE = "qemux86" MACHINE_ESSENTIAL_EXTRA_RRECOMMENDS += "kernel-modules"
Create a Layer for Patches: You need to create a layer to hold patches created for the kernel image. You can use the
bitbake-layers create-layer
command as follows:$ cd poky/build $ bitbake-layers create-layer ../../meta-mylayer NOTE: Starting bitbake server... Add your new layer with 'bitbake-layers add-layer ../../meta-mylayer'
Note
For background information on working with common and BSP layers, see the “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual and the “BSP Layers” section in the Yocto Project Board Support (BSP) Developer’s Guide, respectively. For information on how to use the
bitbake-layers create-layer
command to quickly set up a layer, see the “Creating a General Layer Using the bitbake-layers Script” section in the Yocto Project Development Tasks Manual.Inform the BitBake Build Environment About Your Layer: As directed when you created your layer, you need to add the layer to the BBLAYERS variable in the
bblayers.conf
file as follows:$ cd poky/build $ bitbake-layers add-layer ../../meta-mylayer NOTE: Starting bitbake server ... $
Create a Local Copy of the Kernel Git Repository: You can find Git repositories of supported Yocto Project kernels organized under “Yocto Linux Kernel” in the Yocto Project Source Repositories at https://git.yoctoproject.org/cgit/cgit.cgi/.
For simplicity, it is recommended that you create your copy of the kernel Git repository outside of the Source Directory, which is usually named
poky
. Also, be sure you are in thestandard/base
branch.The following commands show how to create a local copy of the
linux-yocto-4.12
kernel and be in thestandard/base
branch.Note
The
linux-yocto-4.12
kernel can be used with the Yocto Project 2.4 release and forward. You cannot use thelinux-yocto-4.12
kernel with releases prior to Yocto Project 2.4.$ cd ~ $ git clone git://git.yoctoproject.org/linux-yocto-4.12 --branch standard/base Cloning into 'linux-yocto-4.12'... remote: Counting objects: 6097195, done. remote: Compressing objects: 100% (901026/901026), done. remote: Total 6097195 (delta 5152604), reused 6096847 (delta 5152256) Receiving objects: 100% (6097195/6097195), 1.24 GiB | 7.81 MiB/s, done. Resolving deltas: 100% (5152604/5152604), done. Checking connectivity... done. Checking out files: 100% (59846/59846), done.
Create a Local Copy of the Kernel Cache Git Repository: For simplicity, it is recommended that you create your copy of the kernel cache Git repository outside of the Source Directory, which is usually named
poky
. Also, for this example, be sure you are in theyocto-4.12
branch.The following commands show how to create a local copy of the
yocto-kernel-cache
and be in theyocto-4.12
branch:$ cd ~ $ git clone git://git.yoctoproject.org/yocto-kernel-cache --branch yocto-4.12 Cloning into 'yocto-kernel-cache'... remote: Counting objects: 22639, done. remote: Compressing objects: 100% (9761/9761), done. remote: Total 22639 (delta 12400), reused 22586 (delta 12347) Receiving objects: 100% (22639/22639), 22.34 MiB | 6.27 MiB/s, done. Resolving deltas: 100% (12400/12400), done. Checking connectivity... done.
At this point, you are ready to start making modifications to the kernel using traditional kernel development steps. For a continued example, see the “Using Traditional Kernel Development to Patch the Kernel” section.
2.2 Creating and Preparing a Layer
If you are going to be modifying kernel recipes, it is recommended that
you create and prepare your own layer in which to do your work. Your
layer contains its own BitBake
append files (.bbappend
) and provides a convenient mechanism to
create your own recipe files (.bb
) as well as store and use kernel
patch files. For background information on working with layers, see the
“Understanding and Creating Layers”
section in the Yocto Project Development Tasks Manual.
Note
The Yocto Project comes with many tools that simplify tasks you need
to perform. One such tool is the bitbake-layers create-layer
command, which simplifies creating a new layer. See the
“Creating a General Layer Using the bitbake-layers Script”
section in the Yocto Project Development Tasks Manual for
information on how to use this script to quick set up a new layer.
To better understand the layer you create for kernel development, the
following section describes how to create a layer without the aid of
tools. These steps assume creation of a layer named mylayer
in your
home directory:
Create Structure: Create the layer’s structure:
$ mkdir meta-mylayer $ mkdir meta-mylayer/conf $ mkdir meta-mylayer/recipes-kernel $ mkdir meta-mylayer/recipes-kernel/linux $ mkdir meta-mylayer/recipes-kernel/linux/linux-yocto
The
conf
directory holds your configuration files, while therecipes-kernel
directory holds your append file and eventual patch files.Create the Layer Configuration File: Move to the
meta-mylayer/conf
directory and create thelayer.conf
file as follows:# We have a conf and classes directory, add to BBPATH BBPATH .= ":${LAYERDIR}" # We have recipes-* directories, add to BBFILES BBFILES += "${LAYERDIR}/recipes-*/*/*.bb \ ${LAYERDIR}/recipes-*/*/*.bbappend" BBFILE_COLLECTIONS += "mylayer" BBFILE_PATTERN_mylayer = "^${LAYERDIR}/" BBFILE_PRIORITY_mylayer = "5"
Notice
mylayer
as part of the last three statements.Create the Kernel Recipe Append File: Move to the
meta-mylayer/recipes-kernel/linux
directory and create the kernel’s append file. This example uses thelinux-yocto-4.12
kernel. Thus, the name of the append file islinux-yocto_4.12.bbappend
:FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:" SRC_URI_append = " file://patch-file-one.patch" SRC_URI_append = " file://patch-file-two.patch" SRC_URI_append = " file://patch-file-three.patch"
The FILESEXTRAPATHS and SRC_URI statements enable the OpenEmbedded build system to find patch files. For more information on using append files, see the “Using .bbappend Files in Your Layer” section in the Yocto Project Development Tasks Manual.
2.3 Modifying an Existing Recipe
In many cases, you can customize an existing linux-yocto recipe to meet
the needs of your project. Each release of the Yocto Project provides a
few Linux kernel recipes from which you can choose. These are located in
the Source Directory in
meta/recipes-kernel/linux
.
Modifying an existing recipe can consist of the following:
Before modifying an existing recipe, be sure that you have created a minimal, custom layer from which you can work. See the “Creating and Preparing a Layer” section for information.
2.3.1 Creating the Append File
You create this file in your custom layer. You also name it accordingly
based on the linux-yocto recipe you are using. For example, if you are
modifying the meta/recipes-kernel/linux/linux-yocto_4.12.bb
recipe,
the append file will typically be located as follows within your custom
layer:
your-layer/recipes-kernel/linux/linux-yocto_4.12.bbappend
The append file should initially extend the FILESPATH search path by prepending the directory that contains your files to the FILESEXTRAPATHS variable as follows:
FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:"
The path ${
THISDIR}/${
PN}
expands to “linux-yocto” in the current directory for this example. If
you add any new files that modify the kernel recipe and you have
extended FILESPATH
as described above, you must place the files in
your layer in the following area:
your-layer/recipes-kernel/linux/linux-yocto/
Note
If you are working on a new machine Board Support Package (BSP), be sure to refer to the Yocto Project Board Support Package Developer’s Guide.
As an example, consider the following append file used by the BSPs in
meta-yocto-bsp
:
meta-yocto-bsp/recipes-kernel/linux/linux-yocto_4.12.bbappend
The following listing shows the file. Be aware that the actual commit ID
strings in this example listing might be different than the actual
strings in the file from the meta-yocto-bsp
layer upstream.
KBRANCH_genericx86 = "standard/base"
KBRANCH_genericx86-64 = "standard/base"
KMACHINE_genericx86 ?= "common-pc"
KMACHINE_genericx86-64 ?= "common-pc-64"
KBRANCH_edgerouter = "standard/edgerouter"
KBRANCH_beaglebone = "standard/beaglebone"
SRCREV_machine_genericx86 ?= "d09f2ce584d60ecb7890550c22a80c48b83c2e19"
SRCREV_machine_genericx86-64 ?= "d09f2ce584d60ecb7890550c22a80c48b83c2e19"
SRCREV_machine_edgerouter ?= "b5c8cfda2dfe296410d51e131289fb09c69e1e7d"
SRCREV_machine_beaglebone ?= "b5c8cfda2dfe296410d51e131289fb09c69e1e7d"
COMPATIBLE_MACHINE_genericx86 = "genericx86"
COMPATIBLE_MACHINE_genericx86-64 = "genericx86-64"
COMPATIBLE_MACHINE_edgerouter = "edgerouter"
COMPATIBLE_MACHINE_beaglebone = "beaglebone"
LINUX_VERSION_genericx86 = "4.12.7"
LINUX_VERSION_genericx86-64 = "4.12.7"
LINUX_VERSION_edgerouter = "4.12.10"
LINUX_VERSION_beaglebone = "4.12.10"
This append file contains statements used to support several BSPs that ship with the Yocto Project. The file defines machines using the COMPATIBLE_MACHINE variable and uses the KMACHINE variable to ensure the machine name used by the OpenEmbedded build system maps to the machine name used by the Linux Yocto kernel. The file also uses the optional KBRANCH variable to ensure the build process uses the appropriate kernel branch.
Although this particular example does not use it, the
KERNEL_FEATURES
variable could be used to enable features specific to the kernel. The
append file points to specific commits in the
Source Directory Git repository and
the meta
Git repository branches to identify the exact kernel needed
to build the BSP.
One thing missing in this particular BSP, which you will typically need
when developing a BSP, is the kernel configuration file (.config
)
for your BSP. When developing a BSP, you probably have a kernel
configuration file or a set of kernel configuration files that, when
taken together, define the kernel configuration for your BSP. You can
accomplish this definition by putting the configurations in a file or a
set of files inside a directory located at the same level as your
kernel’s append file and having the same name as the kernel’s main
recipe file. With all these conditions met, simply reference those files
in the SRC_URI statement in
the append file.
For example, suppose you had some configuration options in a file called
network_configs.cfg
. You can place that file inside a directory
named linux-yocto
and then add a SRC_URI
statement such as the
following to the append file. When the OpenEmbedded build system builds
the kernel, the configuration options are picked up and applied.
SRC_URI += "file://network_configs.cfg"
To group related configurations into multiple files, you perform a
similar procedure. Here is an example that groups separate
configurations specifically for Ethernet and graphics into their own
files and adds the configurations by using a SRC_URI
statement like
the following in your append file:
SRC_URI += "file://myconfig.cfg \
file://eth.cfg \
file://gfx.cfg"
Another variable you can use in your kernel recipe append file is the FILESEXTRAPATHS variable. When you use this statement, you are extending the locations used by the OpenEmbedded system to look for files and patches as the recipe is processed.
Note
Other methods exist to accomplish grouping and defining configuration
options. For example, if you are working with a local clone of the
kernel repository, you could checkout the kernel’s meta
branch,
make your changes, and then push the changes to the local bare clone
of the kernel. The result is that you directly add configuration
options to the meta
branch for your BSP. The configuration
options will likely end up in that location anyway if the BSP gets
added to the Yocto Project.
In general, however, the Yocto Project maintainers take care of
moving the SRC_URI
-specified configuration options to the
kernel’s meta
branch. Not only is it easier for BSP developers to
not have to worry about putting those configurations in the branch,
but having the maintainers do it allows them to apply ‘global’
knowledge about the kinds of common configuration options multiple
BSPs in the tree are typically using. This allows for promotion of
common configurations into common features.
2.3.2 Applying Patches
If you have a single patch or a small series of patches that you want to
apply to the Linux kernel source, you can do so just as you would with
any other recipe. You first copy the patches to the path added to
FILESEXTRAPATHS in
your .bbappend
file as described in the previous section, and then
reference them in SRC_URI
statements.
For example, you can apply a three-patch series by adding the following
lines to your linux-yocto .bbappend
file in your layer:
SRC_URI += "file://0001-first-change.patch"
SRC_URI += "file://0002-second-change.patch"
SRC_URI += "file://0003-third-change.patch"
The next time you run BitBake to build the Linux kernel, BitBake detects the change in the recipe and fetches and applies the patches before building the kernel.
For a detailed example showing how to patch the kernel using
devtool
, see the
“Using devtool to Patch the Kernel”
and
“Using Traditional Kernel Development to Patch the Kernel”
sections.
2.3.3 Changing the Configuration
You can make wholesale or incremental changes to the final .config
file used for the eventual Linux kernel configuration by including a
defconfig
file and by specifying configuration fragments in the
SRC_URI to be applied to that
file.
If you have a complete, working Linux kernel .config
file you want
to use for the configuration, as before, copy that file to the
appropriate ${PN}
directory in your layer’s recipes-kernel/linux
directory, and rename the copied file to “defconfig”. Then, add the
following lines to the linux-yocto .bbappend
file in your layer:
FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:"
SRC_URI += "file://defconfig"
The SRC_URI
tells the build system how to search
for the file, while the
FILESEXTRAPATHS
extends the FILESPATH
variable (search directories) to include the ${PN}
directory you
created to hold the configuration changes.
Note
The build system applies the configurations from the defconfig
file before applying any subsequent configuration fragments. The
final kernel configuration is a combination of the configurations in
the defconfig
file and any configuration fragments you provide. You need
to realize that if you have any configuration fragments, the build system
applies these on top of and after applying the existing defconfig
file
configurations.
Generally speaking, the preferred approach is to determine the
incremental change you want to make and add that as a configuration
fragment. For example, if you want to add support for a basic serial
console, create a file named 8250.cfg
in the ${PN}
directory
with the following content (without indentation):
CONFIG_SERIAL_8250=y
CONFIG_SERIAL_8250_CONSOLE=y
CONFIG_SERIAL_8250_PCI=y
CONFIG_SERIAL_8250_NR_UARTS=4
CONFIG_SERIAL_8250_RUNTIME_UARTS=4
CONFIG_SERIAL_CORE=y
CONFIG_SERIAL_CORE_CONSOLE=y
Next, include this
configuration fragment and extend the FILESPATH
variable in your
.bbappend
file:
FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:"
SRC_URI += "file://8250.cfg"
The next time you run BitBake to build the Linux kernel, BitBake detects the change in the recipe and fetches and applies the new configuration before building the kernel.
For a detailed example showing how to configure the kernel, see the “Configuring the Kernel” section.
2.3.4 Using an “In-Tree” defconfig
File
It might be desirable to have kernel configuration fragment support
through a defconfig
file that is pulled from the kernel source tree
for the configured machine. By default, the OpenEmbedded build system
looks for defconfig
files in the layer used for Metadata, which is
“out-of-tree”, and then configures them using the following:
SRC_URI += "file://defconfig"
If you do not want to maintain copies of
defconfig
files in your layer but would rather allow users to use
the default configuration from the kernel tree and still be able to add
configuration fragments to the
SRC_URI through, for example,
append files, you can direct the OpenEmbedded build system to use a
defconfig
file that is “in-tree”.
To specify an “in-tree” defconfig
file, use the following statement
form:
KBUILD_DEFCONFIG_KMACHINE ?= "defconfig_file"
Here is an example
that assigns the KBUILD_DEFCONFIG
variable based on “raspberrypi2”
and provides the path to the “in-tree” defconfig
file to be used for
a Raspberry Pi 2, which is based on the Broadcom 2708/2709 chipset:
KBUILD_DEFCONFIG_raspberrypi2 ?= "bcm2709_defconfig"
Aside from modifying your kernel recipe and providing your own
defconfig
file, you need to be sure no files or statements set
SRC_URI
to use a defconfig
other than your “in-tree” file (e.g.
a kernel’s linux-
machine.inc
file). In other words, if the
build system detects a statement that identifies an “out-of-tree”
defconfig
file, that statement will override your
KBUILD_DEFCONFIG
variable.
See the KBUILD_DEFCONFIG variable description for more information.
2.4 Using devtool
to Patch the Kernel
The steps in this procedure show you how you can patch the kernel using
the extensible SDK and devtool
.
Note
Before attempting this procedure, be sure you have performed the steps to get ready for updating the kernel as described in the “Getting Ready to Develop Using devtool” section.
Patching the kernel involves changing or adding configurations to an existing kernel, changing or adding recipes to the kernel that are needed to support specific hardware features, or even altering the source code itself.
This example creates a simple patch by adding some QEMU emulator console
output at boot time through printk
statements in the kernel’s
calibrate.c
source code file. Applying the patch and booting the
modified image causes the added messages to appear on the emulator’s
console. The example is a continuation of the setup procedure found in
the “Getting Ready to Develop Using devtool” Section.
Check Out the Kernel Source Files: First you must use
devtool
to checkout the kernel source code in its workspace. Be sure you are in the terminal set up to do work with the extensible SDK.Note
See this step in the “Getting Ready to Develop Using devtool” section for more information.
Use the following
devtool
command to check out the code:$ devtool modify linux-yocto
Note
During the checkout operation, a bug exists that could cause errors such as the following to appear:
ERROR: Taskhash mismatch 2c793438c2d9f8c3681fd5f7bc819efa versus be3a89ce7c47178880ba7bf6293d7404 for /path/to/esdk/layers/poky/meta/recipes-kernel/linux/linux-yocto_4.10.bb.do_unpack
You can safely ignore these messages. The source code is correctly checked out.
Edit the Source Files Follow these steps to make some simple changes to the source files:
Change the working directory: In the previous step, the output noted where you can find the source files (e.g.
poky_sdk/workspace/sources/linux-yocto
). Change to where the kernel source code is before making your edits to thecalibrate.c
file:$ cd poky_sdk/workspace/sources/linux-yocto
Edit the source file: Edit the
init/calibrate.c
file to have the following changes:void calibrate_delay(void) { unsigned long lpj; static bool printed; int this_cpu = smp_processor_id(); printk("*************************************\n"); printk("* *\n"); printk("* HELLO YOCTO KERNEL *\n"); printk("* *\n"); printk("*************************************\n"); if (per_cpu(cpu_loops_per_jiffy, this_cpu)) { . . .
Build the Updated Kernel Source: To build the updated kernel source, use
devtool
:$ devtool build linux-yocto
Create the Image With the New Kernel: Use the
devtool build-image
command to create a new image that has the new kernel.Note
If the image you originally created resulted in a Wic file, you can use an alternate method to create the new image with the updated kernel. For an example, see the steps in the TipsAndTricks/KernelDevelopmentWithEsdk Wiki Page.
$ cd ~ $ devtool build-image core-image-minimal
Test the New Image: For this example, you can run the new image using QEMU to verify your changes:
Boot the image: Boot the modified image in the QEMU emulator using this command:
$ runqemu qemux86
Verify the changes: Log into the machine using
root
with no password and then use the following shell command to scroll through the console’s boot output.# dmesg | less
You should see the results of your
printk
statements as part of the output when you scroll down the console window.
Stage and commit your changes: Within your eSDK terminal, change your working directory to where you modified the
calibrate.c
file and use these Git commands to stage and commit your changes:$ cd poky_sdk/workspace/sources/linux-yocto $ git status $ git add init/calibrate.c $ git commit -m "calibrate: Add printk example"
Export the Patches and Create an Append File: To export your commits as patches and create a
.bbappend
file, use the following command in the terminal used to work with the extensible SDK. This example uses the previously established layer namedmeta-mylayer
.$ devtool finish linux-yocto ~/meta-mylayer
Note
See Step 3 of the “Getting Ready to Develop Using devtool” section for information on setting up this layer.
Once the command finishes, the patches and the
.bbappend
file are located in the~/meta-mylayer/recipes-kernel/linux
directory.Build the Image With Your Modified Kernel: You can now build an image that includes your kernel patches. Execute the following command from your Build Directory in the terminal set up to run BitBake:
$ cd poky/build $ bitbake core-image-minimal
2.5 Using Traditional Kernel Development to Patch the Kernel
The steps in this procedure show you how you can patch the kernel using
traditional kernel development (i.e. not using devtool
and the
extensible SDK as described in the
“Using devtool to Patch the Kernel”
section).
Note
Before attempting this procedure, be sure you have performed the steps to get ready for updating the kernel as described in the “Getting Ready for Traditional Kernel Development” section.
Patching the kernel involves changing or adding configurations to an existing kernel, changing or adding recipes to the kernel that are needed to support specific hardware features, or even altering the source code itself.
The example in this section creates a simple patch by adding some QEMU
emulator console output at boot time through printk
statements in
the kernel’s calibrate.c
source code file. Applying the patch and
booting the modified image causes the added messages to appear on the
emulator’s console. The example is a continuation of the setup procedure
found in the
“Getting Ready for Traditional Kernel Development”
Section.
Edit the Source Files Prior to this step, you should have used Git to create a local copy of the repository for your kernel. Assuming you created the repository as directed in the “Getting Ready for Traditional Kernel Development” section, use the following commands to edit the
calibrate.c
file:Change the working directory: You need to locate the source files in the local copy of the kernel Git repository. Change to where the kernel source code is before making your edits to the
calibrate.c
file:$ cd ~/linux-yocto-4.12/init
Edit the source file: Edit the
calibrate.c
file to have the following changes:void calibrate_delay(void) { unsigned long lpj; static bool printed; int this_cpu = smp_processor_id(); printk("*************************************\n"); printk("* *\n"); printk("* HELLO YOCTO KERNEL *\n"); printk("* *\n"); printk("*************************************\n"); if (per_cpu(cpu_loops_per_jiffy, this_cpu)) { . . .
Stage and Commit Your Changes: Use standard Git commands to stage and commit the changes you just made:
$ git add calibrate.c $ git commit -m "calibrate.c - Added some printk statements"
If you do not stage and commit your changes, the OpenEmbedded Build System will not pick up the changes.
Update Your local.conf File to Point to Your Source Files: In addition to your
local.conf
file specifying to use “kernel-modules” and the “qemux86” machine, it must also point to the updated kernel source files. Add SRC_URI and SRCREV statements similar to the following to yourlocal.conf
:$ cd poky/build/conf
Add the following to the
local.conf
:SRC_URI_pn-linux-yocto = "git:///path-to/linux-yocto-4.12;protocol=file;name=machine;branch=standard/base; \ git:///path-to/yocto-kernel-cache;protocol=file;type=kmeta;name=meta;branch=yocto-4.12;destsuffix=${KMETA}" SRCREV_meta_qemux86 = "${AUTOREV}" SRCREV_machine_qemux86 = "${AUTOREV}"
Note
Be sure to replace path-to with the pathname to your local Git repositories. Also, you must be sure to specify the correct branch and machine types. For this example, the branch is
standard/base
and the machine isqemux86
.Build the Image: With the source modified, your changes staged and committed, and the
local.conf
file pointing to the kernel files, you can now use BitBake to build the image:$ cd poky/build $ bitbake core-image-minimal
Boot the image: Boot the modified image in the QEMU emulator using this command. When prompted to login to the QEMU console, use “root” with no password:
$ cd poky/build $ runqemu qemux86
Look for Your Changes: As QEMU booted, you might have seen your changes rapidly scroll by. If not, use these commands to see your changes:
# dmesg | less
You should see the results of your
printk
statements as part of the output when you scroll down the console window.Generate the Patch File: Once you are sure that your patch works correctly, you can generate a
*.patch
file in the kernel source repository:$ cd ~/linux-yocto-4.12/init $ git format-patch -1 0001-calibrate.c-Added-some-printk-statements.patch
Move the Patch File to Your Layer: In order for subsequent builds to pick up patches, you need to move the patch file you created in the previous step to your layer
meta-mylayer
. For this example, the layer created earlier is located in your home directory asmeta-mylayer
. When the layer was created using theyocto-create
script, no additional hierarchy was created to support patches. Before moving the patch file, you need to add additional structure to your layer using the following commands:$ cd ~/meta-mylayer $ mkdir recipes-kernel $ mkdir recipes-kernel/linux $ mkdir recipes-kernel/linux/linux-yocto
Once you have created this hierarchy in your layer, you can move the patch file using the following command:
$ mv ~/linux-yocto-4.12/init/0001-calibrate.c-Added-some-printk-statements.patch ~/meta-mylayer/recipes-kernel/linux/linux-yocto
Create the Append File: Finally, you need to create the
linux-yocto_4.12.bbappend
file and insert statements that allow the OpenEmbedded build system to find the patch. The append file needs to be in your layer’srecipes-kernel/linux
directory and it must be namedlinux-yocto_4.12.bbappend
and have the following contents:FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:" SRC_URI_append = "file://0001-calibrate.c-Added-some-printk-statements.patch"
The FILESEXTRAPATHS and SRC_URI statements enable the OpenEmbedded build system to find the patch file.
For more information on append files and patches, see the “Creating the Append File” and “Applying Patches” sections. You can also see the “Using .bbappend Files in Your Layer” section in the Yocto Project Development Tasks Manual.
Note
To build
core-image-minimal
again and see the effects of your patch, you can essentially eliminate the temporary source files saved inpoky/build/tmp/work/...
and residual effects of the build by entering the following sequence of commands:$ cd poky/build $ bitbake -c cleanall yocto-linux $ bitbake core-image-minimal -c cleanall $ bitbake core-image-minimal $ runqemu qemux86
2.6 Configuring the Kernel
Configuring the Yocto Project kernel consists of making sure the
.config
file has all the right information in it for the image you
are building. You can use the menuconfig
tool and configuration
fragments to make sure your .config
file is just how you need it.
You can also save known configurations in a defconfig
file that the
build system can use for kernel configuration.
This section describes how to use menuconfig
, create and use
configuration fragments, and how to interactively modify your
.config
file to create the leanest kernel configuration file
possible.
For more information on kernel configuration, see the “Changing the Configuration” section.
2.6.2 Creating a defconfig
File
A defconfig
file in the context of the Yocto Project is often a
.config
file that is copied from a build or a defconfig
taken
from the kernel tree and moved into recipe space. You can use a
defconfig
file to retain a known set of kernel configurations from
which the OpenEmbedded build system can draw to create the final
.config
file.
Note
Out-of-the-box, the Yocto Project never ships a defconfig
or .config
file. The OpenEmbedded build system creates the final .config
file used
to configure the kernel.
To create a defconfig
, start with a complete, working Linux kernel
.config
file. Copy that file to the appropriate
${
PN}
directory in
your layer’s recipes-kernel/linux
directory, and rename the copied
file to “defconfig” (e.g.
~/meta-mylayer/recipes-kernel/linux/linux-yocto/defconfig
). Then,
add the following lines to the linux-yocto .bbappend
file in your
layer:
FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:"
SRC_URI += "file://defconfig"
The SRC_URI tells the build system how to search for the file, while the
FILESEXTRAPATHS extends the FILESPATH
variable (search directories) to include the ${PN}
directory you
created to hold the configuration changes.
Note
The build system applies the configurations from the defconfig
file before applying any subsequent configuration fragments. The
final kernel configuration is a combination of the configurations in
the defconfig
file and any configuration fragments you provide. You need
to realize that if you have any configuration fragments, the build system
applies these on top of and after applying the existing defconfig
file
configurations.
For more information on configuring the kernel, see the “Changing the Configuration” section.
2.6.3 Creating Configuration Fragments
Configuration fragments are simply kernel options that appear in a file
placed where the OpenEmbedded build system can find and apply them. The
build system applies configuration fragments after applying
configurations from a defconfig
file. Thus, the final kernel
configuration is a combination of the configurations in the
defconfig
file and then any configuration fragments you provide. The
build system applies fragments on top of and after applying the existing
defconfig file configurations.
Syntactically, the configuration statement is identical to what would
appear in the .config
file, which is in the Build Directory.
Note
For more information about where the .config
file is located, see the
example in the
“Using menuconfig”
section.
It is simple to create a configuration fragment. One method is to use
shell commands. For example, issuing the following from the shell
creates a configuration fragment file named my_smp.cfg
that enables
multi-processor support within the kernel:
$ echo "CONFIG_SMP=y" >> my_smp.cfg
Note
All configuration fragment files must use the .cfg
extension in order
for the OpenEmbedded build system to recognize them as a configuration
fragment.
Another method is to create a configuration fragment using the
differences between two configuration files: one previously created and
saved, and one freshly created using the menuconfig
tool.
To create a configuration fragment using this method, follow these steps:
Complete a Build Through Kernel Configuration: Complete a build at least through the kernel configuration task as follows:
$ bitbake linux-yocto -c kernel_configme -f
This step ensures that you create a
.config
file from a known state. Because situations exist where your build state might become unknown, it is best to run this task prior to startingmenuconfig
.Launch menuconfig: Run the
menuconfig
command:$ bitbake linux-yocto -c menuconfig
Create the Configuration Fragment: Run the
diffconfig
command to prepare a configuration fragment. The resulting filefragment.cfg
is placed in the${
WORKDIR}
directory:$ bitbake linux-yocto -c diffconfig
The diffconfig
command creates a file that is a list of Linux kernel
CONFIG_
assignments. See the
“Changing the Configuration” section for additional
information on how to use the output as a configuration fragment.
Note
You can also use this method to create configuration fragments for a BSP. See the “BSP Descriptions” section for more information.
Where do you put your configuration fragment files? You can place these
files in an area pointed to by
SRC_URI as directed by your
bblayers.conf
file, which is located in your layer. The OpenEmbedded
build system picks up the configuration and adds it to the kernel’s
configuration. For example, suppose you had a set of configuration
options in a file called myconfig.cfg
. If you put that file inside a
directory named linux-yocto
that resides in the same directory as
the kernel’s append file within your layer and then add the following
statements to the kernel’s append file, those configuration options will
be picked up and applied when the kernel is built:
FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:"
SRC_URI += "file://myconfig.cfg"
As mentioned earlier, you can group related configurations into multiple
files and name them all in the SRC_URI
statement as well. For
example, you could group separate configurations specifically for
Ethernet and graphics into their own files and add those by using a
SRC_URI
statement like the following in your append file:
SRC_URI += "file://myconfig.cfg \
file://eth.cfg \
file://gfx.cfg"
2.6.4 Validating Configuration
You can use the do_kernel_configcheck task to provide configuration validation:
$ bitbake linux-yocto -c kernel_configcheck -f
Running this task produces warnings for when a
requested configuration does not appear in the final .config
file or
when you override a policy configuration in a hardware configuration
fragment.
In order to run this task, you must have an existing .config
file.
See the “Using menuconfig” section for
information on how to create a configuration file.
Following is sample output from the do_kernel_configcheck
task:
Loading cache: 100% |########################################################| Time: 0:00:00
Loaded 1275 entries from dependency cache.
NOTE: Resolving any missing task queue dependencies
Build Configuration:
.
.
.
NOTE: Executing SetScene Tasks
NOTE: Executing RunQueue Tasks
WARNING: linux-yocto-4.12.12+gitAUTOINC+eda4d18ce4_16de014967-r0 do_kernel_configcheck:
[kernel config]: specified values did not make it into the kernel's final configuration:
---------- CONFIG_X86_TSC -----------------
Config: CONFIG_X86_TSC
From: /home/scottrif/poky/build/tmp/work-shared/qemux86/kernel-source/.kernel-meta/configs/standard/bsp/common-pc/common-pc-cpu.cfg
Requested value: CONFIG_X86_TSC=y
Actual value:
---------- CONFIG_X86_BIGSMP -----------------
Config: CONFIG_X86_BIGSMP
From: /home/scottrif/poky/build/tmp/work-shared/qemux86/kernel-source/.kernel-meta/configs/standard/cfg/smp.cfg
/home/scottrif/poky/build/tmp/work-shared/qemux86/kernel-source/.kernel-meta/configs/standard/defconfig
Requested value: # CONFIG_X86_BIGSMP is not set
Actual value:
---------- CONFIG_NR_CPUS -----------------
Config: CONFIG_NR_CPUS
From: /home/scottrif/poky/build/tmp/work-shared/qemux86/kernel-source/.kernel-meta/configs/standard/cfg/smp.cfg
/home/scottrif/poky/build/tmp/work-shared/qemux86/kernel-source/.kernel-meta/configs/standard/bsp/common-pc/common-pc.cfg
/home/scottrif/poky/build/tmp/work-shared/qemux86/kernel-source/.kernel-meta/configs/standard/defconfig
Requested value: CONFIG_NR_CPUS=8
Actual value: CONFIG_NR_CPUS=1
---------- CONFIG_SCHED_SMT -----------------
Config: CONFIG_SCHED_SMT
From: /home/scottrif/poky/build/tmp/work-shared/qemux86/kernel-source/.kernel-meta/configs/standard/cfg/smp.cfg
/home/scottrif/poky/build/tmp/work-shared/qemux86/kernel-source/.kernel-meta/configs/standard/defconfig
Requested value: CONFIG_SCHED_SMT=y
Actual value:
NOTE: Tasks Summary: Attempted 288 tasks of which 285 didn't need to be rerun and all succeeded.
Summary: There were 3 WARNING messages shown.
Note
The previous output example has artificial line breaks to make it more readable.
The output describes the various problems that you can encounter along with where to find the offending configuration items. You can use the information in the logs to adjust your configuration files and then repeat the do_kernel_configme and do_kernel_configcheck tasks until they produce no warnings.
For more information on how to use the menuconfig
tool, see the
Using menuconfig section.
2.6.5 Fine-Tuning the Kernel Configuration File
You can make sure the .config
file is as lean or efficient as
possible by reading the output of the kernel configuration fragment
audit, noting any issues, making changes to correct the issues, and then
repeating.
As part of the kernel build process, the do_kernel_configcheck
task
runs. This task validates the kernel configuration by checking the final
.config
file against the input files. During the check, the task
produces warning messages for the following issues:
Requested options that did not make the final
.config
file.Configuration items that appear twice in the same configuration fragment.
Configuration items tagged as “required” that were overridden.
A board overrides a non-board specific option.
Listed options not valid for the kernel being processed. In other words, the option does not appear anywhere.
Note
The do_kernel_configcheck task can also optionally report if an option is overridden during processing.
For each output warning, a message points to the file that contains a list of the options and a pointer to the configuration fragment that defines them. Collectively, the files are the key to streamlining the configuration.
To streamline the configuration, do the following:
Use a Working Configuration: Start with a full configuration that you know works. Be sure the configuration builds and boots successfully. Use this configuration file as your baseline.
Run Configure and Check Tasks: Separately run the
do_kernel_configme
anddo_kernel_configcheck
tasks:$ bitbake linux-yocto -c kernel_configme -f $ bitbake linux-yocto -c kernel_configcheck -f
Process the Results: Take the resulting list of files from the
do_kernel_configcheck
task warnings and do the following:Drop values that are redefined in the fragment but do not change the final
.config
file.Analyze and potentially drop values from the
.config
file that override required configurations.Analyze and potentially remove non-board specific options.
Remove repeated and invalid options.
Re-Run Configure and Check Tasks: After you have worked through the output of the kernel configuration audit, you can re-run the
do_kernel_configme
anddo_kernel_configcheck
tasks to see the results of your changes. If you have more issues, you can deal with them as described in the previous step.
Iteratively working through steps two through four eventually yields a
minimal, streamlined configuration file. Once you have the best
.config
, you can build the Linux Yocto kernel.
2.7 Expanding Variables
Sometimes it is helpful to determine what a variable expands to during a
build. You can examine the values of variables by examining the
output of the bitbake -e
command. The output is long and is more
easily managed in a text file, which allows for easy searches:
$ bitbake -e virtual/kernel > some_text_file
Within the text file, you can see exactly how each variable is expanded and used by the OpenEmbedded build system.
2.8 Working with a “Dirty” Kernel Version String
If you build a kernel image and the version string has a “+” or a “-dirty” at the end, uncommitted modifications exist in the kernel’s source directory. Follow these steps to clean up the version string:
Discover the Uncommitted Changes: Go to the kernel’s locally cloned Git repository (source directory) and use the following Git command to list the files that have been changed, added, or removed:
$ git status
Commit the Changes: You should commit those changes to the kernel source tree regardless of whether or not you will save, export, or use the changes:
$ git add $ git commit -s -a -m "getting rid of -dirty"
Rebuild the Kernel Image: Once you commit the changes, rebuild the kernel.
Depending on your particular kernel development workflow, the commands you use to rebuild the kernel might differ. For information on building the kernel image when using
devtool
, see the “Using devtool to Patch the Kernel” section. For information on building the kernel image when using Bitbake, see the “Using Traditional Kernel Development to Patch the Kernel” section.
2.9 Working With Your Own Sources
If you cannot work with one of the Linux kernel versions supported by existing linux-yocto recipes, you can still make use of the Yocto Project Linux kernel tooling by working with your own sources. When you use your own sources, you will not be able to leverage the existing kernel Metadata and stabilization work of the linux-yocto sources. However, you will be able to manage your own Metadata in the same format as the linux-yocto sources. Maintaining format compatibility facilitates converging with linux-yocto on a future, mutually-supported kernel version.
To help you use your own sources, the Yocto Project provides a
linux-yocto custom recipe (linux-yocto-custom.bb
) that uses
kernel.org
sources and the Yocto Project Linux kernel tools for
managing kernel Metadata. You can find this recipe in the poky
Git
repository of the Yocto Project Source Repository
at:
poky/meta-skeleton/recipes-kernel/linux/linux-yocto-custom.bb
Here are some basic steps you can use to work with your own sources:
Create a Copy of the Kernel Recipe: Copy the
linux-yocto-custom.bb
recipe to your layer and give it a meaningful name. The name should include the version of the Yocto Linux kernel you are using (e.g.linux-yocto-myproject_4.12.bb
, where “4.12” is the base version of the Linux kernel with which you would be working).Create a Directory for Your Patches: In the same directory inside your layer, create a matching directory to store your patches and configuration files (e.g.
linux-yocto-myproject
).Ensure You Have Configurations: Make sure you have either a
defconfig
file or configuration fragment files in your layer. When you use thelinux-yocto-custom.bb
recipe, you must specify a configuration. If you do not have adefconfig
file, you can run the following:$ make defconfig
After running the command, copy the resulting
.config
file to thefiles
directory in your layer as “defconfig” and then add it to the SRC_URI variable in the recipe.Running the
make defconfig
command results in the default configuration for your architecture as defined by your kernel. However, no guarantee exists that this configuration is valid for your use case, or that your board will even boot. This is particularly true for non-x86 architectures.To use non-x86
defconfig
files, you need to be more specific and find one that matches your board (i.e. for arm, you look inarch/arm/configs
and use the one that is the best starting point for your board).Edit the Recipe: Edit the following variables in your recipe as appropriate for your project:
SRC_URI: The
SRC_URI
should specify a Git repository that uses one of the supported Git fetcher protocols (i.e.file
,git
,http
, and so forth). TheSRC_URI
variable should also specify either adefconfig
file or some configuration fragment files. The skeleton recipe provides an exampleSRC_URI
as a syntax reference.LINUX_VERSION: The Linux kernel version you are using (e.g. “4.12”).
LINUX_VERSION_EXTENSION: The Linux kernel
CONFIG_LOCALVERSION
that is compiled into the resulting kernel and visible through theuname
command.SRCREV: The commit ID from which you want to build.
PR: Treat this variable the same as you would in any other recipe. Increment the variable to indicate to the OpenEmbedded build system that the recipe has changed.
PV: The default
PV
assignment is typically adequate. It combines theLINUX_VERSION
with the Source Control Manager (SCM) revision as derived from the SRCPV variable. The combined results are a string with the following form:3.19.11+git1+68a635bf8dfb64b02263c1ac80c948647cc76d5f_1+218bd8d2022b9852c60d32f0d770931e3cf343e2
While lengthy, the extra verbosity in
PV
helps ensure you are using the exact sources from which you intend to build.COMPATIBLE_MACHINE: A list of the machines supported by your new recipe. This variable in the example recipe is set by default to a regular expression that matches only the empty string, “(^$)”. This default setting triggers an explicit build failure. You must change it to match a list of the machines that your new recipe supports. For example, to support the
qemux86
andqemux86-64
machines, use the following form:COMPATIBLE_MACHINE = "qemux86|qemux86-64"
Customize Your Recipe as Needed: Provide further customizations to your recipe as needed just as you would customize an existing linux-yocto recipe. See the “Modifying an Existing Recipe” section for information.
2.10 Working with Out-of-Tree Modules
This section describes steps to build out-of-tree modules on your target and describes how to incorporate out-of-tree modules in the build.
2.10.1 Building Out-of-Tree Modules on the Target
While the traditional Yocto Project development model would be to include kernel modules as part of the normal build process, you might find it useful to build modules on the target. This could be the case if your target system is capable and powerful enough to handle the necessary compilation. Before deciding to build on your target, however, you should consider the benefits of using a proper cross-development environment from your build host.
If you want to be able to build out-of-tree modules on the target, there
are some steps you need to take on the target that is running your SDK
image. Briefly, the kernel-dev
package is installed by default on
all *.sdk
images and the kernel-devsrc
package is installed on
many of the *.sdk
images. However, you need to create some scripts
prior to attempting to build the out-of-tree modules on the target that
is running that image.
Prior to attempting to build the out-of-tree modules, you need to be on
the target as root and you need to change to the /usr/src/kernel
directory. Next, make
the scripts:
# cd /usr/src/kernel
# make scripts
Because all SDK image recipes include dev-pkgs
, the
kernel-dev
packages will be installed as part of the SDK image and
the kernel-devsrc
packages will be installed as part of applicable
SDK images. The SDK uses the scripts when building out-of-tree modules.
Once you have switched to that directory and created the scripts, you
should be able to build your out-of-tree modules on the target.
2.10.2 Incorporating Out-of-Tree Modules
While it is always preferable to work with sources integrated into the
Linux kernel sources, if you need an external kernel module, the
hello-mod.bb
recipe is available as a template from which you can
create your own out-of-tree Linux kernel module recipe.
This template recipe is located in the poky
Git repository of the
Yocto Project Source Repository at:
poky/meta-skeleton/recipes-kernel/hello-mod/hello-mod_0.1.bb
To get started, copy this recipe to your layer and give it a meaningful
name (e.g. mymodule_1.0.bb
). In the same directory, create a new
directory named files
where you can store any source files, patches,
or other files necessary for building the module that do not come with
the sources. Finally, update the recipe as needed for the module.
Typically, you will need to set the following variables:
Depending on the build system used by the module sources, you might need
to make some adjustments. For example, a typical module Makefile
looks much like the one provided with the hello-mod
template:
obj-m := hello.o
SRC := $(shell pwd)
all:
$(MAKE) -C $(KERNEL_SRC) M=$(SRC)
modules_install:
$(MAKE) -C $(KERNEL_SRC) M=$(SRC) modules_install
...
The important point to note here is the KERNEL_SRC variable. The
module class sets this variable and the
KERNEL_PATH variable to
${STAGING_KERNEL_DIR}
with the necessary Linux kernel build
information to build modules. If your module Makefile
uses a
different variable, you might want to override the
do_compile step, or
create a patch to the Makefile
to work with the more typical
KERNEL_SRC
or KERNEL_PATH
variables.
After you have prepared your recipe, you will likely want to include the module in your images. To do this, see the documentation for the following variables in the Yocto Project Reference Manual and set one of them appropriately for your machine configuration file:
Modules are often not required for boot and can be excluded from certain build configurations. The following allows for the most flexibility:
MACHINE_EXTRA_RRECOMMENDS += "kernel-module-mymodule"
The value is
derived by appending the module filename without the .ko
extension
to the string “kernel-module-“.
Because the variable is RRECOMMENDS and not a RDEPENDS variable, the build will not fail if this module is not available to include in the image.
2.11 Inspecting Changes and Commits
A common question when working with a kernel is: “What changes have been applied to this tree?” Rather than using “grep” across directories to see what has changed, you can use Git to inspect or search the kernel tree. Using Git is an efficient way to see what has changed in the tree.
2.11.1 What Changed in a Kernel?
Following are a few examples that show how to use Git commands to examine changes. These examples are by no means the only way to see changes.
Note
In the following examples, unless you provide a commit range, kernel.org
history is blended with Yocto Project kernel changes. You can form
ranges by using branch names from the kernel tree as the upper and
lower commit markers with the Git commands. You can see the branch
names through the web interface to the Yocto Project source
repositories at https://git.yoctoproject.org/cgit/cgit.cgi/.
To see a full range of the changes, use the git whatchanged
command
and specify a commit range for the branch (commit..
commit).
Here is an example that looks at what has changed in the emenlow
branch of the linux-yocto-3.19
kernel. The lower commit range is the
commit associated with the standard/base
branch, while the upper
commit range is the commit associated with the standard/emenlow
branch.
$ git whatchanged origin/standard/base..origin/standard/emenlow
To see short, one line summaries of changes use the git log
command:
$ git log --oneline origin/standard/base..origin/standard/emenlow
Use this command to see code differences for the changes:
$ git diff origin/standard/base..origin/standard/emenlow
Use this command to see the commit log messages and the text differences:
$ git show origin/standard/base..origin/standard/emenlow
Use this command to create individual patches for each change. Here is
an example that creates patch files for each commit and places them
in your Documents
directory:
$ git format-patch -o $HOME/Documents origin/standard/base..origin/standard/emenlow
2.11.2 Showing a Particular Feature or Branch Change
Tags in the Yocto Project kernel tree divide changes for significant
features or branches. The git show
tag command shows changes based
on a tag. Here is an example that shows systemtap
changes:
$ git show systemtap
You can use the git branch --contains
tag command to
show the branches that contain a particular feature. This command shows
the branches that contain the systemtap
feature:
$ git branch --contains systemtap
2.12 Adding Recipe-Space Kernel Features
You can add kernel features in the
recipe-space
by using the KERNEL_FEATURES
variable and by specifying the feature’s .scc
file path in the
SRC_URI statement. When you
add features using this method, the OpenEmbedded build system checks to
be sure the features are present. If the features are not present, the
build stops. Kernel features are the last elements processed for
configuring and patching the kernel. Therefore, adding features in this
manner is a way to enforce specific features are present and enabled
without needing to do a full audit of any other layer’s additions to the
SRC_URI
statement.
You add a kernel feature by providing the feature as part of the
KERNEL_FEATURES
variable and by providing the path to the feature’s
.scc
file, which is relative to the root of the kernel Metadata. The
OpenEmbedded build system searches all forms of kernel Metadata on the
SRC_URI
statement regardless of whether the Metadata is in the
“kernel-cache”, system kernel Metadata, or a recipe-space Metadata (i.e.
part of the kernel recipe). See the
“Kernel Metadata Location” section for
additional information.
When you specify the feature’s .scc
file on the SRC_URI
statement, the OpenEmbedded build system adds the directory of that
.scc
file along with all its subdirectories to the kernel feature
search path. Because subdirectories are searched, you can reference a
single .scc
file in the SRC_URI
statement to reference multiple
kernel features.
Consider the following example that adds the “test.scc” feature to the build.
Create the Feature File: Create a
.scc
file and locate it just as you would any other patch file,.cfg
file, or fetcher item you specify in theSRC_URI
statement.Note
You must add the directory of the
.scc
file to the fetcher’s search path in the same manner as you would add a.patch
file.You can create additional
.scc
files beneath the directory that contains the file you are adding. All subdirectories are searched during the build as potential feature directories.
Continuing with the example, suppose the “test.scc” feature you are adding has a
test.scc
file in the following directory:my_recipe | +-linux-yocto | +-test.cfg +-test.scc
In this example, the
linux-yocto
directory has both the featuretest.scc
file and a similarly named configuration fragment filetest.cfg
.Add the Feature File to SRC_URI: Add the
.scc
file to the recipe’sSRC_URI
statement:SRC_URI_append = " file://test.scc"
The leading space before the path is important as the path is appended to the existing path.
Specify the Feature as a Kernel Feature: Use the
KERNEL_FEATURES
statement to specify the feature as a kernel feature:KERNEL_FEATURES_append = " test.scc"
The OpenEmbedded build system processes the kernel feature when it builds the kernel.
Note
If other features are contained below “test.scc”, then their directories are relative to the directory containing the
test.scc
file.
3 Working with Advanced Metadata (yocto-kernel-cache
)
3.1 Overview
In addition to supporting configuration fragments and patches, the Yocto Project kernel tools also support rich Metadata that you can use to define complex policies and Board Support Package (BSP) support. The purpose of the Metadata and the tools that manage it is to help you manage the complexity of the configuration and sources used to support multiple BSPs and Linux kernel types.
Kernel Metadata exists in many places. One area in the
Yocto Project Source Repositories
is the yocto-kernel-cache
Git repository. You can find this repository
grouped under the “Yocto Linux Kernel” heading in the
Yocto Project Source Repositories.
Kernel development tools (“kern-tools”) exist also in the Yocto Project
Source Repositories under the “Yocto Linux Kernel” heading in the
yocto-kernel-tools
Git repository. The recipe that builds these
tools is meta/recipes-kernel/kern-tools/kern-tools-native_git.bb
in
the Source Directory (e.g.
poky
).
3.2 Using Kernel Metadata in a Recipe
As mentioned in the introduction, the Yocto Project contains kernel
Metadata, which is located in the yocto-kernel-cache
Git repository.
This Metadata defines Board Support Packages (BSPs) that correspond to
definitions in linux-yocto recipes for corresponding BSPs. A BSP
consists of an aggregation of kernel policy and enabled
hardware-specific features. The BSP can be influenced from within the
linux-yocto recipe.
Note
A Linux kernel recipe that contains kernel Metadata (e.g. inherits
from the linux-yocto.inc
file) is said to be a “linux-yocto style” recipe.
Every linux-yocto style recipe must define the
KMACHINE variable. This
variable is typically set to the same value as the MACHINE
variable,
which is used by BitBake.
However, in some cases, the variable might instead refer to the
underlying platform of the MACHINE
.
Multiple BSPs can reuse the same KMACHINE
name if they are built
using the same BSP description. Multiple Corei7-based BSPs could share
the same “intel-corei7-64” value for KMACHINE
. It is important to
realize that KMACHINE
is just for kernel mapping, while MACHINE
is the machine type within a BSP Layer. Even with this distinction,
however, these two variables can hold the same value. See the
“BSP Descriptions” section for more information.
Every linux-yocto style recipe must also indicate the Linux kernel source repository branch used to build the Linux kernel. The KBRANCH variable must be set to indicate the branch.
Note
You can use the KBRANCH
value to define an alternate branch typically
with a machine override as shown here from the meta-yocto-bsp
layer:
KBRANCH_edgerouter = "standard/edgerouter"
The linux-yocto style recipes can optionally define the following variables:
LINUX_KERNEL_TYPE
defines the kernel type to be used in assembling the configuration. If
you do not specify a LINUX_KERNEL_TYPE
, it defaults to “standard”.
Together with KMACHINE
, LINUX_KERNEL_TYPE
defines the search
arguments used by the kernel tools to find the appropriate description
within the kernel Metadata with which to build out the sources and
configuration. The linux-yocto recipes define “standard”, “tiny”, and
“preempt-rt” kernel types. See the “Kernel Types”
section for more information on kernel types.
During the build, the kern-tools search for the BSP description file
that most closely matches the KMACHINE
and LINUX_KERNEL_TYPE
variables passed in from the recipe. The tools use the first BSP
description they find that matches both variables. If the tools cannot find
a match, they issue a warning.
The tools first search for the KMACHINE
and then for the
LINUX_KERNEL_TYPE
. If the tools cannot find a partial match, they
will use the sources from the KBRANCH
and any configuration
specified in the SRC_URI.
You can use the
KERNEL_FEATURES
variable to include features (configuration fragments, patches, or both)
that are not already included by the KMACHINE
and
LINUX_KERNEL_TYPE
variable combination. For example, to include a
feature specified as “features/netfilter/netfilter.scc”, specify:
KERNEL_FEATURES += "features/netfilter/netfilter.scc"
To include a
feature called “cfg/sound.scc” just for the qemux86
machine,
specify:
KERNEL_FEATURES_append_qemux86 = " cfg/sound.scc"
The value of
the entries in KERNEL_FEATURES
are dependent on their location
within the kernel Metadata itself. The examples here are taken from the
yocto-kernel-cache
repository. Each branch of this repository
contains “features” and “cfg” subdirectories at the top-level. For more
information, see the “Kernel Metadata Syntax”
section.
3.3 Kernel Metadata Syntax
The kernel Metadata consists of three primary types of files: scc
[1] description files, configuration fragments, and patches. The
scc
files define variables and include or otherwise reference any of
the three file types. The description files are used to aggregate all
types of kernel Metadata into what ultimately describes the sources and
the configuration required to build a Linux kernel tailored to a
specific machine.
The scc
description files are used to define two fundamental types
of kernel Metadata:
Features
Board Support Packages (BSPs)
Features aggregate sources in the form of patches and configuration fragments into a modular reusable unit. You can use features to implement conceptually separate kernel Metadata descriptions such as pure configuration fragments, simple patches, complex features, and kernel types. Kernel Types define general kernel features and policy to be reused in the BSPs.
BSPs define hardware-specific features and aggregate them with kernel types to form the final description of what will be assembled and built.
While the kernel Metadata syntax does not enforce any logical separation of configuration fragments, patches, features or kernel types, best practices dictate a logical separation of these types of Metadata. The following Metadata file hierarchy is recommended:
base/
bsp/
cfg/
features/
ktypes/
patches/
The bsp
directory contains the BSP Descriptions.
The remaining directories all contain “features”. Separating bsp
from the
rest of the structure aids conceptualizing intended usage.
Use these guidelines to help place your scc
description files within
the structure:
If your file contains only configuration fragments, place the file in the
cfg
directory.If your file contains only source-code fixes, place the file in the
patches
directory.If your file encapsulates a major feature, often combining sources and configurations, place the file in
features
directory.If your file aggregates non-hardware configuration and patches in order to define a base kernel policy or major kernel type to be reused across multiple BSPs, place the file in
ktypes
directory.
These distinctions can easily become blurred - especially as out-of-tree
features slowly merge upstream over time. Also, remember that how the
description files are placed is a purely logical organization and has no
impact on the functionality of the kernel Metadata. There is no impact
because all of cfg
, features
, patches
, and ktypes
,
contain “features” as far as the kernel tools are concerned.
Paths used in kernel Metadata files are relative to base, which is either FILESEXTRAPATHS if you are creating Metadata in recipe-space, or the top level of yocto-kernel-cache if you are creating Metadata Outside the Recipe-Space.
3.3.1 Configuration
The simplest unit of kernel Metadata is the configuration-only feature.
This feature consists of one or more Linux kernel configuration
parameters in a configuration fragment file (.cfg
) and a .scc
file that describes the fragment.
As an example, consider the Symmetric Multi-Processing (SMP) fragment
used with the linux-yocto-4.12
kernel as defined outside of the
recipe space (i.e. yocto-kernel-cache
). This Metadata consists of
two files: smp.scc
and smp.cfg
. You can find these files in the
cfg
directory of the yocto-4.12
branch in the
yocto-kernel-cache
Git repository:
cfg/smp.scc:
define KFEATURE_DESCRIPTION "Enable SMP for 32 bit builds"
define KFEATURE_COMPATIBILITY all
kconf hardware smp.cfg
cfg/smp.cfg:
CONFIG_SMP=y
CONFIG_SCHED_SMT=y
# Increase default NR_CPUS from 8 to 64 so that platform with
# more than 8 processors can be all activated at boot time
CONFIG_NR_CPUS=64
# The following is needed when setting NR_CPUS to something
# greater than 8 on x86 architectures, it should be automatically
# disregarded by Kconfig when using a different arch
CONFIG_X86_BIGSMP=y
You can find general information on configuration fragment files in the “Creating Configuration Fragments” section.
Within the smp.scc
file, the
KFEATURE_DESCRIPTION
statement provides a short description of the fragment. Higher level
kernel tools use this description.
Also within the smp.scc
file, the kconf
command includes the
actual configuration fragment in an .scc
file, and the “hardware”
keyword identifies the fragment as being hardware enabling, as opposed
to general policy, which would use the “non-hardware” keyword. The
distinction is made for the benefit of the configuration validation
tools, which warn you if a hardware fragment overrides a policy set by a
non-hardware fragment.
Note
The description file can include multiple kconf
statements, one per
fragment.
As described in the “Validating Configuration” section, you can use the following BitBake command to audit your configuration:
$ bitbake linux-yocto -c kernel_configcheck -f
3.3.2 Patches
Patch descriptions are very similar to configuration fragment
descriptions, which are described in the previous section. However,
instead of a .cfg
file, these descriptions work with source patches
(i.e. .patch
files).
A typical patch includes a description file and the patch itself. As an
example, consider the build patches used with the linux-yocto-4.12
kernel as defined outside of the recipe space (i.e.
yocto-kernel-cache
). This Metadata consists of several files:
build.scc
and a set of *.patch
files. You can find these files
in the patches/build
directory of the yocto-4.12
branch in the
yocto-kernel-cache
Git repository.
The following listings show the build.scc
file and part of the
modpost-mask-trivial-warnings.patch
file:
patches/build/build.scc:
patch arm-serialize-build-targets.patch
patch powerpc-serialize-image-targets.patch
patch kbuild-exclude-meta-directory-from-distclean-processi.patch
# applied by kgit
# patch kbuild-add-meta-files-to-the-ignore-li.patch
patch modpost-mask-trivial-warnings.patch
patch menuconfig-check-lxdiaglog.sh-Allow-specification-of.patch
patches/build/modpost-mask-trivial-warnings.patch:
From bd48931bc142bdd104668f3a062a1f22600aae61 Mon Sep 17 00:00:00 2001
From: Paul Gortmaker <paul.gortmaker@windriver.com>
Date: Sun, 25 Jan 2009 17:58:09 -0500
Subject: [PATCH] modpost: mask trivial warnings
Newer HOSTCC will complain about various stdio fcns because
.
.
.
char *dump_write = NULL, *files_source = NULL;
int opt;
--
2.10.1
generated by cgit v0.10.2 at 2017-09-28 15:23:23 (GMT)
The description file can
include multiple patch statements where each statement handles a single
patch. In the example build.scc
file, five patch statements exist
for the five patches in the directory.
You can create a typical .patch
file using diff -Nurp
or
git format-patch
commands. For information on how to create patches,
see the “Using devtool to Patch the Kernel”
and “Using Traditional Kernel Development to Patch the Kernel”
sections.
3.3.3 Features
Features are complex kernel Metadata types that consist of configuration fragments, patches, and possibly other feature description files. As an example, consider the following generic listing:
features/myfeature.scc
define KFEATURE_DESCRIPTION "Enable myfeature"
patch 0001-myfeature-core.patch
patch 0002-myfeature-interface.patch
include cfg/myfeature_dependency.scc
kconf non-hardware myfeature.cfg
This example shows how the patch
and kconf
commands are used as well
as how an additional feature description file is included with the
include
command.
Typically, features are less granular than configuration fragments and
are more likely than configuration fragments and patches to be the types
of things you want to specify in the KERNEL_FEATURES
variable of the
Linux kernel recipe. See the
“Using Kernel Metadata in a Recipe” section earlier
in the manual.
3.3.4 Kernel Types
A kernel type defines a high-level kernel policy by aggregating
non-hardware configuration fragments with patches you want to use when
building a Linux kernel of a specific type (e.g. a real-time kernel).
Syntactically, kernel types are no different than features as described
in the “Features” section. The
LINUX_KERNEL_TYPE
variable in the kernel recipe selects the kernel type. For example, in
the linux-yocto_4.12.bb
kernel recipe found in
poky/meta/recipes-kernel/linux
, a
require directive
includes the poky/meta/recipes-kernel/linux/linux-yocto.inc
file,
which has the following statement that defines the default kernel type:
LINUX_KERNEL_TYPE ??= "standard"
Another example would be the real-time kernel (i.e.
linux-yocto-rt_4.12.bb
). This kernel recipe directly sets the kernel
type as follows:
LINUX_KERNEL_TYPE = "preempt-rt"
Note
You can find kernel recipes in the meta/recipes-kernel/linux
directory
of the Yocto Project Source Repositories
(e.g. poky/meta/recipes-kernel/linux/linux-yocto_4.12.bb
). See the
“Using Kernel Metadata in a Recipe”
section for more information.
Three kernel types (“standard”, “tiny”, and “preempt-rt”) are supported for Linux Yocto kernels:
“standard”: Includes the generic Linux kernel policy of the Yocto Project linux-yocto kernel recipes. This policy includes, among other things, which file systems, networking options, core kernel features, and debugging and tracing options are supported.
“preempt-rt”: Applies the
PREEMPT_RT
patches and the configuration options required to build a real-time Linux kernel. This kernel type inherits from the “standard” kernel type.“tiny”: Defines a bare minimum configuration meant to serve as a base for very small Linux kernels. The “tiny” kernel type is independent from the “standard” configuration. Although the “tiny” kernel type does not currently include any source changes, it might in the future.
For any given kernel type, the Metadata is defined by the .scc
(e.g.
standard.scc
). Here is a partial listing for the standard.scc
file, which is found in the ktypes/standard
directory of the
yocto-kernel-cache
Git repository:
# Include this kernel type fragment to get the standard features and
# configuration values.
# Note: if only the features are desired, but not the configuration
# then this should be included as:
# include ktypes/standard/standard.scc nocfg
# if no chained configuration is desired, include it as:
# include ktypes/standard/standard.scc nocfg inherit
include ktypes/base/base.scc
branch standard
kconf non-hardware standard.cfg
include features/kgdb/kgdb.scc
.
.
.
include cfg/net/ip6_nf.scc
include cfg/net/bridge.scc
include cfg/systemd.scc
include features/rfkill/rfkill.scc
As with any .scc
file, a kernel type definition can aggregate other
.scc
files with include
commands. These definitions can also
directly pull in configuration fragments and patches with the kconf
and patch
commands, respectively.
Note
It is not strictly necessary to create a kernel type .scc
file. The Board Support Package (BSP) file can implicitly define the
kernel type using a define
KTYPE myktype
line. See the
“BSP Descriptions” section for more
information.
3.3.5 BSP Descriptions
BSP descriptions (i.e. *.scc
files) combine kernel types with
hardware-specific features. The hardware-specific Metadata is typically
defined independently in the BSP layer, and then aggregated with each
supported kernel type.
Note
For BSPs supported by the Yocto Project, the BSP description files
are located in the bsp
directory of the yocto-kernel-cache
repository organized under the “Yocto Linux Kernel” heading in the
Yocto Project Source Repositories.
This section overviews the BSP description structure, the aggregation concepts, and presents a detailed example using a BSP supported by the Yocto Project (i.e. BeagleBone Board). For complete information on BSP layer file hierarchy, see the Yocto Project Board Support Package Developer’s Guide.
3.3.5.1 Description Overview
For simplicity, consider the following root BSP layer description files for the BeagleBone board. These files employ both a structure and naming convention for consistency. The naming convention for the file is as follows:
bsp_root_name-kernel_type.scc
Here are some example root layer BSP filenames for the BeagleBone Board BSP, which is supported by the Yocto Project:
beaglebone-standard.scc
beaglebone-preempt-rt.scc
Each file uses the root name (i.e “beaglebone”) BSP name followed by the kernel type.
Examine the beaglebone-standard.scc
file:
define KMACHINE beaglebone
define KTYPE standard
define KARCH arm
include ktypes/standard/standard.scc
branch beaglebone
include beaglebone.scc
# default policy for standard kernels
include features/latencytop/latencytop.scc
include features/profiling/profiling.scc
Every top-level BSP description file should define the KMACHINE, KTYPE, and KARCH variables. These variables allow the OpenEmbedded build system to identify the description as meeting the criteria set by the recipe being built. This example supports the “beaglebone” machine for the “standard” kernel and the “arm” architecture.
Be aware that a hard link between the KTYPE
variable and a kernel
type description file does not exist. Thus, if you do not have the
kernel type defined in your kernel Metadata as it is here, you only need
to ensure that the
LINUX_KERNEL_TYPE
variable in the kernel recipe and the KTYPE
variable in the BSP
description file match.
To separate your kernel policy from your hardware configuration, you
include a kernel type (ktype
), such as “standard”. In the previous
example, this is done using the following:
include ktypes/standard/standard.scc
This file aggregates all the configuration fragments, patches, and features that make up your standard kernel policy. See the “Kernel Types” section for more information.
To aggregate common configurations and features specific to the kernel for mybsp, use the following:
include mybsp.scc
You can see that in the BeagleBone example with the following:
include beaglebone.scc
For information on how to break a complete .config
file into the various
configuration fragments, see the “Creating Configuration Fragments” section.
Finally, if you have any configurations specific to the hardware that
are not in a *.scc
file, you can include them as follows:
kconf hardware mybsp-extra.cfg
The BeagleBone example does not include these
types of configurations. However, the Malta 32-bit board does
(“mti-malta32”). Here is the mti-malta32-le-standard.scc
file:
define KMACHINE mti-malta32-le
define KMACHINE qemumipsel
define KTYPE standard
define KARCH mips
include ktypes/standard/standard.scc
branch mti-malta32
include mti-malta32.scc
kconf hardware mti-malta32-le.cfg
3.3.5.2 Example
Many real-world examples are more complex. Like any other .scc
file,
BSP descriptions can aggregate features. Consider the Minnow BSP
definition given the linux-yocto-4.4
branch of the
yocto-kernel-cache
(i.e.
yocto-kernel-cache/bsp/minnow/minnow.scc
):
Note
Although the Minnow Board BSP is unused, the Metadata remains and is being used here just as an example.
include cfg/x86.scc
include features/eg20t/eg20t.scc
include cfg/dmaengine.scc
include features/power/intel.scc
include cfg/efi.scc
include features/usb/ehci-hcd.scc
include features/usb/ohci-hcd.scc
include features/usb/usb-gadgets.scc
include features/usb/touchscreen-composite.scc
include cfg/timer/hpet.scc
include features/leds/leds.scc
include features/spi/spidev.scc
include features/i2c/i2cdev.scc
include features/mei/mei-txe.scc
# Earlyprintk and port debug requires 8250
kconf hardware cfg/8250.cfg
kconf hardware minnow.cfg
kconf hardware minnow-dev.cfg
The minnow.scc
description file includes a hardware configuration
fragment (minnow.cfg
) specific to the Minnow BSP as well as several
more general configuration fragments and features enabling hardware
found on the machine. This minnow.scc
description file is then
included in each of the three “minnow” description files for the
supported kernel types (i.e. “standard”, “preempt-rt”, and “tiny”).
Consider the “minnow” description for the “standard” kernel type (i.e.
minnow-standard.scc
):
define KMACHINE minnow
define KTYPE standard
define KARCH i386
include ktypes/standard
include minnow.scc
# Extra minnow configs above the minimal defined in minnow.scc
include cfg/efi-ext.scc
include features/media/media-all.scc
include features/sound/snd_hda_intel.scc
# The following should really be in standard.scc
# USB live-image support
include cfg/usb-mass-storage.scc
include cfg/boot-live.scc
# Basic profiling
include features/latencytop/latencytop.scc
include features/profiling/profiling.scc
# Requested drivers that don't have an existing scc
kconf hardware minnow-drivers-extra.cfg
The include
command midway through the file includes the minnow.scc
description
that defines all enabled hardware for the BSP that is common to all
kernel types. Using this command significantly reduces duplication.
Now consider the “minnow” description for the “tiny” kernel type (i.e.
minnow-tiny.scc
):
define KMACHINE minnow
define KTYPE tiny
define KARCH i386
include ktypes/tiny
include minnow.scc
As you might expect, the “tiny” description includes quite a bit less. In fact, it includes only the minimal policy defined by the “tiny” kernel type and the hardware-specific configuration required for booting the machine along with the most basic functionality of the system as defined in the base “minnow” description file.
Notice again the three critical variables:
KMACHINE,
KTYPE, and
KARCH. Of these variables, only
KTYPE
has changed to specify the “tiny” kernel type.
3.4 Kernel Metadata Location
Kernel Metadata always exists outside of the kernel tree either defined in a kernel recipe (recipe-space) or outside of the recipe. Where you choose to define the Metadata depends on what you want to do and how you intend to work. Regardless of where you define the kernel Metadata, the syntax used applies equally.
If you are unfamiliar with the Linux kernel and only wish to apply a configuration and possibly a couple of patches provided to you by others, the recipe-space method is recommended. This method is also a good approach if you are working with Linux kernel sources you do not control or if you just do not want to maintain a Linux kernel Git repository on your own. For partial information on how you can define kernel Metadata in the recipe-space, see the “Modifying an Existing Recipe” section.
Conversely, if you are actively developing a kernel and are already maintaining a Linux kernel Git repository of your own, you might find it more convenient to work with kernel Metadata kept outside the recipe-space. Working with Metadata in this area can make iterative development of the Linux kernel more efficient outside of the BitBake environment.
3.4.1 Recipe-Space Metadata
When stored in recipe-space, the kernel Metadata files reside in a
directory hierarchy below
FILESEXTRAPATHS. For
a linux-yocto recipe or for a Linux kernel recipe derived by copying and
modifying
oe-core/meta-skeleton/recipes-kernel/linux/linux-yocto-custom.bb
to
a recipe in your layer, FILESEXTRAPATHS
is typically set to
${
THISDIR}/${
PN}
.
See the “Modifying an Existing Recipe”
section for more information.
Here is an example that shows a trivial tree of kernel Metadata stored in recipe-space within a BSP layer:
meta-my_bsp_layer/
`-- recipes-kernel
`-- linux
`-- linux-yocto
|-- bsp-standard.scc
|-- bsp.cfg
`-- standard.cfg
When the Metadata is stored in recipe-space, you must take steps to
ensure BitBake has the necessary information to decide what files to
fetch and when they need to be fetched again. It is only necessary to
specify the .scc
files on the
SRC_URI. BitBake parses them
and fetches any files referenced in the .scc
files by the
include
, patch
, or kconf
commands. Because of this, it is
necessary to bump the recipe PR
value when changing the content of files not explicitly listed in the
SRC_URI
.
If the BSP description is in recipe space, you cannot simply list the
*.scc
in the SRC_URI
statement. You need to use the following
form from your kernel append file:
SRC_URI_append_myplatform = " \
file://myplatform;type=kmeta;destsuffix=myplatform \
"
3.4.2 Metadata Outside the Recipe-Space
When stored outside of the recipe-space, the kernel Metadata files
reside in a separate repository. The OpenEmbedded build system adds the
Metadata to the build as a “type=kmeta” repository through the
SRC_URI variable. As an
example, consider the following SRC_URI
statement from the
linux-yocto_4.12.bb
kernel recipe:
SRC_URI = "git://git.yoctoproject.org/linux-yocto-4.12.git;name=machine;branch=${KBRANCH}; \
git://git.yoctoproject.org/yocto-kernel-cache;type=kmeta;name=meta;branch=yocto-4.12;destsuffix=${KMETA}"
${KMETA}
, in this context, is simply used to name the directory into
which the Git fetcher places the Metadata. This behavior is no different
than any multi-repository SRC_URI
statement used in a recipe (e.g.
see the previous section).
You can keep kernel Metadata in a “kernel-cache”, which is a directory
containing configuration fragments. As with any Metadata kept outside
the recipe-space, you simply need to use the SRC_URI
statement with
the “type=kmeta” attribute. Doing so makes the kernel Metadata available
during the configuration phase.
If you modify the Metadata, you must not forget to update the SRCREV
statements in the kernel’s recipe. In particular, you need to update the
SRCREV_meta
variable to match the commit in the KMETA
branch you
wish to use. Changing the data in these branches and not updating the
SRCREV
statements to match will cause the build to fetch an older
commit.
3.5 Organizing Your Source
Many recipes based on the linux-yocto-custom.bb
recipe use Linux
kernel sources that have only a single branch - “master”. This type of
repository structure is fine for linear development supporting a single
machine and architecture. However, if you work with multiple boards and
architectures, a kernel source repository with multiple branches is more
efficient. For example, suppose you need a series of patches for one
board to boot. Sometimes, these patches are works-in-progress or
fundamentally wrong, yet they are still necessary for specific boards.
In these situations, you most likely do not want to include these
patches in every kernel you build (i.e. have the patches as part of the
lone “master” branch). It is situations like these that give rise to
multiple branches used within a Linux kernel sources Git repository.
Repository organization strategies exist that maximize source reuse, remove redundancy, and logically order your changes. This section presents strategies for the following cases:
Encapsulating patches in a feature description and only including the patches in the BSP descriptions of the applicable boards.
Creating a machine branch in your kernel source repository and applying the patches on that branch only.
Creating a feature branch in your kernel source repository and merging that branch into your BSP when needed.
The approach you take is entirely up to you and depends on what works best for your development model.
3.5.1 Encapsulating Patches
If you are reusing patches from an external tree and are not working on the patches, you might find the encapsulated feature to be appropriate. Given this scenario, you do not need to create any branches in the source repository. Rather, you just take the static patches you need and encapsulate them within a feature description. Once you have the feature description, you simply include that into the BSP description as described in the “BSP Descriptions” section.
You can find information on how to create patches and BSP descriptions in the “Patches” and “BSP Descriptions” sections.
3.5.2 Machine Branches
When you have multiple machines and architectures to support, or you are actively working on board support, it is more efficient to create branches in the repository based on individual machines. Having machine branches allows common source to remain in the “master” branch with any features specific to a machine stored in the appropriate machine branch. This organization method frees you from continually reintegrating your patches into a feature.
Once you have a new branch, you can set up your kernel Metadata to use
the branch a couple different ways. In the recipe, you can specify the
new branch as the KBRANCH
to use for the board as follows:
KBRANCH = "mynewbranch"
Another method is to use the branch
command in the BSP
description:
mybsp.scc:
define KMACHINE mybsp
define KTYPE standard
define KARCH i386
include standard.scc
branch mynewbranch
include mybsp-hw.scc
If you find yourself with numerous branches, you might consider using a hierarchical branching system similar to what the Yocto Linux Kernel Git repositories use:
common/kernel_type/machine
If you had two kernel types, “standard” and “small” for instance, three
machines, and common as mydir
, the branches in your Git repository
might look like this:
mydir/base
mydir/standard/base
mydir/standard/machine_a
mydir/standard/machine_b
mydir/standard/machine_c
mydir/small/base
mydir/small/machine_a
This organization can help clarify the branch relationships. In this
case, mydir/standard/machine_a
includes everything in mydir/base
and mydir/standard/base
. The “standard” and “small” branches add
sources specific to those kernel types that for whatever reason are not
appropriate for the other branches.
Note
The “base” branches are an artifact of the way Git manages its data
internally on the filesystem: Git will not allow you to use
mydir/standard
and mydir/standard/machine_a
because it would have to
create a file and a directory named “standard”.
3.5.3 Feature Branches
When you are actively developing new features, it can be more efficient
to work with that feature as a branch, rather than as a set of patches
that have to be regularly updated. The Yocto Project Linux kernel tools
provide for this with the git merge
command.
To merge a feature branch into a BSP, insert the git merge
command
after any branch
commands:
mybsp.scc:
define KMACHINE mybsp
define KTYPE standard
define KARCH i386
include standard.scc
branch mynewbranch
git merge myfeature
include mybsp-hw.scc
3.6 SCC Description File Reference
This section provides a brief reference for the commands you can use
within an SCC description file (.scc
):
branch [ref]
: Creates a new branch relative to the current branch (typically${KTYPE}
) using the currently checked-out branch, or “ref” if specified.define
: Defines variables, such as KMACHINE, KTYPE, KARCH, and KFEATURE_DESCRIPTION.include SCC_FILE
: Includes an SCC file in the current file. The file is parsed as if you had inserted it inline.kconf [hardware|non-hardware] CFG_FILE
: Queues a configuration fragment for merging into the final Linux.config
file.git merge GIT_BRANCH
: Merges the feature branch into the current branch.patch PATCH_FILE
: Applies the patch to the current Git branch.
4 Advanced Kernel Concepts
4.1 Yocto Project Kernel Development and Maintenance
Kernels available through the Yocto Project (Yocto Linux kernels), like
other kernels, are based off the Linux kernel releases from
https://www.kernel.org. At the beginning of a major Linux kernel
development cycle, the Yocto Project team chooses a Linux kernel based
on factors such as release timing, the anticipated release timing of
final upstream kernel.org
versions, and Yocto Project feature
requirements. Typically, the Linux kernel chosen is in the final stages
of development by the Linux community. In other words, the Linux kernel
is in the release candidate or “rc” phase and has yet to reach final
release. But, by being in the final stages of external development, the
team knows that the kernel.org
final release will clearly be within
the early stages of the Yocto Project development window.
This balance allows the Yocto Project team to deliver the most up-to-date Yocto Linux kernel possible, while still ensuring that the team has a stable official release for the baseline Linux kernel version.
As implied earlier, the ultimate source for Yocto Linux kernels are
released kernels from kernel.org
. In addition to a foundational
kernel from kernel.org
, the available Yocto Linux kernels contain a
mix of important new mainline developments, non-mainline developments
(when no alternative exists), Board Support Package (BSP) developments,
and custom features. These additions result in a commercially released
Yocto Project Linux kernel that caters to specific embedded designer
needs for targeted hardware.
You can find a web interface to the Yocto Linux kernels in the Yocto Project Source Repositories at https://git.yoctoproject.org/cgit/cgit.cgi/. If you look at the interface, you will see to the left a grouping of Git repositories titled “Yocto Linux Kernel”. Within this group, you will find several Linux Yocto kernels developed and included with Yocto Project releases:
linux-yocto-4.1: The stable Yocto Project kernel to use with the Yocto Project Release 2.0. This kernel is based on the Linux 4.1 released kernel.
linux-yocto-4.4: The stable Yocto Project kernel to use with the Yocto Project Release 2.1. This kernel is based on the Linux 4.4 released kernel.
linux-yocto-4.6: A temporary kernel that is not tied to any Yocto Project release.
linux-yocto-4.8: The stable yocto Project kernel to use with the Yocto Project Release 2.2.
linux-yocto-4.9: The stable Yocto Project kernel to use with the Yocto Project Release 2.3. This kernel is based on the Linux 4.9 released kernel.
linux-yocto-4.10: The default stable Yocto Project kernel to use with the Yocto Project Release 2.3. This kernel is based on the Linux 4.10 released kernel.
linux-yocto-4.12: The default stable Yocto Project kernel to use with the Yocto Project Release 2.4. This kernel is based on the Linux 4.12 released kernel.
yocto-kernel-cache: The
linux-yocto-cache
contains patches and configurations for the linux-yocto kernel tree. This repository is useful when working on the linux-yocto kernel. For more information on this “Advanced Kernel Metadata”, see the “Working with Advanced Metadata (yocto-kernel-cache)” Chapter.linux-yocto-dev: A development kernel based on the latest upstream release candidate available.
Note
Long Term Support Initiative (LTSI) for Yocto Linux kernels is as follows:
For Yocto Project releases 1.7, 1.8, and 2.0, the LTSI kernel is
linux-yocto-3.14
.For Yocto Project releases 2.1, 2.2, and 2.3, the LTSI kernel is
linux-yocto-4.1
.For Yocto Project release 2.4, the LTSI kernel is
linux-yocto-4.9
linux-yocto-4.4
is an LTS kernel.
Once a Yocto Linux kernel is officially released, the Yocto Project team goes into their next development cycle, or upward revision (uprev) cycle, while still continuing maintenance on the released kernel. It is important to note that the most sustainable and stable way to include feature development upstream is through a kernel uprev process. Back-porting hundreds of individual fixes and minor features from various kernel versions is not sustainable and can easily compromise quality.
During the uprev cycle, the Yocto Project team uses an ongoing analysis
of Linux kernel development, BSP support, and release timing to select
the best possible kernel.org
Linux kernel version on which to base
subsequent Yocto Linux kernel development. The team continually monitors
Linux community kernel development to look for significant features of
interest. The team does consider back-porting large features if they
have a significant advantage. User or community demand can also trigger
a back-port or creation of new functionality in the Yocto Project
baseline kernel during the uprev cycle.
Generally speaking, every new Linux kernel both adds features and introduces new bugs. These consequences are the basic properties of upstream Linux kernel development and are managed by the Yocto Project team’s Yocto Linux kernel development strategy. It is the Yocto Project team’s policy to not back-port minor features to the released Yocto Linux kernel. They only consider back-porting significant technological jumps - and, that is done after a complete gap analysis. The reason for this policy is that back-porting any small to medium sized change from an evolving Linux kernel can easily create mismatches, incompatibilities and very subtle errors.
The policies described in this section result in both a stable and a cutting edge Yocto Linux kernel that mixes forward ports of existing Linux kernel features and significant and critical new functionality. Forward porting Linux kernel functionality into the Yocto Linux kernels available through the Yocto Project can be thought of as a “micro uprev”. The many “micro uprevs” produce a Yocto Linux kernel version with a mix of important new mainline, non-mainline, BSP developments and feature integrations. This Yocto Linux kernel gives insight into new features and allows focused amounts of testing to be done on the kernel, which prevents surprises when selecting the next major uprev. The quality of these cutting edge Yocto Linux kernels is evolving and the kernels are used in leading edge feature and BSP development.
4.2 Yocto Linux Kernel Architecture and Branching Strategies
As mentioned earlier, a key goal of the Yocto Project is to present the
developer with a kernel that has a clear and continuous history that is
visible to the user. The architecture and mechanisms, in particular the
branching strategies, used achieve that goal in a manner similar to
upstream Linux kernel development in kernel.org
.
You can think of a Yocto Linux kernel as consisting of a baseline Linux kernel with added features logically structured on top of the baseline. The features are tagged and organized by way of a branching strategy implemented by the Yocto Project team using the Source Code Manager (SCM) Git.
Note
Git is the obvious SCM for meeting the Yocto Linux kernel organizational and structural goals described in this section. Not only is Git the SCM for Linux kernel development in
kernel.org
but, Git continues to grow in popularity and supports many different work flows, front-ends and management techniques.You can find documentation on Git at https://git-scm.com/doc. You can also get an introduction to Git as it applies to the Yocto Project in the “Git” section in the Yocto Project Overview and Concepts Manual. The latter reference provides an overview of Git and presents a minimal set of Git commands that allows you to be functional using Git. You can use as much, or as little, of what Git has to offer to accomplish what you need for your project. You do not have to be a “Git Expert” in order to use it with the Yocto Project.
Using Git’s tagging and branching features, the Yocto Project team creates kernel branches at points where functionality is no longer shared and thus, needs to be isolated. For example, board-specific incompatibilities would require different functionality and would require a branch to separate the features. Likewise, for specific kernel features, the same branching strategy is used.
This “tree-like” architecture results in a structure that has features organized to be specific for particular functionality, single kernel types, or a subset of kernel types. Thus, the user has the ability to see the added features and the commits that make up those features. In addition to being able to see added features, the user can also view the history of what made up the baseline Linux kernel.
Another consequence of this strategy results in not having to store the same feature twice internally in the tree. Rather, the kernel team stores the unique differences required to apply the feature onto the kernel type in question.
Note
The Yocto Project team strives to place features in the tree such that features can be shared by all boards and kernel types where possible. However, during development cycles or when large features are merged, the team cannot always follow this practice. In those cases, the team uses isolated branches to merge features.
BSP-specific code additions are handled in a similar manner to kernel-specific additions. Some BSPs only make sense given certain kernel types. So, for these types, the team creates branches off the end of that kernel type for all of the BSPs that are supported on that kernel type. From the perspective of the tools that create the BSP branch, the BSP is really no different than a feature. Consequently, the same branching strategy applies to BSPs as it does to kernel features. So again, rather than store the BSP twice, the team only stores the unique differences for the BSP across the supported multiple kernels.
While this strategy can result in a tree with a significant number of
branches, it is important to realize that from the developer’s point of
view, there is a linear path that travels from the baseline
kernel.org
, through a select group of features and ends with their
BSP-specific commits. In other words, the divisions of the kernel are
transparent and are not relevant to the developer on a day-to-day basis.
From the developer’s perspective, this path is the “master” branch in
Git terms. The developer does not need to be aware of the existence of
any other branches at all. Of course, value exists in the having these
branches in the tree, should a person decide to explore them. For
example, a comparison between two BSPs at either the commit level or at
the line-by-line code diff
level is now a trivial operation.
The following illustration shows the conceptual Yocto Linux kernel.
In the illustration, the “Kernel.org Branch Point” marks the specific spot (or Linux kernel release) from which the Yocto Linux kernel is created. From this point forward in the tree, features and differences are organized and tagged.
The “Yocto Project Baseline Kernel” contains functionality that is common to every kernel type and BSP that is organized further along in the tree. Placing these common features in the tree this way means features do not have to be duplicated along individual branches of the tree structure.
From the “Yocto Project Baseline Kernel”, branch points represent specific functionality for individual Board Support Packages (BSPs) as well as real-time kernels. The illustration represents this through three BSP-specific branches and a real-time kernel branch. Each branch represents some unique functionality for the BSP or for a real-time Yocto Linux kernel.
In this example structure, the “Real-time (rt) Kernel” branch has common features for all real-time Yocto Linux kernels and contains more branches for individual BSP-specific real-time kernels. The illustration shows three branches as an example. Each branch points the way to specific, unique features for a respective real-time kernel as they apply to a given BSP.
The resulting tree structure presents a clear path of markers (or branches) to the developer that, for all practical purposes, is the Yocto Linux kernel needed for any given set of requirements.
Note
Keep in mind the figure does not take into account all the supported Yocto Linux kernels, but rather shows a single generic kernel just for conceptual purposes. Also keep in mind that this structure represents the Yocto Project Source Repositories that are either pulled from during the build or established on the host development system prior to the build by either cloning a particular kernel’s Git repository or by downloading and unpacking a tarball.
Working with the kernel as a structured tree follows recognized community best practices. In particular, the kernel as shipped with the product, should be considered an “upstream source” and viewed as a series of historical and documented modifications (commits). These modifications represent the development and stabilization done by the Yocto Project kernel development team.
Because commits only change at significant release points in the product life cycle, developers can work on a branch created from the last relevant commit in the shipped Yocto Project Linux kernel. As mentioned previously, the structure is transparent to the developer because the kernel tree is left in this state after cloning and building the kernel.
4.3 Kernel Build File Hierarchy
Upstream storage of all the available kernel source code is one thing, while representing and using the code on your host development system is another. Conceptually, you can think of the kernel source repositories as all the source files necessary for all the supported Yocto Linux kernels. As a developer, you are just interested in the source files for the kernel on which you are working. And, furthermore, you need them available on your host system.
Kernel source code is available on your host system several different ways:
Files Accessed While using devtool:
devtool
, which is available with the Yocto Project, is the preferred method by which to modify the kernel. See the “Kernel Modification Workflow” section.Cloned Repository: If you are working in the kernel all the time, you probably would want to set up your own local Git repository of the Yocto Linux kernel tree. For information on how to clone a Yocto Linux kernel Git repository, see the “Preparing the Build Host to Work on the Kernel” section.
Temporary Source Files from a Build: If you just need to make some patches to the kernel using a traditional BitBake workflow (i.e. not using the
devtool
), you can access temporary kernel source files that were extracted and used during a kernel build.
The temporary kernel source files resulting from a build using BitBake have a particular hierarchy. When you build the kernel on your development system, all files needed for the build are taken from the source repositories pointed to by the SRC_URI variable and gathered in a temporary work area where they are subsequently used to create the unique kernel. Thus, in a sense, the process constructs a local source tree specific to your kernel from which to generate the new kernel image.
The following figure shows the temporary file structure created on your host system when you build the kernel using Bitbake. This Build Directory contains all the source files used during the build.
Again, for additional information on the Yocto Project kernel’s architecture and its branching strategy, see the “Yocto Linux Kernel Architecture and Branching Strategies” section. You can also reference the “Using devtool to Patch the Kernel” and “Using Traditional Kernel Development to Patch the Kernel” sections for detailed example that modifies the kernel.
4.4 Determining Hardware and Non-Hardware Features for the Kernel Configuration Audit Phase
This section describes part of the kernel configuration audit phase that
most developers can ignore. For general information on kernel
configuration including menuconfig
, defconfig
files, and
configuration fragments, see the
“Configuring the Kernel” section.
During this part of the audit phase, the contents of the final
.config
file are compared against the fragments specified by the
system. These fragments can be system fragments, distro fragments, or
user-specified configuration elements. Regardless of their origin, the
OpenEmbedded build system warns the user if a specific option is not
included in the final kernel configuration.
By default, in order to not overwhelm the user with configuration warnings, the system only reports missing “hardware” options as they could result in a boot failure or indicate that important hardware is not available.
To determine whether or not a given option is “hardware” or
“non-hardware”, the kernel Metadata in yocto-kernel-cache
contains
files that classify individual or groups of options as either hardware
or non-hardware. To better show this, consider a situation where the
yocto-kernel-cache
contains the following files:
yocto-kernel-cache/features/drm-psb/hardware.cfg
yocto-kernel-cache/features/kgdb/hardware.cfg
yocto-kernel-cache/ktypes/base/hardware.cfg
yocto-kernel-cache/bsp/mti-malta32/hardware.cfg
yocto-kernel-cache/bsp/qemu-ppc32/hardware.cfg
yocto-kernel-cache/bsp/qemuarma9/hardware.cfg
yocto-kernel-cache/bsp/mti-malta64/hardware.cfg
yocto-kernel-cache/bsp/arm-versatile-926ejs/hardware.cfg
yocto-kernel-cache/bsp/common-pc/hardware.cfg
yocto-kernel-cache/bsp/common-pc-64/hardware.cfg
yocto-kernel-cache/features/rfkill/non-hardware.cfg
yocto-kernel-cache/ktypes/base/non-hardware.cfg
yocto-kernel-cache/features/aufs/non-hardware.kcf
yocto-kernel-cache/features/ocf/non-hardware.kcf
yocto-kernel-cache/ktypes/base/non-hardware.kcf
yocto-kernel-cache/ktypes/base/hardware.kcf
yocto-kernel-cache/bsp/qemu-ppc32/hardware.kcf
The following list provides explanations for the various files:
hardware.kcf
: Specifies a list of kernel Kconfig files that contain hardware options only.non-hardware.kcf
: Specifies a list of kernel Kconfig files that contain non-hardware options only.hardware.cfg
: Specifies a list of kernelCONFIG_
options that are hardware, regardless of whether or not they are within a Kconfig file specified by a hardware or non-hardware Kconfig file (i.e.hardware.kcf
ornon-hardware.kcf
).non-hardware.cfg
: Specifies a list of kernelCONFIG_
options that are not hardware, regardless of whether or not they are within a Kconfig file specified by a hardware or non-hardware Kconfig file (i.e.hardware.kcf
ornon-hardware.kcf
).
Here is a specific example using the
kernel-cache/bsp/mti-malta32/hardware.cfg
:
CONFIG_SERIAL_8250
CONFIG_SERIAL_8250_CONSOLE
CONFIG_SERIAL_8250_NR_UARTS
CONFIG_SERIAL_8250_PCI
CONFIG_SERIAL_CORE
CONFIG_SERIAL_CORE_CONSOLE
CONFIG_VGA_ARB
The kernel configuration audit automatically detects
these files (hence the names must be exactly the ones discussed here),
and uses them as inputs when generating warnings about the final
.config
file.
A user-specified kernel Metadata repository, or recipe space feature,
can use these same files to classify options that are found within its
.cfg
files as hardware or non-hardware, to prevent the OpenEmbedded
build system from producing an error or warning when an option is not in
the final .config
file.
5 Kernel Maintenance
5.1 Tree Construction
This section describes construction of the Yocto Project kernel source
repositories as accomplished by the Yocto Project team to create Yocto
Linux kernel repositories. These kernel repositories are found under the
heading “Yocto Linux Kernel” at https://git.yoctoproject.org/cgit/cgit.cgi/ and
are shipped as part of a Yocto Project release. The team creates these
repositories by compiling and executing the set of feature descriptions
for every BSP and feature in the product. Those feature descriptions
list all necessary patches, configurations, branches, tags, and feature
divisions found in a Yocto Linux kernel. Thus, the Yocto Project Linux
kernel repository (or tree) and accompanying Metadata in the
yocto-kernel-cache
are built.
The existence of these repositories allow you to access and clone a particular Yocto Project Linux kernel repository and use it to build images based on their configurations and features.
You can find the files used to describe all the valid features and BSPs
in the Yocto Project Linux kernel in any clone of the Yocto Project
Linux kernel source repository and yocto-kernel-cache
Git trees. For
example, the following commands clone the Yocto Project baseline Linux
kernel that branches off linux.org
version 4.12 and the
yocto-kernel-cache
, which contains stores of kernel Metadata:
$ git clone git://git.yoctoproject.org/linux-yocto-4.12
$ git clone git://git.yoctoproject.org/linux-kernel-cache
For more information on how to set up a local Git repository of the Yocto Project Linux kernel files, see the “Preparing the Build Host to Work on the Kernel” section.
Once you have cloned the kernel Git repository and the cache of Metadata on your local machine, you can discover the branches that are available in the repository using the following Git command:
$ git branch -a
Checking out a branch allows you to work with a particular Yocto Linux
kernel. For example, the following commands check out the
“standard/beagleboard” branch of the Yocto Linux kernel repository and
the “yocto-4.12” branch of the yocto-kernel-cache
repository:
$ cd ~/linux-yocto-4.12
$ git checkout -b my-kernel-4.12 remotes/origin/standard/beagleboard
$ cd ~/linux-kernel-cache
$ git checkout -b my-4.12-metadata remotes/origin/yocto-4.12
Note
Branches in the yocto-kernel-cache
repository correspond to Yocto Linux
kernel versions (e.g. “yocto-4.12”, “yocto-4.10”, “yocto-4.9”, and so forth).
Once you have checked out and switched to appropriate branches, you can see a snapshot of all the kernel source files used to build that particular Yocto Linux kernel for a particular board.
To see the features and configurations for a particular Yocto Linux
kernel, you need to examine the yocto-kernel-cache
Git repository.
As mentioned, branches in the yocto-kernel-cache
repository
correspond to Yocto Linux kernel versions (e.g. yocto-4.12
).
Branches contain descriptions in the form of .scc
and .cfg
files.
You should realize, however, that browsing your local
yocto-kernel-cache
repository for feature descriptions and patches
is not an effective way to determine what is in a particular kernel
branch. Instead, you should use Git directly to discover the changes in
a branch. Using Git is an efficient and flexible way to inspect changes
to the kernel.
Note
Ground up reconstruction of the complete kernel tree is an action only taken by the Yocto Project team during an active development cycle. When you create a clone of the kernel Git repository, you are simply making it efficiently available for building and development.
The following steps describe what happens when the Yocto Project Team constructs the Yocto Project kernel source Git repository (or tree) found at https://git.yoctoproject.org/cgit/cgit.cgi/ given the introduction of a new top-level kernel feature or BSP. The following actions effectively provide the Metadata and create the tree that includes the new feature, patch, or BSP:
Pass Feature to the OpenEmbedded Build System: A top-level kernel feature is passed to the kernel build subsystem. Normally, this feature is a BSP for a particular kernel type.
Locate Feature: The file that describes the top-level feature is located by searching these system directories:
The in-tree kernel-cache directories, which are located in the yocto-kernel-cache repository organized under the “Yocto Linux Kernel” heading in the Yocto Project Source Repositories.
Areas pointed to by
SRC_URI
statements found in kernel recipes.
For a typical build, the target of the search is a feature description in an
.scc
file whose name follows this format (e.g.beaglebone-standard.scc
andbeaglebone-preempt-rt.scc
):bsp_root_name-kernel_type.scc
Expand Feature: Once located, the feature description is either expanded into a simple script of actions, or into an existing equivalent script that is already part of the shipped kernel.
Append Extra Features: Extra features are appended to the top-level feature description. These features can come from the KERNEL_FEATURES variable in recipes.
Locate, Expand, and Append Each Feature: Each extra feature is located, expanded and appended to the script as described in step three.
Execute the Script: The script is executed to produce files
.scc
and.cfg
files in appropriate directories of theyocto-kernel-cache
repository. These files are descriptions of all the branches, tags, patches and configurations that need to be applied to the base Git repository to completely create the source (build) branch for the new BSP or feature.Clone Base Repository: The base repository is cloned, and the actions listed in the
yocto-kernel-cache
directories are applied to the tree.Perform Cleanup: The Git repositories are left with the desired branches checked out and any required branching, patching and tagging has been performed.
The kernel tree and cache are ready for developer consumption to be locally cloned, configured, and built into a Yocto Project kernel specific to some target hardware.
Note
The generated
yocto-kernel-cache
repository adds to the kernel as shipped with the Yocto Project release. Any add-ons and configuration data are applied to the end of an existing branch. The full repository generation that is found in the official Yocto Project kernel repositories at https://git.yoctoproject.org/cgit/cgit.cgi/ is the combination of all supported boards and configurations.The technique the Yocto Project team uses is flexible and allows for seamless blending of an immutable history with additional patches specific to a deployment. Any additions to the kernel become an integrated part of the branches.
The full kernel tree that you see on https://git.yoctoproject.org/cgit/cgit.cgi/ is generated through repeating the above steps for all valid BSPs. The end result is a branched, clean history tree that makes up the kernel for a given release. You can see the script (
kgit-scc
) responsible for this in the yocto-kernel-tools repository.The steps used to construct the full kernel tree are the same steps that BitBake uses when it builds a kernel image.
5.2 Build Strategy
Once you have cloned a Yocto Linux kernel repository and the cache
repository (yocto-kernel-cache
) onto your development system, you
can consider the compilation phase of kernel development, which is
building a kernel image. Some prerequisites exist that are validated by
the build process before compilation starts:
The SRC_URI points to the kernel Git repository.
A BSP build branch with Metadata exists in the
yocto-kernel-cache
repository. The branch is based on the Yocto Linux kernel version and has configurations and features grouped under theyocto-kernel-cache/bsp
directory. For example, features and configurations for the BeagleBone Board assuming alinux-yocto_4.12
kernel reside in the following area of theyocto-kernel-cache
repository: yocto-kernel-cache/bsp/beagleboneNote
In the previous example, the “yocto-4.12” branch is checked out in the
yocto-kernel-cache
repository.
The OpenEmbedded build system makes sure these conditions exist before attempting compilation. Other means, however, do exist, such as bootstrapping a BSP.
Before building a kernel, the build process verifies the tree and
configures the kernel by processing all of the configuration “fragments”
specified by feature descriptions in the .scc
files. As the features
are compiled, associated kernel configuration fragments are noted and
recorded in the series of directories in their compilation order. The
fragments are migrated, pre-processed and passed to the Linux Kernel
Configuration subsystem (lkc
) as raw input in the form of a
.config
file. The lkc
uses its own internal dependency
constraints to do the final processing of that information and generates
the final .config
file that is used during compilation.
Using the board’s architecture and other relevant values from the board’s template, kernel compilation is started and a kernel image is produced.
The other thing that you notice once you configure a kernel is that the
build process generates a build tree that is separate from your kernel’s
local Git source repository tree. This build tree has a name that uses
the following form, where ${MACHINE}
is the metadata name of the
machine (BSP) and “kernel_type” is one of the Yocto Project supported
kernel types (e.g. “standard”):
linux-${MACHINE}-kernel_type-build
The existing support in the kernel.org
tree achieves this default
functionality.
This behavior means that all the generated files for a particular
machine or BSP are now in the build tree directory. The files include
the final .config
file, all the .o
files, the .a
files, and
so forth. Since each machine or BSP has its own separate
Build Directory in its own separate
branch of the Git repository, you can easily switch between different
builds.
6 Kernel Development FAQ
6.1 Common Questions and Solutions
The following lists some solutions for common questions.
6.1.1 How do I use my own Linux kernel .config
file?
Refer to the “Changing the Configuration” section for information.
6.1.2 How do I create configuration fragments?
A: Refer to the “Creating Configuration Fragments” section for information.
6.1.3 How do I use my own Linux kernel sources?
Refer to the “Working With Your Own Sources” section for information.
6.1.4 How do I install/not-install the kernel image on the rootfs?
The kernel image (e.g. vmlinuz
) is provided by the
kernel-image
package. Image recipes depend on kernel-base
. To
specify whether or not the kernel image is installed in the generated
root filesystem, override RDEPENDS_${KERNEL_PACKAGE_NAME}-base
to include or not
include “kernel-image”. See the
“Using .bbappend Files in Your Layer”
section in the
Yocto Project Development Tasks Manual for information on how to use an
append file to override metadata.
6.1.5 How do I install a specific kernel module?
Linux kernel modules are packaged individually. To ensure a specific kernel module is included in an image, include it in the appropriate machine RRECOMMENDS variable. These other variables are useful for installing specific modules: - MACHINE_ESSENTIAL_EXTRA_RDEPENDS - MACHINE_ESSENTIAL_EXTRA_RRECOMMENDS - MACHINE_EXTRA_RDEPENDS - MACHINE_EXTRA_RRECOMMENDS
For example, set the following in the qemux86.conf
file to include
the ab123
kernel modules with images built for the qemux86
machine:
MACHINE_EXTRA_RRECOMMENDS += "kernel-module-ab123"
For more information, see the “Incorporating Out-of-Tree Modules” section.
6.1.6 How do I change the Linux kernel command line?
The Linux kernel command line is
typically specified in the machine config using the APPEND
variable.
For example, you can add some helpful debug information doing the
following:
APPEND += "printk.time=y initcall_debug debug"
7 Manual Revision History
Revision |
Date |
Note |
---|---|---|
1.4 |
April 2013 |
The initial document released with the Yocto Project 1.4 Release |
1.5 |
October 2013 |
Released with the Yocto Project 1.5 Release. |
1.6 |
April 2014 |
Released with the Yocto Project 1.6 Release. |
1.7 |
October 2014 |
Released with the Yocto Project 1.7 Release. |
1.8 |
April 2015 |
Released with the Yocto Project 1.8 Release. |
2.0 |
October 2015 |
Released with the Yocto Project 2.0 Release. |
2.1 |
April 2016 |
Released with the Yocto Project 2.1 Release. |
2.2 |
October 2016 |
Released with the Yocto Project 2.2 Release. |
2.3 |
May 2017 |
Released with the Yocto Project 2.3 Release. |
2.4 |
October 2017 |
Released with the Yocto Project 2.4 Release. |
2.5 |
May 2018 |
Released with the Yocto Project 2.5 Release. |
2.6 |
November 2018 |
Released with the Yocto Project 2.6 Release. |
2.7 |
May 2019 |
Released with the Yocto Project 2.7 Release. |
3.0 |
October 2019 |
Released with the Yocto Project 3.0 Release. |
3.1 |
April 2020 |
Released with the Yocto Project 3.1 Release. |
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Yocto Project Profiling and Tracing Manual
1 Yocto Project Profiling and Tracing Manual
1.1 Introduction
Yocto bundles a number of tracing and profiling tools - this ‘HOWTO’ describes their basic usage and shows by example how to make use of them to examine application and system behavior.
The tools presented are for the most part completely open-ended and have quite good and/or extensive documentation of their own which can be used to solve just about any problem you might come across in Linux. Each section that describes a particular tool has links to that tool’s documentation and website.
The purpose of this ‘HOWTO’ is to present a set of common and generally useful tracing and profiling idioms along with their application (as appropriate) to each tool, in the context of a general-purpose ‘drill-down’ methodology that can be applied to solving a large number (90%?) of problems. For help with more advanced usages and problems, please see the documentation and/or websites listed for each tool.
The final section of this ‘HOWTO’ is a collection of real-world examples which we’ll be continually adding to as we solve more problems using the tools - feel free to add your own examples to the list!
1.2 General Setup
Most of the tools are available only in ‘sdk’ images or in images built after adding ‘tools-profile’ to your local.conf. So, in order to be able to access all of the tools described here, please first build and boot an ‘sdk’ image e.g.
$ bitbake core-image-sato-sdk
or alternatively by adding ‘tools-profile’ to the EXTRA_IMAGE_FEATURES line in your local.conf:
EXTRA_IMAGE_FEATURES = "debug-tweaks tools-profile"
If you use the ‘tools-profile’ method, you don’t need to build an sdk image - the tracing and profiling tools will be included in non-sdk images as well e.g.:
$ bitbake core-image-sato
Note
By default, the Yocto build system strips symbols from the binaries it packages, which makes it difficult to use some of the tools.
You can prevent that by setting the
INHIBIT_PACKAGE_STRIP
variable to “1” in your local.conf
when you build the image:
INHIBIT_PACKAGE_STRIP = "1"
The above setting will noticeably increase the size of your image.
If you’ve already built a stripped image, you can generate debug packages (xxx-dbg) which you can manually install as needed.
To generate debug info for packages, you can add dbg-pkgs to EXTRA_IMAGE_FEATURES in local.conf. For example:
EXTRA_IMAGE_FEATURES = "debug-tweaks tools-profile dbg-pkgs"
Additionally, in order to generate the right type of debuginfo, we also need to
set PACKAGE_DEBUG_SPLIT_STYLE in the local.conf
file:
PACKAGE_DEBUG_SPLIT_STYLE = 'debug-file-directory'
2 Overall Architecture of the Linux Tracing and Profiling Tools
2.1 Architecture of the Tracing and Profiling Tools
It may seem surprising to see a section covering an ‘overall architecture’ for what seems to be a random collection of tracing tools that together make up the Linux tracing and profiling space. The fact is, however, that in recent years this seemingly disparate set of tools has started to converge on a ‘core’ set of underlying mechanisms:
static tracepoints
dynamic tracepoints
kprobes
uprobes
the perf_events subsystem
debugfs
3 Basic Usage (with examples) for each of the Yocto Tracing Tools
This chapter presents basic usage examples for each of the tracing tools.
3.1 perf
The ‘perf’ tool is the profiling and tracing tool that comes bundled with the Linux kernel.
Don’t let the fact that it’s part of the kernel fool you into thinking that it’s only for tracing and profiling the kernel - you can indeed use it to trace and profile just the kernel, but you can also use it to profile specific applications separately (with or without kernel context), and you can also use it to trace and profile the kernel and all applications on the system simultaneously to gain a system-wide view of what’s going on.
In many ways, perf aims to be a superset of all the tracing and profiling tools available in Linux today, including all the other tools covered in this HOWTO. The past couple of years have seen perf subsume a lot of the functionality of those other tools and, at the same time, those other tools have removed large portions of their previous functionality and replaced it with calls to the equivalent functionality now implemented by the perf subsystem. Extrapolation suggests that at some point those other tools will simply become completely redundant and go away; until then, we’ll cover those other tools in these pages and in many cases show how the same things can be accomplished in perf and the other tools when it seems useful to do so.
The coverage below details some of the most common ways you’ll likely want to apply the tool; full documentation can be found either within the tool itself or in the man pages at perf(1).
3.1.1 Perf Setup
For this section, we’ll assume you’ve already performed the basic setup outlined in the “General Setup” section.
In particular, you’ll get the most mileage out of perf if you profile an
image built with the following in your local.conf
file:
INHIBIT_PACKAGE_STRIP = "1"
perf runs on the target system for the most part. You can archive profile data and copy it to the host for analysis, but for the rest of this document we assume you’ve ssh’ed to the host and will be running the perf commands on the target.
3.1.2 Basic Perf Usage
The perf tool is pretty much self-documenting. To remind yourself of the available commands, simply type ‘perf’, which will show you basic usage along with the available perf subcommands:
root@crownbay:~# perf
usage: perf [--version] [--help] COMMAND [ARGS]
The most commonly used perf commands are:
annotate Read perf.data (created by perf record) and display annotated code
archive Create archive with object files with build-ids found in perf.data file
bench General framework for benchmark suites
buildid-cache Manage build-id cache.
buildid-list List the buildids in a perf.data file
diff Read two perf.data files and display the differential profile
evlist List the event names in a perf.data file
inject Filter to augment the events stream with additional information
kmem Tool to trace/measure kernel memory(slab) properties
kvm Tool to trace/measure kvm guest os
list List all symbolic event types
lock Analyze lock events
probe Define new dynamic tracepoints
record Run a command and record its profile into perf.data
report Read perf.data (created by perf record) and display the profile
sched Tool to trace/measure scheduler properties (latencies)
script Read perf.data (created by perf record) and display trace output
stat Run a command and gather performance counter statistics
test Runs sanity tests.
timechart Tool to visualize total system behavior during a workload
top System profiling tool.
See 'perf help COMMAND' for more information on a specific command.
3.1.2.1 Using perf to do Basic Profiling
As a simple test case, we’ll profile the ‘wget’ of a fairly large file, which is a minimally interesting case because it has both file and network I/O aspects, and at least in the case of standard Yocto images, it’s implemented as part of BusyBox, so the methods we use to analyze it can be used in a very similar way to the whole host of supported BusyBox applets in Yocto.
root@crownbay:~# rm linux-2.6.19.2.tar.bz2; \
wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2
The quickest and easiest way to get some basic overall data about what’s going on for a particular workload is to profile it using ‘perf stat’. ‘perf stat’ basically profiles using a few default counters and displays the summed counts at the end of the run:
root@crownbay:~# perf stat wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2
Connecting to downloads.yoctoproject.org (140.211.169.59:80)
linux-2.6.19.2.tar.b 100% |***************************************************| 41727k 0:00:00 ETA
Performance counter stats for 'wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2':
4597.223902 task-clock # 0.077 CPUs utilized
23568 context-switches # 0.005 M/sec
68 CPU-migrations # 0.015 K/sec
241 page-faults # 0.052 K/sec
3045817293 cycles # 0.663 GHz
<not supported> stalled-cycles-frontend
<not supported> stalled-cycles-backend
858909167 instructions # 0.28 insns per cycle
165441165 branches # 35.987 M/sec
19550329 branch-misses # 11.82% of all branches
59.836627620 seconds time elapsed
Many times such a simple-minded test doesn’t yield much of interest, but sometimes it does (see Real-world Yocto bug (slow loop-mounted write speed)).
Also, note that ‘perf stat’ isn’t restricted to a fixed set of counters - basically any event listed in the output of ‘perf list’ can be tallied by ‘perf stat’. For example, suppose we wanted to see a summary of all the events related to kernel memory allocation/freeing along with cache hits and misses:
root@crownbay:~# perf stat -e kmem:* -e cache-references -e cache-misses wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2
Connecting to downloads.yoctoproject.org (140.211.169.59:80)
linux-2.6.19.2.tar.b 100% |***************************************************| 41727k 0:00:00 ETA
Performance counter stats for 'wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2':
5566 kmem:kmalloc
125517 kmem:kmem_cache_alloc
0 kmem:kmalloc_node
0 kmem:kmem_cache_alloc_node
34401 kmem:kfree
69920 kmem:kmem_cache_free
133 kmem:mm_page_free
41 kmem:mm_page_free_batched
11502 kmem:mm_page_alloc
11375 kmem:mm_page_alloc_zone_locked
0 kmem:mm_page_pcpu_drain
0 kmem:mm_page_alloc_extfrag
66848602 cache-references
2917740 cache-misses # 4.365 % of all cache refs
44.831023415 seconds time elapsed
So ‘perf stat’ gives us a nice easy way to get a quick overview of what might be happening for a set of events, but normally we’d need a little more detail in order to understand what’s going on in a way that we can act on in a useful way.
To dive down into a next level of detail, we can use ‘perf record’/’perf report’ which will collect profiling data and present it to use using an interactive text-based UI (or simply as text if we specify –stdio to ‘perf report’).
As our first attempt at profiling this workload, we’ll simply run ‘perf record’, handing it the workload we want to profile (everything after ‘perf record’ and any perf options we hand it - here none - will be executed in a new shell). perf collects samples until the process exits and records them in a file named ‘perf.data’ in the current working directory.
root@crownbay:~# perf record wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2
Connecting to downloads.yoctoproject.org (140.211.169.59:80)
linux-2.6.19.2.tar.b 100% |************************************************| 41727k 0:00:00 ETA
[ perf record: Woken up 1 times to write data ]
[ perf record: Captured and wrote 0.176 MB perf.data (~7700 samples) ]
To see the results in a ‘text-based UI’ (tui), simply run ‘perf report’, which will read the perf.data file in the current working directory and display the results in an interactive UI:
root@crownbay:~# perf report
The above screenshot displays a ‘flat’ profile, one entry for each ‘bucket’ corresponding to the functions that were profiled during the profiling run, ordered from the most popular to the least (perf has options to sort in various orders and keys as well as display entries only above a certain threshold and so on - see the perf documentation for details). Note that this includes both userspace functions (entries containing a [.]) and kernel functions accounted to the process (entries containing a [k]). (perf has command-line modifiers that can be used to restrict the profiling to kernel or userspace, among others).
Notice also that the above report shows an entry for ‘busybox’, which is the executable that implements ‘wget’ in Yocto, but that instead of a useful function name in that entry, it displays a not-so-friendly hex value instead. The steps below will show how to fix that problem.
Before we do that, however, let’s try running a different profile, one which shows something a little more interesting. The only difference between the new profile and the previous one is that we’ll add the -g option, which will record not just the address of a sampled function, but the entire callchain to the sampled function as well:
root@crownbay:~# perf record -g wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2
Connecting to downloads.yoctoproject.org (140.211.169.59:80)
linux-2.6.19.2.tar.b 100% |************************************************| 41727k 0:00:00 ETA
[ perf record: Woken up 3 times to write data ]
[ perf record: Captured and wrote 0.652 MB perf.data (~28476 samples) ]
root@crownbay:~# perf report
Using the callgraph view, we can actually see not only which functions took the most time, but we can also see a summary of how those functions were called and learn something about how the program interacts with the kernel in the process.
Notice that each entry in the above screenshot now contains a ‘+’ on the left-hand side. This means that we can expand the entry and drill down into the callchains that feed into that entry. Pressing ‘enter’ on any one of them will expand the callchain (you can also press ‘E’ to expand them all at the same time or ‘C’ to collapse them all).
In the screenshot above, we’ve toggled the __copy_to_user_ll()
entry
and several subnodes all the way down. This lets us see which callchains
contributed to the profiled __copy_to_user_ll()
function which
contributed 1.77% to the total profile.
As a bit of background explanation for these callchains, think about what happens at a high level when you run wget to get a file out on the network. Basically what happens is that the data comes into the kernel via the network connection (socket) and is passed to the userspace program ‘wget’ (which is actually a part of BusyBox, but that’s not important for now), which takes the buffers the kernel passes to it and writes it to a disk file to save it.
The part of this process that we’re looking at in the above call stacks is the part where the kernel passes the data it has read from the socket down to wget i.e. a copy-to-user.
Notice also that here there’s also a case where the hex value is
displayed in the callstack, here in the expanded sys_clock_gettime()
function. Later we’ll see it resolve to a userspace function call in
busybox.
The above screenshot shows the other half of the journey for the data -
from the wget program’s userspace buffers to disk. To get the buffers to
disk, the wget program issues a write(2)
, which does a copy-from-user
to
the kernel, which then takes care via some circuitous path (probably
also present somewhere in the profile data), to get it safely to disk.
Now that we’ve seen the basic layout of the profile data and the basics of how to extract useful information out of it, let’s get back to the task at hand and see if we can get some basic idea about where the time is spent in the program we’re profiling, wget. Remember that wget is actually implemented as an applet in BusyBox, so while the process name is ‘wget’, the executable we’re actually interested in is BusyBox. So let’s expand the first entry containing BusyBox:
Again, before we expanded we saw that the function was labeled with a hex value instead of a symbol as with most of the kernel entries. Expanding the BusyBox entry doesn’t make it any better.
The problem is that perf can’t find the symbol information for the busybox binary, which is actually stripped out by the Yocto build system.
One way around that is to put the following in your local.conf
file
when you build the image:
INHIBIT_PACKAGE_STRIP = "1"
However, we already have an image with the binaries stripped, so what can we do to get perf to resolve the symbols? Basically we need to install the debuginfo for the BusyBox package.
To generate the debug info for the packages in the image, we can add
dbg-pkgs
to EXTRA_IMAGE_FEATURES in local.conf
. For example:
EXTRA_IMAGE_FEATURES = "debug-tweaks tools-profile dbg-pkgs"
Additionally, in order to generate the type of debuginfo that perf
understands, we also need to set
PACKAGE_DEBUG_SPLIT_STYLE
in the local.conf
file:
PACKAGE_DEBUG_SPLIT_STYLE = 'debug-file-directory'
Once we’ve done that, we can install the
debuginfo for BusyBox. The debug packages once built can be found in
build/tmp/deploy/rpm/*
on the host system. Find the busybox-dbg-…rpm
file and copy it to the target. For example:
[trz@empanada core2]$ scp /home/trz/yocto/crownbay-tracing-dbg/build/tmp/deploy/rpm/core2_32/busybox-dbg-1.20.2-r2.core2_32.rpm root@192.168.1.31:
busybox-dbg-1.20.2-r2.core2_32.rpm 100% 1826KB 1.8MB/s 00:01
Now install the debug rpm on the target:
root@crownbay:~# rpm -i busybox-dbg-1.20.2-r2.core2_32.rpm
Now that the debuginfo is installed, we see that the BusyBox entries now display their functions symbolically:
If we expand one of the entries and press ‘enter’ on a leaf node, we’re presented with a menu of actions we can take to get more information related to that entry:
One of these actions allows us to show a view that displays a busybox-centric view of the profiled functions (in this case we’ve also expanded all the nodes using the ‘E’ key):
Finally, we can see that now that the BusyBox debuginfo is installed,
the previously unresolved symbol in the sys_clock_gettime()
entry
mentioned previously is now resolved, and shows that the
sys_clock_gettime system call that was the source of 6.75% of the
copy-to-user overhead was initiated by the handle_input()
BusyBox
function:
At the lowest level of detail, we can dive down to the assembly level and see which instructions caused the most overhead in a function. Pressing ‘enter’ on the ‘udhcpc_main’ function, we’re again presented with a menu:
Selecting ‘Annotate udhcpc_main’, we get a detailed listing of percentages by instruction for the udhcpc_main function. From the display, we can see that over 50% of the time spent in this function is taken up by a couple tests and the move of a constant (1) to a register:
As a segue into tracing, let’s try another profile using a different counter, something other than the default ‘cycles’.
The tracing and profiling infrastructure in Linux has become unified in a way that allows us to use the same tool with a completely different set of counters, not just the standard hardware counters that traditional tools have had to restrict themselves to (of course the traditional tools can also make use of the expanded possibilities now available to them, and in some cases have, as mentioned previously).
We can get a list of the available events that can be used to profile a workload via ‘perf list’:
root@crownbay:~# perf list
List of pre-defined events (to be used in -e):
cpu-cycles OR cycles [Hardware event]
stalled-cycles-frontend OR idle-cycles-frontend [Hardware event]
stalled-cycles-backend OR idle-cycles-backend [Hardware event]
instructions [Hardware event]
cache-references [Hardware event]
cache-misses [Hardware event]
branch-instructions OR branches [Hardware event]
branch-misses [Hardware event]
bus-cycles [Hardware event]
ref-cycles [Hardware event]
cpu-clock [Software event]
task-clock [Software event]
page-faults OR faults [Software event]
minor-faults [Software event]
major-faults [Software event]
context-switches OR cs [Software event]
cpu-migrations OR migrations [Software event]
alignment-faults [Software event]
emulation-faults [Software event]
L1-dcache-loads [Hardware cache event]
L1-dcache-load-misses [Hardware cache event]
L1-dcache-prefetch-misses [Hardware cache event]
L1-icache-loads [Hardware cache event]
L1-icache-load-misses [Hardware cache event]
.
.
.
rNNN [Raw hardware event descriptor]
cpu/t1=v1[,t2=v2,t3 ...]/modifier [Raw hardware event descriptor]
(see 'perf list --help' on how to encode it)
mem:<addr>[:access] [Hardware breakpoint]
sunrpc:rpc_call_status [Tracepoint event]
sunrpc:rpc_bind_status [Tracepoint event]
sunrpc:rpc_connect_status [Tracepoint event]
sunrpc:rpc_task_begin [Tracepoint event]
skb:kfree_skb [Tracepoint event]
skb:consume_skb [Tracepoint event]
skb:skb_copy_datagram_iovec [Tracepoint event]
net:net_dev_xmit [Tracepoint event]
net:net_dev_queue [Tracepoint event]
net:netif_receive_skb [Tracepoint event]
net:netif_rx [Tracepoint event]
napi:napi_poll [Tracepoint event]
sock:sock_rcvqueue_full [Tracepoint event]
sock:sock_exceed_buf_limit [Tracepoint event]
udp:udp_fail_queue_rcv_skb [Tracepoint event]
hda:hda_send_cmd [Tracepoint event]
hda:hda_get_response [Tracepoint event]
hda:hda_bus_reset [Tracepoint event]
scsi:scsi_dispatch_cmd_start [Tracepoint event]
scsi:scsi_dispatch_cmd_error [Tracepoint event]
scsi:scsi_eh_wakeup [Tracepoint event]
drm:drm_vblank_event [Tracepoint event]
drm:drm_vblank_event_queued [Tracepoint event]
drm:drm_vblank_event_delivered [Tracepoint event]
random:mix_pool_bytes [Tracepoint event]
random:mix_pool_bytes_nolock [Tracepoint event]
random:credit_entropy_bits [Tracepoint event]
gpio:gpio_direction [Tracepoint event]
gpio:gpio_value [Tracepoint event]
block:block_rq_abort [Tracepoint event]
block:block_rq_requeue [Tracepoint event]
block:block_rq_issue [Tracepoint event]
block:block_bio_bounce [Tracepoint event]
block:block_bio_complete [Tracepoint event]
block:block_bio_backmerge [Tracepoint event]
.
.
writeback:writeback_wake_thread [Tracepoint event]
writeback:writeback_wake_forker_thread [Tracepoint event]
writeback:writeback_bdi_register [Tracepoint event]
.
.
writeback:writeback_single_inode_requeue [Tracepoint event]
writeback:writeback_single_inode [Tracepoint event]
kmem:kmalloc [Tracepoint event]
kmem:kmem_cache_alloc [Tracepoint event]
kmem:mm_page_alloc [Tracepoint event]
kmem:mm_page_alloc_zone_locked [Tracepoint event]
kmem:mm_page_pcpu_drain [Tracepoint event]
kmem:mm_page_alloc_extfrag [Tracepoint event]
vmscan:mm_vmscan_kswapd_sleep [Tracepoint event]
vmscan:mm_vmscan_kswapd_wake [Tracepoint event]
vmscan:mm_vmscan_wakeup_kswapd [Tracepoint event]
vmscan:mm_vmscan_direct_reclaim_begin [Tracepoint event]
.
.
module:module_get [Tracepoint event]
module:module_put [Tracepoint event]
module:module_request [Tracepoint event]
sched:sched_kthread_stop [Tracepoint event]
sched:sched_wakeup [Tracepoint event]
sched:sched_wakeup_new [Tracepoint event]
sched:sched_process_fork [Tracepoint event]
sched:sched_process_exec [Tracepoint event]
sched:sched_stat_runtime [Tracepoint event]
rcu:rcu_utilization [Tracepoint event]
workqueue:workqueue_queue_work [Tracepoint event]
workqueue:workqueue_execute_end [Tracepoint event]
signal:signal_generate [Tracepoint event]
signal:signal_deliver [Tracepoint event]
timer:timer_init [Tracepoint event]
timer:timer_start [Tracepoint event]
timer:hrtimer_cancel [Tracepoint event]
timer:itimer_state [Tracepoint event]
timer:itimer_expire [Tracepoint event]
irq:irq_handler_entry [Tracepoint event]
irq:irq_handler_exit [Tracepoint event]
irq:softirq_entry [Tracepoint event]
irq:softirq_exit [Tracepoint event]
irq:softirq_raise [Tracepoint event]
printk:console [Tracepoint event]
task:task_newtask [Tracepoint event]
task:task_rename [Tracepoint event]
syscalls:sys_enter_socketcall [Tracepoint event]
syscalls:sys_exit_socketcall [Tracepoint event]
.
.
.
syscalls:sys_enter_unshare [Tracepoint event]
syscalls:sys_exit_unshare [Tracepoint event]
raw_syscalls:sys_enter [Tracepoint event]
raw_syscalls:sys_exit [Tracepoint event]
Only a subset of these would be of interest to us when looking at this workload, so let’s choose the most likely subsystems (identified by the string before the colon in the Tracepoint events) and do a ‘perf stat’ run using only those wildcarded subsystems:
root@crownbay:~# perf stat -e skb:* -e net:* -e napi:* -e sched:* -e workqueue:* -e irq:* -e syscalls:* wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2
Performance counter stats for 'wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2':
23323 skb:kfree_skb
0 skb:consume_skb
49897 skb:skb_copy_datagram_iovec
6217 net:net_dev_xmit
6217 net:net_dev_queue
7962 net:netif_receive_skb
2 net:netif_rx
8340 napi:napi_poll
0 sched:sched_kthread_stop
0 sched:sched_kthread_stop_ret
3749 sched:sched_wakeup
0 sched:sched_wakeup_new
0 sched:sched_switch
29 sched:sched_migrate_task
0 sched:sched_process_free
1 sched:sched_process_exit
0 sched:sched_wait_task
0 sched:sched_process_wait
0 sched:sched_process_fork
1 sched:sched_process_exec
0 sched:sched_stat_wait
2106519415641 sched:sched_stat_sleep
0 sched:sched_stat_iowait
147453613 sched:sched_stat_blocked
12903026955 sched:sched_stat_runtime
0 sched:sched_pi_setprio
3574 workqueue:workqueue_queue_work
3574 workqueue:workqueue_activate_work
0 workqueue:workqueue_execute_start
0 workqueue:workqueue_execute_end
16631 irq:irq_handler_entry
16631 irq:irq_handler_exit
28521 irq:softirq_entry
28521 irq:softirq_exit
28728 irq:softirq_raise
1 syscalls:sys_enter_sendmmsg
1 syscalls:sys_exit_sendmmsg
0 syscalls:sys_enter_recvmmsg
0 syscalls:sys_exit_recvmmsg
14 syscalls:sys_enter_socketcall
14 syscalls:sys_exit_socketcall
.
.
.
16965 syscalls:sys_enter_read
16965 syscalls:sys_exit_read
12854 syscalls:sys_enter_write
12854 syscalls:sys_exit_write
.
.
.
58.029710972 seconds time elapsed
Let’s pick one of these tracepoints and tell perf to do a profile using it as the sampling event:
root@crownbay:~# perf record -g -e sched:sched_wakeup wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2
The screenshot above shows the results of running a profile using sched:sched_switch tracepoint, which shows the relative costs of various paths to sched_wakeup (note that sched_wakeup is the name of the tracepoint - it’s actually defined just inside ttwu_do_wakeup(), which accounts for the function name actually displayed in the profile:
/*
* Mark the task runnable and perform wakeup-preemption.
*/
static void
ttwu_do_wakeup(struct rq *rq, struct task_struct *p, int wake_flags)
{
trace_sched_wakeup(p, true);
.
.
.
}
A couple of the more interesting callchains are expanded and displayed above, basically some network receive paths that presumably end up waking up wget (busybox) when network data is ready.
Note that because tracepoints are normally used for tracing, the default sampling period for tracepoints is 1 i.e. for tracepoints perf will sample on every event occurrence (this can be changed using the -c option). This is in contrast to hardware counters such as for example the default ‘cycles’ hardware counter used for normal profiling, where sampling periods are much higher (in the thousands) because profiling should have as low an overhead as possible and sampling on every cycle would be prohibitively expensive.
3.1.2.2 Using perf to do Basic Tracing
Profiling is a great tool for solving many problems or for getting a high-level view of what’s going on with a workload or across the system. It is however by definition an approximation, as suggested by the most prominent word associated with it, ‘sampling’. On the one hand, it allows a representative picture of what’s going on in the system to be cheaply taken, but on the other hand, that cheapness limits its utility when that data suggests a need to ‘dive down’ more deeply to discover what’s really going on. In such cases, the only way to see what’s really going on is to be able to look at (or summarize more intelligently) the individual steps that go into the higher-level behavior exposed by the coarse-grained profiling data.
As a concrete example, we can trace all the events we think might be applicable to our workload:
root@crownbay:~# perf record -g -e skb:* -e net:* -e napi:* -e sched:sched_switch -e sched:sched_wakeup -e irq:*
-e syscalls:sys_enter_read -e syscalls:sys_exit_read -e syscalls:sys_enter_write -e syscalls:sys_exit_write
wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2
We can look at the raw trace output using ‘perf script’ with no arguments:
root@crownbay:~# perf script
perf 1262 [000] 11624.857082: sys_exit_read: 0x0
perf 1262 [000] 11624.857193: sched_wakeup: comm=migration/0 pid=6 prio=0 success=1 target_cpu=000
wget 1262 [001] 11624.858021: softirq_raise: vec=1 [action=TIMER]
wget 1262 [001] 11624.858074: softirq_entry: vec=1 [action=TIMER]
wget 1262 [001] 11624.858081: softirq_exit: vec=1 [action=TIMER]
wget 1262 [001] 11624.858166: sys_enter_read: fd: 0x0003, buf: 0xbf82c940, count: 0x0200
wget 1262 [001] 11624.858177: sys_exit_read: 0x200
wget 1262 [001] 11624.858878: kfree_skb: skbaddr=0xeb248d80 protocol=0 location=0xc15a5308
wget 1262 [001] 11624.858945: kfree_skb: skbaddr=0xeb248000 protocol=0 location=0xc15a5308
wget 1262 [001] 11624.859020: softirq_raise: vec=1 [action=TIMER]
wget 1262 [001] 11624.859076: softirq_entry: vec=1 [action=TIMER]
wget 1262 [001] 11624.859083: softirq_exit: vec=1 [action=TIMER]
wget 1262 [001] 11624.859167: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400
wget 1262 [001] 11624.859192: sys_exit_read: 0x1d7
wget 1262 [001] 11624.859228: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400
wget 1262 [001] 11624.859233: sys_exit_read: 0x0
wget 1262 [001] 11624.859573: sys_enter_read: fd: 0x0003, buf: 0xbf82c580, count: 0x0200
wget 1262 [001] 11624.859584: sys_exit_read: 0x200
wget 1262 [001] 11624.859864: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400
wget 1262 [001] 11624.859888: sys_exit_read: 0x400
wget 1262 [001] 11624.859935: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400
wget 1262 [001] 11624.859944: sys_exit_read: 0x400
This gives us a detailed timestamped sequence of events that occurred within the workload with respect to those events.
In many ways, profiling can be viewed as a subset of tracing - theoretically, if you have a set of trace events that’s sufficient to capture all the important aspects of a workload, you can derive any of the results or views that a profiling run can.
Another aspect of traditional profiling is that while powerful in many ways, it’s limited by the granularity of the underlying data. Profiling tools offer various ways of sorting and presenting the sample data, which make it much more useful and amenable to user experimentation, but in the end it can’t be used in an open-ended way to extract data that just isn’t present as a consequence of the fact that conceptually, most of it has been thrown away.
Full-blown detailed tracing data does however offer the opportunity to manipulate and present the information collected during a tracing run in an infinite variety of ways.
Another way to look at it is that there are only so many ways that the ‘primitive’ counters can be used on their own to generate interesting output; to get anything more complicated than simple counts requires some amount of additional logic, which is typically very specific to the problem at hand. For example, if we wanted to make use of a ‘counter’ that maps to the value of the time difference between when a process was scheduled to run on a processor and the time it actually ran, we wouldn’t expect such a counter to exist on its own, but we could derive one called say ‘wakeup_latency’ and use it to extract a useful view of that metric from trace data. Likewise, we really can’t figure out from standard profiling tools how much data every process on the system reads and writes, along with how many of those reads and writes fail completely. If we have sufficient trace data, however, we could with the right tools easily extract and present that information, but we’d need something other than pre-canned profiling tools to do that.
Luckily, there is a general-purpose way to handle such needs, called ‘programming languages’. Making programming languages easily available to apply to such problems given the specific format of data is called a ‘programming language binding’ for that data and language. Perf supports two programming language bindings, one for Python and one for Perl.
Now that we have the trace data in perf.data, we can use ‘perf script -g’ to generate a skeleton script with handlers for the read/write entry/exit events we recorded:
root@crownbay:~# perf script -g python
generated Python script: perf-script.py
The skeleton script simply creates a python function for each event type in the perf.data file. The body of each function simply prints the event name along with its parameters. For example:
def net__netif_rx(event_name, context, common_cpu,
common_secs, common_nsecs, common_pid, common_comm,
skbaddr, len, name):
print_header(event_name, common_cpu, common_secs, common_nsecs,
common_pid, common_comm)
print "skbaddr=%u, len=%u, name=%s\n" % (skbaddr, len, name),
We can run that script directly to print all of the events contained in the perf.data file:
root@crownbay:~# perf script -s perf-script.py
in trace_begin
syscalls__sys_exit_read 0 11624.857082795 1262 perf nr=3, ret=0
sched__sched_wakeup 0 11624.857193498 1262 perf comm=migration/0, pid=6, prio=0, success=1, target_cpu=0
irq__softirq_raise 1 11624.858021635 1262 wget vec=TIMER
irq__softirq_entry 1 11624.858074075 1262 wget vec=TIMER
irq__softirq_exit 1 11624.858081389 1262 wget vec=TIMER
syscalls__sys_enter_read 1 11624.858166434 1262 wget nr=3, fd=3, buf=3213019456, count=512
syscalls__sys_exit_read 1 11624.858177924 1262 wget nr=3, ret=512
skb__kfree_skb 1 11624.858878188 1262 wget skbaddr=3945041280, location=3243922184, protocol=0
skb__kfree_skb 1 11624.858945608 1262 wget skbaddr=3945037824, location=3243922184, protocol=0
irq__softirq_raise 1 11624.859020942 1262 wget vec=TIMER
irq__softirq_entry 1 11624.859076935 1262 wget vec=TIMER
irq__softirq_exit 1 11624.859083469 1262 wget vec=TIMER
syscalls__sys_enter_read 1 11624.859167565 1262 wget nr=3, fd=3, buf=3077701632, count=1024
syscalls__sys_exit_read 1 11624.859192533 1262 wget nr=3, ret=471
syscalls__sys_enter_read 1 11624.859228072 1262 wget nr=3, fd=3, buf=3077701632, count=1024
syscalls__sys_exit_read 1 11624.859233707 1262 wget nr=3, ret=0
syscalls__sys_enter_read 1 11624.859573008 1262 wget nr=3, fd=3, buf=3213018496, count=512
syscalls__sys_exit_read 1 11624.859584818 1262 wget nr=3, ret=512
syscalls__sys_enter_read 1 11624.859864562 1262 wget nr=3, fd=3, buf=3077701632, count=1024
syscalls__sys_exit_read 1 11624.859888770 1262 wget nr=3, ret=1024
syscalls__sys_enter_read 1 11624.859935140 1262 wget nr=3, fd=3, buf=3077701632, count=1024
syscalls__sys_exit_read 1 11624.859944032 1262 wget nr=3, ret=1024
That in itself isn’t very useful; after all, we can accomplish pretty much the same thing by simply running ‘perf script’ without arguments in the same directory as the perf.data file.
We can however replace the print statements in the generated function bodies with whatever we want, and thereby make it infinitely more useful.
As a simple example, let’s just replace the print statements in the function bodies with a simple function that does nothing but increment a per-event count. When the program is run against a perf.data file, each time a particular event is encountered, a tally is incremented for that event. For example:
def net__netif_rx(event_name, context, common_cpu,
common_secs, common_nsecs, common_pid, common_comm,
skbaddr, len, name):
inc_counts(event_name)
Each event handler function in the generated code is modified to do this. For convenience, we define a common function called inc_counts() that each handler calls; inc_counts() simply tallies a count for each event using the ‘counts’ hash, which is a specialized hash function that does Perl-like autovivification, a capability that’s extremely useful for kinds of multi-level aggregation commonly used in processing traces (see perf’s documentation on the Python language binding for details):
counts = autodict()
def inc_counts(event_name):
try:
counts[event_name] += 1
except TypeError:
counts[event_name] = 1
Finally, at the end of the trace processing run, we want to print the result of all the per-event tallies. For that, we use the special ‘trace_end()’ function:
def trace_end():
for event_name, count in counts.iteritems():
print "%-40s %10s\n" % (event_name, count)
The end result is a summary of all the events recorded in the trace:
skb__skb_copy_datagram_iovec 13148
irq__softirq_entry 4796
irq__irq_handler_exit 3805
irq__softirq_exit 4795
syscalls__sys_enter_write 8990
net__net_dev_xmit 652
skb__kfree_skb 4047
sched__sched_wakeup 1155
irq__irq_handler_entry 3804
irq__softirq_raise 4799
net__net_dev_queue 652
syscalls__sys_enter_read 17599
net__netif_receive_skb 1743
syscalls__sys_exit_read 17598
net__netif_rx 2
napi__napi_poll 1877
syscalls__sys_exit_write 8990
Note that this is pretty much exactly the same information we get from ‘perf stat’, which goes a little way to support the idea mentioned previously that given the right kind of trace data, higher-level profiling-type summaries can be derived from it.
Documentation on using the ‘perf script’ python binding.
3.1.2.3 System-Wide Tracing and Profiling
The examples so far have focused on tracing a particular program or workload - in other words, every profiling run has specified the program to profile in the command-line e.g. ‘perf record wget …’.
It’s also possible, and more interesting in many cases, to run a system-wide profile or trace while running the workload in a separate shell.
To do system-wide profiling or tracing, you typically use the -a flag to ‘perf record’.
To demonstrate this, open up one window and start the profile using the -a flag (press Ctrl-C to stop tracing):
root@crownbay:~# perf record -g -a
^C[ perf record: Woken up 6 times to write data ]
[ perf record: Captured and wrote 1.400 MB perf.data (~61172 samples) ]
In another window, run the wget test:
root@crownbay:~# wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2
Connecting to downloads.yoctoproject.org (140.211.169.59:80)
linux-2.6.19.2.tar.b 100% \|*******************************\| 41727k 0:00:00 ETA
Here we see entries not only for our wget load, but for other processes running on the system as well:
In the snapshot above, we can see callchains that originate in libc, and a callchain from Xorg that demonstrates that we’re using a proprietary X driver in userspace (notice the presence of ‘PVR’ and some other unresolvable symbols in the expanded Xorg callchain).
Note also that we have both kernel and userspace entries in the above snapshot. We can also tell perf to focus on userspace but providing a modifier, in this case ‘u’, to the ‘cycles’ hardware counter when we record a profile:
root@crownbay:~# perf record -g -a -e cycles:u
^C[ perf record: Woken up 2 times to write data ]
[ perf record: Captured and wrote 0.376 MB perf.data (~16443 samples) ]
Notice in the screenshot above, we see only userspace entries ([.])
Finally, we can press ‘enter’ on a leaf node and select the ‘Zoom into DSO’ menu item to show only entries associated with a specific DSO. In the screenshot below, we’ve zoomed into the ‘libc’ DSO which shows all the entries associated with the libc-xxx.so DSO.
We can also use the system-wide -a switch to do system-wide tracing. Here we’ll trace a couple of scheduler events:
root@crownbay:~# perf record -a -e sched:sched_switch -e sched:sched_wakeup
^C[ perf record: Woken up 38 times to write data ]
[ perf record: Captured and wrote 9.780 MB perf.data (~427299 samples) ]
We can look at the raw output using ‘perf script’ with no arguments:
root@crownbay:~# perf script
perf 1383 [001] 6171.460045: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
perf 1383 [001] 6171.460066: sched_switch: prev_comm=perf prev_pid=1383 prev_prio=120 prev_state=R+ ==> next_comm=kworker/1:1 next_pid=21 next_prio=120
kworker/1:1 21 [001] 6171.460093: sched_switch: prev_comm=kworker/1:1 prev_pid=21 prev_prio=120 prev_state=S ==> next_comm=perf next_pid=1383 next_prio=120
swapper 0 [000] 6171.468063: sched_wakeup: comm=kworker/0:3 pid=1209 prio=120 success=1 target_cpu=000
swapper 0 [000] 6171.468107: sched_switch: prev_comm=swapper/0 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/0:3 next_pid=1209 next_prio=120
kworker/0:3 1209 [000] 6171.468143: sched_switch: prev_comm=kworker/0:3 prev_pid=1209 prev_prio=120 prev_state=S ==> next_comm=swapper/0 next_pid=0 next_prio=120
perf 1383 [001] 6171.470039: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
perf 1383 [001] 6171.470058: sched_switch: prev_comm=perf prev_pid=1383 prev_prio=120 prev_state=R+ ==> next_comm=kworker/1:1 next_pid=21 next_prio=120
kworker/1:1 21 [001] 6171.470082: sched_switch: prev_comm=kworker/1:1 prev_pid=21 prev_prio=120 prev_state=S ==> next_comm=perf next_pid=1383 next_prio=120
perf 1383 [001] 6171.480035: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
3.1.2.3.1 Filtering
Notice that there are a lot of events that don’t really have anything to do with what we’re interested in, namely events that schedule ‘perf’ itself in and out or that wake perf up. We can get rid of those by using the ‘–filter’ option - for each event we specify using -e, we can add a –filter after that to filter out trace events that contain fields with specific values:
root@crownbay:~# perf record -a -e sched:sched_switch --filter 'next_comm != perf && prev_comm != perf' -e sched:sched_wakeup --filter 'comm != perf'
^C[ perf record: Woken up 38 times to write data ]
[ perf record: Captured and wrote 9.688 MB perf.data (~423279 samples) ]
root@crownbay:~# perf script
swapper 0 [000] 7932.162180: sched_switch: prev_comm=swapper/0 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/0:3 next_pid=1209 next_prio=120
kworker/0:3 1209 [000] 7932.162236: sched_switch: prev_comm=kworker/0:3 prev_pid=1209 prev_prio=120 prev_state=S ==> next_comm=swapper/0 next_pid=0 next_prio=120
perf 1407 [001] 7932.170048: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
perf 1407 [001] 7932.180044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
perf 1407 [001] 7932.190038: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
perf 1407 [001] 7932.200044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
perf 1407 [001] 7932.210044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
perf 1407 [001] 7932.220044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
swapper 0 [001] 7932.230111: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
swapper 0 [001] 7932.230146: sched_switch: prev_comm=swapper/1 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/1:1 next_pid=21 next_prio=120
kworker/1:1 21 [001] 7932.230205: sched_switch: prev_comm=kworker/1:1 prev_pid=21 prev_prio=120 prev_state=S ==> next_comm=swapper/1 next_pid=0 next_prio=120
swapper 0 [000] 7932.326109: sched_wakeup: comm=kworker/0:3 pid=1209 prio=120 success=1 target_cpu=000
swapper 0 [000] 7932.326171: sched_switch: prev_comm=swapper/0 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/0:3 next_pid=1209 next_prio=120
kworker/0:3 1209 [000] 7932.326214: sched_switch: prev_comm=kworker/0:3 prev_pid=1209 prev_prio=120 prev_state=S ==> next_comm=swapper/0 next_pid=0 next_prio=120
In this case, we’ve filtered out all events that have ‘perf’ in their ‘comm’ or ‘comm_prev’ or ‘comm_next’ fields. Notice that there are still events recorded for perf, but notice that those events don’t have values of ‘perf’ for the filtered fields. To completely filter out anything from perf will require a bit more work, but for the purpose of demonstrating how to use filters, it’s close enough.
3.1.2.4 Using Dynamic Tracepoints
perf isn’t restricted to the fixed set of static tracepoints listed by ‘perf list’. Users can also add their own ‘dynamic’ tracepoints anywhere in the kernel. For instance, suppose we want to define our own tracepoint on do_fork(). We can do that using the ‘perf probe’ perf subcommand:
root@crownbay:~# perf probe do_fork
Added new event:
probe:do_fork (on do_fork)
You can now use it in all perf tools, such as:
perf record -e probe:do_fork -aR sleep 1
Adding a new tracepoint via ‘perf probe’ results in an event with all the expected files and format in /sys/kernel/debug/tracing/events, just the same as for static tracepoints (as discussed in more detail in the trace events subsystem section:
root@crownbay:/sys/kernel/debug/tracing/events/probe/do_fork# ls -al
drwxr-xr-x 2 root root 0 Oct 28 11:42 .
drwxr-xr-x 3 root root 0 Oct 28 11:42 ..
-rw-r--r-- 1 root root 0 Oct 28 11:42 enable
-rw-r--r-- 1 root root 0 Oct 28 11:42 filter
-r--r--r-- 1 root root 0 Oct 28 11:42 format
-r--r--r-- 1 root root 0 Oct 28 11:42 id
root@crownbay:/sys/kernel/debug/tracing/events/probe/do_fork# cat format
name: do_fork
ID: 944
format:
field:unsigned short common_type; offset:0; size:2; signed:0;
field:unsigned char common_flags; offset:2; size:1; signed:0;
field:unsigned char common_preempt_count; offset:3; size:1; signed:0;
field:int common_pid; offset:4; size:4; signed:1;
field:int common_padding; offset:8; size:4; signed:1;
field:unsigned long __probe_ip; offset:12; size:4; signed:0;
print fmt: "(%lx)", REC->__probe_ip
We can list all dynamic tracepoints currently in existence:
root@crownbay:~# perf probe -l
probe:do_fork (on do_fork)
probe:schedule (on schedule)
Let’s record system-wide (‘sleep 30’ is a trick for recording system-wide but basically do nothing and then wake up after 30 seconds):
root@crownbay:~# perf record -g -a -e probe:do_fork sleep 30
[ perf record: Woken up 1 times to write data ]
[ perf record: Captured and wrote 0.087 MB perf.data (~3812 samples) ]
Using ‘perf script’ we can see each do_fork event that fired:
root@crownbay:~# perf script
# ========
# captured on: Sun Oct 28 11:55:18 2012
# hostname : crownbay
# os release : 3.4.11-yocto-standard
# perf version : 3.4.11
# arch : i686
# nrcpus online : 2
# nrcpus avail : 2
# cpudesc : Intel(R) Atom(TM) CPU E660 @ 1.30GHz
# cpuid : GenuineIntel,6,38,1
# total memory : 1017184 kB
# cmdline : /usr/bin/perf record -g -a -e probe:do_fork sleep 30
# event : name = probe:do_fork, type = 2, config = 0x3b0, config1 = 0x0, config2 = 0x0, excl_usr = 0, excl_kern
= 0, id = { 5, 6 }
# HEADER_CPU_TOPOLOGY info available, use -I to display
# ========
#
matchbox-deskto 1197 [001] 34211.378318: do_fork: (c1028460)
matchbox-deskto 1295 [001] 34211.380388: do_fork: (c1028460)
pcmanfm 1296 [000] 34211.632350: do_fork: (c1028460)
pcmanfm 1296 [000] 34211.639917: do_fork: (c1028460)
matchbox-deskto 1197 [001] 34217.541603: do_fork: (c1028460)
matchbox-deskto 1299 [001] 34217.543584: do_fork: (c1028460)
gthumb 1300 [001] 34217.697451: do_fork: (c1028460)
gthumb 1300 [001] 34219.085734: do_fork: (c1028460)
gthumb 1300 [000] 34219.121351: do_fork: (c1028460)
gthumb 1300 [001] 34219.264551: do_fork: (c1028460)
pcmanfm 1296 [000] 34219.590380: do_fork: (c1028460)
matchbox-deskto 1197 [001] 34224.955965: do_fork: (c1028460)
matchbox-deskto 1306 [001] 34224.957972: do_fork: (c1028460)
matchbox-termin 1307 [000] 34225.038214: do_fork: (c1028460)
matchbox-termin 1307 [001] 34225.044218: do_fork: (c1028460)
matchbox-termin 1307 [000] 34225.046442: do_fork: (c1028460)
matchbox-deskto 1197 [001] 34237.112138: do_fork: (c1028460)
matchbox-deskto 1311 [001] 34237.114106: do_fork: (c1028460)
gaku 1312 [000] 34237.202388: do_fork: (c1028460)
And using ‘perf report’ on the same file, we can see the callgraphs from starting a few programs during those 30 seconds:
3.1.3 Perf Documentation
Online versions of the man pages for the commands discussed in this section can be found here:
The ‘perf stat’ manpage.
The ‘perf probe’ manpage.
Documentation on using the ‘perf script’ python binding.
The top-level perf(1) manpage.
Normally, you should be able to invoke the man pages via perf itself e.g. ‘perf help’ or ‘perf help record’.
However, by default Yocto doesn’t install man pages, but perf invokes the man pages for most help functionality. This is a bug and is being addressed by a Yocto bug: Bug 3388 - perf: enable man pages for basic ‘help’ functionality.
The man pages in text form, along with some other files, such as a set of examples, can be found in the ‘perf’ directory of the kernel tree:
tools/perf/Documentation
There’s also a nice perf tutorial on the perf wiki that goes into more detail than we do here in certain areas: Perf Tutorial
3.2 ftrace
‘ftrace’ literally refers to the ‘ftrace function tracer’ but in reality this encompasses a number of related tracers along with the infrastructure that they all make use of.
3.2.1 ftrace Setup
For this section, we’ll assume you’ve already performed the basic setup outlined in the “General Setup” section.
ftrace, trace-cmd, and kernelshark run on the target system, and are ready to go out-of-the-box - no additional setup is necessary. For the rest of this section we assume you’ve ssh’ed to the host and will be running ftrace on the target. kernelshark is a GUI application and if you use the ‘-X’ option to ssh you can have the kernelshark GUI run on the target but display remotely on the host if you want.
3.2.2 Basic ftrace usage
‘ftrace’ essentially refers to everything included in the /tracing directory of the mounted debugfs filesystem (Yocto follows the standard convention and mounts it at /sys/kernel/debug). Here’s a listing of all the files found in /sys/kernel/debug/tracing on a Yocto system:
root@sugarbay:/sys/kernel/debug/tracing# ls
README kprobe_events trace
available_events kprobe_profile trace_clock
available_filter_functions options trace_marker
available_tracers per_cpu trace_options
buffer_size_kb printk_formats trace_pipe
buffer_total_size_kb saved_cmdlines tracing_cpumask
current_tracer set_event tracing_enabled
dyn_ftrace_total_info set_ftrace_filter tracing_on
enabled_functions set_ftrace_notrace tracing_thresh
events set_ftrace_pid
free_buffer set_graph_function
The files listed above are used for various purposes - some relate directly to the tracers themselves, others are used to set tracing options, and yet others actually contain the tracing output when a tracer is in effect. Some of the functions can be guessed from their names, others need explanation; in any case, we’ll cover some of the files we see here below but for an explanation of the others, please see the ftrace documentation.
We’ll start by looking at some of the available built-in tracers.
cat’ing the ‘available_tracers’ file lists the set of available tracers:
root@sugarbay:/sys/kernel/debug/tracing# cat available_tracers
blk function_graph function nop
The ‘current_tracer’ file contains the tracer currently in effect:
root@sugarbay:/sys/kernel/debug/tracing# cat current_tracer
nop
The above listing of current_tracer shows that the ‘nop’ tracer is in effect, which is just another way of saying that there’s actually no tracer currently in effect.
echo’ing one of the available_tracers into current_tracer makes the specified tracer the current tracer:
root@sugarbay:/sys/kernel/debug/tracing# echo function > current_tracer
root@sugarbay:/sys/kernel/debug/tracing# cat current_tracer
function
The above sets the current tracer to be the ‘function tracer’. This tracer traces every function call in the kernel and makes it available as the contents of the ‘trace’ file. Reading the ‘trace’ file lists the currently buffered function calls that have been traced by the function tracer:
root@sugarbay:/sys/kernel/debug/tracing# cat trace | less
# tracer: function
#
# entries-in-buffer/entries-written: 310629/766471 #P:8
#
# _-----=> irqs-off
# / _----=> need-resched
# | / _---=> hardirq/softirq
# || / _--=> preempt-depth
# ||| / delay
# TASK-PID CPU# |||| TIMESTAMP FUNCTION
# | | | |||| | |
<idle>-0 [004] d..1 470.867169: ktime_get_real <-intel_idle
<idle>-0 [004] d..1 470.867170: getnstimeofday <-ktime_get_real
<idle>-0 [004] d..1 470.867171: ns_to_timeval <-intel_idle
<idle>-0 [004] d..1 470.867171: ns_to_timespec <-ns_to_timeval
<idle>-0 [004] d..1 470.867172: smp_apic_timer_interrupt <-apic_timer_interrupt
<idle>-0 [004] d..1 470.867172: native_apic_mem_write <-smp_apic_timer_interrupt
<idle>-0 [004] d..1 470.867172: irq_enter <-smp_apic_timer_interrupt
<idle>-0 [004] d..1 470.867172: rcu_irq_enter <-irq_enter
<idle>-0 [004] d..1 470.867173: rcu_idle_exit_common.isra.33 <-rcu_irq_enter
<idle>-0 [004] d..1 470.867173: local_bh_disable <-irq_enter
<idle>-0 [004] d..1 470.867173: add_preempt_count <-local_bh_disable
<idle>-0 [004] d.s1 470.867174: tick_check_idle <-irq_enter
<idle>-0 [004] d.s1 470.867174: tick_check_oneshot_broadcast <-tick_check_idle
<idle>-0 [004] d.s1 470.867174: ktime_get <-tick_check_idle
<idle>-0 [004] d.s1 470.867174: tick_nohz_stop_idle <-tick_check_idle
<idle>-0 [004] d.s1 470.867175: update_ts_time_stats <-tick_nohz_stop_idle
<idle>-0 [004] d.s1 470.867175: nr_iowait_cpu <-update_ts_time_stats
<idle>-0 [004] d.s1 470.867175: tick_do_update_jiffies64 <-tick_check_idle
<idle>-0 [004] d.s1 470.867175: _raw_spin_lock <-tick_do_update_jiffies64
<idle>-0 [004] d.s1 470.867176: add_preempt_count <-_raw_spin_lock
<idle>-0 [004] d.s2 470.867176: do_timer <-tick_do_update_jiffies64
<idle>-0 [004] d.s2 470.867176: _raw_spin_lock <-do_timer
<idle>-0 [004] d.s2 470.867176: add_preempt_count <-_raw_spin_lock
<idle>-0 [004] d.s3 470.867177: ntp_tick_length <-do_timer
<idle>-0 [004] d.s3 470.867177: _raw_spin_lock_irqsave <-ntp_tick_length
.
.
.
Each line in the trace above shows what was happening in the kernel on a given cpu, to the level of detail of function calls. Each entry shows the function called, followed by its caller (after the arrow).
The function tracer gives you an extremely detailed idea of what the kernel was doing at the point in time the trace was taken, and is a great way to learn about how the kernel code works in a dynamic sense.
It is a little more difficult to follow the call chains than it needs to be - luckily there’s a variant of the function tracer that displays the callchains explicitly, called the ‘function_graph’ tracer:
root@sugarbay:/sys/kernel/debug/tracing# echo function_graph > current_tracer
root@sugarbay:/sys/kernel/debug/tracing# cat trace | less
tracer: function_graph
CPU DURATION FUNCTION CALLS
| | | | | | |
7) 0.046 us | pick_next_task_fair();
7) 0.043 us | pick_next_task_stop();
7) 0.042 us | pick_next_task_rt();
7) 0.032 us | pick_next_task_fair();
7) 0.030 us | pick_next_task_idle();
7) | _raw_spin_unlock_irq() {
7) 0.033 us | sub_preempt_count();
7) 0.258 us | }
7) 0.032 us | sub_preempt_count();
7) + 13.341 us | } /* __schedule */
7) 0.095 us | } /* sub_preempt_count */
7) | schedule() {
7) | __schedule() {
7) 0.060 us | add_preempt_count();
7) 0.044 us | rcu_note_context_switch();
7) | _raw_spin_lock_irq() {
7) 0.033 us | add_preempt_count();
7) 0.247 us | }
7) | idle_balance() {
7) | _raw_spin_unlock() {
7) 0.031 us | sub_preempt_count();
7) 0.246 us | }
7) | update_shares() {
7) 0.030 us | __rcu_read_lock();
7) 0.029 us | __rcu_read_unlock();
7) 0.484 us | }
7) 0.030 us | __rcu_read_lock();
7) | load_balance() {
7) | find_busiest_group() {
7) 0.031 us | idle_cpu();
7) 0.029 us | idle_cpu();
7) 0.035 us | idle_cpu();
7) 0.906 us | }
7) 1.141 us | }
7) 0.022 us | msecs_to_jiffies();
7) | load_balance() {
7) | find_busiest_group() {
7) 0.031 us | idle_cpu();
.
.
.
4) 0.062 us | msecs_to_jiffies();
4) 0.062 us | __rcu_read_unlock();
4) | _raw_spin_lock() {
4) 0.073 us | add_preempt_count();
4) 0.562 us | }
4) + 17.452 us | }
4) 0.108 us | put_prev_task_fair();
4) 0.102 us | pick_next_task_fair();
4) 0.084 us | pick_next_task_stop();
4) 0.075 us | pick_next_task_rt();
4) 0.062 us | pick_next_task_fair();
4) 0.066 us | pick_next_task_idle();
------------------------------------------
4) kworker-74 => <idle>-0
------------------------------------------
4) | finish_task_switch() {
4) | _raw_spin_unlock_irq() {
4) 0.100 us | sub_preempt_count();
4) 0.582 us | }
4) 1.105 us | }
4) 0.088 us | sub_preempt_count();
4) ! 100.066 us | }
.
.
.
3) | sys_ioctl() {
3) 0.083 us | fget_light();
3) | security_file_ioctl() {
3) 0.066 us | cap_file_ioctl();
3) 0.562 us | }
3) | do_vfs_ioctl() {
3) | drm_ioctl() {
3) 0.075 us | drm_ut_debug_printk();
3) | i915_gem_pwrite_ioctl() {
3) | i915_mutex_lock_interruptible() {
3) 0.070 us | mutex_lock_interruptible();
3) 0.570 us | }
3) | drm_gem_object_lookup() {
3) | _raw_spin_lock() {
3) 0.080 us | add_preempt_count();
3) 0.620 us | }
3) | _raw_spin_unlock() {
3) 0.085 us | sub_preempt_count();
3) 0.562 us | }
3) 2.149 us | }
3) 0.133 us | i915_gem_object_pin();
3) | i915_gem_object_set_to_gtt_domain() {
3) 0.065 us | i915_gem_object_flush_gpu_write_domain();
3) 0.065 us | i915_gem_object_wait_rendering();
3) 0.062 us | i915_gem_object_flush_cpu_write_domain();
3) 1.612 us | }
3) | i915_gem_object_put_fence() {
3) 0.097 us | i915_gem_object_flush_fence.constprop.36();
3) 0.645 us | }
3) 0.070 us | add_preempt_count();
3) 0.070 us | sub_preempt_count();
3) 0.073 us | i915_gem_object_unpin();
3) 0.068 us | mutex_unlock();
3) 9.924 us | }
3) + 11.236 us | }
3) + 11.770 us | }
3) + 13.784 us | }
3) | sys_ioctl() {
As you can see, the function_graph display is much easier to follow. Also note that in addition to the function calls and associated braces, other events such as scheduler events are displayed in context. In fact, you can freely include any tracepoint available in the trace events subsystem described in the next section by simply enabling those events, and they’ll appear in context in the function graph display. Quite a powerful tool for understanding kernel dynamics.
Also notice that there are various annotations on the left hand side of the display. For example if the total time it took for a given function to execute is above a certain threshold, an exclamation point or plus sign appears on the left hand side. Please see the ftrace documentation for details on all these fields.
3.2.3 The ‘trace events’ Subsystem
One especially important directory contained within the /sys/kernel/debug/tracing directory is the ‘events’ subdirectory, which contains representations of every tracepoint in the system. Listing out the contents of the ‘events’ subdirectory, we see mainly another set of subdirectories:
root@sugarbay:/sys/kernel/debug/tracing# cd events
root@sugarbay:/sys/kernel/debug/tracing/events# ls -al
drwxr-xr-x 38 root root 0 Nov 14 23:19 .
drwxr-xr-x 5 root root 0 Nov 14 23:19 ..
drwxr-xr-x 19 root root 0 Nov 14 23:19 block
drwxr-xr-x 32 root root 0 Nov 14 23:19 btrfs
drwxr-xr-x 5 root root 0 Nov 14 23:19 drm
-rw-r--r-- 1 root root 0 Nov 14 23:19 enable
drwxr-xr-x 40 root root 0 Nov 14 23:19 ext3
drwxr-xr-x 79 root root 0 Nov 14 23:19 ext4
drwxr-xr-x 14 root root 0 Nov 14 23:19 ftrace
drwxr-xr-x 8 root root 0 Nov 14 23:19 hda
-r--r--r-- 1 root root 0 Nov 14 23:19 header_event
-r--r--r-- 1 root root 0 Nov 14 23:19 header_page
drwxr-xr-x 25 root root 0 Nov 14 23:19 i915
drwxr-xr-x 7 root root 0 Nov 14 23:19 irq
drwxr-xr-x 12 root root 0 Nov 14 23:19 jbd
drwxr-xr-x 14 root root 0 Nov 14 23:19 jbd2
drwxr-xr-x 14 root root 0 Nov 14 23:19 kmem
drwxr-xr-x 7 root root 0 Nov 14 23:19 module
drwxr-xr-x 3 root root 0 Nov 14 23:19 napi
drwxr-xr-x 6 root root 0 Nov 14 23:19 net
drwxr-xr-x 3 root root 0 Nov 14 23:19 oom
drwxr-xr-x 12 root root 0 Nov 14 23:19 power
drwxr-xr-x 3 root root 0 Nov 14 23:19 printk
drwxr-xr-x 8 root root 0 Nov 14 23:19 random
drwxr-xr-x 4 root root 0 Nov 14 23:19 raw_syscalls
drwxr-xr-x 3 root root 0 Nov 14 23:19 rcu
drwxr-xr-x 6 root root 0 Nov 14 23:19 rpm
drwxr-xr-x 20 root root 0 Nov 14 23:19 sched
drwxr-xr-x 7 root root 0 Nov 14 23:19 scsi
drwxr-xr-x 4 root root 0 Nov 14 23:19 signal
drwxr-xr-x 5 root root 0 Nov 14 23:19 skb
drwxr-xr-x 4 root root 0 Nov 14 23:19 sock
drwxr-xr-x 10 root root 0 Nov 14 23:19 sunrpc
drwxr-xr-x 538 root root 0 Nov 14 23:19 syscalls
drwxr-xr-x 4 root root 0 Nov 14 23:19 task
drwxr-xr-x 14 root root 0 Nov 14 23:19 timer
drwxr-xr-x 3 root root 0 Nov 14 23:19 udp
drwxr-xr-x 21 root root 0 Nov 14 23:19 vmscan
drwxr-xr-x 3 root root 0 Nov 14 23:19 vsyscall
drwxr-xr-x 6 root root 0 Nov 14 23:19 workqueue
drwxr-xr-x 26 root root 0 Nov 14 23:19 writeback
Each one of these subdirectories corresponds to a ‘subsystem’ and contains yet again more subdirectories, each one of those finally corresponding to a tracepoint. For example, here are the contents of the ‘kmem’ subsystem:
root@sugarbay:/sys/kernel/debug/tracing/events# cd kmem
root@sugarbay:/sys/kernel/debug/tracing/events/kmem# ls -al
drwxr-xr-x 14 root root 0 Nov 14 23:19 .
drwxr-xr-x 38 root root 0 Nov 14 23:19 ..
-rw-r--r-- 1 root root 0 Nov 14 23:19 enable
-rw-r--r-- 1 root root 0 Nov 14 23:19 filter
drwxr-xr-x 2 root root 0 Nov 14 23:19 kfree
drwxr-xr-x 2 root root 0 Nov 14 23:19 kmalloc
drwxr-xr-x 2 root root 0 Nov 14 23:19 kmalloc_node
drwxr-xr-x 2 root root 0 Nov 14 23:19 kmem_cache_alloc
drwxr-xr-x 2 root root 0 Nov 14 23:19 kmem_cache_alloc_node
drwxr-xr-x 2 root root 0 Nov 14 23:19 kmem_cache_free
drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_alloc
drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_alloc_extfrag
drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_alloc_zone_locked
drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_free
drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_free_batched
drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_pcpu_drain
Let’s see what’s inside the subdirectory for a specific tracepoint, in this case the one for kmalloc:
root@sugarbay:/sys/kernel/debug/tracing/events/kmem# cd kmalloc
root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# ls -al
drwxr-xr-x 2 root root 0 Nov 14 23:19 .
drwxr-xr-x 14 root root 0 Nov 14 23:19 ..
-rw-r--r-- 1 root root 0 Nov 14 23:19 enable
-rw-r--r-- 1 root root 0 Nov 14 23:19 filter
-r--r--r-- 1 root root 0 Nov 14 23:19 format
-r--r--r-- 1 root root 0 Nov 14 23:19 id
The ‘format’ file for the tracepoint describes the event in memory, which is used by the various tracing tools that now make use of these tracepoint to parse the event and make sense of it, along with a ‘print fmt’ field that allows tools like ftrace to display the event as text. Here’s what the format of the kmalloc event looks like:
root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# cat format
name: kmalloc
ID: 313
format:
field:unsigned short common_type; offset:0; size:2; signed:0;
field:unsigned char common_flags; offset:2; size:1; signed:0;
field:unsigned char common_preempt_count; offset:3; size:1; signed:0;
field:int common_pid; offset:4; size:4; signed:1;
field:int common_padding; offset:8; size:4; signed:1;
field:unsigned long call_site; offset:16; size:8; signed:0;
field:const void * ptr; offset:24; size:8; signed:0;
field:size_t bytes_req; offset:32; size:8; signed:0;
field:size_t bytes_alloc; offset:40; size:8; signed:0;
field:gfp_t gfp_flags; offset:48; size:4; signed:0;
print fmt: "call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s", REC->call_site, REC->ptr, REC->bytes_req, REC->bytes_alloc,
(REC->gfp_flags) ? __print_flags(REC->gfp_flags, "|", {(unsigned long)(((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | ((
gfp_t)0x20000u) | (( gfp_t)0x02u) | (( gfp_t)0x08u)) | (( gfp_t)0x4000u) | (( gfp_t)0x10000u) | (( gfp_t)0x1000u) | (( gfp_t)0x200u) | ((
gfp_t)0x400000u)), "GFP_TRANSHUGE"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | (( gfp_t)0x20000u) | ((
gfp_t)0x02u) | (( gfp_t)0x08u)), "GFP_HIGHUSER_MOVABLE"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | ((
gfp_t)0x20000u) | (( gfp_t)0x02u)), "GFP_HIGHUSER"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | ((
gfp_t)0x20000u)), "GFP_USER"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | (( gfp_t)0x80000u)), GFP_TEMPORARY"},
{(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u)), "GFP_KERNEL"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u)),
"GFP_NOFS"}, {(unsigned long)((( gfp_t)0x20u)), "GFP_ATOMIC"}, {(unsigned long)((( gfp_t)0x10u)), "GFP_NOIO"}, {(unsigned long)((
gfp_t)0x20u), "GFP_HIGH"}, {(unsigned long)(( gfp_t)0x10u), "GFP_WAIT"}, {(unsigned long)(( gfp_t)0x40u), "GFP_IO"}, {(unsigned long)((
gfp_t)0x100u), "GFP_COLD"}, {(unsigned long)(( gfp_t)0x200u), "GFP_NOWARN"}, {(unsigned long)(( gfp_t)0x400u), "GFP_REPEAT"}, {(unsigned
long)(( gfp_t)0x800u), "GFP_NOFAIL"}, {(unsigned long)(( gfp_t)0x1000u), "GFP_NORETRY"}, {(unsigned long)(( gfp_t)0x4000u), "GFP_COMP"},
{(unsigned long)(( gfp_t)0x8000u), "GFP_ZERO"}, {(unsigned long)(( gfp_t)0x10000u), "GFP_NOMEMALLOC"}, {(unsigned long)(( gfp_t)0x20000u),
"GFP_HARDWALL"}, {(unsigned long)(( gfp_t)0x40000u), "GFP_THISNODE"}, {(unsigned long)(( gfp_t)0x80000u), "GFP_RECLAIMABLE"}, {(unsigned
long)(( gfp_t)0x08u), "GFP_MOVABLE"}, {(unsigned long)(( gfp_t)0), "GFP_NOTRACK"}, {(unsigned long)(( gfp_t)0x400000u), "GFP_NO_KSWAPD"},
{(unsigned long)(( gfp_t)0x800000u), "GFP_OTHER_NODE"} ) : "GFP_NOWAIT"
The ‘enable’ file in the tracepoint directory is what allows the user (or tools such as trace-cmd) to actually turn the tracepoint on and off. When enabled, the corresponding tracepoint will start appearing in the ftrace ‘trace’ file described previously. For example, this turns on the kmalloc tracepoint:
root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# echo 1 > enable
At the moment, we’re not interested in the function tracer or some other tracer that might be in effect, so we first turn it off, but if we do that, we still need to turn tracing on in order to see the events in the output buffer:
root@sugarbay:/sys/kernel/debug/tracing# echo nop > current_tracer
root@sugarbay:/sys/kernel/debug/tracing# echo 1 > tracing_on
Now, if we look at the ‘trace’ file, we see nothing but the kmalloc events we just turned on:
root@sugarbay:/sys/kernel/debug/tracing# cat trace | less
# tracer: nop
#
# entries-in-buffer/entries-written: 1897/1897 #P:8
#
# _-----=> irqs-off
# / _----=> need-resched
# | / _---=> hardirq/softirq
# || / _--=> preempt-depth
# ||| / delay
# TASK-PID CPU# |||| TIMESTAMP FUNCTION
# | | | |||| | |
dropbear-1465 [000] ...1 18154.620753: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
<idle>-0 [000] ..s3 18154.621640: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
<idle>-0 [000] ..s3 18154.621656: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
matchbox-termin-1361 [001] ...1 18154.755472: kmalloc: call_site=ffffffff81614050 ptr=ffff88006d5f0e00 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_KERNEL|GFP_REPEAT
Xorg-1264 [002] ...1 18154.755581: kmalloc: call_site=ffffffff8141abe8 ptr=ffff8800734f4cc0 bytes_req=168 bytes_alloc=192 gfp_flags=GFP_KERNEL|GFP_NOWARN|GFP_NORETRY
Xorg-1264 [002] ...1 18154.755583: kmalloc: call_site=ffffffff814192a3 ptr=ffff88001f822520 bytes_req=24 bytes_alloc=32 gfp_flags=GFP_KERNEL|GFP_ZERO
Xorg-1264 [002] ...1 18154.755589: kmalloc: call_site=ffffffff81419edb ptr=ffff8800721a2f00 bytes_req=64 bytes_alloc=64 gfp_flags=GFP_KERNEL|GFP_ZERO
matchbox-termin-1361 [001] ...1 18155.354594: kmalloc: call_site=ffffffff81614050 ptr=ffff88006db35400 bytes_req=576 bytes_alloc=1024 gfp_flags=GFP_KERNEL|GFP_REPEAT
Xorg-1264 [002] ...1 18155.354703: kmalloc: call_site=ffffffff8141abe8 ptr=ffff8800734f4cc0 bytes_req=168 bytes_alloc=192 gfp_flags=GFP_KERNEL|GFP_NOWARN|GFP_NORETRY
Xorg-1264 [002] ...1 18155.354705: kmalloc: call_site=ffffffff814192a3 ptr=ffff88001f822520 bytes_req=24 bytes_alloc=32 gfp_flags=GFP_KERNEL|GFP_ZERO
Xorg-1264 [002] ...1 18155.354711: kmalloc: call_site=ffffffff81419edb ptr=ffff8800721a2f00 bytes_req=64 bytes_alloc=64 gfp_flags=GFP_KERNEL|GFP_ZERO
<idle>-0 [000] ..s3 18155.673319: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
dropbear-1465 [000] ...1 18155.673525: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
<idle>-0 [000] ..s3 18155.674821: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
<idle>-0 [000] ..s3 18155.793014: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
dropbear-1465 [000] ...1 18155.793219: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
<idle>-0 [000] ..s3 18155.794147: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
<idle>-0 [000] ..s3 18155.936705: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
dropbear-1465 [000] ...1 18155.936910: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
<idle>-0 [000] ..s3 18155.937869: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
matchbox-termin-1361 [001] ...1 18155.953667: kmalloc: call_site=ffffffff81614050 ptr=ffff88006d5f2000 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_KERNEL|GFP_REPEAT
Xorg-1264 [002] ...1 18155.953775: kmalloc: call_site=ffffffff8141abe8 ptr=ffff8800734f4cc0 bytes_req=168 bytes_alloc=192 gfp_flags=GFP_KERNEL|GFP_NOWARN|GFP_NORETRY
Xorg-1264 [002] ...1 18155.953777: kmalloc: call_site=ffffffff814192a3 ptr=ffff88001f822520 bytes_req=24 bytes_alloc=32 gfp_flags=GFP_KERNEL|GFP_ZERO
Xorg-1264 [002] ...1 18155.953783: kmalloc: call_site=ffffffff81419edb ptr=ffff8800721a2f00 bytes_req=64 bytes_alloc=64 gfp_flags=GFP_KERNEL|GFP_ZERO
<idle>-0 [000] ..s3 18156.176053: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
dropbear-1465 [000] ...1 18156.176257: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
<idle>-0 [000] ..s3 18156.177717: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
<idle>-0 [000] ..s3 18156.399229: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
dropbear-1465 [000] ...1 18156.399434: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_http://rostedt.homelinux.com/kernelshark/req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
<idle>-0 [000] ..s3 18156.400660: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
matchbox-termin-1361 [001] ...1 18156.552800: kmalloc: call_site=ffffffff81614050 ptr=ffff88006db34800 bytes_req=576 bytes_alloc=1024 gfp_flags=GFP_KERNEL|GFP_REPEAT
To again disable the kmalloc event, we need to send 0 to the enable file:
root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# echo 0 > enable
You can enable any number of events or complete subsystems (by using the ‘enable’ file in the subsystem directory) and get an arbitrarily fine-grained idea of what’s going on in the system by enabling as many of the appropriate tracepoints as applicable.
A number of the tools described in this HOWTO do just that, including trace-cmd and kernelshark in the next section.
3.2.4 trace-cmd/kernelshark
trace-cmd is essentially an extensive command-line ‘wrapper’ interface that hides the details of all the individual files in /sys/kernel/debug/tracing, allowing users to specify specific particular events within the /sys/kernel/debug/tracing/events/ subdirectory and to collect traces and avoid having to deal with those details directly.
As yet another layer on top of that, kernelshark provides a GUI that allows users to start and stop traces and specify sets of events using an intuitive interface, and view the output as both trace events and as a per-CPU graphical display. It directly uses ‘trace-cmd’ as the plumbing that accomplishes all that underneath the covers (and actually displays the trace-cmd command it uses, as we’ll see).
To start a trace using kernelshark, first start kernelshark:
root@sugarbay:~# kernelshark
Then bring up the ‘Capture’ dialog by choosing from the kernelshark menu:
Capture | Record
That will display the following dialog, which allows you to choose one or more events (or even one or more complete subsystems) to trace:
Note that these are exactly the same sets of events described in the previous trace events subsystem section, and in fact is where trace-cmd gets them for kernelshark.
In the above screenshot, we’ve decided to explore the graphics subsystem a bit and so have chosen to trace all the tracepoints contained within the ‘i915’ and ‘drm’ subsystems.
After doing that, we can start and stop the trace using the ‘Run’ and ‘Stop’ button on the lower right corner of the dialog (the same button will turn into the ‘Stop’ button after the trace has started):
Notice that the right-hand pane shows the exact trace-cmd command-line that’s used to run the trace, along with the results of the trace-cmd run.
Once the ‘Stop’ button is pressed, the graphical view magically fills up with a colorful per-cpu display of the trace data, along with the detailed event listing below that:
Here’s another example, this time a display resulting from tracing ‘all events’:
The tool is pretty self-explanatory, but for more detailed information on navigating through the data, see the kernelshark website.
3.2.5 ftrace Documentation
The documentation for ftrace can be found in the kernel Documentation directory:
Documentation/trace/ftrace.txt
The documentation for the trace event subsystem can also be found in the kernel Documentation directory:
Documentation/trace/events.txt
There is a nice series of articles on using ftrace and trace-cmd at LWN:
There’s more detailed documentation kernelshark usage here: KernelShark
An amusing yet useful README (a tracing mini-HOWTO) can be found in
/sys/kernel/debug/tracing/README
.
3.3 systemtap
SystemTap is a system-wide script-based tracing and profiling tool.
SystemTap scripts are C-like programs that are executed in the kernel to gather/print/aggregate data extracted from the context they end up being invoked under.
For example, this probe from the SystemTap tutorial simply prints a line every time any process on the system open()s a file. For each line, it prints the executable name of the program that opened the file, along with its PID, and the name of the file it opened (or tried to open), which it extracts from the open syscall’s argstr.
probe syscall.open
{
printf ("%s(%d) open (%s)\n", execname(), pid(), argstr)
}
probe timer.ms(4000) # after 4 seconds
{
exit ()
}
Normally, to execute this probe, you’d simply install systemtap on the system you want to probe, and directly run the probe on that system e.g. assuming the name of the file containing the above text is trace_open.stp:
# stap trace_open.stp
What systemtap does under the covers to run this probe is 1) parse and convert the probe to an equivalent ‘C’ form, 2) compile the ‘C’ form into a kernel module, 3) insert the module into the kernel, which arms it, and 4) collect the data generated by the probe and display it to the user.
In order to accomplish steps 1 and 2, the ‘stap’ program needs access to the kernel build system that produced the kernel that the probed system is running. In the case of a typical embedded system (the ‘target’), the kernel build system unfortunately isn’t typically part of the image running on the target. It is normally available on the ‘host’ system that produced the target image however; in such cases, steps 1 and 2 are executed on the host system, and steps 3 and 4 are executed on the target system, using only the systemtap ‘runtime’.
The systemtap support in Yocto assumes that only steps 3 and 4 are run on the target; it is possible to do everything on the target, but this section assumes only the typical embedded use-case.
So basically what you need to do in order to run a systemtap script on the target is to 1) on the host system, compile the probe into a kernel module that makes sense to the target, 2) copy the module onto the target system and 3) insert the module into the target kernel, which arms it, and 4) collect the data generated by the probe and display it to the user.
3.3.1 systemtap Setup
Those are a lot of steps and a lot of details, but fortunately Yocto includes a script called ‘crosstap’ that will take care of those details, allowing you to simply execute a systemtap script on the remote target, with arguments if necessary.
In order to do this from a remote host, however, you need to have access to the build for the image you booted. The ‘crosstap’ script provides details on how to do this if you run the script on the host without having done a build:
$ crosstap root@192.168.1.88 trace_open.stp
Error: No target kernel build found.
Did you forget to create a local build of your image?
'crosstap' requires a local sdk build of the target system
(or a build that includes 'tools-profile') in order to build
kernel modules that can probe the target system.
Practically speaking, that means you need to do the following:
- If you're running a pre-built image, download the release
and/or BSP tarballs used to build the image.
- If you're working from git sources, just clone the metadata
and BSP layers needed to build the image you'll be booting.
- Make sure you're properly set up to build a new image (see
the BSP README and/or the widely available basic documentation
that discusses how to build images).
- Build an -sdk version of the image e.g.:
$ bitbake core-image-sato-sdk
OR
- Build a non-sdk image but include the profiling tools:
[ edit local.conf and add 'tools-profile' to the end of
the EXTRA_IMAGE_FEATURES variable ]
$ bitbake core-image-sato
Once you've build the image on the host system, you're ready to
boot it (or the equivalent pre-built image) and use 'crosstap'
to probe it (you need to source the environment as usual first):
$ source oe-init-build-env
$ cd ~/my/systemtap/scripts
$ crosstap root@192.168.1.xxx myscript.stp
Note
SystemTap, which uses ‘crosstap’, assumes you can establish an ssh connection to the remote target. Please refer to the crosstap wiki page for details on verifying ssh connections at . Also, the ability to ssh into the target system is not enabled by default in *-minimal images.
So essentially what you need to do is build an SDK image or image with ‘tools-profile’ as detailed in the “General Setup” section of this manual, and boot the resulting target image.
Note
If you have a build directory containing multiple machines, you need to have the MACHINE you’re connecting to selected in local.conf, and the kernel in that machine’s build directory must match the kernel on the booted system exactly, or you’ll get the above ‘crosstap’ message when you try to invoke a script.
3.3.2 Running a Script on a Target
Once you’ve done that, you should be able to run a systemtap script on the target:
$ cd /path/to/yocto
$ source oe-init-build-env
### Shell environment set up for builds. ###
You can now run 'bitbake <target>'
Common targets are:
core-image-minimal
core-image-sato
meta-toolchain
meta-ide-support
You can also run generated QEMU images with a command like 'runqemu qemux86-64'
Once you’ve done that, you can cd to whatever directory contains your scripts and use ‘crosstap’ to run the script:
$ cd /path/to/my/systemap/script
$ crosstap root@192.168.7.2 trace_open.stp
If you get an error connecting to the target e.g.:
$ crosstap root@192.168.7.2 trace_open.stp
error establishing ssh connection on remote 'root@192.168.7.2'
Try ssh’ing to the target and see what happens:
$ ssh root@192.168.7.2
A lot of the time, connection problems are due specifying a wrong IP address or having a ‘host key verification error’.
If everything worked as planned, you should see something like this (enter the password when prompted, or press enter if it’s set up to use no password):
$ crosstap root@192.168.7.2 trace_open.stp
root@192.168.7.2's password:
matchbox-termin(1036) open ("/tmp/vte3FS2LW", O_RDWR|O_CREAT|O_EXCL|O_LARGEFILE, 0600)
matchbox-termin(1036) open ("/tmp/vteJMC7LW", O_RDWR|O_CREAT|O_EXCL|O_LARGEFILE, 0600)
3.3.3 systemtap Documentation
The SystemTap language reference can be found here: SystemTap Language Reference
Links to other SystemTap documents, tutorials, and examples can be found here: SystemTap documentation page
3.4 Sysprof
Sysprof is a very easy to use system-wide profiler that consists of a single window with three panes and a few buttons which allow you to start, stop, and view the profile from one place.
3.4.1 Sysprof Setup
For this section, we’ll assume you’ve already performed the basic setup outlined in the “General Setup” section.
Sysprof is a GUI-based application that runs on the target system. For the rest of this document we assume you’ve ssh’ed to the host and will be running Sysprof on the target (you can use the ‘-X’ option to ssh and have the Sysprof GUI run on the target but display remotely on the host if you want).
3.4.2 Basic Sysprof Usage
To start profiling the system, you simply press the ‘Start’ button. To stop profiling and to start viewing the profile data in one easy step, press the ‘Profile’ button.
Once you’ve pressed the profile button, the three panes will fill up with profiling data:
The left pane shows a list of functions and processes. Selecting one of those expands that function in the right pane, showing all its callees. Note that this caller-oriented display is essentially the inverse of perf’s default callee-oriented callchain display.
In the screenshot above, we’re focusing on __copy_to_user_ll()
and
looking up the callchain we can see that one of the callers of
__copy_to_user_ll
is sys_read() and the complete callpath between them.
Notice that this is essentially a portion of the same information we saw
in the perf display shown in the perf section of this page.
Similarly, the above is a snapshot of the Sysprof display of a copy-from-user callchain.
Finally, looking at the third Sysprof pane in the lower left, we can see
a list of all the callers of a particular function selected in the top
left pane. In this case, the lower pane is showing all the callers of
__mark_inode_dirty
:
Double-clicking on one of those functions will in turn change the focus to the selected function, and so on.
3.4.3 Sysprof Documentation
There doesn’t seem to be any documentation for Sysprof, but maybe that’s because it’s pretty self-explanatory. The Sysprof website, however, is here: Sysprof, System-wide Performance Profiler for Linux
3.5 LTTng (Linux Trace Toolkit, next generation)
3.5.1 LTTng Setup
For this section, we’ll assume you’ve already performed the basic setup outlined in the “General Setup” section. LTTng is run on the target system by ssh’ing to it.
3.5.2 Collecting and Viewing Traces
Once you’ve applied the above commits and built and booted your image (you need to build the core-image-sato-sdk image or use one of the other methods described in the “General Setup” section), you’re ready to start tracing.
3.5.2.1 Collecting and viewing a trace on the target (inside a shell)
First, from the host, ssh to the target:
$ ssh -l root 192.168.1.47
The authenticity of host '192.168.1.47 (192.168.1.47)' can't be established.
RSA key fingerprint is 23:bd:c8:b1:a8:71:52:00:ee:00:4f:64:9e:10:b9:7e.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added '192.168.1.47' (RSA) to the list of known hosts.
root@192.168.1.47's password:
Once on the target, use these steps to create a trace:
root@crownbay:~# lttng create
Spawning a session daemon
Session auto-20121015-232120 created.
Traces will be written in /home/root/lttng-traces/auto-20121015-232120
Enable the events you want to trace (in this case all kernel events):
root@crownbay:~# lttng enable-event --kernel --all
All kernel events are enabled in channel channel0
Start the trace:
root@crownbay:~# lttng start
Tracing started for session auto-20121015-232120
And then stop the trace after awhile or after running a particular workload that you want to trace:
root@crownbay:~# lttng stop
Tracing stopped for session auto-20121015-232120
You can now view the trace in text form on the target:
root@crownbay:~# lttng view
[23:21:56.989270399] (+?.?????????) sys_geteuid: { 1 }, { }
[23:21:56.989278081] (+0.000007682) exit_syscall: { 1 }, { ret = 0 }
[23:21:56.989286043] (+0.000007962) sys_pipe: { 1 }, { fildes = 0xB77B9E8C }
[23:21:56.989321802] (+0.000035759) exit_syscall: { 1 }, { ret = 0 }
[23:21:56.989329345] (+0.000007543) sys_mmap_pgoff: { 1 }, { addr = 0x0, len = 10485760, prot = 3, flags = 131362, fd = 4294967295, pgoff = 0 }
[23:21:56.989351694] (+0.000022349) exit_syscall: { 1 }, { ret = -1247805440 }
[23:21:56.989432989] (+0.000081295) sys_clone: { 1 }, { clone_flags = 0x411, newsp = 0xB5EFFFE4, parent_tid = 0xFFFFFFFF, child_tid = 0x0 }
[23:21:56.989477129] (+0.000044140) sched_stat_runtime: { 1 }, { comm = "lttng-consumerd", tid = 1193, runtime = 681660, vruntime = 43367983388 }
[23:21:56.989486697] (+0.000009568) sched_migrate_task: { 1 }, { comm = "lttng-consumerd", tid = 1193, prio = 20, orig_cpu = 1, dest_cpu = 1 }
[23:21:56.989508418] (+0.000021721) hrtimer_init: { 1 }, { hrtimer = 3970832076, clockid = 1, mode = 1 }
[23:21:56.989770462] (+0.000262044) hrtimer_cancel: { 1 }, { hrtimer = 3993865440 }
[23:21:56.989771580] (+0.000001118) hrtimer_cancel: { 0 }, { hrtimer = 3993812192 }
[23:21:56.989776957] (+0.000005377) hrtimer_expire_entry: { 1 }, { hrtimer = 3993865440, now = 79815980007057, function = 3238465232 }
[23:21:56.989778145] (+0.000001188) hrtimer_expire_entry: { 0 }, { hrtimer = 3993812192, now = 79815980008174, function = 3238465232 }
[23:21:56.989791695] (+0.000013550) softirq_raise: { 1 }, { vec = 1 }
[23:21:56.989795396] (+0.000003701) softirq_raise: { 0 }, { vec = 1 }
[23:21:56.989800635] (+0.000005239) softirq_raise: { 0 }, { vec = 9 }
[23:21:56.989807130] (+0.000006495) sched_stat_runtime: { 1 }, { comm = "lttng-consumerd", tid = 1193, runtime = 330710, vruntime = 43368314098 }
[23:21:56.989809993] (+0.000002863) sched_stat_runtime: { 0 }, { comm = "lttng-sessiond", tid = 1181, runtime = 1015313, vruntime = 36976733240 }
[23:21:56.989818514] (+0.000008521) hrtimer_expire_exit: { 0 }, { hrtimer = 3993812192 }
[23:21:56.989819631] (+0.000001117) hrtimer_expire_exit: { 1 }, { hrtimer = 3993865440 }
[23:21:56.989821866] (+0.000002235) hrtimer_start: { 0 }, { hrtimer = 3993812192, function = 3238465232, expires = 79815981000000, softexpires = 79815981000000 }
[23:21:56.989822984] (+0.000001118) hrtimer_start: { 1 }, { hrtimer = 3993865440, function = 3238465232, expires = 79815981000000, softexpires = 79815981000000 }
[23:21:56.989832762] (+0.000009778) softirq_entry: { 1 }, { vec = 1 }
[23:21:56.989833879] (+0.000001117) softirq_entry: { 0 }, { vec = 1 }
[23:21:56.989838069] (+0.000004190) timer_cancel: { 1 }, { timer = 3993871956 }
[23:21:56.989839187] (+0.000001118) timer_cancel: { 0 }, { timer = 3993818708 }
[23:21:56.989841492] (+0.000002305) timer_expire_entry: { 1 }, { timer = 3993871956, now = 79515980, function = 3238277552 }
[23:21:56.989842819] (+0.000001327) timer_expire_entry: { 0 }, { timer = 3993818708, now = 79515980, function = 3238277552 }
[23:21:56.989854831] (+0.000012012) sched_stat_runtime: { 1 }, { comm = "lttng-consumerd", tid = 1193, runtime = 49237, vruntime = 43368363335 }
[23:21:56.989855949] (+0.000001118) sched_stat_runtime: { 0 }, { comm = "lttng-sessiond", tid = 1181, runtime = 45121, vruntime = 36976778361 }
[23:21:56.989861257] (+0.000005308) sched_stat_sleep: { 1 }, { comm = "kworker/1:1", tid = 21, delay = 9451318 }
[23:21:56.989862374] (+0.000001117) sched_stat_sleep: { 0 }, { comm = "kworker/0:0", tid = 4, delay = 9958820 }
[23:21:56.989868241] (+0.000005867) sched_wakeup: { 0 }, { comm = "kworker/0:0", tid = 4, prio = 120, success = 1, target_cpu = 0 }
[23:21:56.989869358] (+0.000001117) sched_wakeup: { 1 }, { comm = "kworker/1:1", tid = 21, prio = 120, success = 1, target_cpu = 1 }
[23:21:56.989877460] (+0.000008102) timer_expire_exit: { 1 }, { timer = 3993871956 }
[23:21:56.989878577] (+0.000001117) timer_expire_exit: { 0 }, { timer = 3993818708 }
.
.
.
You can now safely destroy the trace session (note that this doesn’t delete the trace - it’s still there in ~/lttng-traces):
root@crownbay:~# lttng destroy
Session auto-20121015-232120 destroyed at /home/root
Note that the trace is saved in a directory of the same name as returned by ‘lttng create’, under the ~/lttng-traces directory (note that you can change this by supplying your own name to ‘lttng create’):
root@crownbay:~# ls -al ~/lttng-traces
drwxrwx--- 3 root root 1024 Oct 15 23:21 .
drwxr-xr-x 5 root root 1024 Oct 15 23:57 ..
drwxrwx--- 3 root root 1024 Oct 15 23:21 auto-20121015-232120
3.5.2.2 Collecting and viewing a userspace trace on the target (inside a shell)
For LTTng userspace tracing, you need to have a properly instrumented userspace program. For this example, we’ll use the ‘hello’ test program generated by the lttng-ust build.
The ‘hello’ test program isn’t installed on the rootfs by the lttng-ust build, so we need to copy it over manually. First cd into the build directory that contains the hello executable:
$ cd build/tmp/work/core2_32-poky-linux/lttng-ust/2.0.5-r0/git/tests/hello/.libs
Copy that over to the target machine:
$ scp hello root@192.168.1.20:
You now have the instrumented lttng ‘hello world’ test program on the target, ready to test.
First, from the host, ssh to the target:
$ ssh -l root 192.168.1.47
The authenticity of host '192.168.1.47 (192.168.1.47)' can't be established.
RSA key fingerprint is 23:bd:c8:b1:a8:71:52:00:ee:00:4f:64:9e:10:b9:7e.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added '192.168.1.47' (RSA) to the list of known hosts.
root@192.168.1.47's password:
Once on the target, use these steps to create a trace:
root@crownbay:~# lttng create
Session auto-20190303-021943 created.
Traces will be written in /home/root/lttng-traces/auto-20190303-021943
Enable the events you want to trace (in this case all userspace events):
root@crownbay:~# lttng enable-event --userspace --all
All UST events are enabled in channel channel0
Start the trace:
root@crownbay:~# lttng start
Tracing started for session auto-20190303-021943
Run the instrumented hello world program:
root@crownbay:~# ./hello
Hello, World!
Tracing... done.
And then stop the trace after awhile or after running a particular workload that you want to trace:
root@crownbay:~# lttng stop
Tracing stopped for session auto-20190303-021943
You can now view the trace in text form on the target:
root@crownbay:~# lttng view
[02:31:14.906146544] (+?.?????????) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 0, intfield2 = 0x0, longfield = 0, netintfield = 0, netintfieldhex = 0x0, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 }
[02:31:14.906170360] (+0.000023816) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 1, intfield2 = 0x1, longfield = 1, netintfield = 1, netintfieldhex = 0x1, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 }
[02:31:14.906183140] (+0.000012780) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 2, intfield2 = 0x2, longfield = 2, netintfield = 2, netintfieldhex = 0x2, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 }
[02:31:14.906194385] (+0.000011245) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 3, intfield2 = 0x3, longfield = 3, netintfield = 3, netintfieldhex = 0x3, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 }
.
.
.
You can now safely destroy the trace session (note that this doesn’t delete the trace - it’s still there in ~/lttng-traces):
root@crownbay:~# lttng destroy
Session auto-20190303-021943 destroyed at /home/root
3.5.3 LTTng Documentation
You can find the primary LTTng Documentation on the LTTng Documentation site. The documentation on this site is appropriate for intermediate to advanced software developers who are working in a Linux environment and are interested in efficient software tracing.
For information on LTTng in general, visit the LTTng Project site. You can find a “Getting Started” link on this site that takes you to an LTTng Quick Start.
3.6 blktrace
blktrace is a tool for tracing and reporting low-level disk I/O. blktrace provides the tracing half of the equation; its output can be piped into the blkparse program, which renders the data in a human-readable form and does some basic analysis:
3.6.1 blktrace Setup
For this section, we’ll assume you’ve already performed the basic setup outlined in the “General Setup” section.
blktrace is an application that runs on the target system. You can run the entire blktrace and blkparse pipeline on the target, or you can run blktrace in ‘listen’ mode on the target and have blktrace and blkparse collect and analyze the data on the host (see the “Using blktrace Remotely” section below). For the rest of this section we assume you’ve ssh’ed to the host and will be running blkrace on the target.
3.6.2 Basic blktrace Usage
To record a trace, simply run the ‘blktrace’ command, giving it the name of the block device you want to trace activity on:
root@crownbay:~# blktrace /dev/sdc
In another shell, execute a workload you want to trace.
root@crownbay:/media/sdc# rm linux-2.6.19.2.tar.bz2; wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2; sync
Connecting to downloads.yoctoproject.org (140.211.169.59:80)
linux-2.6.19.2.tar.b 100% \|*******************************\| 41727k 0:00:00 ETA
Press Ctrl-C in the blktrace shell to stop the trace. It will display how many events were logged, along with the per-cpu file sizes (blktrace records traces in per-cpu kernel buffers and simply dumps them to userspace for blkparse to merge and sort later).
^C=== sdc ===
CPU 0: 7082 events, 332 KiB data
CPU 1: 1578 events, 74 KiB data
Total: 8660 events (dropped 0), 406 KiB data
If you examine the files saved to disk, you see multiple files, one per CPU and with the device name as the first part of the filename:
root@crownbay:~# ls -al
drwxr-xr-x 6 root root 1024 Oct 27 22:39 .
drwxr-sr-x 4 root root 1024 Oct 26 18:24 ..
-rw-r--r-- 1 root root 339938 Oct 27 22:40 sdc.blktrace.0
-rw-r--r-- 1 root root 75753 Oct 27 22:40 sdc.blktrace.1
To view the trace events, simply invoke ‘blkparse’ in the directory containing the trace files, giving it the device name that forms the first part of the filenames:
root@crownbay:~# blkparse sdc
8,32 1 1 0.000000000 1225 Q WS 3417048 + 8 [jbd2/sdc-8]
8,32 1 2 0.000025213 1225 G WS 3417048 + 8 [jbd2/sdc-8]
8,32 1 3 0.000033384 1225 P N [jbd2/sdc-8]
8,32 1 4 0.000043301 1225 I WS 3417048 + 8 [jbd2/sdc-8]
8,32 1 0 0.000057270 0 m N cfq1225 insert_request
8,32 1 0 0.000064813 0 m N cfq1225 add_to_rr
8,32 1 5 0.000076336 1225 U N [jbd2/sdc-8] 1
8,32 1 0 0.000088559 0 m N cfq workload slice:150
8,32 1 0 0.000097359 0 m N cfq1225 set_active wl_prio:0 wl_type:1
8,32 1 0 0.000104063 0 m N cfq1225 Not idling. st->count:1
8,32 1 0 0.000112584 0 m N cfq1225 fifo= (null)
8,32 1 0 0.000118730 0 m N cfq1225 dispatch_insert
8,32 1 0 0.000127390 0 m N cfq1225 dispatched a request
8,32 1 0 0.000133536 0 m N cfq1225 activate rq, drv=1
8,32 1 6 0.000136889 1225 D WS 3417048 + 8 [jbd2/sdc-8]
8,32 1 7 0.000360381 1225 Q WS 3417056 + 8 [jbd2/sdc-8]
8,32 1 8 0.000377422 1225 G WS 3417056 + 8 [jbd2/sdc-8]
8,32 1 9 0.000388876 1225 P N [jbd2/sdc-8]
8,32 1 10 0.000397886 1225 Q WS 3417064 + 8 [jbd2/sdc-8]
8,32 1 11 0.000404800 1225 M WS 3417064 + 8 [jbd2/sdc-8]
8,32 1 12 0.000412343 1225 Q WS 3417072 + 8 [jbd2/sdc-8]
8,32 1 13 0.000416533 1225 M WS 3417072 + 8 [jbd2/sdc-8]
8,32 1 14 0.000422121 1225 Q WS 3417080 + 8 [jbd2/sdc-8]
8,32 1 15 0.000425194 1225 M WS 3417080 + 8 [jbd2/sdc-8]
8,32 1 16 0.000431968 1225 Q WS 3417088 + 8 [jbd2/sdc-8]
8,32 1 17 0.000435251 1225 M WS 3417088 + 8 [jbd2/sdc-8]
8,32 1 18 0.000440279 1225 Q WS 3417096 + 8 [jbd2/sdc-8]
8,32 1 19 0.000443911 1225 M WS 3417096 + 8 [jbd2/sdc-8]
8,32 1 20 0.000450336 1225 Q WS 3417104 + 8 [jbd2/sdc-8]
8,32 1 21 0.000454038 1225 M WS 3417104 + 8 [jbd2/sdc-8]
8,32 1 22 0.000462070 1225 Q WS 3417112 + 8 [jbd2/sdc-8]
8,32 1 23 0.000465422 1225 M WS 3417112 + 8 [jbd2/sdc-8]
8,32 1 24 0.000474222 1225 I WS 3417056 + 64 [jbd2/sdc-8]
8,32 1 0 0.000483022 0 m N cfq1225 insert_request
8,32 1 25 0.000489727 1225 U N [jbd2/sdc-8] 1
8,32 1 0 0.000498457 0 m N cfq1225 Not idling. st->count:1
8,32 1 0 0.000503765 0 m N cfq1225 dispatch_insert
8,32 1 0 0.000512914 0 m N cfq1225 dispatched a request
8,32 1 0 0.000518851 0 m N cfq1225 activate rq, drv=2
.
.
.
8,32 0 0 58.515006138 0 m N cfq3551 complete rqnoidle 1
8,32 0 2024 58.516603269 3 C WS 3156992 + 16 [0]
8,32 0 0 58.516626736 0 m N cfq3551 complete rqnoidle 1
8,32 0 0 58.516634558 0 m N cfq3551 arm_idle: 8 group_idle: 0
8,32 0 0 58.516636933 0 m N cfq schedule dispatch
8,32 1 0 58.516971613 0 m N cfq3551 slice expired t=0
8,32 1 0 58.516982089 0 m N cfq3551 sl_used=13 disp=6 charge=13 iops=0 sect=80
8,32 1 0 58.516985511 0 m N cfq3551 del_from_rr
8,32 1 0 58.516990819 0 m N cfq3551 put_queue
CPU0 (sdc):
Reads Queued: 0, 0KiB Writes Queued: 331, 26,284KiB
Read Dispatches: 0, 0KiB Write Dispatches: 485, 40,484KiB
Reads Requeued: 0 Writes Requeued: 0
Reads Completed: 0, 0KiB Writes Completed: 511, 41,000KiB
Read Merges: 0, 0KiB Write Merges: 13, 160KiB
Read depth: 0 Write depth: 2
IO unplugs: 23 Timer unplugs: 0
CPU1 (sdc):
Reads Queued: 0, 0KiB Writes Queued: 249, 15,800KiB
Read Dispatches: 0, 0KiB Write Dispatches: 42, 1,600KiB
Reads Requeued: 0 Writes Requeued: 0
Reads Completed: 0, 0KiB Writes Completed: 16, 1,084KiB
Read Merges: 0, 0KiB Write Merges: 40, 276KiB
Read depth: 0 Write depth: 2
IO unplugs: 30 Timer unplugs: 1
Total (sdc):
Reads Queued: 0, 0KiB Writes Queued: 580, 42,084KiB
Read Dispatches: 0, 0KiB Write Dispatches: 527, 42,084KiB
Reads Requeued: 0 Writes Requeued: 0
Reads Completed: 0, 0KiB Writes Completed: 527, 42,084KiB
Read Merges: 0, 0KiB Write Merges: 53, 436KiB
IO unplugs: 53 Timer unplugs: 1
Throughput (R/W): 0KiB/s / 719KiB/s
Events (sdc): 6,592 entries
Skips: 0 forward (0 - 0.0%)
Input file sdc.blktrace.0 added
Input file sdc.blktrace.1 added
The report shows each event that was found in the blktrace data, along with a summary of the overall block I/O traffic during the run. You can look at the blkparse manpage to learn the meaning of each field displayed in the trace listing.
3.6.2.1 Live Mode
blktrace and blkparse are designed from the ground up to be able to operate together in a ‘pipe mode’ where the stdout of blktrace can be fed directly into the stdin of blkparse:
root@crownbay:~# blktrace /dev/sdc -o - | blkparse -i -
This enables long-lived tracing sessions to run without writing anything to disk, and allows the user to look for certain conditions in the trace data in ‘real-time’ by viewing the trace output as it scrolls by on the screen or by passing it along to yet another program in the pipeline such as grep which can be used to identify and capture conditions of interest.
There’s actually another blktrace command that implements the above pipeline as a single command, so the user doesn’t have to bother typing in the above command sequence:
root@crownbay:~# btrace /dev/sdc
3.6.2.2 Using blktrace Remotely
Because blktrace traces block I/O and at the same time normally writes its trace data to a block device, and in general because it’s not really a great idea to make the device being traced the same as the device the tracer writes to, blktrace provides a way to trace without perturbing the traced device at all by providing native support for sending all trace data over the network.
To have blktrace operate in this mode, start blktrace on the target system being traced with the -l option, along with the device to trace:
root@crownbay:~# blktrace -l /dev/sdc
server: waiting for connections...
On the host system, use the -h option to connect to the target system, also passing it the device to trace:
$ blktrace -d /dev/sdc -h 192.168.1.43
blktrace: connecting to 192.168.1.43
blktrace: connected!
On the target system, you should see this:
server: connection from 192.168.1.43
In another shell, execute a workload you want to trace.
root@crownbay:/media/sdc# rm linux-2.6.19.2.tar.bz2; wget http://downloads.yoctoproject.org/mirror/sources/linux-2.6.19.2.tar.bz2; sync
Connecting to downloads.yoctoproject.org (140.211.169.59:80)
linux-2.6.19.2.tar.b 100% \|*******************************\| 41727k 0:00:00 ETA
When it’s done, do a Ctrl-C on the host system to stop the trace:
^C=== sdc ===
CPU 0: 7691 events, 361 KiB data
CPU 1: 4109 events, 193 KiB data
Total: 11800 events (dropped 0), 554 KiB data
On the target system, you should also see a trace summary for the trace just ended:
server: end of run for 192.168.1.43:sdc
=== sdc ===
CPU 0: 7691 events, 361 KiB data
CPU 1: 4109 events, 193 KiB data
Total: 11800 events (dropped 0), 554 KiB data
The blktrace instance on the host will save the target output inside a hostname-timestamp directory:
$ ls -al
drwxr-xr-x 10 root root 1024 Oct 28 02:40 .
drwxr-sr-x 4 root root 1024 Oct 26 18:24 ..
drwxr-xr-x 2 root root 1024 Oct 28 02:40 192.168.1.43-2012-10-28-02:40:56
cd into that directory to see the output files:
$ ls -l
-rw-r--r-- 1 root root 369193 Oct 28 02:44 sdc.blktrace.0
-rw-r--r-- 1 root root 197278 Oct 28 02:44 sdc.blktrace.1
And run blkparse on the host system using the device name:
$ blkparse sdc
8,32 1 1 0.000000000 1263 Q RM 6016 + 8 [ls]
8,32 1 0 0.000036038 0 m N cfq1263 alloced
8,32 1 2 0.000039390 1263 G RM 6016 + 8 [ls]
8,32 1 3 0.000049168 1263 I RM 6016 + 8 [ls]
8,32 1 0 0.000056152 0 m N cfq1263 insert_request
8,32 1 0 0.000061600 0 m N cfq1263 add_to_rr
8,32 1 0 0.000075498 0 m N cfq workload slice:300
.
.
.
8,32 0 0 177.266385696 0 m N cfq1267 arm_idle: 8 group_idle: 0
8,32 0 0 177.266388140 0 m N cfq schedule dispatch
8,32 1 0 177.266679239 0 m N cfq1267 slice expired t=0
8,32 1 0 177.266689297 0 m N cfq1267 sl_used=9 disp=6 charge=9 iops=0 sect=56
8,32 1 0 177.266692649 0 m N cfq1267 del_from_rr
8,32 1 0 177.266696560 0 m N cfq1267 put_queue
CPU0 (sdc):
Reads Queued: 0, 0KiB Writes Queued: 270, 21,708KiB
Read Dispatches: 59, 2,628KiB Write Dispatches: 495, 39,964KiB
Reads Requeued: 0 Writes Requeued: 0
Reads Completed: 90, 2,752KiB Writes Completed: 543, 41,596KiB
Read Merges: 0, 0KiB Write Merges: 9, 344KiB
Read depth: 2 Write depth: 2
IO unplugs: 20 Timer unplugs: 1
CPU1 (sdc):
Reads Queued: 688, 2,752KiB Writes Queued: 381, 20,652KiB
Read Dispatches: 31, 124KiB Write Dispatches: 59, 2,396KiB
Reads Requeued: 0 Writes Requeued: 0
Reads Completed: 0, 0KiB Writes Completed: 11, 764KiB
Read Merges: 598, 2,392KiB Write Merges: 88, 448KiB
Read depth: 2 Write depth: 2
IO unplugs: 52 Timer unplugs: 0
Total (sdc):
Reads Queued: 688, 2,752KiB Writes Queued: 651, 42,360KiB
Read Dispatches: 90, 2,752KiB Write Dispatches: 554, 42,360KiB
Reads Requeued: 0 Writes Requeued: 0
Reads Completed: 90, 2,752KiB Writes Completed: 554, 42,360KiB
Read Merges: 598, 2,392KiB Write Merges: 97, 792KiB
IO unplugs: 72 Timer unplugs: 1
Throughput (R/W): 15KiB/s / 238KiB/s
Events (sdc): 9,301 entries
Skips: 0 forward (0 - 0.0%)
You should see the trace events and summary just as you would have if you’d run the same command on the target.
3.6.2.3 Tracing Block I/O via ‘ftrace’
It’s also possible to trace block I/O using only The ‘trace events’ Subsystem, which can be useful for casual tracing if you don’t want to bother dealing with the userspace tools.
To enable tracing for a given device, use /sys/block/xxx/trace/enable, where xxx is the device name. This for example enables tracing for /dev/sdc:
root@crownbay:/sys/kernel/debug/tracing# echo 1 > /sys/block/sdc/trace/enable
Once you’ve selected the device(s) you want to trace, selecting the ‘blk’ tracer will turn the blk tracer on:
root@crownbay:/sys/kernel/debug/tracing# cat available_tracers
blk function_graph function nop
root@crownbay:/sys/kernel/debug/tracing# echo blk > current_tracer
Execute the workload you’re interested in:
root@crownbay:/sys/kernel/debug/tracing# cat /media/sdc/testfile.txt
And look at the output (note here that we’re using ‘trace_pipe’ instead of trace to capture this trace - this allows us to wait around on the pipe for data to appear):
root@crownbay:/sys/kernel/debug/tracing# cat trace_pipe
cat-3587 [001] d..1 3023.276361: 8,32 Q R 1699848 + 8 [cat]
cat-3587 [001] d..1 3023.276410: 8,32 m N cfq3587 alloced
cat-3587 [001] d..1 3023.276415: 8,32 G R 1699848 + 8 [cat]
cat-3587 [001] d..1 3023.276424: 8,32 P N [cat]
cat-3587 [001] d..2 3023.276432: 8,32 I R 1699848 + 8 [cat]
cat-3587 [001] d..1 3023.276439: 8,32 m N cfq3587 insert_request
cat-3587 [001] d..1 3023.276445: 8,32 m N cfq3587 add_to_rr
cat-3587 [001] d..2 3023.276454: 8,32 U N [cat] 1
cat-3587 [001] d..1 3023.276464: 8,32 m N cfq workload slice:150
cat-3587 [001] d..1 3023.276471: 8,32 m N cfq3587 set_active wl_prio:0 wl_type:2
cat-3587 [001] d..1 3023.276478: 8,32 m N cfq3587 fifo= (null)
cat-3587 [001] d..1 3023.276483: 8,32 m N cfq3587 dispatch_insert
cat-3587 [001] d..1 3023.276490: 8,32 m N cfq3587 dispatched a request
cat-3587 [001] d..1 3023.276497: 8,32 m N cfq3587 activate rq, drv=1
cat-3587 [001] d..2 3023.276500: 8,32 D R 1699848 + 8 [cat]
And this turns off tracing for the specified device:
root@crownbay:/sys/kernel/debug/tracing# echo 0 > /sys/block/sdc/trace/enable
3.6.3 blktrace Documentation
Online versions of the man pages for the commands discussed in this section can be found here:
The above manpages, along with manpages for the other blktrace utilities (btt, blkiomon, etc) can be found in the /doc directory of the blktrace tools git repo:
$ git clone git://git.kernel.dk/blktrace.git
4 Real-World Examples
This chapter contains real-world examples.
4.1 Slow Write Speed on Live Images
In one of our previous releases (denzil), users noticed that booting off of a live image and writing to disk was noticeably slower. This included the boot itself, especially the first one, since first boots tend to do a significant amount of writing due to certain post-install scripts.
The problem (and solution) was discovered by using the Yocto tracing tools, in this case ‘perf stat’, ‘perf script’, ‘perf record’ and ‘perf report’.
See all the unvarnished details of how this bug was diagnosed and solved here: Yocto Bug #3049
5 Manual Revision History
Revision |
Date |
Note |
---|---|---|
1.4 |
April 2013 |
The initial document released with the Yocto Project 1.4 Release |
1.5 |
October 2013 |
Released with the Yocto Project 1.5 Release. |
1.6 |
April 2014 |
Released with the Yocto Project 1.6 Release. |
1.7 |
October 2014 |
Released with the Yocto Project 1.7 Release. |
1.8 |
April 2015 |
Released with the Yocto Project 1.8 Release. |
2.0 |
October 2015 |
Released with the Yocto Project 2.0 Release. |
2.1 |
April 2016 |
Released with the Yocto Project 2.1 Release. |
2.2 |
October 2016 |
Released with the Yocto Project 2.2 Release. |
2.3 |
May 2017 |
Released with the Yocto Project 2.3 Release. |
2.4 |
October 2017 |
Released with the Yocto Project 2.4 Release. |
2.5 |
May 2018 |
Released with the Yocto Project 2.5 Release. |
2.6 |
November 2018 |
Released with the Yocto Project 2.6 Release. |
2.7 |
May 2019 |
Released with the Yocto Project 2.7 Release. |
3.0 |
October 2019 |
Released with the Yocto Project 3.0 Release. |
3.1 |
April 2020 |
Released with the Yocto Project 3.1 Release. |
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Yocto Project Application Development and the Extensible Software Development Kit (eSDK)
1 Introduction
1.1 eSDK Introduction
Welcome to the Yocto Project Application Development and the Extensible Software Development Kit (eSDK) manual. This manual provides information that explains how to use both the Yocto Project extensible and standard SDKs to develop applications and images.
Note
Prior to the 2.0 Release of the Yocto Project, application development was primarily accomplished through the use of the Application Development Toolkit (ADT) and the availability of stand-alone cross-development toolchains and other tools. With the 2.1 Release of the Yocto Project, application development has transitioned to within a tool-rich extensible SDK and the more traditional standard SDK.
All SDKs consist of the following:
Cross-Development Toolchain: This toolchain contains a compiler, debugger, and various miscellaneous tools.
Libraries, Headers, and Symbols: The libraries, headers, and symbols are specific to the image (i.e. they match the image).
Environment Setup Script: This
*.sh
file, once run, sets up the cross-development environment by defining variables and preparing for SDK use.
Additionally, an extensible SDK has tools that allow you to easily add new applications and libraries to an image, modify the source of an existing component, test changes on the target hardware, and easily integrate an application into the OpenEmbedded Build System.
You can use an SDK to independently develop and test code that is
destined to run on some target machine. SDKs are completely
self-contained. The binaries are linked against their own copy of
libc
, which results in no dependencies on the target system. To
achieve this, the pointer to the dynamic loader is configured at install
time since that path cannot be dynamically altered. This is the reason
for a wrapper around the populate_sdk
and populate_sdk_ext
archives.
Another feature for the SDKs is that only one set of cross-compiler
toolchain binaries are produced for any given architecture. This feature
takes advantage of the fact that the target hardware can be passed to
gcc
as a set of compiler options. Those options are set up by the
environment script and contained in variables such as
CC and
LD. This reduces the space needed
for the tools. Understand, however, that every target still needs a
sysroot because those binaries are target-specific.
The SDK development environment consists of the following:
The self-contained SDK, which is an architecture-specific cross-toolchain and matching sysroots (target and native) all built by the OpenEmbedded build system (e.g. the SDK). The toolchain and sysroots are based on a Metadata configuration and extensions, which allows you to cross-develop on the host machine for the target hardware. Additionally, the extensible SDK contains the
devtool
functionality.The Quick EMUlator (QEMU), which lets you simulate target hardware. QEMU is not literally part of the SDK. You must build and include this emulator separately. However, QEMU plays an important role in the development process that revolves around use of the SDK.
In summary, the extensible and standard SDK share many features. However, the extensible SDK has powerful development tools to help you more quickly develop applications. Following is a table that summarizes the primary differences between the standard and extensible SDK types when considering which to build:
Feature |
Standard SDK |
Extensible SDK |
---|---|---|
Toolchain |
Yes |
Yes [1] |
Debugger |
Yes |
Yes [1] |
Size |
100+ MBytes |
1+ GBytes (or 300+ MBytes for minimal w/toolchain) |
|
No |
Yes |
Build Images |
No |
Yes |
Updateable |
No |
Yes |
Managed Sysroot [2] |
No |
Yes |
Installed Packages |
No [3] |
Yes [4] |
Construction |
Packages |
Shared State |
Extensible SDK contains the toolchain and debugger if SDK_EXT_TYPE is “full” or SDK_INCLUDE_TOOLCHAIN is “1”, which is the default.
Sysroot is managed through the use of devtool
. Thus, it is less
likely that you will corrupt your SDK sysroot when you try to add
additional libraries.
You can add runtime package management to the standard SDK but it is not supported by default.
You must build and make the shared state available to extensible SDK users for “packages” you want to enable users to install.
1.1.1 The Cross-Development Toolchain
The Cross-Development Toolchain consists
of a cross-compiler, cross-linker, and cross-debugger that are used to
develop user-space applications for targeted hardware. Additionally, for
an extensible SDK, the toolchain also has built-in devtool
functionality. This toolchain is created by running a SDK installer
script or through a Build Directory that is based on
your metadata configuration or extension for your targeted device. The
cross-toolchain works with a matching target sysroot.
1.1.2 Sysroots
The native and target sysroots contain needed headers and libraries for generating binaries that run on the target architecture. The target sysroot is based on the target root filesystem image that is built by the OpenEmbedded build system and uses the same metadata configuration used to build the cross-toolchain.
1.1.3 The QEMU Emulator
The QEMU emulator allows you to simulate your hardware while running your application or image. QEMU is not part of the SDK but is made available a number of different ways:
If you have cloned the
poky
Git repository to create a Source Directory and you have sourced the environment setup script, QEMU is installed and automatically available.If you have downloaded a Yocto Project release and unpacked it to create a Source Directory and you have sourced the environment setup script, QEMU is installed and automatically available.
If you have installed the cross-toolchain tarball and you have sourced the toolchain’s setup environment script, QEMU is also installed and automatically available.
1.2 SDK Development Model
Fundamentally, the SDK fits into the development process as follows:
The SDK is installed on any machine and can be used to develop applications, images, and kernels. An SDK can even be used by a QA Engineer or Release Engineer. The fundamental concept is that the machine that has the SDK installed does not have to be associated with the machine that has the Yocto Project installed. A developer can independently compile and test an object on their machine and then, when the object is ready for integration into an image, they can simply make it available to the machine that has the Yocto Project. Once the object is available, the image can be rebuilt using the Yocto Project to produce the modified image.
You just need to follow these general steps:
Install the SDK for your target hardware: For information on how to install the SDK, see the “Installing the SDK” section.
Download or Build the Target Image: The Yocto Project supports several target architectures and has many pre-built kernel images and root filesystem images.
If you are going to develop your application on hardware, go to the machines download area and choose a target machine area from which to download the kernel image and root filesystem. This download area could have several files in it that support development using actual hardware. For example, the area might contain
.hddimg
files that combine the kernel image with the filesystem, boot loaders, and so forth. Be sure to get the files you need for your particular development process.If you are going to develop your application and then run and test it using the QEMU emulator, go to the machines/qemu download area. From this area, go down into the directory for your target architecture (e.g.
qemux86_64
for an Intel-based 64-bit architecture). Download the kernel, root filesystem, and any other files you need for your process.Note
To use the root filesystem in QEMU, you need to extract it. See the ” Extracting the Root Filesystem “ section for information on how to extract the root filesystem.
Develop and Test your Application: At this point, you have the tools to develop your application. If you need to separately install and use the QEMU emulator, you can go to QEMU Home Page to download and learn about the emulator. See the “Using the Quick EMUlator (QEMU)” chapter in the Yocto Project Development Tasks Manual for information on using QEMU within the Yocto Project.
The remainder of this manual describes how to use the extensible and standard SDKs. Information also exists in appendix form that describes how you can build, install, and modify an SDK.
2 Using the Extensible SDK
This chapter describes the extensible SDK and how to install it.
Information covers the pieces of the SDK, how to install it, and
presents a look at using the devtool
functionality. The extensible
SDK makes it easy to add new applications and libraries to an image,
modify the source for an existing component, test changes on the target
hardware, and ease integration into the rest of the
OpenEmbedded Build System.
Note
For a side-by-side comparison of main features supported for an extensible SDK as compared to a standard SDK, see the ” Introduction “ section.
In addition to the functionality available through devtool
, you can
alternatively make use of the toolchain directly, for example from
Makefile and Autotools. See the
“Using the SDK Toolchain Directly” chapter
for more information.
2.1 Why use the Extensible SDK and What is in It?
The extensible SDK provides a cross-development toolchain and libraries
tailored to the contents of a specific image. You would use the
Extensible SDK if you want a toolchain experience supplemented with the
powerful set of devtool
commands tailored for the Yocto Project
environment.
The installed extensible SDK consists of several files and directories.
Basically, it contains an SDK environment setup script, some
configuration files, an internal build system, and the devtool
functionality.
2.2 Installing the Extensible SDK
The first thing you need to do is install the SDK on your Build
Host by running the *.sh
installation script.
You can download a tarball installer, which includes the pre-built
toolchain, the runqemu
script, the internal build system,
devtool
, and support files from the appropriate
toolchain directory within the Index of
Releases. Toolchains are available for several 32-bit and 64-bit
architectures with the x86_64
directories, respectively. The
toolchains the Yocto Project provides are based off the
core-image-sato
and core-image-minimal
images and contain
libraries appropriate for developing against that image.
The names of the tarball installer scripts are such that a string representing the host system appears first in the filename and then is immediately followed by a string representing the target architecture. An extensible SDK has the string “-ext” as part of the name. Following is the general form:
poky-glibc-host_system-image_type-arch-toolchain-ext-release_version.sh
Where:
host_system is a string representing your development system:
i686 or x86_64.
image_type is the image for which the SDK was built:
core-image-sato or core-image-minimal
arch is a string representing the tuned target architecture:
aarch64, armv5e, core2-64, i586, mips32r2, mips64, ppc7400, or cortexa8hf-neon
release_version is a string representing the release number of the Yocto Project:
3.3.1, 3.3.1+snapshot
For example, the following SDK installer is for a 64-bit
development host system and a i586-tuned target architecture based off
the SDK for core-image-sato
and using the current 3.3.1 snapshot:
poky-glibc-x86_64-core-image-sato-i586-toolchain-ext-3.3.1.sh
Note
As an alternative to downloading an SDK, you can build the SDK installer. For information on building the installer, see the ” Building an SDK Installer “ section.
The SDK and toolchains are self-contained and by default are installed
into the poky_sdk
folder in your home directory. You can choose to
install the extensible SDK in any location when you run the installer.
However, because files need to be written under that directory during
the normal course of operation, the location you choose for installation
must be writable for whichever users need to use the SDK.
The following command shows how to run the installer given a toolchain
tarball for a 64-bit x86 development host system and a 64-bit x86 target
architecture. The example assumes the SDK installer is located in
~/Downloads/
and has execution rights.
Note
If you do not have write permissions for the directory into which you are installing the SDK, the installer notifies you and exits. For that case, set up the proper permissions in the directory and run the installer again.
$ ./Downloads/poky-glibc-x86_64-core-image-minimal-core2-64-toolchain-ext-2.5.sh
Poky (Yocto Project Reference Distro) Extensible SDK installer version 2.5
==========================================================================
Enter target directory for SDK (default: poky_sdk):
You are about to install the SDK to "/home/scottrif/poky_sdk". Proceed [Y/n]? Y
Extracting SDK..............done
Setting it up...
Extracting buildtools...
Preparing build system...
Parsing recipes: 100% |##################################################################| Time: 0:00:52
Initialising tasks: 100% |###############################################################| Time: 0:00:00
Checking sstate mirror object availability: 100% |#######################################| Time: 0:00:00
Loading cache: 100% |####################################################################| Time: 0:00:00
Initialising tasks: 100% |###############################################################| Time: 0:00:00
done
SDK has been successfully set up and is ready to be used.
Each time you wish to use the SDK in a new shell session, you need to source the environment setup script e.g.
$ . /home/scottrif/poky_sdk/environment-setup-core2-64-poky-linux
2.3 Running the Extensible SDK Environment Setup Script
Once you have the SDK installed, you must run the SDK environment setup
script before you can actually use the SDK. This setup script resides in
the directory you chose when you installed the SDK, which is either the
default poky_sdk
directory or the directory you chose during
installation.
Before running the script, be sure it is the one that matches the
architecture for which you are developing. Environment setup scripts
begin with the string “environment-setup
” and include as part of
their name the tuned target architecture. As an example, the following
commands set the working directory to where the SDK was installed and
then source the environment setup script. In this example, the setup
script is for an IA-based target machine using i586 tuning:
$ cd /home/scottrif/poky_sdk
$ source environment-setup-core2-64-poky-linux
SDK environment now set up; additionally you may now run devtool to perform development tasks.
Run devtool --help for further details.
Running the setup script defines many environment variables needed in
order to use the SDK (e.g. PATH
,
CC,
LD, and so forth). If you want to
see all the environment variables the script exports, examine the
installation file itself.
2.4 Using devtool
in Your SDK Workflow
The cornerstone of the extensible SDK is a command-line tool called
devtool
. This tool provides a number of features that help you
build, test and package software within the extensible SDK, and
optionally integrate it into an image built by the OpenEmbedded build
system.
Note
The use of devtool is not limited to the extensible SDK. You can use devtool to help you easily develop any project whose build output must be part of an image built using the build system.
The devtool
command line is organized similarly to
Git in that it has a number of
sub-commands for each function. You can run devtool --help
to see
all the commands.
Note
See the ” devtool Quick Reference “ in the Yocto Project Reference Manual for a devtool quick reference.
Three devtool
subcommands exist that provide entry-points into
development:
devtool add: Assists in adding new software to be built.
devtool modify: Sets up an environment to enable you to modify the source of an existing component.
devtool upgrade: Updates an existing recipe so that you can build it for an updated set of source files.
As with the build system, “recipes” represent software packages within
devtool
. When you use devtool add
, a recipe is automatically
created. When you use devtool modify
, the specified existing recipe
is used in order to determine where to get the source code and how to
patch it. In both cases, an environment is set up so that when you build
the recipe a source tree that is under your control is used in order to
allow you to make changes to the source as desired. By default, new
recipes and the source go into a “workspace” directory under the SDK.
The remainder of this section presents the devtool add
,
devtool modify
, and devtool upgrade
workflows.
2.4.1 Use devtool add
to Add an Application
The devtool add
command generates a new recipe based on existing
source code. This command takes advantage of the
The Workspace Layer Structure
layer that many devtool
commands use. The command is flexible enough
to allow you to extract source code into both the workspace or a
separate local Git repository and to use existing code that does not
need to be extracted.
Depending on your particular scenario, the arguments and options you use
with devtool add
form different combinations. The following diagram
shows common development flows you would use with the devtool add
command:
Generating the New Recipe: The top part of the flow shows three scenarios by which you could use
devtool add
to generate a recipe based on existing source code.In a shared development environment, it is typical for other developers to be responsible for various areas of source code. As a developer, you are probably interested in using that source code as part of your development within the Yocto Project. All you need is access to the code, a recipe, and a controlled area in which to do your work.
Within the diagram, three possible scenarios feed into the
devtool add
workflow:Left: The left scenario in the figure represents a common situation where the source code does not exist locally and needs to be extracted. In this situation, the source code is extracted to the default workspace - you do not want the files in some specific location outside of the workspace. Thus, everything you need will be located in the workspace:
$ devtool add recipe fetchuri
With this command,
devtool
extracts the upstream source files into a local Git repository within thesources
folder. The command then creates a recipe named recipe and a corresponding append file in the workspace. If you do not provide recipe, the command makes an attempt to determine the recipe name.Middle: The middle scenario in the figure also represents a situation where the source code does not exist locally. In this case, the code is again upstream and needs to be extracted to some local area - this time outside of the default workspace.
Note
If required, devtool always creates a Git repository locally during the extraction.
Furthermore, the first positional argument srctree in this case identifies where the
devtool add
command will locate the extracted code outside of the workspace. You need to specify an empty directory:$ devtool add recipe srctree fetchuri
In summary, the source code is pulled from fetchuri and extracted into the location defined by srctree as a local Git repository.
Within workspace,
devtool
creates a recipe named recipe along with an associated append file.Right: The right scenario in the figure represents a situation where the srctree has been previously prepared outside of the
devtool
workspace.The following command provides a new recipe name and identifies the existing source tree location:
$ devtool add recipe srctree
The command examines the source code and creates a recipe named recipe for the code and places the recipe into the workspace.
Because the extracted source code already exists,
devtool
does not try to relocate the source code into the workspace - only the new recipe is placed in the workspace.Aside from a recipe folder, the command also creates an associated append folder and places an initial
*.bbappend
file within.
Edit the Recipe: You can use
devtool edit-recipe
to open up the editor as defined by the$EDITOR
environment variable and modify the file:$ devtool edit-recipe recipe
From within the editor, you can make modifications to the recipe that take affect when you build it later.
Build the Recipe or Rebuild the Image: The next step you take depends on what you are going to do with the new code.
If you need to eventually move the build output to the target hardware, use the following
devtool
command: :;$ devtool build recipe
On the other hand, if you want an image to contain the recipe’s packages from the workspace for immediate deployment onto a device (e.g. for testing purposes), you can use the
devtool build-image
command:$ devtool build-image image
Deploy the Build Output: When you use the
devtool build
command to build out your recipe, you probably want to see if the resulting build output works as expected on the target hardware.Note
This step assumes you have a previously built image that is already either running in QEMU or is running on actual hardware. Also, it is assumed that for deployment of the image to the target, SSH is installed in the image and, if the image is running on real hardware, you have network access to and from your development machine.
You can deploy your build output to that target hardware by using the
devtool deploy-target
command: $ devtool deploy-target recipe target The target is a live target machine running as an SSH server.You can, of course, also deploy the image you build to actual hardware by using the
devtool build-image
command. However,devtool
does not provide a specific command that allows you to deploy the image to actual hardware.Finish Your Work With the Recipe: The
devtool finish
command creates any patches corresponding to commits in the local Git repository, moves the new recipe to a more permanent layer, and then resets the recipe so that the recipe is built normally rather than from the workspace.$ devtool finish recipe layer
Note
Any changes you want to turn into patches must be committed to the Git repository in the source tree.
As mentioned, the
devtool finish
command moves the final recipe to its permanent layer.As a final process of the
devtool finish
command, the state of the standard layers and the upstream source is restored so that you can build the recipe from those areas rather than the workspace.Note
You can use the devtool reset command to put things back should you decide you do not want to proceed with your work. If you do use this command, realize that the source tree is preserved.
2.4.2 Use devtool modify
to Modify the Source of an Existing Component
The devtool modify
command prepares the way to work on existing code
that already has a local recipe in place that is used to build the
software. The command is flexible enough to allow you to extract code
from an upstream source, specify the existing recipe, and keep track of
and gather any patch files from other developers that are associated
with the code.
Depending on your particular scenario, the arguments and options you use
with devtool modify
form different combinations. The following
diagram shows common development flows for the devtool modify
command:
Preparing to Modify the Code: The top part of the flow shows three scenarios by which you could use
devtool modify
to prepare to work on source files. Each scenario assumes the following:The recipe exists locally in a layer external to the
devtool
workspace.The source files exist either upstream in an un-extracted state or locally in a previously extracted state.
The typical situation is where another developer has created a layer for use with the Yocto Project and their recipe already resides in that layer. Furthermore, their source code is readily available either upstream or locally.
Left: The left scenario in the figure represents a common situation where the source code does not exist locally and it needs to be extracted from an upstream source. In this situation, the source is extracted into the default
devtool
workspace location. The recipe, in this scenario, is in its own layer outside the workspace (i.e.meta-
layername).The following command identifies the recipe and, by default, extracts the source files:
$ devtool modify recipe
Once
devtool
locates the recipe,devtool
uses the recipe’s SRC_URI statements to locate the source code and any local patch files from other developers.With this scenario, no srctree argument exists. Consequently, the default behavior of the
devtool modify
command is to extract the source files pointed to by theSRC_URI
statements into a local Git structure. Furthermore, the location for the extracted source is the default area within thedevtool
workspace. The result is that the command sets up both the source code and an append file within the workspace while the recipe remains in its original location.Additionally, if you have any non-patch local files (i.e. files referred to with
file://
entries inSRC_URI
statement excluding*.patch/
or*.diff
), these files are copied to anoe-local-files
folder under the newly created source tree. Copying the files here gives you a convenient area from which you can modify the files. Any changes or additions you make to those files are incorporated into the build the next time you build the software just as are other changes you might have made to the source.Middle: The middle scenario in the figure represents a situation where the source code also does not exist locally. In this case, the code is again upstream and needs to be extracted to some local area as a Git repository. The recipe, in this scenario, is again local and in its own layer outside the workspace.
The following command tells
devtool
the recipe with which to work and, in this case, identifies a local area for the extracted source files that exists outside of the defaultdevtool
workspace:$ devtool modify recipe srctree
Note
You cannot provide a URL for srctree using the devtool command.
As with all extractions, the command uses the recipe’s
SRC_URI
statements to locate the source files and any associated patch files. Non-patch files are copied to anoe-local-files
folder under the newly created source tree.Once the files are located, the command by default extracts them into srctree.
Within workspace,
devtool
creates an append file for the recipe. The recipe remains in its original location but the source files are extracted to the location you provide with srctree.Right: The right scenario in the figure represents a situation where the source tree (srctree) already exists locally as a previously extracted Git structure outside of the
devtool
workspace. In this example, the recipe also exists elsewhere locally in its own layer.The following command tells
devtool
the recipe with which to work, uses the “-n” option to indicate source does not need to be extracted, and uses srctree to point to the previously extracted source files:$ devtool modify -n recipe srctree
If an
oe-local-files
subdirectory happens to exist and it contains non-patch files, the files are used. However, if the subdirectory does not exist and you run thedevtool finish
command, any non-patch files that might exist next to the recipe are removed because it appears todevtool
that you have deleted those files.Once the
devtool modify
command finishes, it creates only an append file for the recipe in thedevtool
workspace. The recipe and the source code remain in their original locations.
Edit the Source: Once you have used the
devtool modify
command, you are free to make changes to the source files. You can use any editor you like to make and save your source code modifications.Build the Recipe or Rebuild the Image: The next step you take depends on what you are going to do with the new code.
If you need to eventually move the build output to the target hardware, use the following
devtool
command:$ devtool build recipe
On the other hand, if you want an image to contain the recipe’s packages from the workspace for immediate deployment onto a device (e.g. for testing purposes), you can use the
devtool build-image
command: $ devtool build-image imageDeploy the Build Output: When you use the
devtool build
command to build out your recipe, you probably want to see if the resulting build output works as expected on target hardware.Note
This step assumes you have a previously built image that is already either running in QEMU or running on actual hardware. Also, it is assumed that for deployment of the image to the target, SSH is installed in the image and if the image is running on real hardware that you have network access to and from your development machine.
You can deploy your build output to that target hardware by using the
devtool deploy-target
command:$ devtool deploy-target recipe target
The target is a live target machine running as an SSH server.
You can, of course, use other methods to deploy the image you built using the
devtool build-image
command to actual hardware.devtool
does not provide a specific command to deploy the image to actual hardware.Finish Your Work With the Recipe: The
devtool finish
command creates any patches corresponding to commits in the local Git repository, updates the recipe to point to them (or creates a.bbappend
file to do so, depending on the specified destination layer), and then resets the recipe so that the recipe is built normally rather than from the workspace.$ devtool finish recipe layer
Note
Any changes you want to turn into patches must be staged and committed within the local Git repository before you use the devtool finish command.
Because there is no need to move the recipe,
devtool finish
either updates the original recipe in the original layer or the command creates a.bbappend
file in a different layer as provided by layer. Any work you did in theoe-local-files
directory is preserved in the original files next to the recipe during thedevtool finish
command.As a final process of the
devtool finish
command, the state of the standard layers and the upstream source is restored so that you can build the recipe from those areas rather than from the workspace.Note
You can use the devtool reset command to put things back should you decide you do not want to proceed with your work. If you do use this command, realize that the source tree is preserved.
2.4.3 Use devtool upgrade
to Create a Version of the Recipe that Supports a Newer Version of the Software
The devtool upgrade
command upgrades an existing recipe to that of a
more up-to-date version found upstream. Throughout the life of software,
recipes continually undergo version upgrades by their upstream
publishers. You can use the devtool upgrade
workflow to make sure
your recipes you are using for builds are up-to-date with their upstream
counterparts.
Note
Several methods exist by which you can upgrade recipes - devtool upgrade happens to be one. You can read about all the methods by which you can upgrade recipes in the ” Upgrading Recipes “ section of the Yocto Project Development Tasks Manual.
The devtool upgrade
command is flexible enough to allow you to
specify source code revision and versioning schemes, extract code into
or out of the devtool
The Workspace Layer Structure,
and work with any source file forms that the
fetchers support.
The following diagram shows the common development flow used with the
devtool upgrade
command:
Initiate the Upgrade: The top part of the flow shows the typical scenario by which you use the
devtool upgrade
command. The following conditions exist:The recipe exists in a local layer external to the
devtool
workspace.The source files for the new release exist in the same location pointed to by SRC_URI in the recipe (e.g. a tarball with the new version number in the name, or as a different revision in the upstream Git repository).
A common situation is where third-party software has undergone a revision so that it has been upgraded. The recipe you have access to is likely in your own layer. Thus, you need to upgrade the recipe to use the newer version of the software:
$ devtool upgrade -V version recipe
By default, the
devtool upgrade
command extracts source code into thesources
directory in the The Workspace Layer Structure. If you want the code extracted to any other location, you need to provide the srctree positional argument with the command as follows: $ devtool upgrade -V version recipe srctreeNote
In this example, the “-V” option specifies the new version. If you don’t use “-V”, the command upgrades the recipe to the latest version.
If the source files pointed to by the
SRC_URI
statement in the recipe are in a Git repository, you must provide the “-S” option and specify a revision for the software.Once
devtool
locates the recipe, it uses theSRC_URI
variable to locate the source code and any local patch files from other developers. The result is that the command sets up the source code, the new version of the recipe, and an append file all within the workspace.Additionally, if you have any non-patch local files (i.e. files referred to with
file://
entries inSRC_URI
statement excluding*.patch/
or*.diff
), these files are copied to anoe-local-files
folder under the newly created source tree. Copying the files here gives you a convenient area from which you can modify the files. Any changes or additions you make to those files are incorporated into the build the next time you build the software just as are other changes you might have made to the source.Resolve any Conflicts created by the Upgrade: Conflicts could exist due to the software being upgraded to a new version. Conflicts occur if your recipe specifies some patch files in
SRC_URI
that conflict with changes made in the new version of the software. For such cases, you need to resolve the conflicts by editing the source and following the normalgit rebase
conflict resolution process.Before moving onto the next step, be sure to resolve any such conflicts created through use of a newer or different version of the software.
Build the Recipe or Rebuild the Image: The next step you take depends on what you are going to do with the new code.
If you need to eventually move the build output to the target hardware, use the following
devtool
command:$ devtool build recipe
On the other hand, if you want an image to contain the recipe’s packages from the workspace for immediate deployment onto a device (e.g. for testing purposes), you can use the
devtool build-image
command:$ devtool build-image image
Deploy the Build Output: When you use the
devtool build
command orbitbake
to build your recipe, you probably want to see if the resulting build output works as expected on target hardware.Note
This step assumes you have a previously built image that is already either running in QEMU or running on actual hardware. Also, it is assumed that for deployment of the image to the target, SSH is installed in the image and if the image is running on real hardware that you have network access to and from your development machine.
You can deploy your build output to that target hardware by using the
devtool deploy-target
command: $ devtool deploy-target recipe target The target is a live target machine running as an SSH server.You can, of course, also deploy the image you build using the
devtool build-image
command to actual hardware. However,devtool
does not provide a specific command that allows you to do this.Finish Your Work With the Recipe: The
devtool finish
command creates any patches corresponding to commits in the local Git repository, moves the new recipe to a more permanent layer, and then resets the recipe so that the recipe is built normally rather than from the workspace.Any work you did in the
oe-local-files
directory is preserved in the original files next to the recipe during thedevtool finish
command.If you specify a destination layer that is the same as the original source, then the old version of the recipe and associated files are removed prior to adding the new version.
$ devtool finish recipe layer
Note
Any changes you want to turn into patches must be committed to the Git repository in the source tree.
As a final process of the
devtool finish
command, the state of the standard layers and the upstream source is restored so that you can build the recipe from those areas rather than the workspace.Note
You can use the devtool reset command to put things back should you decide you do not want to proceed with your work. If you do use this command, realize that the source tree is preserved.
2.5 A Closer Look at devtool add
The devtool add
command automatically creates a recipe based on the
source tree you provide with the command. Currently, the command has
support for the following:
Autotools (
autoconf
andautomake
)CMake
Scons
qmake
Plain
Makefile
Out-of-tree kernel module
Binary package (i.e. “-b” option)
Node.js module
Python modules that use
setuptools
ordistutils
Apart from binary packages, the determination of how a source tree
should be treated is automatic based on the files present within that
source tree. For example, if a CMakeLists.txt
file is found, then
the source tree is assumed to be using CMake and is treated accordingly.
Note
In most cases, you need to edit the automatically generated recipe in order to make it build properly. Typically, you would go through several edit and build cycles until the recipe successfully builds. Once the recipe builds, you could use possible further iterations to test the recipe on the target device.
The remainder of this section covers specifics regarding how parts of the recipe are generated.
2.5.1 Name and Version
If you do not specify a name and version on the command line,
devtool add
uses various metadata within the source tree in an
attempt to determine the name and version of the software being built.
Based on what the tool determines, devtool
sets the name of the
created recipe file accordingly.
If devtool
cannot determine the name and version, the command prints
an error. For such cases, you must re-run the command and provide the
name and version, just the name, or just the version as part of the
command line.
Sometimes the name or version determined from the source tree might be incorrect. For such a case, you must reset the recipe:
$ devtool reset -n recipename
After running the devtool reset
command, you need to
run devtool add
again and provide the name or the version.
2.5.2 Dependency Detection and Mapping
The devtool add
command attempts to detect build-time dependencies
and map them to other recipes in the system. During this mapping, the
command fills in the names of those recipes as part of the
DEPENDS variable within the
recipe. If a dependency cannot be mapped, devtool
places a comment
in the recipe indicating such. The inability to map a dependency can
result from naming not being recognized or because the dependency simply
is not available. For cases where the dependency is not available, you
must use the devtool add
command to add an additional recipe that
satisfies the dependency. Once you add that recipe, you need to update
the DEPENDS
variable in the original recipe to include the new
recipe.
If you need to add runtime dependencies, you can do so by adding the following to your recipe:
RDEPENDS_${PN} += "dependency1 dependency2 ..."
Note
The devtool add command often cannot distinguish between mandatory and optional dependencies. Consequently, some of the detected dependencies might in fact be optional. When in doubt, consult the documentation or the configure script for the software the recipe is building for further details. In some cases, you might find you can substitute the dependency with an option that disables the associated functionality passed to the configure script.
2.5.3 License Detection
The devtool add
command attempts to determine if the software you
are adding is able to be distributed under a common, open-source
license. If so, the command sets the
LICENSE value accordingly.
You should double-check the value added by the command against the
documentation or source files for the software you are building and, if
necessary, update that LICENSE
value.
The devtool add
command also sets the
LIC_FILES_CHKSUM
value to point to all files that appear to be license-related. Realize
that license statements often appear in comments at the top of source
files or within the documentation. In such cases, the command does not
recognize those license statements. Consequently, you might need to
amend the LIC_FILES_CHKSUM
variable to point to one or more of those
comments if present. Setting LIC_FILES_CHKSUM
is particularly
important for third-party software. The mechanism attempts to ensure
correct licensing should you upgrade the recipe to a newer upstream
version in future. Any change in licensing is detected and you receive
an error prompting you to check the license text again.
If the devtool add
command cannot determine licensing information,
devtool
sets the LICENSE
value to “CLOSED” and leaves the
LIC_FILES_CHKSUM
value unset. This behavior allows you to continue
with development even though the settings are unlikely to be correct in
all cases. You should check the documentation or source files for the
software you are building to determine the actual license.
2.5.4 Adding Makefile-Only Software
The use of Make by itself is very common in both proprietary and
open-source software. Unfortunately, Makefiles are often not written
with cross-compilation in mind. Thus, devtool add
often cannot do
very much to ensure that these Makefiles build correctly. It is very
common, for example, to explicitly call gcc
instead of using the
CC variable. Usually, in a
cross-compilation environment, gcc
is the compiler for the build
host and the cross-compiler is named something similar to
arm-poky-linux-gnueabi-gcc
and might require arguments (e.g. to
point to the associated sysroot for the target machine).
When writing a recipe for Makefile-only software, keep the following in mind:
You probably need to patch the Makefile to use variables instead of hardcoding tools within the toolchain such as
gcc
andg++
.The environment in which Make runs is set up with various standard variables for compilation (e.g.
CC
,CXX
, and so forth) in a similar manner to the environment set up by the SDK’s environment setup script. One easy way to see these variables is to run thedevtool build
command on the recipe and then look inoe-logs/run.do_compile
. Towards the top of this file, a list of environment variables exists that are being set. You can take advantage of these variables within the Makefile.If the Makefile sets a default for a variable using “=”, that default overrides the value set in the environment, which is usually not desirable. For this case, you can either patch the Makefile so it sets the default using the “?=” operator, or you can alternatively force the value on the
make
command line. To force the value on the command line, add the variable setting to EXTRA_OEMAKE or PACKAGECONFIG_CONFARGS within the recipe. Here is an example usingEXTRA_OEMAKE
:EXTRA_OEMAKE += "'CC=${CC}' 'CXX=${CXX}'"
In the above example, single quotes are used around the variable settings as the values are likely to contain spaces because required default options are passed to the compiler.
Hardcoding paths inside Makefiles is often problematic in a cross-compilation environment. This is particularly true because those hardcoded paths often point to locations on the build host and thus will either be read-only or will introduce contamination into the cross-compilation because they are specific to the build host rather than the target. Patching the Makefile to use prefix variables or other path variables is usually the way to handle this situation.
Sometimes a Makefile runs target-specific commands such as
ldconfig
. For such cases, you might be able to apply patches that remove these commands from the Makefile.
2.5.5 Adding Native Tools
Often, you need to build additional tools that run on the Build
Host as opposed to
the target. You should indicate this requirement by using one of the
following methods when you run devtool add
:
Specify the name of the recipe such that it ends with “-native”. Specifying the name like this produces a recipe that only builds for the build host.
Specify the “DASHDASHalso-native” option with the
devtool add
command. Specifying this option creates a recipe file that still builds for the target but also creates a variant with a “-native” suffix that builds for the build host.
Note
If you need to add a tool that is shipped as part of a source tree that builds code for the target, you can typically accomplish this by building the native and target parts separately rather than within the same compilation process. Realize though that with the “DASHDASHalso-native” option, you can add the tool using just one recipe file.
2.5.6 Adding Node.js Modules
You can use the devtool add
command two different ways to add
Node.js modules: 1) Through npm
and, 2) from a repository or local
source.
Use the following form to add Node.js modules through npm
:
$ devtool add "npm://registry.npmjs.org;name=forever;version=0.15.1"
The name and version parameters are mandatory. Lockdown and shrinkwrap files are generated and pointed to by the recipe in order to freeze the version that is fetched for the dependencies according to the first time. This also saves checksums that are verified on future fetches. Together, these behaviors ensure the reproducibility and integrity of the build.
Note
You must use quotes around the URL. The
devtool add
does not require the quotes, but the shell considers “;” as a splitter between multiple commands. Thus, without the quotes,devtool add
does not receive the other parts, which results in several “command not found” errors.In order to support adding Node.js modules, a
nodejs
recipe must be part of your SDK.
As mentioned earlier, you can also add Node.js modules directly from a
repository or local source tree. To add modules this way, use
devtool add
in the following form:
$ devtool add https://github.com/diversario/node-ssdp
In this example, devtool
fetches the specified Git repository, detects the code as Node.js code,
fetches dependencies using npm
, and sets
SRC_URI accordingly.
2.6 Working With Recipes
When building a recipe using the devtool build
command, the typical
build progresses as follows:
Fetch the source
Unpack the source
Configure the source
Compile the source
Install the build output
Package the installed output
For recipes in the workspace, fetching and unpacking is disabled as the source tree has already been prepared and is persistent. Each of these build steps is defined as a function (task), usually with a “do_” prefix (e.g. do_fetch, do_unpack, and so forth). These functions are typically shell scripts but can instead be written in Python.
If you look at the contents of a recipe, you will see that the recipe
does not include complete instructions for building the software.
Instead, common functionality is encapsulated in classes inherited with
the inherit
directive. This technique leaves the recipe to describe
just the things that are specific to the software being built. A
base class exists that
is implicitly inherited by all recipes and provides the functionality
that most recipes typically need.
The remainder of this section presents information useful when working with recipes.
2.6.1 Finding Logs and Work Files
After the first run of the devtool build
command, recipes that were
previously created using the devtool add
command or whose sources
were modified using the devtool modify
command contain symbolic
links created within the source tree:
oe-logs
: This link points to the directory in which log files and run scripts for each build step are created.oe-workdir
: This link points to the temporary work area for the recipe. The following locations underoe-workdir
are particularly useful:image/
: Contains all of the files installed during the do_install stage. Within a recipe, this directory is referred to by the expression${
D}
.sysroot-destdir/
: Contains a subset of files installed withindo_install
that have been put into the shared sysroot. For more information, see the “Sharing Files Between Recipes” section.packages-split/
: Contains subdirectories for each package produced by the recipe. For more information, see the “Packaging” section.
You can use these links to get more information on what is happening at each build step.
2.6.2 Setting Configure Arguments
If the software your recipe is building uses GNU autoconf, then a fixed
set of arguments is passed to it to enable cross-compilation plus any
extras specified by
EXTRA_OECONF or
PACKAGECONFIG_CONFARGS
set within the recipe. If you wish to pass additional options, add them
to EXTRA_OECONF
or PACKAGECONFIG_CONFARGS
. Other supported build
tools have similar variables (e.g.
EXTRA_OECMAKE for
CMake, EXTRA_OESCONS
for Scons, and so forth). If you need to pass anything on the make
command line, you can use EXTRA_OEMAKE
or the
PACKAGECONFIG_CONFARGS
variables to do so.
You can use the devtool configure-help
command to help you set the
arguments listed in the previous paragraph. The command determines the
exact options being passed, and shows them to you along with any custom
arguments specified through EXTRA_OECONF
or
PACKAGECONFIG_CONFARGS
. If applicable, the command also shows you
the output of the configure script’s “DASHDASHhelp” option as a
reference.
2.6.4 Packaging
Packaging is not always particularly relevant within the extensible SDK. However, if you examine how build output gets into the final image on the target device, it is important to understand packaging because the contents of the image are expressed in terms of packages and not recipes.
During the do_package
task, files installed during the
do_install task are
split into one main package, which is almost always named the same as
the recipe, and into several other packages. This separation exists
because not all of those installed files are useful in every image. For
example, you probably do not need any of the documentation installed in
a production image. Consequently, for each recipe the documentation
files are separated into a -doc
package. Recipes that package
software containing optional modules or plugins might undergo additional
package splitting as well.
After building a recipe, you can see where files have gone by looking in
the oe-workdir/packages-split
directory, which contains a
subdirectory for each package. Apart from some advanced cases, the
PACKAGES and
FILES variables controls
splitting. The PACKAGES
variable lists all of the packages to be
produced, while the FILES
variable specifies which files to include
in each package by using an override to specify the package. For
example, FILES_${PN}
specifies the files to go into the main package
(i.e. the main package has the same name as the recipe and
${
PN}
evaluates to the
recipe name). The order of the PACKAGES
value is significant. For
each installed file, the first package whose FILES
value matches the
file is the package into which the file goes. Defaults exist for both
the PACKAGES
and FILES
variables. Consequently, you might find
you do not even need to set these variables in your recipe unless the
software the recipe is building installs files into non-standard
locations.
2.7 Restoring the Target Device to its Original State
If you use the devtool deploy-target
command to write a recipe’s
build output to the target, and you are working on an existing component
of the system, then you might find yourself in a situation where you
need to restore the original files that existed prior to running the
devtool deploy-target
command. Because the devtool deploy-target
command backs up any files it overwrites, you can use the
devtool undeploy-target
command to restore those files and remove
any other files the recipe deployed. Consider the following example:
$ devtool undeploy-target lighttpd root@192.168.7.2
If you have deployed multiple applications, you can remove them all using the “-a” option thus restoring the target device to its original state:
$ devtool undeploy-target -a root@192.168.7.2
Information about files deployed to the target as well as any backed up files are stored on the target itself. This storage, of course, requires some additional space on the target machine.
Note
The devtool deploy-target and devtool undeploy-target commands do not currently interact with any package management system on the target device (e.g. RPM or OPKG). Consequently, you should not intermingle devtool deploy-target and package manager operations on the target device. Doing so could result in a conflicting set of files.
2.8 Installing Additional Items Into the Extensible SDK
Out of the box the extensible SDK typically only comes with a small
number of tools and libraries. A minimal SDK starts mostly empty and is
populated on-demand. Sometimes you must explicitly install extra items
into the SDK. If you need these extra items, you can first search for
the items using the devtool search
command. For example, suppose you
need to link to libGL but you are not sure which recipe provides libGL.
You can use the following command to find out:
$ devtool search libGL mesa
A free implementation of the OpenGL API Once you know the recipe
(i.e. mesa
in this example), you can install it:
$ devtool sdk-install mesa
By default, the devtool sdk-install
command assumes
the item is available in pre-built form from your SDK provider. If the
item is not available and it is acceptable to build the item from
source, you can add the “-s” option as follows:
$ devtool sdk-install -s mesa
It is important to remember that building the item from source
takes significantly longer than installing the pre-built artifact. Also,
if no recipe exists for the item you want to add to the SDK, you must
instead add the item using the devtool add
command.
2.9 Applying Updates to an Installed Extensible SDK
If you are working with an installed extensible SDK that gets occasionally updated (e.g. a third-party SDK), then you will need to manually “pull down” the updates into the installed SDK.
To update your installed SDK, use devtool
as follows:
$ devtool sdk-update
The previous command assumes your SDK provider has set the default update URL for you through the SDK_UPDATE_URL variable as described in the “Providing Updates to the Extensible SDK After Installation” section. If the SDK provider has not set that default URL, you need to specify it yourself in the command as follows: $ devtool sdk-update path_to_update_directory
Note
The URL needs to point specifically to a published SDK and not to an SDK installer that you would download and install.
2.10 Creating a Derivative SDK With Additional Components
You might need to produce an SDK that contains your own custom libraries. A good example would be if you were a vendor with customers that use your SDK to build their own platform-specific software and those customers need an SDK that has custom libraries. In such a case, you can produce a derivative SDK based on the currently installed SDK fairly easily by following these steps:
If necessary, install an extensible SDK that you want to use as a base for your derivative SDK.
Source the environment script for the SDK.
Add the extra libraries or other components you want by using the
devtool add
command.Run the
devtool build-sdk
command.
The previous steps take the recipes added to the workspace and construct a new SDK installer that contains those recipes and the resulting binary artifacts. The recipes go into their own separate layer in the constructed derivative SDK, which leaves the workspace clean and ready for users to add their own recipes.
3 Using the Standard SDK
This chapter describes the standard SDK and how to install it. Information includes unique installation and setup aspects for the standard SDK.
Note
For a side-by-side comparison of main features supported for a standard SDK as compared to an extensible SDK, see the ” Introduction “ section.
You can use a standard SDK to work on Makefile and Autotools-based projects. See the “Using the SDK Toolchain Directly” chapter for more information.
3.1 Why use the Standard SDK and What is in It?
The Standard SDK provides a cross-development toolchain and libraries
tailored to the contents of a specific image. You would use the Standard
SDK if you want a more traditional toolchain experience as compared to
the extensible SDK, which provides an internal build system and the
devtool
functionality.
The installed Standard SDK consists of several files and directories. Basically, it contains an SDK environment setup script, some configuration files, and host and target root filesystems to support usage. You can see the directory structure in the “Installed Standard SDK Directory Structure” section.
3.2 Installing the SDK
The first thing you need to do is install the SDK on your Build
Host by running the *.sh
installation script.
You can download a tarball installer, which includes the pre-built
toolchain, the runqemu
script, and support files from the
appropriate toolchain directory within
the Index of Releases. Toolchains are available for several 32-bit and
64-bit architectures with the x86_64
directories, respectively. The
toolchains the Yocto Project provides are based off the
core-image-sato
and core-image-minimal
images and contain
libraries appropriate for developing against that image.
The names of the tarball installer scripts are such that a string representing the host system appears first in the filename and then is immediately followed by a string representing the target architecture.
poky-glibc-host_system-image_type-arch-toolchain-release_version.sh
Where:
host_system is a string representing your development system:
i686 or x86_64.
image_type is the image for which the SDK was built:
core-image-minimal or core-image-sato.
arch is a string representing the tuned target architecture:
aarch64, armv5e, core2-64, i586, mips32r2, mips64, ppc7400, or cortexa8hf-neon.
release_version is a string representing the release number of the Yocto Project:
3.3.1, 3.3.1+snapshot
For example, the following SDK installer is for a 64-bit
development host system and a i586-tuned target architecture based off
the SDK for core-image-sato
and using the current DISTRO snapshot:
poky-glibc-x86_64-core-image-sato-i586-toolchain-DISTRO.sh
Note
As an alternative to downloading an SDK, you can build the SDK installer. For information on building the installer, see the ” Building an SDK Installer “ section.
The SDK and toolchains are self-contained and by default are installed
into the poky_sdk
folder in your home directory. You can choose to
install the extensible SDK in any location when you run the installer.
However, because files need to be written under that directory during
the normal course of operation, the location you choose for installation
must be writable for whichever users need to use the SDK.
The following command shows how to run the installer given a toolchain
tarball for a 64-bit x86 development host system and a 64-bit x86 target
architecture. The example assumes the SDK installer is located in
~/Downloads/
and has execution rights.
Note
If you do not have write permissions for the directory into which you are installing the SDK, the installer notifies you and exits. For that case, set up the proper permissions in the directory and run the installer again.
$ ./Downloads/poky-glibc-x86_64-core-image-sato-i586-toolchain-3.3.1.sh
Poky (Yocto Project Reference Distro) SDK installer version 3.3.1
===============================================================
Enter target directory for SDK (default: /opt/poky/3.3.1):
You are about to install the SDK to "/opt/poky/3.3.1". Proceed [Y/n]? Y
Extracting SDK........................................ ..............................done
Setting it up...done
SDK has been successfully set up and is ready to be used.
Each time you wish to use the SDK in a new shell session, you need to source the environment setup script e.g.
$ . /opt/poky/3.3.1/environment-setup-i586-poky-linux
Again, reference the “Installed Standard SDK Directory Structure” section for more details on the resulting directory structure of the installed SDK.
3.3 Running the SDK Environment Setup Script
Once you have the SDK installed, you must run the SDK environment setup
script before you can actually use the SDK. This setup script resides in
the directory you chose when you installed the SDK, which is either the
default /opt/poky/3.3.1
directory or the directory you chose during
installation.
Before running the script, be sure it is the one that matches the
architecture for which you are developing. Environment setup scripts
begin with the string “environment-setup
” and include as part of
their name the tuned target architecture. As an example, the following
commands set the working directory to where the SDK was installed and
then source the environment setup script. In this example, the setup
script is for an IA-based target machine using i586 tuning:
$ source /opt/poky/3.3.1/environment-setup-i586-poky-linux
When you run the setup script, the same environment variables are defined as are when you run the setup script for an extensible SDK. See the “Installed Extensible SDK Directory Structure” section for more information.
4 Using the SDK Toolchain Directly
You can use the SDK toolchain directly with Makefile and Autotools-based projects.
4.1 Autotools-Based Projects
Once you have a suitable The Cross-Development Toolchain installed, it is very easy to develop a project using the GNU Autotools-based workflow, which is outside of the OpenEmbedded Build System.
The following figure presents a simple Autotools workflow.
Follow these steps to create a simple Autotools-based “Hello World” project:
Note
For more information on the GNU Autotools workflow, see the same example on the GNOME Developer site.
Create a Working Directory and Populate It: Create a clean directory for your project and then make that directory your working location.
$ mkdir $HOME/helloworld $ cd $HOME/helloworld
After setting up the directory, populate it with files needed for the flow. You need a project source file, a file to help with configuration, and a file to help create the Makefile, and a README file:
hello.c
,configure.ac
,Makefile.am
, andREADME
, respectively.Use the following command to create an empty README file, which is required by GNU Coding Standards:
$ touch README
Create the remaining three files as follows:
hello.c
:#include <stdio.h> main() { printf("Hello World!\n"); }
configure.ac
:AC_INIT(hello,0.1) AM_INIT_AUTOMAKE([foreign]) AC_PROG_CC AC_CONFIG_FILES(Makefile) AC_OUTPUT
Makefile.am
:bin_PROGRAMS = hello hello_SOURCES = hello.c
Source the Cross-Toolchain Environment Setup File: As described earlier in the manual, installing the cross-toolchain creates a cross-toolchain environment setup script in the directory that the SDK was installed. Before you can use the tools to develop your project, you must source this setup script. The script begins with the string “environment-setup” and contains the machine architecture, which is followed by the string “poky-linux”. For this example, the command sources a script from the default SDK installation directory that uses the 32-bit Intel x86 Architecture and the 3.3.1 Yocto Project release:
$ source /opt/poky/3.3.1/environment-setup-i586-poky-linux
Create the configure Script: Use the
autoreconf
command to generate theconfigure
script.$ autoreconf
The
autoreconf
tool takes care of running the other Autotools such asaclocal
,autoconf
, andautomake
.Note
If you get errors from
configure.ac
, whichautoreconf
runs, that indicate missing files, you can use the “-i” option, which ensures missing auxiliary files are copied to the build host.Cross-Compile the Project: This command compiles the project using the cross-compiler. The CONFIGURE_FLAGS environment variable provides the minimal arguments for GNU configure:
$ ./configure ${CONFIGURE_FLAGS}
For an Autotools-based project, you can use the cross-toolchain by just passing the appropriate host option to
configure.sh
. The host option you use is derived from the name of the environment setup script found in the directory in which you installed the cross-toolchain. For example, the host option for an ARM-based target that uses the GNU EABI isarmv5te-poky-linux-gnueabi
. You will notice that the name of the script isenvironment-setup-armv5te-poky-linux-gnueabi
. Thus, the following command works to update your project and rebuild it using the appropriate cross-toolchain tools:$ ./configure --host=armv5te-poky-linux-gnueabi --with-libtool-sysroot=sysroot_dir
Make and Install the Project: These two commands generate and install the project into the destination directory:
$ make $ make install DESTDIR=./tmp
Note
To learn about environment variables established when you run the cross-toolchain environment setup script and how they are used or overridden when the Makefile, see the ” Makefile-Based Projects “ section.
This next command is a simple way to verify the installation of your project. Running the command prints the architecture on which the binary file can run. This architecture should be the same architecture that the installed cross-toolchain supports.
$ file ./tmp/usr/local/bin/hello
Execute Your Project: To execute the project, you would need to run it on your target hardware. If your target hardware happens to be your build host, you could run the project as follows:
$ ./tmp/usr/local/bin/hello
As expected, the project displays the “Hello World!” message.
4.2 Makefile-Based Projects
Simple Makefile-based projects use and interact with the cross-toolchain
environment variables established when you run the cross-toolchain
environment setup script. The environment variables are subject to
general make
rules.
This section presents a simple Makefile development flow and provides an example that lets you see how you can use cross-toolchain environment variables and Makefile variables during development.
The main point of this section is to explain the following three cases regarding variable behavior:
Case 1 - No Variables Set in the Makefile Map to Equivalent Environment Variables Set in the SDK Setup Script: Because matching variables are not specifically set in the
Makefile
, the variables retain their values based on the environment setup script.Case 2 - Variables Are Set in the Makefile that Map to Equivalent Environment Variables from the SDK Setup Script: Specifically setting matching variables in the
Makefile
during the build results in the environment settings of the variables being overwritten. In this case, the variables you set in theMakefile
are used.Case 3 - Variables Are Set Using the Command Line that Map to Equivalent Environment Variables from the SDK Setup Script: Executing the
Makefile
from the command line results in the environment variables being overwritten. In this case, the command-line content is used.
Note
Regardless of how you set your variables, if you use the “-e” option
with make
, the variables from the SDK setup script take precedence:
$ make -e target
The remainder of this section presents a simple Makefile example that demonstrates these variable behaviors.
In a new shell environment variables are not established for the SDK until you run the setup script. For example, the following commands show a null value for the compiler variable (i.e. CC).
$ echo ${CC}
$
Running the
SDK setup script for a 64-bit build host and an i586-tuned target
architecture for a core-image-sato
image using the current 3.3.1
Yocto Project release and then echoing that variable shows the value
established through the script:
$ source /opt/poky/3.3.1/environment-setup-i586-poky-linux
$ echo ${CC}
i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/3.3.1/sysroots/i586-poky-linux
To illustrate variable use, work through this simple “Hello World!” example:
Create a Working Directory and Populate It: Create a clean directory for your project and then make that directory your working location.
$ mkdir $HOME/helloworld $ cd $HOME/helloworld
After setting up the directory, populate it with files needed for the flow. You need a
main.c
file from which you call your function, amodule.h
file to contain headers, and amodule.c
that defines your function.Create the three files as follows:
main.c
:#include "module.h" void sample_func(); int main() { sample_func(); return 0; }
module.h
:#include <stdio.h> void sample_func();
module.c
:#include "module.h" void sample_func() { printf("Hello World!"); printf("\n"); }
Source the Cross-Toolchain Environment Setup File: As described earlier in the manual, installing the cross-toolchain creates a cross-toolchain environment setup script in the directory that the SDK was installed. Before you can use the tools to develop your project, you must source this setup script. The script begins with the string “environment-setup” and contains the machine architecture, which is followed by the string “poky-linux”. For this example, the command sources a script from the default SDK installation directory that uses the 32-bit Intel x86 Architecture and the Hardknott Yocto Project release:
$ source /opt/poky/3.3.1/environment-setup-i586-poky-linux
Create the Makefile: For this example, the Makefile contains two lines that can be used to set the
CC
variable. One line is identical to the value that is set when you run the SDK environment setup script, and the other line setsCC
to “gcc”, the default GNU compiler on the build host:# CC=i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux # CC="gcc" all: main.o module.o ${CC} main.o module.o -o target_bin main.o: main.c module.h ${CC} -I . -c main.c module.o: module.c module.h ${CC} -I . -c module.c clean: rm -rf *.o rm target_bin
Make the Project: Use the
make
command to create the binary output file. Because variables are commented out in the Makefile, the value used forCC
is the value set when the SDK environment setup file was run:$ make i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux -I . -c main.c i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux -I . -c module.c i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux main.o module.o -o target_bin
From the results of the previous command, you can see that the compiler used was the compiler established through the
CC
variable defined in the setup script.You can override the
CC
environment variable with the same variable as set from the Makefile by uncommenting the line in the Makefile and runningmake
again.$ make clean rm -rf *.o rm target_bin # # Edit the Makefile by uncommenting the line that sets CC to "gcc" # $ make gcc -I . -c main.c gcc -I . -c module.c gcc main.o module.o -o target_bin
As shown in the previous example, the cross-toolchain compiler is not used. Rather, the default compiler is used.
This next case shows how to override a variable by providing the variable as part of the command line. Go into the Makefile and re-insert the comment character so that running
make
uses the established SDK compiler. However, when you runmake
, use a command-line argument to setCC
to “gcc”:$ make clean rm -rf *.o rm target_bin # # Edit the Makefile to comment out the line setting CC to "gcc" # $ make i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux -I . -c main.c i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux -I . -c module.c i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux main.o module.o -o target_bin $ make clean rm -rf *.o rm target_bin $ make CC="gcc" gcc -I . -c main.c gcc -I . -c module.c gcc main.o module.o -o target_bin
In the previous case, the command-line argument overrides the SDK environment variable.
In this last case, edit Makefile again to use the “gcc” compiler but then use the “-e” option on the
make
command line:$ make clean rm -rf *.o rm target_bin # # Edit the Makefile to use "gcc" # $ make gcc -I . -c main.c gcc -I . -c module.c gcc main.o module.o -o target_bin $ make clean rm -rf *.o rm target_bin $ make -e i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux -I . -c main.c i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux -I . -c module.c i586-poky-linux-gcc -m32 -march=i586 --sysroot=/opt/poky/2.5/sysroots/i586-poky-linux main.o module.o -o target_bin
In the previous case, the “-e” option forces
make
to use the SDK environment variables regardless of the values in the Makefile.Execute Your Project: To execute the project (i.e.
target_bin
), use the following command:$ ./target_bin Hello World!
Note
If you used the cross-toolchain compiler to build target_bin and your build host differs in architecture from that of the target machine, you need to run your project on the target device.
As expected, the project displays the “Hello World!” message.
5 Obtaining the SDK
5.1 Locating Pre-Built SDK Installers
You can use existing, pre-built toolchains by locating and running an SDK installer script that ships with the Yocto Project. Using this method, you select and download an architecture-specific SDK installer and then run the script to hand-install the toolchain.
Follow these steps to locate and hand-install the toolchain:
Go to the Installers Directory: Go to https://downloads.yoctoproject.org/releases/yocto/yocto-3.3.1/toolchain/
Open the Folder for Your Build Host: Open the folder that matches your Build Host (i.e.
i686
for 32-bit machines orx86_64
for 64-bit machines).Locate and Download the SDK Installer: You need to find and download the installer appropriate for your build host, target hardware, and image type.
The installer files (
*.sh
) follow this naming convention:poky-glibc-host_system-core-image-type-arch-toolchain[-ext]-release.sh Where: host_system is a string representing your development system: "i686" or "x86_64" type is a string representing the image: "sato" or "minimal" arch is a string representing the target architecture: "aarch64", "armv5e", "core2-64", "cortexa8hf-neon", "i586", "mips32r2", "mips64", or "ppc7400" release is the version of Yocto Project. NOTE: The standard SDK installer does not have the "-ext" string as part of the filename.
The toolchains provided by the Yocto Project are based off of the
core-image-sato
andcore-image-minimal
images and contain libraries appropriate for developing against those images.For example, if your build host is a 64-bit x86 system and you need an extended SDK for a 64-bit core2 target, go into the
x86_64
folder and download the following installer:poky-glibc-x86_64-core-image-sato-core2-64-toolchain-ext-3.3.1.sh
Run the Installer: Be sure you have execution privileges and run the installer. Following is an example from the
Downloads
directory:$ ~/Downloads/poky-glibc-x86_64-core-image-sato-core2-64-toolchain-ext-3.3.1.sh
During execution of the script, you choose the root location for the toolchain. See the “Installed Standard SDK Directory Structure” section and the “Installed Extensible SDK Directory Structure” section for more information.
5.2 Building an SDK Installer
As an alternative to locating and downloading an SDK installer, you can build the SDK installer. Follow these steps:
Set Up the Build Environment: Be sure you are set up to use BitBake in a shell. See the “Preparing the Build Host” section in the Yocto Project Development Tasks Manual for information on how to get a build host ready that is either a native Linux machine or a machine that uses CROPS.
Clone the ``poky`` Repository: You need to have a local copy of the Yocto Project Source Directory (i.e. a local
poky
repository). See the “Cloning the poky Repository” and possibly the “Checking Out by Branch in Poky” and “Checking Out by Tag in Poky” sections all in the Yocto Project Development Tasks Manual for information on how to clone thepoky
repository and check out the appropriate branch for your work.Initialize the Build Environment: While in the root directory of the Source Directory (i.e.
poky
), run the oe-init-build-env environment setup script to define the OpenEmbedded build environment on your build host.$ source oe-init-build-env
Among other things, the script creates the Build Directory, which is
build
in this case and is located in the Source Directory. After the script runs, your current working directory is set to thebuild
directory.Make Sure You Are Building an Installer for the Correct Machine: Check to be sure that your MACHINE variable in the
local.conf
file in your Build Directory matches the architecture for which you are building.Make Sure Your SDK Machine is Correctly Set: If you are building a toolchain designed to run on an architecture that differs from your current development host machine (i.e. the build host), be sure that the SDKMACHINE variable in the
local.conf
file in your Build Directory is correctly set.Note
If you are building an SDK installer for the Extensible SDK, the SDKMACHINE value must be set for the architecture of the machine you are using to build the installer. If SDKMACHINE is not set appropriately, the build fails and provides an error message similar to the following:
The extensible SDK can currently only be built for the same architecture as the machine being built on - SDK_ARCH is set to i686 (likely via setting SDKMACHINE) which is different from the architecture of the build machine (x86_64). Unable to continue.
Build the SDK Installer: To build the SDK installer for a standard SDK and populate the SDK image, use the following command form. Be sure to replace image with an image (e.g. “core-image-sato”): $ bitbake image -c populate_sdk You can do the same for the extensible SDK using this command form:
$ bitbake image -c populate_sdk_ext
These commands produce an SDK installer that contains the sysroot that matches your target root filesystem.
When the
bitbake
command completes, the SDK installer will be intmp/deploy/sdk
in the Build Directory.Note
By default, the previous BitBake command does not build static binaries. If you want to use the toolchain to build these types of libraries, you need to be sure your SDK has the appropriate static development libraries. Use the TOOLCHAIN_TARGET_TASK variable inside your
local.conf
file before building the SDK installer. Doing so ensures that the eventual SDK installation process installs the appropriate library packages as part of the SDK. Following is an example usinglibc
static development libraries: TOOLCHAIN_TARGET_TASK_append = ” libc-staticdev”
Run the Installer: You can now run the SDK installer from
tmp/deploy/sdk
in the Build Directory. Following is an example:$ cd poky/build/tmp/deploy/sdk $ ./poky-glibc-x86_64-core-image-sato-core2-64-toolchain-ext-3.3.1.sh
During execution of the script, you choose the root location for the toolchain. See the “Installed Standard SDK Directory Structure” section and the “Installed Extensible SDK Directory Structure” section for more information.
5.3 Extracting the Root Filesystem
After installing the toolchain, for some use cases you might need to separately extract a root filesystem:
You want to boot the image using NFS.
You want to use the root filesystem as the target sysroot.
You want to develop your target application using the root filesystem as the target sysroot.
Follow these steps to extract the root filesystem:
Locate and Download the Tarball for the Pre-Built Root Filesystem Image File: You need to find and download the root filesystem image file that is appropriate for your target system. These files are kept in machine-specific folders in the Index of Releases in the “machines” directory.
The machine-specific folders of the “machines” directory contain tarballs (
*.tar.bz2
) for supported machines. These directories also contain flattened root filesystem image files (*.ext4
), which you can use with QEMU directly.The pre-built root filesystem image files follow these naming conventions:
core-image-profile-arch.tar.bz2 Where: profile is the filesystem image's profile: lsb, lsb-dev, lsb-sdk, minimal, minimal-dev, minimal-initramfs, sato, sato-dev, sato-sdk, sato-sdk-ptest. For information on these types of image profiles, see the "Images" chapter in the Yocto Project Reference Manual. arch is a string representing the target architecture: beaglebone-yocto, beaglebone-yocto-lsb, edgerouter, edgerouter-lsb, genericx86, genericx86-64, genericx86-64-lsb, genericx86-lsb and qemu*.
The root filesystems provided by the Yocto Project are based off of the
core-image-sato
andcore-image-minimal
images.For example, if you plan on using a BeagleBone device as your target hardware and your image is a
core-image-sato-sdk
image, you can download the following file:core-image-sato-sdk-beaglebone-yocto.tar.bz2
Initialize the Cross-Development Environment: You must
source
the cross-development environment setup script to establish necessary environment variables.This script is located in the top-level directory in which you installed the toolchain (e.g.
poky_sdk
).Following is an example based on the toolchain installed in the “Locating Pre-Built SDK Installers” section:
$ source poky_sdk/environment-setup-core2-64-poky-linux
Extract the Root Filesystem: Use the
runqemu-extract-sdk
command and provide the root filesystem image.Following is an example command that extracts the root filesystem from a previously built root filesystem image that was downloaded from the Index of Releases. This command extracts the root filesystem into the
core2-64-sato
directory:$ runqemu-extract-sdk ~/Downloads/core-image-sato-sdk-beaglebone-yocto.tar.bz2 ~/beaglebone-sato
You could now point to the target sysroot at
beablebone-sato
.
5.4 Installed Standard SDK Directory Structure
The following figure shows the resulting directory structure after you
install the Standard SDK by running the *.sh
SDK installation
script:
The installed SDK consists of an environment setup script for the SDK, a
configuration file for the target, a version file for the target, and
the root filesystem (sysroots
) needed to develop objects for the
target system.
Within the figure, italicized text is used to indicate replaceable
portions of the file or directory name. For example, install_dir/version
is the directory where the SDK is installed. By default, this directory
is /opt/poky/
. And, version represents the specific snapshot of the
SDK (e.g. 3.3.1). Furthermore, target represents the target architecture
(e.g. i586
) and host represents the development system’s
architecture (e.g. x86_64
). Thus, the complete names of the two
directories within the sysroots
could be i586-poky-linux
and
x86_64-pokysdk-linux
for the target and host, respectively.
5.5 Installed Extensible SDK Directory Structure
The following figure shows the resulting directory structure after you
install the Extensible SDK by running the *.sh
SDK installation
script:
The installed directory structure for the extensible SDK is quite different than the installed structure for the standard SDK. The extensible SDK does not separate host and target parts in the same manner as does the standard SDK. The extensible SDK uses an embedded copy of the OpenEmbedded build system, which has its own sysroots.
Of note in the directory structure are an environment setup script for the SDK, a configuration file for the target, a version file for the target, and log files for the OpenEmbedded build system preparation script run by the installer and BitBake.
Within the figure, italicized text is used to indicate replaceable
portions of the file or directory name. For example, install_dir is the
directory where the SDK is installed, which is poky_sdk
by default,
and target represents the target architecture (e.g. i586
).
6 Customizing the Extensible SDK
This appendix describes customizations you can apply to the extensible SDK.
6.1 Configuring the Extensible SDK
The extensible SDK primarily consists of a pre-configured copy of the
OpenEmbedded build system from which it was produced. Thus, the SDK’s
configuration is derived using that build system and the filters shown
in the following list. When these filters are present, the OpenEmbedded
build system applies them against local.conf
and auto.conf
:
Variables whose values start with “/” are excluded since the assumption is that those values are paths that are likely to be specific to the Build Host.
Variables listed in SDK_LOCAL_CONF_BLACKLIST are excluded. These variables are not allowed through from the OpenEmbedded build system configuration into the extensible SDK configuration. Typically, these variables are specific to the machine on which the build system is running and could be problematic as part of the extensible SDK configuration.
For a list of the variables excluded by default, see the SDK_LOCAL_CONF_BLACKLIST in the glossary of the Yocto Project Reference Manual.
Variables listed in SDK_LOCAL_CONF_WHITELIST are included. Including a variable in the value of
SDK_LOCAL_CONF_WHITELIST
overrides either of the previous two filters. The default value is blank.Classes inherited globally with INHERIT that are listed in SDK_INHERIT_BLACKLIST are disabled. Using
SDK_INHERIT_BLACKLIST
to disable these classes is the typical method to disable classes that are problematic or unnecessary in the SDK context. The default value blacklists the buildhistory and icecc classes.
Additionally, the contents of conf/sdk-extra.conf
, when present, are
appended to the end of conf/local.conf
within the produced SDK,
without any filtering. The sdk-extra.conf
file is particularly
useful if you want to set a variable value just for the SDK and not the
OpenEmbedded build system used to create the SDK.
6.2 Adjusting the Extensible SDK to Suit Your Build Host’s Setup
In most cases, the extensible SDK defaults should work with your Build Host’s setup. However, some cases exist for which you might consider making adjustments:
If your SDK configuration inherits additional classes using the INHERIT variable and you do not need or want those classes enabled in the SDK, you can blacklist them by adding them to the SDK_INHERIT_BLACKLIST variable as described in the fourth bullet of the previous section.
Note
The default value of SDK_INHERIT_BLACKLIST is set using the “?=” operator. Consequently, you will need to either define the entire list by using the “=” operator, or you will need to append a value using either “_append” or the “+=” operator. You can learn more about these operators in the ” Basic Syntax “ section of the BitBake User Manual.
.
If you have classes or recipes that add additional tasks to the standard build flow (i.e. the tasks execute as the recipe builds as opposed to being called explicitly), then you need to do one of the following:
After ensuring the tasks are shared state tasks (i.e. the output of the task is saved to and can be restored from the shared state cache) or ensuring the tasks are able to be produced quickly from a task that is a shared state task, add the task name to the value of SDK_RECRDEP_TASKS.
Disable the tasks if they are added by a class and you do not need the functionality the class provides in the extensible SDK. To disable the tasks, add the class to the
SDK_INHERIT_BLACKLIST
variable as described in the previous section.
Generally, you want to have a shared state mirror set up so users of the SDK can add additional items to the SDK after installation without needing to build the items from source. See the “Providing Additional Installable Extensible SDK Content” section for information.
If you want users of the SDK to be able to easily update the SDK, you need to set the SDK_UPDATE_URL variable. For more information, see the “Providing Updates to the Extensible SDK After Installation” section.
If you have adjusted the list of files and directories that appear in COREBASE (other than layers that are enabled through
bblayers.conf
), then you must list these files in COREBASE_FILES so that the files are copied into the SDK.If your OpenEmbedded build system setup uses a different environment setup script other than oe-init-build-env, then you must set OE_INIT_ENV_SCRIPT to point to the environment setup script you use.
Note
You must also reflect this change in the value used for the COREBASE_FILES variable as previously described.
6.3 Changing the Extensible SDK Installer Title
You can change the displayed title for the SDK installer by setting the SDK_TITLE variable and then rebuilding the SDK installer. For information on how to build an SDK installer, see the “Building an SDK Installer” section.
By default, this title is derived from
DISTRO_NAME when it is
set. If the DISTRO_NAME
variable is not set, the title is derived
from the DISTRO variable.
The
populate_sdk_base
class defines the default value of the SDK_TITLE
variable as
follows:
SDK_TITLE ??= "${@d.getVar('DISTRO_NAME') or d.getVar('DISTRO')} SDK"
While several ways exist to change this variable, an efficient method is
to set the variable in your distribution’s configuration file. Doing so
creates an SDK installer title that applies across your distribution. As
an example, assume you have your own layer for your distribution named
“meta-mydistro” and you are using the same type of file hierarchy as
does the default “poky” distribution. If so, you could update the
SDK_TITLE
variable in the
~/meta-mydistro/conf/distro/mydistro.conf
file using the following
form:
SDK_TITLE = "your_title"
6.4 Providing Updates to the Extensible SDK After Installation
When you make changes to your configuration or to the metadata and if
you want those changes to be reflected in installed SDKs, you need to
perform additional steps. These steps make it possible for anyone using
the installed SDKs to update the installed SDKs by using the
devtool sdk-update
command:
Create a directory that can be shared over HTTP or HTTPS. You can do this by setting up a web server such as an Apache HTTP Server or Nginx server in the cloud to host the directory. This directory must contain the published SDK.
Set the SDK_UPDATE_URL variable to point to the corresponding HTTP or HTTPS URL. Setting this variable causes any SDK built to default to that URL and thus, the user does not have to pass the URL to the
devtool sdk-update
command as described in the “Applying Updates to an Installed Extensible SDK” section.Build the extensible SDK normally (i.e., use the
bitbake -c populate_sdk_ext
imagename command).Publish the SDK using the following command:
$ oe-publish-sdk some_path/sdk-installer.sh path_to_shared_http_directory
You must repeat this step each time you rebuild the SDK with changes that you want to make available through the update mechanism.
Completing the above steps allows users of the existing installed SDKs
to simply run devtool sdk-update
to retrieve and apply the latest
updates. See the
“Applying Updates to an Installed Extensible SDK”
section for further information.
6.5 Changing the Default SDK Installation Directory
When you build the installer for the Extensible SDK, the default installation directory for the SDK is based on the DISTRO and SDKEXTPATH variables from within the populate_sdk_base class as follows:
SDKEXTPATH ??= "~/${@d.getVar('DISTRO')}_sdk"
You can
change this default installation directory by specifically setting the
SDKEXTPATH
variable.
While a number of ways exist through which you can set this variable,
the method that makes the most sense is to set the variable in your
distribution’s configuration file. Doing so creates an SDK installer
default directory that applies across your distribution. As an example,
assume you have your own layer for your distribution named
“meta-mydistro” and you are using the same type of file hierarchy as
does the default “poky” distribution. If so, you could update the
SDKEXTPATH
variable in the
~/meta-mydistro/conf/distro/mydistro.conf
file using the following
form:
SDKEXTPATH = "some_path_for_your_installed_sdk"
After building your installer, running it prompts the user for acceptance of the some_path_for_your_installed_sdk directory as the default location to install the Extensible SDK.
6.6 Providing Additional Installable Extensible SDK Content
If you want the users of an extensible SDK you build to be able to add items to the SDK without requiring the users to build the items from source, you need to do a number of things:
Ensure the additional items you want the user to be able to install are already built:
Build the items explicitly. You could use one or more “meta” recipes that depend on lists of other recipes.
Build the “world” target and set
EXCLUDE_FROM_WORLD_pn-
recipename for the recipes you do not want built. See the EXCLUDE_FROM_WORLD variable for additional information.
Expose the
sstate-cache
directory produced by the build. Typically, you expose this directory by making it available through an Apache HTTP Server or Nginx server.Set the appropriate configuration so that the produced SDK knows how to find the configuration. The variable you need to set is SSTATE_MIRRORS:
SSTATE_MIRRORS = "file://.* http://example.com/some_path/sstate-cache/PATH"
You can set the
SSTATE_MIRRORS
variable in two different places:If the mirror value you are setting is appropriate to be set for both the OpenEmbedded build system that is actually building the SDK and the SDK itself (i.e. the mirror is accessible in both places or it will fail quickly on the OpenEmbedded build system side, and its contents will not interfere with the build), then you can set the variable in your
local.conf
or custom distro configuration file. You can then “whitelist” the variable through to the SDK by adding the following:SDK_LOCAL_CONF_WHITELIST = "SSTATE_MIRRORS"
Alternatively, if you just want to set the
SSTATE_MIRRORS
variable’s value for the SDK alone, create aconf/sdk-extra.conf
file either in your Build Directory or within any layer and put yourSSTATE_MIRRORS
setting within that file.Note
This second option is the safest option should you have any doubts as to which method to use when setting SSTATE_MIRRORS .
6.7 Minimizing the Size of the Extensible SDK Installer Download
By default, the extensible SDK bundles the shared state artifacts for
everything needed to reconstruct the image for which the SDK was built.
This bundling can lead to an SDK installer file that is a Gigabyte or
more in size. If the size of this file causes a problem, you can build
an SDK that has just enough in it to install and provide access to the
devtool command
by setting the following in your configuration:
SDK_EXT_TYPE = "minimal"
Setting
SDK_EXT_TYPE to
“minimal” produces an SDK installer that is around 35 Mbytes in size,
which downloads and installs quickly. You need to realize, though, that
the minimal installer does not install any libraries or tools out of the
box. These libraries and tools must be installed either “on the fly” or
through actions you perform using devtool
or explicitly with the
devtool sdk-install
command.
In most cases, when building a minimal SDK you need to also enable
bringing in the information on a wider range of packages produced by the
system. Requiring this wider range of information is particularly true
so that devtool add
is able to effectively map dependencies it
discovers in a source tree to the appropriate recipes. Additionally, the
information enables the devtool search
command to return useful
results.
To facilitate this wider range of information, you would need to set the following:
SDK_INCLUDE_PKGDATA = "1"
See the SDK_INCLUDE_PKGDATA variable for additional information.
Setting the SDK_INCLUDE_PKGDATA
variable as shown causes the “world”
target to be built so that information for all of the recipes included
within it are available. Having these recipes available increases build
time significantly and increases the size of the SDK installer by 30-80
Mbytes depending on how many recipes are included in your configuration.
You can use EXCLUDE_FROM_WORLD_pn-
recipename for recipes you want
to exclude. However, it is assumed that you would need to be building
the “world” target if you want to provide additional items to the SDK.
Consequently, building for “world” should not represent undue overhead
in most cases.
Note
If you set SDK_EXT_TYPE to “minimal”, then providing a shared state mirror is mandatory so that items can be installed as needed. See the ” Providing Additional Installable Extensible SDK Content “ section for more information.
You can explicitly control whether or not to include the toolchain when
you build an SDK by setting the
SDK_INCLUDE_TOOLCHAIN
variable to “1”. In particular, it is useful to include the toolchain
when you have set SDK_EXT_TYPE
to “minimal”, which by default,
excludes the toolchain. Also, it is helpful if you are building a small
SDK for use with an IDE or some other tool where you do not want to take
extra steps to install a toolchain.
7 Customizing the Standard SDK
This appendix presents customizations you can apply to the standard SDK.
7.1 Adding Individual Packages to the Standard SDK
When you build a standard SDK using the bitbake -c populate_sdk
, a
default set of packages is included in the resulting SDK. The
TOOLCHAIN_HOST_TASK
and
TOOLCHAIN_TARGET_TASK
variables control the set of packages adding to the SDK.
If you want to add individual packages to the toolchain that runs on the
host, simply add those packages to the TOOLCHAIN_HOST_TASK
variable.
Similarly, if you want to add packages to the default set that is part
of the toolchain that runs on the target, add the packages to the
TOOLCHAIN_TARGET_TASK
variable.
7.2 Adding API Documentation to the Standard SDK
You can include API documentation as well as any other documentation provided by recipes with the standard SDK by adding “api-documentation” to the DISTRO_FEATURES variable: DISTRO_FEATURES_append = “ api-documentation” Setting this variable as shown here causes the OpenEmbedded build system to build the documentation and then include it in the standard SDK.
8 Manual Revision History
Revision |
Date |
Note |
---|---|---|
2.1 |
April 2016 |
The initial document released with the Yocto Project 2.1 Release |
2.2 |
October 2016 |
Released with the Yocto Project 2.2 Release. |
2.3 |
May 2017 |
Released with the Yocto Project 2.3 Release. |
2.4 |
October 2017 |
Released with the Yocto Project 2.4 Release. |
2.5 |
May 2018 |
Released with the Yocto Project 2.5 Release. |
2.6 |
November 2018 |
Released with the Yocto Project 2.6 Release. |
2.7 |
May 2019 |
Released with the Yocto Project 2.7 Release. |
3.0 |
October 2019 |
Released with the Yocto Project 3.0 Release. |
3.1 |
April 2020 |
Released with the Yocto Project 3.1 Release. |
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Toaster User Manual
1 Introduction
Toaster is a web interface to the Yocto Project’s OpenEmbedded Build System. The interface enables you to configure and run your builds. Information about builds is collected and stored in a database. You can use Toaster to configure and start builds on multiple remote build servers.
1.1 Toaster Features
Toaster allows you to configure and run builds, and it provides extensive information about the build process.
Configure and Run Builds: You can use the Toaster web interface to configure and start your builds. Builds started using the Toaster web interface are organized into projects. When you create a project, you are asked to select a release, or version of the build system you want to use for the project builds. As shipped, Toaster supports Yocto Project releases 1.8 and beyond. With the Toaster web interface, you can:
Browse layers listed in the various layer sources that are available in your project (e.g. the OpenEmbedded Layer Index at https://layers.openembedded.org/).
Browse images, recipes, and machines provided by those layers.
Import your own layers for building.
Add and remove layers from your configuration.
Set configuration variables.
Select a target or multiple targets to build.
Start your builds.
Toaster also allows you to configure and run your builds from the command line, and switch between the command line and the web interface at any time. Builds started from the command line appear within a special Toaster project called “Command line builds”.
Information About the Build Process: Toaster also records extensive information about your builds. Toaster collects data for builds you start from the web interface and from the command line as long as Toaster is running.
Note
You must start Toaster before the build or it will not collect build data.
With Toaster you can:
See what was built (recipes and packages) and what packages were installed into your final image.
Browse the directory structure of your image.
See the value of all variables in your build configuration, and which files set each value.
Examine error, warning, and trace messages to aid in debugging.
See information about the BitBake tasks executed and reused during your build, including those that used shared state.
See dependency relationships between recipes, packages, and tasks.
See performance information such as build time, task time, CPU usage, and disk I/O.
For an overview of Toaster shipped with the Yocto Project 3.3.1 Release, see the “Toaster - Yocto Project 2.2” video.
1.2 Installation Options
You can set Toaster up to run as a local instance or as a shared hosted service.
When Toaster is set up as a local instance, all the components reside on a single build host. Fundamentally, a local instance of Toaster is suited for a single user developing on a single build host.
Toaster as a hosted service is suited for multiple users developing across several build hosts. When Toaster is set up as a hosted service, its components can be spread across several machines:
2 Preparing to Use Toaster
This chapter describes how you need to prepare your system in order to use Toaster.
2.1 Setting Up the Basic System Requirements
Before you can use Toaster, you need to first set up your build system to run the Yocto Project. To do this, follow the instructions in the “Preparing the Build Host” section of the Yocto Project Development Tasks Manual. For Ubuntu/Debian, you might also need to do an additional install of pip3.
$ sudo apt-get install python3-pip
2.2 Establishing Toaster System Dependencies
Toaster requires extra Python dependencies in order to run. A Toaster
requirements file named toaster-requirements.txt
defines the Python
dependencies. The requirements file is located in the bitbake
directory, which is located in the root directory of the
Source Directory (e.g.
poky/bitbake/toaster-requirements.txt
). The dependencies appear in a
pip
, install-compatible format.
2.2.1 Install Toaster Packages
You need to install the packages that Toaster requires. Use this command:
$ pip3 install --user -r bitbake/toaster-requirements.txt
The previous command installs the necessary Toaster modules into a local
python 3 cache in your $HOME
directory. The caches is actually
located in $HOME/.local
. To see what packages have been installed
into your $HOME
directory, do the following:
$ pip3 list installed --local
If you need to remove something, the following works:
$ pip3 uninstall PackageNameToUninstall
3 Setting Up and Using Toaster
3.1 Starting Toaster for Local Development
Once you have set up the Yocto Project and installed the Toaster system dependencies as described in the “Preparing to Use Toaster” chapter, you are ready to start Toaster.
Navigate to the root of your
Source Directory (e.g. poky
):
$ cd poky
Once in that directory, source the build environment script:
$ source oe-init-build-env
Next, from the build directory (e.g.
poky/build
), start Toaster using this command:
$ source toaster start
You can now run your builds from the command line, or with Toaster as explained in section “Using the Toaster Web Interface”.
To access the Toaster web interface, open your favorite browser and enter the following:
http://127.0.0.1:8000
3.2 Setting a Different Port
By default, Toaster starts on port 8000. You can use the WEBPORT
parameter to set a different port. For example, the following command
sets the port to “8400”:
$ source toaster start webport=8400
3.3 Setting Up Toaster Without a Web Server
You can start a Toaster environment without starting its web server. This is useful for the following:
Capturing a command-line build’s statistics into the Toaster database for examination later.
Capturing a command-line build’s statistics when the Toaster server is already running.
Having one instance of the Toaster web server track and capture multiple command-line builds, where each build is started in its own “noweb” Toaster environment.
The following commands show how to start a Toaster environment without starting its web server, perform BitBake operations, and then shut down the Toaster environment. Once the build is complete, you can close the Toaster environment. Before closing the environment, however, you should allow a few minutes to ensure the complete transfer of its BitBake build statistics to the Toaster database. If you have a separate Toaster web server instance running, you can watch this command-line build’s progress and examine the results as soon as they are posted:
$ source toaster start noweb
$ bitbake target
$ source toaster stop
3.4 Setting Up Toaster Without a Build Server
You can start a Toaster environment with the “New Projects” feature disabled. Doing so is useful for the following:
Sharing your build results over the web server while blocking others from starting builds on your host.
Allowing only local command-line builds to be captured into the Toaster database.
Use the following command to set up Toaster without a build server:
$ source toaster start nobuild webport=port
3.5 Setting up External Access
By default, Toaster binds to the loop back address (i.e. localhost
),
which does not allow access from external hosts. To allow external
access, use the WEBPORT
parameter to open an address that connects
to the network, specifically the IP address that your NIC uses to
connect to the network. You can also bind to all IP addresses the
computer supports by using the shortcut “0.0.0.0:port”.
The following example binds to all IP addresses on the host:
$ source toaster start webport=0.0.0.0:8400
This example binds to a specific IP address on the host’s NIC:
$ source toaster start webport=192.168.1.1:8400
3.6 The Directory for Cloning Layers
Toaster creates a _toaster_clones
directory inside your Source
Directory (i.e. poky
) to clone any layers needed for your builds.
Alternatively, if you would like all of your Toaster related files and
directories to be in a particular location other than the default, you
can set the TOASTER_DIR
environment variable, which takes precedence
over your current working directory. Setting this environment variable
causes Toaster to create and use $TOASTER_DIR./_toaster_clones
.
3.7 The Build Directory
Toaster creates a build directory within your Source Directory (e.g.
poky
) to execute the builds.
Alternatively, if you would like all of your Toaster related files and
directories to be in a particular location, you can set the
TOASTER_DIR
environment variable, which takes precedence over your
current working directory. Setting this environment variable causes
Toaster to use $TOASTER_DIR/build
as the build directory.
3.8 Creating a Django Superuser
Toaster is built on the Django framework. Django provides an administration interface you can use to edit Toaster configuration parameters.
To access the Django administration interface, you must create a superuser by following these steps:
If you used
pip3
, which is recommended, to set up the Toaster system dependencies, you need be sure the local user path is in yourPATH
list. To append the pip3 local user path, use the following command:$ export PATH=$PATH:$HOME/.local/bin
From the directory containing the Toaster database, which by default is the Build Directory, invoke the
createsuperuser
command frommanage.py
:$ cd poky/build $ ../bitbake/lib/toaster/manage.py createsuperuser
Django prompts you for the username, which you need to provide.
Django prompts you for an email address, which is optional.
Django prompts you for a password, which you must provide.
Django prompts you to re-enter your password for verification.
After completing these steps, the following confirmation message appears:
Superuser created successfully.
Creating a superuser allows you to access the Django administration interface through a browser. The URL for this interface is the same as the URL used for the Toaster instance with “/admin” on the end. For example, if you are running Toaster locally, use the following URL:
http://127.0.0.1:8000/admin
You can use the Django administration interface to set Toaster configuration parameters such as the build directory, layer sources, default variable values, and BitBake versions.
3.9 Setting Up a Production Instance of Toaster
You can use a production instance of Toaster to share the Toaster instance with remote users, multiple users, or both. The production instance is also the setup that can handle heavier loads on the web service. Use the instructions in the following sections to set up Toaster to run builds through the Toaster web interface.
3.9.1 Requirements
Be sure you meet the following requirements:
Note
You must comply with all Apache, mod-wsgi
, and Mysql requirements.
Have all the build requirements as described in the “Preparing to Use Toaster” chapter.
Have an Apache webserver.
Have
mod-wsgi
for the Apache webserver.Use the Mysql database server.
If you are using Ubuntu, run the following:
$ sudo apt-get install apache2 libapache2-mod-wsgi-py3 mysql-server python3-pip libmysqlclient-dev
If you are using Fedora or a RedHat distribution, run the following:
$ sudo dnf install httpd python3-mod_wsgi python3-pip mariadb-server mariadb-devel python3-devel
If you are using openSUSE, run the following:
$ sudo zypper install apache2 apache2-mod_wsgi-python3 python3-pip mariadb mariadb-client python3-devel
3.9.2 Installation
Perform the following steps to install Toaster:
Create toaster user and set its home directory to
/var/www/toaster
:$ sudo /usr/sbin/useradd toaster -md /var/www/toaster -s /bin/false $ sudo su - toaster -s /bin/bash
Checkout a copy of
poky
into the web server directory. You will be using/var/www/toaster
:$ git clone git://git.yoctoproject.org/poky $ git checkout hardknott
Install Toaster dependencies using the –user flag which keeps the Python packages isolated from your system-provided packages:
$ cd /var/www/toaster/ $ pip3 install --user -r ./poky/bitbake/toaster-requirements.txt $ pip3 install --user mysqlclient
Note
Isolating these packages is not required but is recommended. Alternatively, you can use your operating system’s package manager to install the packages.
Configure Toaster by editing
/var/www/toaster/poky/bitbake/lib/toaster/toastermain/settings.py
as follows:Edit the DATABASES settings:
DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'toaster_data', 'USER': 'toaster', 'PASSWORD': 'yourpasswordhere', 'HOST': 'localhost', 'PORT': '3306', } }
Edit the SECRET_KEY:
SECRET_KEY = 'your_secret_key'
Edit the STATIC_ROOT:
STATIC_ROOT = '/var/www/toaster/static_files/'
Add the database and user to the
mysql
server defined earlier:$ mysql -u root -p mysql> CREATE DATABASE toaster_data; mysql> CREATE USER 'toaster'@'localhost' identified by 'yourpasswordhere'; mysql> GRANT all on toaster_data.\* to 'toaster'@'localhost'; mysql> quit
Get Toaster to create the database schema, default data, and gather the statically-served files:
$ cd /var/www/toaster/poky/ $ ./bitbake/lib/toaster/manage.py migrate $ TOASTER_DIR=`pwd\` TEMPLATECONF='poky' \ ./bitbake/lib/toaster/manage.py checksettings $ ./bitbake/lib/toaster/manage.py collectstatic
In the previous example, from the
poky
directory, themigrate
command ensures the database schema changes have propagated correctly (i.e. migrations). The next line sets the Toaster root directoryTOASTER_DIR
and the location of the Toaster configuration fileTOASTER_CONF
, which is relative toTOASTER_DIR
. TheTEMPLATECONF
value reflects the contents ofpoky/.templateconf
, and by default, should include the string “poky”. For more information on the Toaster configuration file, see the “Configuring Toaster” section.This line also runs the
checksettings
command, which configures the location of the Toaster Build Directory. The Toaster root directoryTOASTER_DIR
determines where the Toaster build directory is created on the file system. In the example above,TOASTER_DIR
is set as follows:/var/www/toaster/poky
This setting causes the Toaster build directory to be:
/var/www/toaster/poky/build
Finally, the
collectstatic
command is a Django framework command that collects all the statically served files into a designated directory to be served up by the Apache web server as defined bySTATIC_ROOT
.Test and/or use the Mysql integration with Toaster’s Django web server. At this point, you can start up the normal Toaster Django web server with the Toaster database in Mysql. You can use this web server to confirm that the database migration and data population from the Layer Index is complete.
To start the default Toaster Django web server with the Toaster database now in Mysql, use the standard start commands:
$ source oe-init-build-env $ source toaster start
Additionally, if Django is sufficient for your requirements, you can use it for your release system and migrate later to Apache as your requirements change.
Add an Apache configuration file for Toaster to your Apache web server’s configuration directory. If you are using Ubuntu or Debian, put the file here:
/etc/apache2/conf-available/toaster.conf
If you are using Fedora or RedHat, put it here:
/etc/httpd/conf.d/toaster.conf
If you are using openSUSE, put it here:
/etc/apache2/conf.d/toaster.conf
Following is a sample Apache configuration for Toaster you can follow:
Alias /static /var/www/toaster/static_files <Directory /var/www/toaster/static_files> <IfModule mod_access_compat.c> Order allow,deny Allow from all </IfModule> <IfModule !mod_access_compat.c> Require all granted </IfModule> </Directory> <Directory /var/www/toaster/poky/bitbake/lib/toaster/toastermain> <Files "wsgi.py"> Require all granted </Files> </Directory> WSGIDaemonProcess toaster_wsgi python-path=/var/www/toaster/poky/bitbake/lib/toaster:/var/www/toaster/.local/lib/python3.4/site-packages WSGIScriptAlias / "/var/www/toaster/poky/bitbake/lib/toaster/toastermain/wsgi.py" <Location /> WSGIProcessGroup toaster_wsgi </Location>
If you are using Ubuntu or Debian, you will need to enable the config and module for Apache:
$ sudo a2enmod wsgi $ sudo a2enconf toaster $ chmod +x bitbake/lib/toaster/toastermain/wsgi.py
Finally, restart Apache to make sure all new configuration is loaded. For Ubuntu, Debian, and openSUSE use:
$ sudo service apache2 restart
For Fedora and RedHat use:
$ sudo service httpd restart
Prepare the systemd service to run Toaster builds. Here is a sample configuration file for the service:
[Unit] Description=Toaster runbuilds [Service] Type=forking User=toaster ExecStart=/usr/bin/screen -d -m -S runbuilds /var/www/toaster/poky/bitbake/lib/toaster/runbuilds-service.sh start ExecStop=/usr/bin/screen -S runbuilds -X quit WorkingDirectory=/var/www/toaster/poky [Install] WantedBy=multi-user.target
Prepare the
runbuilds-service.sh
script that you need to place in the/var/www/toaster/poky/bitbake/lib/toaster/
directory by setting up executable permissions:#!/bin/bash #export http_proxy=http://proxy.host.com:8080 #export https_proxy=http://proxy.host.com:8080 #export GIT_PROXY_COMMAND=$HOME/bin/gitproxy cd poky/ source ./oe-init-build-env build source ../bitbake/bin/toaster $1 noweb [ "$1" == 'start' ] && /bin/bash
Run the service:
$ sudo service runbuilds start
Since the service is running in a detached screen session, you can attach to it using this command:
$ sudo su - toaster $ screen -rS runbuilds
You can detach from the service again using “Ctrl-a” followed by “d” key combination.
You can now open up a browser and start using Toaster.
3.10 Using the Toaster Web Interface
The Toaster web interface allows you to do the following:
Browse published layers in the OpenEmbedded Layer Index that are available for your selected version of the build system.
Import your own layers for building.
Add and remove layers from your configuration.
Set configuration variables.
Select a target or multiple targets to build.
Start your builds.
See what was built (recipes and packages) and what packages were installed into your final image.
Browse the directory structure of your image.
See the value of all variables in your build configuration, and which files set each value.
Examine error, warning and trace messages to aid in debugging.
See information about the BitBake tasks executed and reused during your build, including those that used shared state.
See dependency relationships between recipes, packages and tasks.
See performance information such as build time, task time, CPU usage, and disk I/O.
3.10.1 Toaster Web Interface Videos
Following are several videos that show how to use the Toaster GUI:
Build Configuration: This video overviews and demonstrates build configuration for Toaster.
Build Custom Layers: This video shows you how to build custom layers that are used with Toaster.
Toaster Homepage and Table Controls: This video goes over the Toaster entry page, and provides an overview of the data manipulation capabilities of Toaster, which include search, sorting and filtering by different criteria.
Build Dashboard: This video shows you the build dashboard, a page providing an overview of the information available for a selected build.
Image Information: This video walks through the information Toaster provides about images: packages installed and root file system.
Configuration: This video provides Toaster build configuration information.
Tasks: This video shows the information Toaster provides about the tasks run by the build system.
Recipes and Packages Built: This video shows the information Toaster provides about recipes and packages built.
Performance Data: This video shows the build performance data provided by Toaster.
3.10.2 Additional Information About the Local Yocto Project Release
This section only applies if you have set up Toaster for local development, as explained in the “Starting Toaster for Local Development” section.
When you create a project in Toaster, you will be asked to provide a name and to select a Yocto Project release. One of the release options you will find is called “Local Yocto Project”.
When you select the “Local Yocto Project” release, Toaster will run your builds using the local Yocto Project clone you have in your computer: the same clone you are using to run Toaster. Unless you manually update this clone, your builds will always use the same Git revision.
If you select any of the other release options, Toaster will fetch the tip of your selected release from the upstream Yocto Project repository every time you run a build. Fetching this tip effectively means that if your selected release is updated upstream, the Git revision you are using for your builds will change. If you are doing development locally, you might not want this change to happen. In that case, the “Local Yocto Project” release might be the right choice.
However, the “Local Yocto Project” release will not provide you with any compatible layers, other than the three core layers that come with the Yocto Project:
If you want to build any other layers, you will need to manually import them into your Toaster project, using the “Import layer” page.
3.10.3 Building a Specific Recipe Given Multiple Versions
Occasionally, a layer might provide more than one version of the same
recipe. For example, the openembedded-core
layer provides two
versions of the bash
recipe (i.e. 3.2.48 and 4.3.30-r0) and two
versions of the which
recipe (i.e. 2.21 and 2.18). The following
figure shows this exact scenario:
By default, the OpenEmbedded build system builds one of the two recipes.
For the bash
case, version 4.3.30-r0 is built by default.
Unfortunately, Toaster as it exists, is not able to override the default
recipe version. If you would like to build bash 3.2.48, you need to set
the
PREFERRED_VERSION
variable. You can do so from Toaster, using the “Add variable” form,
which is available in the “BitBake variables” page of the project
configuration section as shown in the following screen:
To specify bash
3.2.48 as the version to build, enter
“PREFERRED_VERSION_bash” in the “Variable” field, and “3.2.48” in the
“Value” field. Next, click the “Add variable” button:
After clicking the “Add variable” button, the settings for
PREFERRED_VERSION
are added to the bottom of the BitBake variables
list. With these settings, the OpenEmbedded build system builds the
desired version of the recipe rather than the default version:
4 Concepts and Reference
In order to configure and use Toaster, you should understand some
concepts and have some basic command reference material available. This
final chapter provides conceptual information on layer sources,
releases, and JSON configuration files. Also provided is a quick look at
some useful manage.py
commands that are Toaster-specific.
Information on manage.py
commands does exist across the Web and the
information in this manual by no means attempts to provide a command
comprehensive reference.
4.1 Layer Source
In general, a “layer source” is a source of information about existing layers. In particular, we are concerned with layers that you can use with the Yocto Project and Toaster. This chapter describes a particular type of layer source called a “layer index.”
A layer index is a web application that contains information about a set of custom layers. A good example of an existing layer index is the OpenEmbedded Layer Index. A public instance of this layer index exists at https://layers.openembedded.org/. You can find the code for this layer index’s web application at https://git.yoctoproject.org/cgit/cgit.cgi/layerindex-web/.
When you tie a layer source into Toaster, it can query the layer source
through a
REST
API, store the information about the layers in the Toaster database, and
then show the information to users. Users are then able to view that
information and build layers from Toaster itself without worrying about
cloning or editing the BitBake layers configuration file
bblayers.conf
.
Tying a layer source into Toaster is convenient when you have many custom layers that need to be built on a regular basis by a community of developers. In fact, Toaster comes pre-configured with the OpenEmbedded Metadata Index.
Note
You do not have to use a layer source to use Toaster. Tying into a layer source is optional.
4.1.1 Setting Up and Using a Layer Source
To use your own layer source, you need to set up the layer source and then tie it into Toaster. This section describes how to tie into a layer index in a manner similar to the way Toaster ties into the OpenEmbedded Metadata Index.
4.1.1.1 Understanding Your Layers
The obvious first step for using a layer index is to have several custom layers that developers build and access using the Yocto Project on a regular basis. This set of layers needs to exist and you need to be familiar with where they reside. You will need that information when you set up the code for the web application that “hooks” into your set of layers.
For general information on layers, see the “The Yocto Project Layer Model” section in the Yocto Project Overview and Concepts Manual. For information on how to create layers, see the “Understanding and Creating Layers” section in the Yocto Project Development Tasks Manual.
4.1.1.2 Configuring Toaster to Hook Into Your Layer Index
If you want Toaster to use your layer index, you must host the web application in a server to which Toaster can connect. You also need to give Toaster the information about your layer index. In other words, you have to configure Toaster to use your layer index. This section describes two methods by which you can configure and use your layer index.
In the previous section, the code for the OpenEmbedded Metadata Index (i.e. https://layers.openembedded.org/) was referenced. You can use this code, which is at https://git.yoctoproject.org/cgit/cgit.cgi/layerindex-web/, as a base to create your own layer index.
4.1.1.2.1 Use the Administration Interface
Access the administration interface through a browser by entering the
URL of your Toaster instance and adding “/admin
” to the end of the
URL. As an example, if you are running Toaster locally, use the
following URL:
http://127.0.0.1:8000/admin
The administration interface has a “Layer sources” section that includes an “Add layer source” button. Click that button and provide the required information. Make sure you select “layerindex” as the layer source type.
4.1.1.2.2 Use the Fixture Feature
The Django fixture feature overrides the default layer server when you
use it to specify a custom URL. To use the fixture feature, create (or
edit) the file bitbake/lib/toaster.orm/fixtures/custom.xml
, and then
set the following Toaster setting to your custom URL:
<?xml version="1.0" ?>
<django-objects version="1.0">
<object model="orm.toastersetting" pk="100">
<field name="name" type="CharField">CUSTOM_LAYERINDEX_SERVER</field>
<field name="value" type="CharField">https://layers.my_organization.org/layerindex/branch/master/layers/</field>
</object>
<django-objects>
When you start Toaster for the first time, or
if you delete the file toaster.sqlite
and restart, the database will
populate cleanly from this layer index server.
Once the information has been updated, verify the new layer information is available by using the Toaster web interface. To do that, visit the “All compatible layers” page inside a Toaster project. The layers from your layer source should be listed there.
If you change the information in your layer index server, refresh the Toaster database by running the following command:
$ bitbake/lib/toaster/manage.py lsupdates
If Toaster can reach the API URL, you should see a message telling you that Toaster is updating the layer source information.
4.2 Releases
When you create a Toaster project using the web interface, you are asked to choose a “Release.” In the context of Toaster, the term “Release” refers to a set of layers and a BitBake version the OpenEmbedded build system uses to build something. As shipped, Toaster is pre-configured with releases that correspond to Yocto Project release branches. However, you can modify, delete, and create new releases according to your needs. This section provides some background information on releases.
4.2.1 Pre-Configured Releases
As shipped, Toaster is configured to use a specific set of releases. Of course, you can always configure Toaster to use any release. For example, you might want your project to build against a specific commit of any of the “out-of-the-box” releases. Or, you might want your project to build against different revisions of OpenEmbedded and BitBake.
As shipped, Toaster is configured to work with the following releases:
Yocto Project 3.3.1 “Hardknott” or OpenEmbedded “Hardknott”: This release causes your Toaster projects to build against the head of the hardknott branch at https://git.yoctoproject.org/cgit/cgit.cgi/poky/log/?h=hardknott or https://git.openembedded.org/openembedded-core/commit/?h=hardknott.
Yocto Project “Master” or OpenEmbedded “Master”: This release causes your Toaster Projects to build against the head of the master branch, which is where active development takes place, at https://git.yoctoproject.org/cgit/cgit.cgi/poky/log/ or https://git.openembedded.org/openembedded-core/log/.
Local Yocto Project or Local OpenEmbedded: This release causes your Toaster Projects to build against the head of the
poky
oropenembedded-core
clone you have local to the machine running Toaster.
4.3 Configuring Toaster
In order to use Toaster, you must configure the database with the default content. The following subsections describe various aspects of Toaster configuration.
4.3.1 Configuring the Workflow
The bldcontrol/management/commands/checksettings.py
file controls
workflow configuration. The following steps outline the process to
initially populate this database.
The default project settings are set from
orm/fixtures/settings.xml
.The default project distro and layers are added from
orm/fixtures/poky.xml
if poky is installed. If poky is not installed, they are added fromorm/fixtures/oe-core.xml
.If the
orm/fixtures/custom.xml
file exists, then its values are added.The layer index is then scanned and added to the database.
Once these steps complete, Toaster is set up and ready to use.
4.3.2 Customizing Pre-Set Data
The pre-set data for Toaster is easily customizable. You can create the
orm/fixtures/custom.xml
file to customize the values that go into
the database. Customization is additive, and can either extend or
completely replace the existing values.
You use the orm/fixtures/custom.xml
file to change the default
project settings for the machine, distro, file images, and layers. When
creating a new project, you can use the file to define the offered
alternate project release selections. For example, you can add one or
more additional selections that present custom layer sets or distros,
and any other local or proprietary content.
Additionally, you can completely disable the content from the
oe-core.xml
and poky.xml
files by defining the section shown
below in the settings.xml
file. For example, this option is
particularly useful if your custom configuration defines fewer releases
or layers than the default fixture files.
The following example sets “name” to “CUSTOM_XML_ONLY” and its value to “True”.
<object model="orm.toastersetting" pk="99">
<field type="CharField" name="name">CUSTOM_XML_ONLY</field>
<field type="CharField" name="value">True</field>
</object>
4.3.3 Understanding Fixture File Format
The following is an overview of the file format used by the
oe-core.xml
, poky.xml
, and custom.xml
files.
The following subsections describe each of the sections in the fixture
files, and outline an example section of the XML code. you can use to
help understand this information and create a local custom.xml
file.
4.3.3.1 Defining the Default Distro and Other Values
This section defines the default distro value for new projects. By default, it reserves the first Toaster Setting record “1”. The following demonstrates how to set the project default value for DISTRO:
<!-- Set the project default value for DISTRO -->
<object model="orm.toastersetting" pk="1">
<field type="CharField" name="name">DEFCONF_DISTRO</field>
<field type="CharField" name="value">poky</field>
</object>
You can override
other default project values by adding additional Toaster Setting
sections such as any of the settings coming from the settings.xml
file. Also, you can add custom values that are included in the BitBake
environment. The “pk” values must be unique. By convention, values that
set default project values have a “DEFCONF” prefix.
4.3.3.2 Defining BitBake Version
The following defines which version of BitBake is used for the following release selection:
<!-- Bitbake versions which correspond to the metadata release -->
<object model="orm.bitbakeversion" pk="1">
<field type="CharField" name="name">hardknott</field>
<field type="CharField" name="giturl">git://git.yoctoproject.org/poky</field>
<field type="CharField" name="branch">hardknott</field>
<field type="CharField" name="dirpath">bitbake</field>
</object>
4.3.3.3 Defining Release
The following defines the releases when you create a new project:
<!-- Releases available -->
<object model="orm.release" pk="1">
<field type="CharField" name="name">hardknott</field>
<field type="CharField" name="description">Yocto Project 3.3.1 "Hardknott"</field>
<field rel="ManyToOneRel" to="orm.bitbakeversion" name="bitbake_version">1</field>
<field type="CharField" name="branch_name">hardknott</field>
<field type="TextField" name="helptext">Toaster will run your builds using the tip of the <a href="http://git.yoctoproject.org/cgit/cgit.cgi/poky/log/?h=hardknott">Yocto Project Hardknott branch</a>.</field>
</object>
The “pk” value must match the above respective BitBake version record.
4.3.3.4 Defining the Release Default Layer Names
The following defines the default layers for each release:
<!-- Default project layers for each release -->
<object model="orm.releasedefaultlayer" pk="1">
<field rel="ManyToOneRel" to="orm.release" name="release">1</field>
<field type="CharField" name="layer_name">openembedded-core</field>
</object>
The ‘pk’ values in the example above should start at “1” and increment uniquely. You can use the same layer name in multiple releases.
4.3.3.5 Defining Layer Definitions
Layer definitions are the most complex. The following defines each of
the layers, and then defines the exact layer version of the layer used
for each respective release. You must have one orm.layer
entry for
each layer. Then, with each entry you need a set of
orm.layer_version
entries that connects the layer with each release
that includes the layer. In general all releases include the layer.
<object model="orm.layer" pk="1">
<field type="CharField" name="name">openembedded-core</field>
<field type="CharField" name="layer_index_url"></field>
<field type="CharField" name="vcs_url">git://git.yoctoproject.org/poky</field>
<field type="CharField" name="vcs_web_url">http://git.yoctoproject.org/cgit/cgit.cgi/poky</field>
<field type="CharField" name="vcs_web_tree_base_url">http://git.yoctoproject.org/cgit/cgit.cgi/poky/tree/%path%?h=%branch%</field>
<field type="CharField" name="vcs_web_file_base_url">http://git.yoctoproject.org/cgit/cgit.cgi/poky/tree/%path%?h=%branch%</field>
</object>
<object model="orm.layer_version" pk="1">
<field rel="ManyToOneRel" to="orm.layer" name="layer">1</field>
<field type="IntegerField" name="layer_source">0</field>
<field rel="ManyToOneRel" to="orm.release" name="release">1</field>
<field type="CharField" name="branch">hardknott</field>
<field type="CharField" name="dirpath">meta</field>
</object> <object model="orm.layer_version" pk="2">
<field rel="ManyToOneRel" to="orm.layer" name="layer">1</field>
<field type="IntegerField" name="layer_source">0</field>
<field rel="ManyToOneRel" to="orm.release" name="release">2</field>
<field type="CharField" name="branch">HEAD</field>
<field type="CharField" name="commit">HEAD</field>
<field type="CharField" name="dirpath">meta</field>
</object>
<object model="orm.layer_version" pk="3">
<field rel="ManyToOneRel" to="orm.layer" name="layer">1</field>
<field type="IntegerField" name="layer_source">0</field>
<field rel="ManyToOneRel" to="orm.release" name="release">3</field>
<field type="CharField" name="branch">master</field>
<field type="CharField" name="dirpath">meta</field>
</object>
The layer “pk” values above must be unique, and typically start at “1”. The layer version “pk” values must also be unique across all layers, and typically start at “1”.
4.4 Remote Toaster Monitoring
Toaster has an API that allows remote management applications to directly query the state of the Toaster server and its builds in a machine-to-machine manner. This API uses the REST interface and the transfer of JSON files. For example, you might monitor a build inside a container through well supported known HTTP ports in order to easily access a Toaster server inside the container. In this example, when you use this direct JSON API, you avoid having web page parsing against the display the user sees.
4.4.1 Checking Health
Before you use remote Toaster monitoring, you should do a health check. To do this, ping the Toaster server using the following call to see if it is still alive:
http://host:port/health
Be sure to provide values for host and port. If the server is alive, you will get the response HTML:
<!DOCTYPE html>
<html lang="en">
<head><title>Toaster Health</title></head>
<body>Ok</body>
</html>
4.4.2 Determining Status of Builds in Progress
Sometimes it is useful to determine the status of a build in progress. To get the status of pending builds, use the following call:
http://host:port/toastergui/api/building
Be sure to provide values for host and port. The output is a JSON file that itemizes all builds in progress. This file includes the time in seconds since each respective build started as well as the progress of the cloning, parsing, and task execution. The following is sample output for a build in progress:
{"count": 1,
"building": [
{"machine": "beaglebone",
"seconds": "463.869",
"task": "927:2384",
"distro": "poky",
"clone": "1:1",
"id": 2,
"start": "2017-09-22T09:31:44.887Z",
"name": "20170922093200",
"parse": "818:818",
"project": "my_rocko",
"target": "core-image-minimal"
}]
}
The JSON data for this query is returned in a single line. In the previous example the line has been artificially split for readability.
4.4.3 Checking Status of Builds Completed
Once a build is completed, you get the status when you use the following call:
http://host:port/toastergui/api/builds
Be sure to provide values for host and port. The output is a JSON file that itemizes all complete builds, and includes build summary information. The following is sample output for a completed build:
{"count": 1,
"builds": [
{"distro": "poky",
"errors": 0,
"machine": "beaglebone",
"project": "my_rocko",
"stop": "2017-09-22T09:26:36.017Z",
"target": "quilt-native",
"seconds": "78.193",
"outcome": "Succeeded",
"id": 1,
"start": "2017-09-22T09:25:17.824Z",
"warnings": 1,
"name": "20170922092618"
}]
}
The JSON data for this query is returned in a single line. In the previous example the line has been artificially split for readability.
4.4.4 Determining Status of a Specific Build
Sometimes it is useful to determine the status of a specific build. To get the status of a specific build, use the following call:
http://host:port/toastergui/api/build/ID
Be sure to provide values for host, port, and ID. You can find the value for ID from the Builds Completed query. See the “Checking Status of Builds Completed” section for more information.
The output is a JSON file that itemizes the specific build and includes build summary information. The following is sample output for a specific build:
{"build":
{"distro": "poky",
"errors": 0,
"machine": "beaglebone",
"project": "my_rocko",
"stop": "2017-09-22T09:26:36.017Z",
"target": "quilt-native",
"seconds": "78.193",
"outcome": "Succeeded",
"id": 1,
"start": "2017-09-22T09:25:17.824Z",
"warnings": 1,
"name": "20170922092618",
"cooker_log": "/opt/user/poky/build-toaster-2/tmp/log/cooker/beaglebone/build_20170922_022607.991.log"
}
}
The JSON data for this query is returned in a single line. In the previous example the line has been artificially split for readability.
4.5 Useful Commands
In addition to the web user interface and the scripts that start and
stop Toaster, command-line commands exist through the manage.py
management script. You can find general documentation on manage.py
at the
Django
site. However, several manage.py
commands have been created that are
specific to Toaster and are used to control configuration and back-end
tasks. You can locate these commands in the
Source Directory (e.g. poky
) at
bitbake/lib/manage.py
. This section documents those commands.
Note
When using
manage.py
commands given a default configuration, you must be sure that your working directory is set to the Build Directory. Usingmanage.py
commands from the Build Directory allows Toaster to find thetoaster.sqlite
file, which is located in the Build Directory.For non-default database configurations, it is possible that you can use
manage.py
commands from a directory other than the Build Directory. To do so, thetoastermain/settings.py
file must be configured to point to the correct database backend.
4.5.1 buildslist
The buildslist
command lists all builds that Toaster has recorded.
Access the command as follows:
$ bitbake/lib/toaster/manage.py buildslist
The command returns a list, which includes numeric identifications, of the builds that Toaster has recorded in the current database.
You need to run the buildslist
command first to identify existing
builds in the database before using the
builddelete command. Here is an
example that assumes default repository and build directory names:
$ cd poky/build
$ python ../bitbake/lib/toaster/manage.py buildslist
If your Toaster database had only one build, the above buildslist command would return something like the following:
1: qemux86 poky core-image-minimal
4.5.2 builddelete
The builddelete
command deletes data associated with a build. Access
the command as follows:
$ bitbake/lib/toaster/manage.py builddelete build_id
The command deletes all the build data for the specified build_id. This command is useful for removing old and unused data from the database.
Prior to running the builddelete
command, you need to get the ID
associated with builds by using the
buildslist command.
4.5.3 perf
The perf
command measures Toaster performance. Access the command as
follows:
$ bitbake/lib/toaster/manage.py perf
The command is a sanity check that returns page loading times in order to identify performance problems.
4.5.4 checksettings
The checksettings
command verifies existing Toaster settings. Access
the command as follows:
$ bitbake/lib/toaster/manage.py checksettings
Toaster uses settings that are based on the database to configure the
building tasks. The checksettings
command verifies that the database
settings are valid in the sense that they have the minimal information
needed to start a build.
In order for the checksettings
command to work, the database must be
correctly set up and not have existing data. To be sure the database is
ready, you can run the following:
$ bitbake/lib/toaster/manage.py syncdb
$ bitbake/lib/toaster/manage.py migrate orm
$ bitbake/lib/toaster/manage.py migrate bldcontrol
After running these commands, you can run the checksettings
command.
4.5.5 runbuilds
The runbuilds
command launches scheduled builds. Access the command
as follows:
$ bitbake/lib/toaster/manage.py runbuilds
The runbuilds
command checks if scheduled builds exist in the database
and then launches them per schedule. The command returns after the builds
start but before they complete. The Toaster Logging Interface records and
updates the database when the builds complete.
5 Manual Revision History
Revision |
Date |
Note |
---|---|---|
1.8 |
April 2015 |
The initial document released with the Yocto Project 1.8 Release |
2.0 |
October 2015 |
Released with the Yocto Project 2.0 Release. |
2.1 |
April 2016 |
Released with the Yocto Project 2.1 Release. |
2.2 |
October 2016 |
Released with the Yocto Project 2.2 Release. |
2.3 |
May 2017 |
Released with the Yocto Project 2.3 Release. |
2.4 |
October 2017 |
Released with the Yocto Project 2.4 Release. |
2.5 |
May 2018 |
Released with the Yocto Project 2.5 Release. |
2.6 |
November 2018 |
Released with the Yocto Project 2.6 Release. |
2.7 |
May 2019 |
Released with the Yocto Project 2.7 Release. |
3.0 |
October 2019 |
Released with the Yocto Project 3.0 Release. |
3.1 |
April 2020 |
Released with the Yocto Project 3.1 Release. |
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
Yocto Project Test Environment Manual
1 The Yocto Project Test Environment Manual
1.1 Welcome
Welcome to the Yocto Project Test Environment Manual! This manual is a work in progress. The manual contains information about the testing environment used by the Yocto Project to make sure each major and minor release works as intended. All the project’s testing infrastructure and processes are publicly visible and available so that the community can see what testing is being performed, how it’s being done and the current status of the tests and the project at any given time. It is intended that Other organizations can leverage off the process and testing environment used by the Yocto Project to create their own automated, production test environment, building upon the foundations from the project core.
Currently, the Yocto Project Test Environment Manual has no projected release date. This manual is a work-in-progress and is being initially loaded with information from the README files and notes from key engineers:
yocto-autobuilder2: This README.md is the main README which details how to set up the Yocto Project Autobuilder. The
yocto-autobuilder2
repository represents the Yocto Project’s console UI plugin to Buildbot and the configuration necessary to configure Buildbot to perform the testing the project requires.yocto-autobuilder-helper: This README and repository contains Yocto Project Autobuilder Helper scripts and configuration. The
yocto-autobuilder-helper
repository contains the “glue” logic that defines which tests to run and how to run them. As a result, it can be used by any Continuous Improvement (CI) system to run builds, support getting the correct code revisions, configure builds and layers, run builds, and collect results. The code is independent of any CI system, which means the code can work Buildbot, Jenkins, or others. This repository has a branch per release of the project defining the tests to run on a per release basis.
1.2 Yocto Project Autobuilder Overview
The Yocto Project Autobuilder collectively refers to the software, tools, scripts, and procedures used by the Yocto Project to test released software across supported hardware in an automated and regular fashion. Basically, during the development of a Yocto Project release, the Autobuilder tests if things work. The Autobuilder builds all test targets and runs all the tests.
The Yocto Project uses now uses standard upstream
Buildbot (version 9) to
drive its integration and testing. Buildbot Nine has a plug-in interface
that the Yocto Project customizes using code from the
yocto-autobuilder2
repository, adding its own console UI plugin. The
resulting UI plug-in allows you to visualize builds in a way suited to
the project’s needs.
A helper
layer provides configuration and job management through
scripts found in the yocto-autobuilder-helper
repository. The
helper
layer contains the bulk of the build configuration
information and is release-specific, which makes it highly customizable
on a per-project basis. The layer is CI system-agnostic and contains a
number of Helper scripts that can generate build configurations from
simple JSON files.
Note
The project uses Buildbot for historical reasons but also because many of the project developers have knowledge of python. It is possible to use the outer layers from another Continuous Integration (CI) system such as Jenkins instead of Buildbot.
The following figure shows the Yocto Project Autobuilder stack with a topology that includes a controller and a cluster of workers:
1.3 Yocto Project Tests - Types of Testing Overview
The Autobuilder tests different elements of the project by using the following types of tests:
Build Testing: Tests whether specific configurations build by varying MACHINE, DISTRO, other configuration options, and the specific target images being built (or world). Used to trigger builds of all the different test configurations on the Autobuilder. Builds usually cover many different targets for different architectures, machines, and distributions, as well as different configurations, such as different init systems. The Autobuilder tests literally hundreds of configurations and targets.
Sanity Checks During the Build Process: Tests initiated through the insane class. These checks ensure the output of the builds are correct. For example, does the ELF architecture in the generated binaries match the target system? ARM binaries would not work in a MIPS system!
Build Performance Testing: Tests whether or not commonly used steps during builds work efficiently and avoid regressions. Tests to time commonly used usage scenarios are run through
oe-build-perf-test
. These tests are run on isolated machines so that the time measurements of the tests are accurate and no other processes interfere with the timing results. The project currently tests performance on two different distributions, Fedora and Ubuntu, to ensure we have no single point of failure and can ensure the different distros work effectively.eSDK Testing: Image tests initiated through the following command:
$ bitbake image -c testsdkext
The tests utilize the
testsdkext
class and thedo_testsdkext
task.Feature Testing: Various scenario-based tests are run through the OpenEmbedded Self test (oe-selftest). We test oe-selftest on each of the main distributions we support.
Image Testing: Image tests initiated through the following command:
$ bitbake image -c testimage
The tests utilize the testimage* classes and the do_testimage task.
Layer Testing: The Autobuilder has the possibility to test whether specific layers work with the test of the system. The layers tested may be selected by members of the project. Some key community layers are also tested periodically.
Package Testing: A Package Test (ptest) runs tests against packages built by the OpenEmbedded build system on the target machine. See the Testing Packages With ptest section in the Yocto Project Development Tasks Manual and the “Ptest” Wiki page for more information on Ptest.
SDK Testing: Image tests initiated through the following command:
$ bitbake image -c testsdk
The tests utilize the testsdk class and the
do_testsdk
task.Unit Testing: Unit tests on various components of the system run through bitbake-selftest and oe-selftest.
Automatic Upgrade Helper: This target tests whether new versions of software are available and whether we can automatically upgrade to those new versions. If so, this target emails the maintainers with a patch to let them know this is possible.
1.4 How Tests Map to Areas of Code
Tests map into the codebase as follows:
bitbake-selftest:
These tests are self-contained and test BitBake as well as its APIs, which include the fetchers. The tests are located in
bitbake/lib/*/tests
.From within the BitBake repository, run the following:
$ bitbake-selftest
To skip tests that access the Internet, use the
BB_SKIP_NETTEST
variable when running “bitbake-selftest” as follows:$ BB_SKIP_NETTEST=yes bitbake-selftest
The default output is quiet and just prints a summary of what was run. To see more information, there is a verbose option:
$ bitbake-selftest -v
Use this option when you wish to skip tests that access the network, which are mostly necessary to test the fetcher modules. To specify individual test modules to run, append the test module name to the “bitbake-selftest” command. For example, to specify the tests for the bb.data.module, run:
$ bitbake-selftest bb.test.data.module
You can also specify individual tests by defining the full name and module plus the class path of the test, for example:
$ bitbake-selftest bb.tests.data.TestOverrides.test_one_override
The tests are based on Python unittest.
oe-selftest:
These tests use OE to test the workflows, which include testing specific features, behaviors of tasks, and API unit tests.
The tests can take advantage of parallelism through the “-j” option, which can specify a number of threads to spread the tests across. Note that all tests from a given class of tests will run in the same thread. To parallelize large numbers of tests you can split the class into multiple units.
The tests are based on Python unittest.
The code for the tests resides in
meta/lib/oeqa/selftest/cases/
.To run all the tests, enter the following command:
$ oe-selftest -a
To run a specific test, use the following command form where testname is the name of the specific test:
$ oe-selftest -r <testname>
For example, the following command would run the tinfoil getVar API test:
$ oe-selftest -r tinfoil.TinfoilTests.test_getvar
It is also possible to run a set of tests. For example the following command will run all of the tinfoil tests:
$ oe-selftest -r tinfoil
testimage:
These tests build an image, boot it, and run tests against the image’s content.
The code for these tests resides in
meta/lib/oeqa/runtime/cases/
.You need to set the IMAGE_CLASSES variable as follows:
IMAGE_CLASSES += "testimage"
Run the tests using the following command form:
$ bitbake image -c testimage
testsdk:
These tests build an SDK, install it, and then run tests against that SDK.
The code for these tests resides in
meta/lib/oeqa/sdk/cases/
.Run the test using the following command form:
$ bitbake image -c testsdk
testsdk_ext:
These tests build an extended SDK (eSDK), install that eSDK, and run tests against the eSDK.
The code for these tests resides in
meta/lib/oeqa/esdk
.To run the tests, use the following command form:
$ bitbake image -c testsdkext
oe-build-perf-test:
These tests run through commonly used usage scenarios and measure the performance times.
The code for these tests resides in
meta/lib/oeqa/buildperf
.To run the tests, use the following command form:
$ oe-build-perf-test <options>
The command takes a number of options, such as where to place the test results. The Autobuilder Helper Scripts include the
build-perf-test-wrapper
script with examples of how to use the oe-build-perf-test from the command line.Use the
oe-git-archive
command to store test results into a Git repository.Use the
oe-build-perf-report
command to generate text reports and HTML reports with graphs of the performance data. For examples, see https://downloads.yoctoproject.org/releases/yocto/yocto-2.7/testresults/buildperf-centos7/perf-centos7.yoctoproject.org_warrior_20190414204758_0e39202.html and https://downloads.yoctoproject.org/releases/yocto/yocto-2.7/testresults/buildperf-centos7/perf-centos7.yoctoproject.org_warrior_20190414204758_0e39202.txt.The tests are contained in
lib/oeqa/buildperf/test_basic.py
.
1.5 Test Examples
This section provides example tests for each of the tests listed in the How Tests Map to Areas of Code section.
For oeqa tests, testcases for each area reside in the main test
directory at meta/lib/oeqa/selftest/cases
directory.
For oe-selftest. bitbake testcases reside in the lib/bb/tests/
directory.
1.5.1 bitbake-selftest
A simple test example from lib/bb/tests/data.py
is:
class DataExpansions(unittest.TestCase):
def setUp(self):
self.d = bb.data.init()
self.d["foo"] = "value_of_foo"
self.d["bar"] = "value_of_bar"
self.d["value_of_foo"] = "value_of_'value_of_foo'"
def test_one_var(self):
val = self.d.expand("${foo}")
self.assertEqual(str(val), "value_of_foo")
In this example, a DataExpansions
class of tests is created,
derived from standard python unittest. The class has a common setUp
function which is shared by all the tests in the class. A simple test is
then added to test that when a variable is expanded, the correct value
is found.
Bitbake selftests are straightforward python unittest. Refer to the Python unittest documentation for additional information on writing these tests at: https://docs.python.org/3/library/unittest.html.
1.5.2 oe-selftest
These tests are more complex due to the setup required behind the scenes for full builds. Rather than directly using Python’s unittest, the code wraps most of the standard objects. The tests can be simple, such as testing a command from within the OE build environment using the following example:
class BitbakeLayers(OESelftestTestCase):
def test_bitbakelayers_showcrossdepends(self):
result = runCmd('bitbake-layers show-cross-depends')
self.assertTrue('aspell' in result.output, msg = "No dependencies were shown. bitbake-layers show-cross-depends output: %s"% result.output)
This example, taken from meta/lib/oeqa/selftest/cases/bblayers.py
,
creates a testcase from the OESelftestTestCase
class, derived
from unittest.TestCase
, which runs the bitbake-layers
command
and checks the output to ensure it contains something we know should be
here.
The oeqa.utils.commands
module contains Helpers which can assist
with common tasks, including:
Obtaining the value of a bitbake variable: Use
oeqa.utils.commands.get_bb_var()
or useoeqa.utils.commands.get_bb_vars()
for more than one variableRunning a bitbake invocation for a build: Use
oeqa.utils.commands.bitbake()
Running a command: Use
oeqa.utils.commandsrunCmd()
There is also a oeqa.utils.commands.runqemu()
function for launching
the runqemu
command for testing things within a running, virtualized
image.
You can run these tests in parallel. Parallelism works per test class,
so tests within a given test class should always run in the same build,
while tests in different classes or modules may be split into different
builds. There is no data store available for these tests since the tests
launch the bitbake
command and exist outside of its context. As a
result, common bitbake library functions (bb.*) are also unavailable.
1.5.3 testimage
These tests are run once an image is up and running, either on target
hardware or under QEMU. As a result, they are assumed to be running in a
target image environment, as opposed to a host build environment. A
simple example from meta/lib/oeqa/runtime/cases/python.py
contains
the following:
class PythonTest(OERuntimeTestCase):
@OETestDepends(['ssh.SSHTest.test_ssh'])
@OEHasPackage(['python3-core'])
def test_python3(self):
cmd = "python3 -c \\"import codecs; print(codecs.encode('Uryyb, jbeyq', 'rot13'))\""
status, output = self.target.run(cmd)
msg = 'Exit status was not 0. Output: %s' % output
self.assertEqual(status, 0, msg=msg)
In this example, the OERuntimeTestCase
class wraps
unittest.TestCase
. Within the test, self.target
represents the
target system, where commands can be run on it using the run()
method.
To ensure certain test or package dependencies are met, you can use the
OETestDepends
and OEHasPackage
decorators. For example, the test
in this example would only make sense if python3-core is installed in
the image.
1.5.4 testsdk_ext
These tests are run against built extensible SDKs (eSDKs). The tests can
assume that the eSDK environment has already been setup. An example from
meta/lib/oeqa/sdk/cases/devtool.py
contains the following:
class DevtoolTest(OESDKExtTestCase):
@classmethod def setUpClass(cls):
myapp_src = os.path.join(cls.tc.esdk_files_dir, "myapp")
cls.myapp_dst = os.path.join(cls.tc.sdk_dir, "myapp")
shutil.copytree(myapp_src, cls.myapp_dst)
subprocess.check_output(['git', 'init', '.'], cwd=cls.myapp_dst)
subprocess.check_output(['git', 'add', '.'], cwd=cls.myapp_dst)
subprocess.check_output(['git', 'commit', '-m', "'test commit'"], cwd=cls.myapp_dst)
@classmethod
def tearDownClass(cls):
shutil.rmtree(cls.myapp_dst)
def _test_devtool_build(self, directory):
self._run('devtool add myapp %s' % directory)
try:
self._run('devtool build myapp')
finally:
self._run('devtool reset myapp')
def test_devtool_build_make(self):
self._test_devtool_build(self.myapp_dst)
In this example, the devtool
command is tested to see whether a sample application can be built with
the devtool build
command within the eSDK.
1.5.5 testsdk
These tests are run against built SDKs. The tests can assume that an SDK
has already been extracted and its environment file has been sourced. A
simple example from meta/lib/oeqa/sdk/cases/python2.py
contains the
following:
class Python3Test(OESDKTestCase):
def setUp(self):
if not (self.tc.hasHostPackage("nativesdk-python3-core") or
self.tc.hasHostPackage("python3-core-native")):
raise unittest.SkipTest("No python3 package in the SDK")
def test_python3(self):
cmd = "python3 -c \\"import codecs; print(codecs.encode('Uryyb, jbeyq', 'rot13'))\""
output = self._run(cmd)
self.assertEqual(output, "Hello, world\n")
In this example, if nativesdk-python3-core has been installed into the SDK, the code runs the python3 interpreter with a basic command to check it is working correctly. The test would only run if python3 is installed in the SDK.
1.5.6 oe-build-perf-test
The performance tests usually measure how long operations take and the
resource utilization as that happens. An example from
meta/lib/oeqa/buildperf/test_basic.py
contains the following:
class Test3(BuildPerfTestCase):
def test3(self):
"""Bitbake parsing (bitbake -p)"""
# Drop all caches and parse
self.rm_cache()
oe.path.remove(os.path.join(self.bb_vars['TMPDIR'], 'cache'), True)
self.measure_cmd_resources(['bitbake', '-p'], 'parse_1',
'bitbake -p (no caches)')
# Drop tmp/cache
oe.path.remove(os.path.join(self.bb_vars['TMPDIR'], 'cache'), True)
self.measure_cmd_resources(['bitbake', '-p'], 'parse_2',
'bitbake -p (no tmp/cache)')
# Parse with fully cached data
self.measure_cmd_resources(['bitbake', '-p'], 'parse_3',
'bitbake -p (cached)')
This example shows how three specific parsing timings are measured, with and without various caches, to show how BitBake’s parsing performance trends over time.
1.6 Considerations When Writing Tests
When writing good tests, there are several things to keep in mind. Since things running on the Autobuilder are accessed concurrently by multiple workers, consider the following:
Running “cleanall” is not permitted.
This can delete files from DL_DIR which would potentially break other builds running in parallel. If this is required, DL_DIR must be set to an isolated directory.
Running “cleansstate” is not permitted.
This can delete files from SSTATE_DIR which would potentially break
other builds running in parallel. If this is required, SSTATE_DIR must
be set to an isolated directory. Alternatively, you can use the “-f”
option with the bitbake
command to “taint” tasks by changing the
sstate checksums to ensure sstate cache items will not be reused.
Tests should not change the metadata.
This is particularly true for oe-selftests since these can run in parallel and changing metadata leads to changing checksums, which confuses BitBake while running in parallel. If this is necessary, copy layers to a temporary location and modify them. Some tests need to change metadata, such as the devtool tests. To protect the metadata from changes, set up temporary copies of that data first.
2 Project Testing and Release Process
2.1 Day to Day Development
This section details how the project tests changes, through automation on the Autobuilder or with the assistance of QA teams, through to making releases.
The project aims to test changes against our test matrix before those
changes are merged into the master branch. As such, changes are queued
up in batches either in the master-next
branch in the main trees, or
in user trees such as ross/mut
in poky-contrib
(Ross Burton
helps review and test patches and this is his testing tree).
We have two broad categories of test builds, including “full” and “quick”. On the Autobuilder, these can be seen as “a-quick” and “a-full”, simply for ease of sorting in the UI. Use our Autobuilder console view to see where me manage most test-related items, available at: https://autobuilder.yoctoproject.org/typhoon/#/console.
Builds are triggered manually when the test branches are ready. The
builds are monitored by the SWAT team. For additional information, see
https://wiki.yoctoproject.org/wiki/Yocto_Build_Failure_Swat_Team.
If successful, the changes would usually be merged to the master
branch. If not successful, someone would respond to the changes on the
mailing list explaining that there was a failure in testing. The choice
of quick or full would depend on the type of changes and the speed with
which the result was required.
The Autobuilder does build the master
branch once daily for several
reasons, in particular, to ensure the current master
branch does
build, but also to keep yocto-testresults
(https://git.yoctoproject.org/cgit/cgit.cgi/yocto-testresults/),
buildhistory
(https://git.yoctoproject.org/cgit/cgit.cgi/poky-buildhistory/), and
our sstate up to date. On the weekend, there is a master-next build
instead to ensure the test results are updated for the less frequently
run targets.
Performance builds (buildperf-* targets in the console) are triggered separately every six hours and automatically push their results to the buildstats repository at: https://git.yoctoproject.org/cgit/cgit.cgi/yocto-buildstats/.
The ‘quick’ targets have been selected to be the ones which catch the most failures or give the most valuable data. We run ‘fast’ ptests in this case for example but not the ones which take a long time. The quick target doesn’t include *-lsb builds for all architectures, some world builds and doesn’t trigger performance tests or ltp testing. The full build includes all these things and is slower but more comprehensive.
2.2 Release Builds
The project typically has two major releases a year with a six month cadence in April and October. Between these there would be a number of milestone releases (usually four) with the final one being stabilization only along with point releases of our stable branches.
The build and release process for these project releases is similar to that in Day to Day Development, in that the a-full target of the Autobuilder is used but in addition the form is configured to generate and publish artifacts and the milestone number, version, release candidate number and other information is entered. The box to “generate an email to QA”is also checked.
When the build completes, an email is sent out using the send-qa-email
script in the yocto-autobuilder-helper
repository to the list of
people configured for that release. Release builds are placed into a
directory in https://autobuilder.yocto.io/pub/releases on the
Autobuilder which is included in the email. The process from here is
more manual and control is effectively passed to release engineering.
The next steps include:
QA teams respond to the email saying which tests they plan to run and when the results will be available.
QA teams run their tests and share their results in the yocto- testresults-contrib repository, along with a summary of their findings.
Release engineering prepare the release as per their process.
Test results from the QA teams are included into the release in separate directories and also uploaded to the yocto-testresults repository alongside the other test results for the given revision.
The QA report in the final release is regenerated using resulttool to include the new test results and the test summaries from the teams (as headers to the generated report).
The release is checked against the release checklist and release readiness criteria.
A final decision on whether to release is made by the YP TSC who have final oversight on release readiness.
3 Understanding the Yocto Project Autobuilder
3.1 Execution Flow within the Autobuilder
The “a-full” and “a-quick” targets are the usual entry points into the Autobuilder and it makes sense to follow the process through the system starting there. This is best visualized from the Autobuilder Console view (https://autobuilder.yoctoproject.org/typhoon/#/console).
Each item along the top of that view represents some “target build” and
these targets are all run in parallel. The ‘full’ build will trigger the
majority of them, the “quick” build will trigger some subset of them.
The Autobuilder effectively runs whichever configuration is defined for
each of those targets on a separate buildbot worker. To understand the
configuration, you need to look at the entry on config.json
file
within the yocto-autobuilder-helper
repository. The targets are
defined in the ‘overrides’ section, a quick example could be qemux86-64
which looks like:
"qemux86-64" : {
"MACHINE" : "qemux86-64",
"TEMPLATE" : "arch-qemu",
"step1" : {
"extravars" : [
"IMAGE_FSTYPES_append = ' wic wic.bmap'"
]
}
},
And to expand that, you need the “arch-qemu” entry from the “templates” section, which looks like:
"arch-qemu" : {
"BUILDINFO" : true,
"BUILDHISTORY" : true,
"step1" : {
"BBTARGETS" : "core-image-sato core-image-sato-dev core-image-sato-sdk core-image-minimal core-image-minimal-dev core-image-sato:do_populate_sdk",
"SANITYTARGETS" : "core-image-minimal:do_testimage core-image-sato:do_testimage core-image-sato-sdk:do_testimage core-image-sato:do_testsdk"
},
"step2" : {
"SDKMACHINE" : "x86_64",
"BBTARGETS" : "core-image-sato:do_populate_sdk core-image-minimal:do_populate_sdk_ext core-image-sato:do_populate_sdk_ext",
"SANITYTARGETS" : "core-image-sato:do_testsdk core-image-minimal:do_testsdkext core-image-sato:do_testsdkext"
},
"step3" : {
"BUILDHISTORY" : false,
"EXTRACMDS" : ["${SCRIPTSDIR}/checkvnc; DISPLAY=:1 oe-selftest ${HELPERSTMACHTARGS} -j 15"],
"ADDLAYER" : ["${BUILDDIR}/../meta-selftest"]
}
},
Combining these two entries you can see that “qemux86-64” is a three step build where the
bitbake BBTARGETS
would be run, then bitbake SANITYTARGETS
for each step; all for
MACHINE="qemx86-64"
but with differing SDKMACHINE settings. In step
1 an extra variable is added to the auto.conf
file to enable wic
image generation.
While not every detail of this is covered here, you can see how the template mechanism allows quite complex configurations to be built up yet allows duplication and repetition to be kept to a minimum.
The different build targets are designed to allow for parallelization, so different machines are usually built in parallel, operations using the same machine and metadata are built sequentially, with the aim of trying to optimize build efficiency as much as possible.
The config.json
file is processed by the scripts in the Helper
repository in the scripts
directory. The following section details
how this works.
3.2 Autobuilder Target Execution Overview
For each given target in a build, the Autobuilder executes several
steps. These are configured in yocto-autobuilder2/builders.py
and
roughly consist of:
Run clobberdir.
This cleans out any previous build. Old builds are left around to allow easier debugging of failed builds. For additional information, see clobberdir.
Obtain yocto-autobuilder-helper
This step clones the
yocto-autobuilder-helper
git repository. This is necessary to prevent the requirement to maintain all the release or project-specific code within Buildbot. The branch chosen matches the release being built so we can support older releases and still make changes in newer ones.Write layerinfo.json
This transfers data in the Buildbot UI when the build was configured to the Helper.
Call scripts/shared-repo-unpack
This is a call into the Helper scripts to set up a checkout of all the pieces this build might need. It might clone the BitBake repository and the OpenEmbedded-Core repository. It may clone the Poky repository, as well as additional layers. It will use the data from the
layerinfo.json
file to help understand the configuration. It will also use a local cache of repositories to speed up the clone checkouts. For additional information, see Autobuilder Clone Cache.This step has two possible modes of operation. If the build is part of a parent build, it’s possible that all the repositories needed may already be available, ready in a pre-prepared directory. An “a-quick” or “a-full” build would prepare this before starting the other sub-target builds. This is done for two reasons:
the upstream may change during a build, for example, from a forced push and this ensures we have matching content for the whole build
if 15 Workers all tried to pull the same data from the same repos, we can hit resource limits on upstream servers as they can think they are under some kind of network attack
This pre-prepared directory is shared among the Workers over NFS. If the build is an individual build and there is no “shared” directory available, it would clone from the cache and the upstreams as necessary. This is considered the fallback mode.
Call scripts/run-config
This is another call into the Helper scripts where it’s expected that the main functionality of this target will be executed.
3.3 Autobuilder Technology
The Autobuilder has Yocto Project-specific functionality to allow builds to operate with increased efficiency and speed.
3.3.1 clobberdir
When deleting files, the Autobuilder uses clobberdir
, which is a
special script that moves files to a special location, rather than
deleting them. Files in this location are deleted by an rm
command,
which is run under ionice -c 3
. For example, the deletion only
happens when there is idle IO capacity on the Worker. The Autobuilder
Worker Janitor runs this deletion. See Autobuilder Worker Janitor.
3.3.2 Autobuilder Clone Cache
Cloning repositories from scratch each time they are required was slow on the Autobuilder. We therefore have a stash of commonly used repositories pre-cloned on the Workers. Data is fetched from these during clones first, then “topped up” with later revisions from any upstream when necessary. The cache is maintained by the Autobuilder Worker Janitor. See Autobuilder Worker Janitor.
3.3.3 Autobuilder Worker Janitor
This is a process running on each Worker that performs two basic operations, including background file deletion at IO idle (see Autobuilder Target Execution Overview: Run clobberdir) and maintenance of a cache of cloned repositories to improve the speed the system can checkout repositories.
3.3.6 Resulttool
All of the different tests run as part of the build generate output into
testresults.json
files. This allows us to determine which tests ran
in a given build and their status. Additional information, such as
failure logs or the time taken to run the tests, may also be included.
Resulttool is part of OpenEmbedded-Core and is used to manipulate these json results files. It has the ability to merge files together, display reports of the test results and compare different result files.
For details, see https://wiki.yoctoproject.org/wiki/Resulttool.
3.4 run-config Target Execution
The scripts/run-config
execution is where most of the work within
the Autobuilder happens. It runs through a number of steps; the first
are general setup steps that are run once and include:
Set up any
buildtools-tarball
if configured.Call “buildhistory-init” if buildhistory is configured.
For each step that is configured in config.json
, it will perform the
following:
Add any layers that are specified using the
bitbake-layers add-layer
command (logging as stepXa)Call the
scripts/setup-config
script to generate the necessaryauto.conf
configuration file for the buildRun the
bitbake BBTARGETS
command (logging as stepXb)Run the
bitbake SANITYTARGETS
command (logging as stepXc)Run the
EXTRACMDS
command, which are run within the BitBake build environment (logging as stepXd)Run the
EXTRAPLAINCMDS
command(s), which are run outside the BitBake build environment (logging as stepXd)Remove any layers added in step 1 using the
bitbake-layers remove-layer
command (logging as stepXa)
Once the execution steps above complete, run-config
executes a set
of post-build steps, including:
Call
scripts/publish-artifacts
to collect any output which is to be saved from the build.Call
scripts/collect-results
to collect any test results to be saved from the build.Call
scripts/upload-error-reports
to send any error reports generated to the remote server.Cleanup the build directory using clobberdir if the build was successful, else rename it to “build-renamed” for potential future debugging.
3.5 Deploying Yocto Autobuilder
The most up to date information about how to setup and deploy your own
Autobuilder can be found in README.md in the yocto-autobuilder2
repository.
We hope that people can use the yocto-autobuilder2
code directly but
it is inevitable that users will end up needing to heavily customise the
yocto-autobuilder-helper
repository, particularly the
config.json
file as they will want to define their own test matrix.
The Autobuilder supports wo customization options:
variable substitution
overlaying configuration files
The standard config.json
minimally attempts to allow substitution of
the paths. The Helper script repository includes a
local-example.json
file to show how you could override these from a
separate configuration file. Pass the following into the environment of
the Autobuilder:
$ ABHELPER_JSON="config.json local-example.json"
As another example, you could also pass the following into the environment:
$ ABHELPER_JSON="config.json /some/location/local.json"
One issue users often run into is validation of the config.json
files. A
tip for minimizing issues from invalid json files is to use a Git
pre-commit-hook.sh
script to verify the JSON file before committing
it. Create a symbolic link as follows:
$ ln -s ../../scripts/pre-commit-hook.sh .git/hooks/pre-commit
4 Manual Revision History
Revision |
Date |
Note |
---|---|---|
3.2 |
October 2020 |
The initial document released with the Yocto Project 3.2 Release |
Permission is granted to copy, distribute and/or modify this document under the terms of the Creative Commons Attribution-Share Alike 2.0 UK: England & Wales as published by Creative Commons.
To report any inaccuracies or problems with this (or any other Yocto Project)
manual, or to send additions or changes, please send email/patches to the Yocto
Project documentation mailing list at docs@lists.yoctoproject.org
or
log into the Freenode #yocto
channel.
BitBake Documentation
BitBake was originally a part of the OpenEmbedded project. It was inspired by the Portage package management system used by the Gentoo Linux distribution. In 2004, the OpenEmbedded project was split the project into two distinct pieces:
BitBake, a generic task executor
OpenEmbedded, a metadata set utilized by BitBake
Today, BitBake is the primary build tool of OpenEmbedded based projects, such as the Yocto Project.
The BitBake documentation can be found here.