mirror of
https://github.com/u-boot/u-boot.git
synced 2024-11-24 04:34:22 +08:00
f7c6ee7fe7
There are a lot of zynq configurations which can be merged together and use
only one for all. The similar change has been done for ZynqMP by commit
be1b6c32d9
("arm64: zynqmp: Use zynqmp_virt platform")
Build SPL with u-boot.img for zc706 like this.
export DEVICE_TREE=zynq-zc706 && make xilinx_zynq_virt_defconfig && make -j8
u-boot.img is generic for all boards.
Tested on Zybo, zc702, zc706, zc770-xm011-x16, cc108 and microzed.
Signed-off-by: Michal Simek <michal.simek@xilinx.com>
433 lines
15 KiB
YAML
433 lines
15 KiB
YAML
variables:
|
||
windows_vm: vs2017-win2016
|
||
ubuntu_vm: ubuntu-18.04
|
||
ci_runner_image: trini/u-boot-gitlab-ci-runner:bionic-20200112-21Feb2020
|
||
# Add '-u 0' options for Azure pipelines, otherwise we get "permission
|
||
# denied" error when it tries to "useradd -m -u 1001 vsts_azpcontainer",
|
||
# since our $(ci_runner_image) user is not root.
|
||
container_option: -u 0
|
||
work_dir: /u
|
||
|
||
jobs:
|
||
- job: tools_only_windows
|
||
displayName: 'Ensure host tools build for Windows'
|
||
pool:
|
||
vmImage: $(windows_vm)
|
||
strategy:
|
||
matrix:
|
||
i686:
|
||
MSYS_DIR: msys32
|
||
BASE_REPO: msys2-ci-base-i686
|
||
x86_64:
|
||
MSYS_DIR: msys64
|
||
BASE_REPO: msys2-ci-base
|
||
steps:
|
||
- script: |
|
||
git clone https://github.com/msys2/$(BASE_REPO).git %CD:~0,2%\$(MSYS_DIR)
|
||
displayName: 'Install MSYS2'
|
||
- script: |
|
||
set PATH=%CD:~0,2%\$(MSYS_DIR)\usr\bin;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem
|
||
%CD:~0,2%\$(MSYS_DIR)\usr\bin\pacman --noconfirm -Syyuu
|
||
displayName: 'Update MSYS2'
|
||
- script: |
|
||
set PATH=%CD:~0,2%\$(MSYS_DIR)\usr\bin;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem
|
||
%CD:~0,2%\$(MSYS_DIR)\usr\bin\pacman --noconfirm --needed -S make gcc bison diffutils openssl-devel
|
||
displayName: 'Install Toolchain'
|
||
- script: |
|
||
set PATH=C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem
|
||
echo make tools-only_defconfig tools-only NO_SDL=1 > build-tools.sh
|
||
%CD:~0,2%\$(MSYS_DIR)\usr\bin\bash -lc "bash build-tools.sh"
|
||
displayName: 'Build Host Tools'
|
||
env:
|
||
# Tell MSYS2 we need a POSIX emulation layer
|
||
MSYSTEM: MSYS
|
||
# Tell MSYS2 not to ‘cd’ our startup directory to HOME
|
||
CHERE_INVOKING: yes
|
||
|
||
- job: cppcheck
|
||
displayName: 'Static code analysis with cppcheck'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
container:
|
||
image: $(ci_runner_image)
|
||
options: $(container_option)
|
||
steps:
|
||
- script: cppcheck --force --quiet --inline-suppr .
|
||
|
||
- job: htmldocs
|
||
displayName: 'Build HTML documentation'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
container:
|
||
image: $(ci_runner_image)
|
||
options: $(container_option)
|
||
steps:
|
||
- script: make htmldocs
|
||
|
||
- job: todo
|
||
displayName: 'Search for TODO within source tree'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
container:
|
||
image: $(ci_runner_image)
|
||
options: $(container_option)
|
||
steps:
|
||
- script: grep -r TODO .
|
||
- script: grep -r FIXME .
|
||
- script: grep -r HACK . | grep -v HACKKIT
|
||
|
||
- job: sloccount
|
||
displayName: 'Some statistics about the code base'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
container:
|
||
image: $(ci_runner_image)
|
||
options: $(container_option)
|
||
steps:
|
||
- script: sloccount .
|
||
|
||
- job: maintainers
|
||
displayName: 'Ensure all configs have MAINTAINERS entries'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
container:
|
||
image: $(ci_runner_image)
|
||
options: $(container_option)
|
||
steps:
|
||
- script: |
|
||
if [ `./tools/genboardscfg.py -f 2>&1 | wc -l` -ne 0 ]; then exit 1; fi
|
||
|
||
- job: tools_only
|
||
displayName: 'Ensure host tools build'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
container:
|
||
image: $(ci_runner_image)
|
||
options: $(container_option)
|
||
steps:
|
||
- script: |
|
||
make tools-only_config tools-only -j$(nproc)
|
||
|
||
- job: envtools
|
||
displayName: 'Ensure env tools build'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
container:
|
||
image: $(ci_runner_image)
|
||
options: $(container_option)
|
||
steps:
|
||
- script: |
|
||
make tools-only_config envtools -j$(nproc)
|
||
|
||
- job: utils
|
||
displayName: 'Run binman, buildman, dtoc and patman testsuites'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
steps:
|
||
- script: |
|
||
cat << EOF > build.sh
|
||
set -ex
|
||
cd ${WORK_DIR}
|
||
EOF
|
||
cat << "EOF" >> build.sh
|
||
git config --global user.name "Azure Pipelines"
|
||
git config --global user.email bmeng.cn@gmail.com
|
||
export USER=azure
|
||
virtualenv -p /usr/bin/python3 /tmp/venv
|
||
. /tmp/venv/bin/activate
|
||
pip install pyelftools
|
||
export UBOOT_TRAVIS_BUILD_DIR=/tmp/.bm-work/sandbox_spl
|
||
export PYTHONPATH=${UBOOT_TRAVIS_BUILD_DIR}/scripts/dtc/pylibfdt
|
||
export PATH=${UBOOT_TRAVIS_BUILD_DIR}/scripts/dtc:${PATH}
|
||
./tools/buildman/buildman -o /tmp -P sandbox_spl
|
||
./tools/binman/binman --toolpath ${UBOOT_TRAVIS_BUILD_DIR}/tools test
|
||
./tools/buildman/buildman -t
|
||
./tools/dtoc/dtoc -t
|
||
./tools/patman/patman --test
|
||
EOF
|
||
cat build.sh
|
||
# We cannot use "container" like other jobs above, as buildman
|
||
# seems to hang forever with pre-configured "container" environment
|
||
docker run -v $PWD:$(work_dir) $(ci_runner_image) /bin/bash $(work_dir)/build.sh
|
||
|
||
- job: test_py
|
||
displayName: 'test.py'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
strategy:
|
||
matrix:
|
||
sandbox:
|
||
TEST_PY_BD: "sandbox"
|
||
BUILDMAN: "^sandbox$"
|
||
sandbox_clang:
|
||
TEST_PY_BD: "sandbox"
|
||
BUILDMAN: "^sandbox$"
|
||
OVERRIDE: "-O clang-7"
|
||
sandbox_spl:
|
||
TEST_PY_BD: "sandbox_spl"
|
||
TEST_PY_TEST_SPEC: "test_ofplatdata"
|
||
BUILDMAN: "^sandbox_spl$"
|
||
sandbox_flattree:
|
||
TEST_PY_BD: "sandbox_flattree"
|
||
BUILDMAN: "^sandbox_flattree$"
|
||
evb_ast2500:
|
||
TEST_PY_BD: "evb-ast2500"
|
||
TEST_PY_ID: "--id qemu"
|
||
BUILDMAN: "^evb-ast2500$"
|
||
vexpress_ca15_tc2:
|
||
TEST_PY_BD: "vexpress_ca15_tc2"
|
||
TEST_PY_ID: "--id qemu"
|
||
BUILDMAN: "^vexpress_ca15_tc2$"
|
||
vexpress_ca9x4:
|
||
TEST_PY_BD: "vexpress_ca9x4"
|
||
TEST_PY_ID: "--id qemu"
|
||
BUILDMAN: "^vexpress_ca9x4$"
|
||
integratorcp_cm926ejs:
|
||
TEST_PY_BD: "integratorcp_cm926ejs"
|
||
TEST_PY_ID: "--id qemu"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^integratorcp_cm926ejs$"
|
||
qemu_arm:
|
||
TEST_PY_BD: "qemu_arm"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu_arm$"
|
||
qemu_arm64:
|
||
TEST_PY_BD: "qemu_arm64"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu_arm64$"
|
||
qemu_mips:
|
||
TEST_PY_BD: "qemu_mips"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu_mips$"
|
||
qemu_mipsel:
|
||
TEST_PY_BD: "qemu_mipsel"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu_mipsel$"
|
||
qemu_mips64:
|
||
TEST_PY_BD: "qemu_mips64"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu_mips64$"
|
||
qemu_mips64el:
|
||
TEST_PY_BD: "qemu_mips64el"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu_mips64el$"
|
||
qemu_ppce500:
|
||
TEST_PY_BD: "qemu-ppce500"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu-ppce500$"
|
||
qemu_riscv64:
|
||
TEST_PY_BD: "qemu-riscv64"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu-riscv64$"
|
||
qemu_x86:
|
||
TEST_PY_BD: "qemu-x86"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu-x86$"
|
||
qemu_x86_64:
|
||
TEST_PY_BD: "qemu-x86_64"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^qemu-x86_64$"
|
||
xilinx_zynq_virt:
|
||
TEST_PY_BD: "xilinx_zynq_virt"
|
||
TEST_PY_ID: "--id qemu"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^xilinx_zynq_virt$"
|
||
xilinx_versal_virt:
|
||
TEST_PY_BD: "xilinx_versal_virt"
|
||
TEST_PY_ID: "--id qemu"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^xilinx_versal_virt$"
|
||
xtfpga:
|
||
TEST_PY_BD: "xtfpga"
|
||
TEST_PY_ID: "--id qemu"
|
||
TEST_PY_TEST_SPEC: "not sleep"
|
||
BUILDMAN: "^xtfpga$"
|
||
steps:
|
||
- script: |
|
||
cat << EOF > test.sh
|
||
set -ex
|
||
# make environment variables available as tests are running inside a container
|
||
export WORK_DIR="${WORK_DIR}"
|
||
export TEST_PY_BD="${TEST_PY_BD}"
|
||
export TEST_PY_ID="${TEST_PY_ID}"
|
||
export TEST_PY_TEST_SPEC="${TEST_PY_TEST_SPEC}"
|
||
export BUILDMAN="${BUILDMAN}"
|
||
export OVERRIDE="${OVERRIDE}"
|
||
EOF
|
||
cat << "EOF" >> test.sh
|
||
# the below corresponds to .gitlab-ci.yml "before_script"
|
||
cd ${WORK_DIR}
|
||
git clone --depth=1 git://github.com/swarren/uboot-test-hooks.git /tmp/uboot-test-hooks
|
||
ln -s travis-ci /tmp/uboot-test-hooks/bin/`hostname`
|
||
ln -s travis-ci /tmp/uboot-test-hooks/py/`hostname`
|
||
grub-mkimage --prefix=\"\" -o ~/grub_x86.efi -O i386-efi normal echo lsefimmap lsefi lsefisystab efinet tftp minicmd
|
||
grub-mkimage --prefix=\"\" -o ~/grub_x64.efi -O x86_64-efi normal echo lsefimmap lsefi lsefisystab efinet tftp minicmd
|
||
cp /opt/grub/grubriscv64.efi ~/grub_riscv64.efi
|
||
cp /opt/grub/grubaa64.efi ~/grub_arm64.efi
|
||
cp /opt/grub/grubarm.efi ~/grub_arm.efi
|
||
# the below corresponds to .gitlab-ci.yml "script"
|
||
cd ${WORK_DIR}
|
||
if [[ "${BUILDMAN}" != "" ]]; then
|
||
ret=0;
|
||
tools/buildman/buildman -o /tmp -P -E ${BUILDMAN} ${OVERRIDE} || ret=$?;
|
||
if [[ $ret -ne 0 && $ret -ne 129 ]]; then
|
||
tools/buildman/buildman -o /tmp -sdeP ${BUILDMAN};
|
||
exit $ret;
|
||
fi;
|
||
fi
|
||
virtualenv -p /usr/bin/python3 /tmp/venv
|
||
. /tmp/venv/bin/activate
|
||
pip install -r test/py/requirements.txt
|
||
export UBOOT_TRAVIS_BUILD_DIR=/tmp/.bm-work/${TEST_PY_BD};
|
||
export PATH=/opt/qemu/bin:/tmp/uboot-test-hooks/bin:${PATH};
|
||
export PYTHONPATH=/tmp/uboot-test-hooks/py/travis-ci;
|
||
if [[ "${TEST_PY_BD}" != "" ]]; then
|
||
./test/py/test.py --bd ${TEST_PY_BD} ${TEST_PY_ID} -k "${TEST_PY_TEST_SPEC:-not a_test_which_does_not_exist}" --build-dir "$UBOOT_TRAVIS_BUILD_DIR";
|
||
ret=$?;
|
||
if [[ $ret -ne 0 ]]; then
|
||
exit $ret;
|
||
fi;
|
||
fi
|
||
# the below corresponds to .gitlab-ci.yml "after_script"
|
||
rm -rf /tmp/uboot-test-hooks /tmp/venv
|
||
EOF
|
||
cat test.sh
|
||
# make current directory writeable to uboot user inside the container
|
||
# as sandbox testing need create files like spi flash images, etc.
|
||
# (TODO: clean up this in the future)
|
||
chmod 777 .
|
||
docker run -v $PWD:$(work_dir) $(ci_runner_image) /bin/bash $(work_dir)/test.sh
|
||
|
||
- job: build_the_world
|
||
displayName: 'Build the World'
|
||
pool:
|
||
vmImage: $(ubuntu_vm)
|
||
strategy:
|
||
# Use almost the same target division in .travis.yml, only merged
|
||
# 4 small build jobs (arc/microblaze/nds32/xtensa) into one.
|
||
matrix:
|
||
arc_microblaze_nds32_xtensa:
|
||
BUILDMAN: "arc microblaze nds32 xtensa"
|
||
arm11_arm7_arm920t_arm946es:
|
||
BUILDMAN: "arm11 arm7 arm920t arm946es"
|
||
arm926ejs:
|
||
BUILDMAN: "arm926ejs -x freescale,siemens,at91,kirkwood,spear,omap"
|
||
at91_non_armv7:
|
||
BUILDMAN: "at91 -x armv7"
|
||
at91_non_arm926ejs:
|
||
BUILDMAN: "at91 -x arm926ejs"
|
||
boundary_engicam_toradex:
|
||
BUILDMAN: "boundary engicam toradex"
|
||
arm_bcm:
|
||
BUILDMAN: "bcm -x mips"
|
||
nxp_arm32:
|
||
BUILDMAN: "freescale -x powerpc,m68k,aarch64,ls101,ls102,ls104,ls108,ls20,lx216"
|
||
nxp_ls101x:
|
||
BUILDMAN: "freescale&ls101"
|
||
nxp_ls102x:
|
||
BUILDMAN: "freescale&ls102"
|
||
nxp_ls104x:
|
||
BUILDMAN: "freescale&ls104"
|
||
nxp_ls108x:
|
||
BUILDMAN: "freescale&ls108"
|
||
nxp_ls20xx:
|
||
BUILDMAN: "freescale&ls20"
|
||
nxp_lx216x:
|
||
BUILDMAN: "freescale&lx216"
|
||
imx6:
|
||
BUILDMAN: "mx6 -x boundary,engicam,freescale,technexion,toradex"
|
||
imx:
|
||
BUILDMAN: "mx -x mx6,freescale,technexion,toradex"
|
||
keystone2_keystone3:
|
||
BUILDMAN: "k2 k3"
|
||
samsung_socfpga:
|
||
BUILDMAN: "samsung socfpga"
|
||
spear:
|
||
BUILDMAN: "spear"
|
||
sun4i:
|
||
BUILDMAN: "sun4i"
|
||
sun5i:
|
||
BUILDMAN: "sun5i"
|
||
sun6i:
|
||
BUILDMAN: "sun6i"
|
||
sun7i:
|
||
BUILDMAN: "sun7i"
|
||
sun8i_32bit:
|
||
BUILDMAN: "sun8i&armv7"
|
||
sun8i_64bit:
|
||
BUILDMAN: "sun8i&aarch64"
|
||
sun9i:
|
||
BUILDMAN: "sun9i"
|
||
sun50i:
|
||
BUILDMAN: "sun50i"
|
||
arm_catch_all:
|
||
BUILDMAN: "arm -x arm11,arm7,arm9,aarch64,at91,bcm,freescale,kirkwood,mvebu,siemens,tegra,uniphier,mx,samsung,sunxi,am33xx,omap,rk,toradex,socfpga,k2,k3,zynq"
|
||
sandbox_x86:
|
||
BUILDMAN: "sandbox x86"
|
||
technexion:
|
||
BUILDMAN: "technexion"
|
||
kirkwood:
|
||
BUILDMAN: "kirkwood"
|
||
mvebu:
|
||
BUILDMAN: "mvebu"
|
||
m68k:
|
||
BUILDMAN: "m68k"
|
||
mips:
|
||
BUILDMAN: "mips"
|
||
non_fsl_ppc:
|
||
BUILDMAN: "powerpc -x freescale"
|
||
mpc85xx_freescale:
|
||
BUILDMAN: "mpc85xx&freescale -x t208xrdb -x t4qds -x t102* -x p1_p2_rdb_pc -x p1010rdb -x corenet_ds -x b4860qds -x bsc91*"
|
||
t208xrdb_corenet_ds:
|
||
BUILDMAN: "t208xrdb corenet_ds"
|
||
fsl_ppc:
|
||
BUILDMAN: "t4qds b4860qds mpc83xx&freescale mpc86xx&freescale"
|
||
t102x:
|
||
BUILDMAN: "t102*"
|
||
p1_p2_rdb_pc:
|
||
BUILDMAN: "p1_p2_rdb_pc"
|
||
p1010rdb_bsc91:
|
||
BUILDMAN: "p1010rdb bsc91"
|
||
siemens:
|
||
BUILDMAN: "siemens"
|
||
tegra:
|
||
BUILDMAN: "tegra -x toradex"
|
||
am33xx_no_siemens:
|
||
BUILDMAN: "am33xx -x siemens"
|
||
omap:
|
||
BUILDMAN: "omap"
|
||
uniphier:
|
||
BUILDMAN: "uniphier"
|
||
aarch64_catch_all:
|
||
BUILDMAN: "aarch64 -x bcm,k3,tegra,ls1,ls2,lx216,mvebu,uniphier,sunxi,samsung,socfpga,rk,versal,zynq"
|
||
rockchip:
|
||
BUILDMAN: "rk"
|
||
sh:
|
||
BUILDMAN: "sh -x arm"
|
||
zynq:
|
||
BUILDMAN: "zynq&armv7"
|
||
zynqmp_versal:
|
||
BUILDMAN: "versal|zynqmp&aarch64"
|
||
riscv:
|
||
BUILDMAN: "riscv"
|
||
steps:
|
||
- script: |
|
||
cat << EOF > build.sh
|
||
set -ex
|
||
cd ${WORK_DIR}
|
||
# make environment variables available as tests are running inside a container
|
||
export BUILDMAN="${BUILDMAN}"
|
||
EOF
|
||
cat << "EOF" >> build.sh
|
||
if [[ "${BUILDMAN}" != "" ]]; then
|
||
ret=0;
|
||
tools/buildman/buildman -o /tmp -P -E ${BUILDMAN} ${OVERRIDE} || ret=$?;
|
||
if [[ $ret -ne 0 && $ret -ne 129 ]]; then
|
||
tools/buildman/buildman -o /tmp -sdeP ${BUILDMAN};
|
||
exit $ret;
|
||
fi;
|
||
fi
|
||
EOF
|
||
cat build.sh
|
||
docker run -v $PWD:$(work_dir) $(ci_runner_image) /bin/bash $(work_dir)/build.sh
|