mirror of
https://github.com/dashpay/dash.git
synced 2024-12-24 19:42:46 +01:00
Add support for Gitlab CI (#3149)
* Add .gitlab-ci.yml * Use | instead of > for multiline commands This honor new-lines and makes ; unnecessary * Use ubuntu:bionic as base image * Move cache initialization before apt-get installs * Cache apt packages * Move installation of wget and unzip up as we need it for the cache * Prevent apt from deleting caches * Collect test logs into artifact * Make combine_logs.py always look for the template in the correct dir * Move final cache stuff into after_script * Reintroduce PYTHON_DEBUG=1, but only for .travis.yml * Install jinja2 in Travis builder image * Enable ChainLocks after quorums have been created Creating 4 quorums causes a lot of blocks to be created and signed by ChainLocks, which then causes timeouts later. * Increase timeout in wallet-dump.py test The first dumpwallet is quite slow sometimes, which then makes the later called dumpwallet throw a wallet locked exception.
This commit is contained in:
parent
b4e19f8dd5
commit
24fee30513
140
.gitlab-ci.yml
Normal file
140
.gitlab-ci.yml
Normal file
@ -0,0 +1,140 @@
|
|||||||
|
image: "ubuntu:bionic"
|
||||||
|
|
||||||
|
variables:
|
||||||
|
DOCKER_DRIVER: overlay2
|
||||||
|
|
||||||
|
cache:
|
||||||
|
# Cache by branch/tag and job name
|
||||||
|
# Gitlab can't use caches from parent pipelines when doing the first build in a PR, so we use artifacts to copy
|
||||||
|
# caches into PRs
|
||||||
|
key: ${CI_COMMIT_REF_SLUG}-${CI_JOB_NAME}${CI_EXTERNAL_PULL_REQUEST_IID}
|
||||||
|
paths:
|
||||||
|
- $CI_PROJECT_DIR/cache
|
||||||
|
|
||||||
|
stages:
|
||||||
|
- build
|
||||||
|
|
||||||
|
.build_template: &build_template
|
||||||
|
stage: build
|
||||||
|
before_script:
|
||||||
|
- export BUILD_TARGET="$CI_JOB_NAME"
|
||||||
|
- echo BUILD_TARGET=$BUILD_TARGET
|
||||||
|
- source ./ci/matrix.sh
|
||||||
|
|
||||||
|
# The ubuntu base image has apt configured to delete caches after each invocation, which is something that is not desirable for us
|
||||||
|
- rm /etc/apt/apt.conf.d/docker-clean
|
||||||
|
- apt-get update
|
||||||
|
- apt-get install -y wget unzip
|
||||||
|
|
||||||
|
# Init cache
|
||||||
|
- export CACHE_DIR=$CI_PROJECT_DIR/cache
|
||||||
|
- mkdir -p $CACHE_DIR
|
||||||
|
- |
|
||||||
|
if [ "$CI_COMMIT_REF_SLUG" != "develop" -a "$CI_COMMIT_TAG" == "" ]; then
|
||||||
|
if [ ! -d $CACHE_DIR/ccache ]; then
|
||||||
|
echo "Downloading cache from develop branch"
|
||||||
|
if wget -O cache-artifact.zip https://gitlab.com/$CI_PROJECT_NAMESPACE/$CI_PROJECT_NAME/-/jobs/artifacts/develop/download?job=$CI_JOB_NAME; then
|
||||||
|
unzip cache-artifact.zip
|
||||||
|
rm cache-artifact.zip
|
||||||
|
mv cache-artifact/* $CACHE_DIR/
|
||||||
|
else
|
||||||
|
echo "Failed to download cache"
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
echo "Not touching cache (was initialized from previous build)"
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
echo "Not touching cache (building develop branch or tag)"
|
||||||
|
fi
|
||||||
|
# Create missing cache dirs
|
||||||
|
- mkdir -p $CACHE_DIR/ccache && mkdir -p $CACHE_DIR/depends && mkdir -p $CACHE_DIR/sdk-sources && mkdir -p $CACHE_DIR/apt
|
||||||
|
# Keep this as it makes caching related debugging easier
|
||||||
|
- ls -lah $CACHE_DIR && ls -lah $CACHE_DIR/depends && ls -lah $CACHE_DIR/ccache && ls -lah $CACHE_DIR/apt
|
||||||
|
- mv $CACHE_DIR/apt/* /var/cache/apt/archives/ || true
|
||||||
|
|
||||||
|
# Install base packages
|
||||||
|
- apt-get dist-upgrade -y
|
||||||
|
- apt-get install -y git g++ autotools-dev libtool m4 automake autoconf pkg-config zlib1g-dev libssl1.0-dev curl ccache bsdmainutils cmake
|
||||||
|
- apt-get install -y python3 python3-dev python3-pip
|
||||||
|
|
||||||
|
# jinja2 is needed for combine_logs.py
|
||||||
|
- pip3 install jinja2
|
||||||
|
|
||||||
|
# Setup some environment variables
|
||||||
|
- if [ "$CI_EXTERNAL_PULL_REQUEST_IID" != "" ]; then export PULL_REQUEST="true"; else export PULL_REQUEST="false"; fi
|
||||||
|
- export COMMIT_RANGE="$CI_COMMIT_BEFORE_SHA..$CI_COMMIT_SHA"
|
||||||
|
- export JOB_NUMBER="$CI_JOB_ID"
|
||||||
|
- export HOST_SRC_DIR=$CI_PROJECT_DIR
|
||||||
|
- echo PULL_REQUEST=$PULL_REQUEST COMMIT_RANGE=$COMMIT_RANGE HOST_SRC_DIR=$HOST_SRC_DIR CACHE_DIR=$CACHE_DIR
|
||||||
|
- echo "Commit log:" && git log --format=fuller -1
|
||||||
|
|
||||||
|
# Build dash_hash
|
||||||
|
- git clone https://github.com/dashpay/dash_hash
|
||||||
|
- cd dash_hash && python3 setup.py install
|
||||||
|
|
||||||
|
# Install build target specific packages
|
||||||
|
- echo PACKAGES=$PACKAGES
|
||||||
|
- if [ -n "$DPKG_ADD_ARCH" ]; then dpkg --add-architecture "$DPKG_ADD_ARCH" ; fi
|
||||||
|
- if [ -n "$PACKAGES" ]; then apt-get update && apt-get install -y --no-install-recommends --no-upgrade $PACKAGES; fi
|
||||||
|
|
||||||
|
# Move apt packages into cache
|
||||||
|
- mv /var/cache/apt/archives/* $CACHE_DIR/apt/ || true
|
||||||
|
|
||||||
|
# Make mingw use correct threading libraries
|
||||||
|
- update-alternatives --set i686-w64-mingw32-gcc /usr/bin/i686-w64-mingw32-gcc-posix || true
|
||||||
|
- update-alternatives --set i686-w64-mingw32-g++ /usr/bin/i686-w64-mingw32-g++-posix || true
|
||||||
|
- update-alternatives --set x86_64-w64-mingw32-gcc /usr/bin/x86_64-w64-mingw32-gcc-posix || true
|
||||||
|
- update-alternatives --set x86_64-w64-mingw32-g++ /usr/bin/x86_64-w64-mingw32-g++-posix || true
|
||||||
|
|
||||||
|
script:
|
||||||
|
- export BUILD_TARGET="$CI_JOB_NAME"
|
||||||
|
- cd $CI_PROJECT_DIR
|
||||||
|
- ./ci/build_depends.sh
|
||||||
|
- ./ci/build_src.sh
|
||||||
|
- ./ci/test_unittests.sh
|
||||||
|
- ./ci/test_integrationtests.sh
|
||||||
|
|
||||||
|
after_script:
|
||||||
|
# Copy all cache files into cache-artifact so that they get uploaded. We only do this for develop so that artifacts
|
||||||
|
# stay minimal for PRs and branches (we never need them)
|
||||||
|
- mkdir $CI_PROJECT_DIR/cache-artifact
|
||||||
|
- mkdir -p $CI_PROJECT_DIR/testlogs
|
||||||
|
- |
|
||||||
|
if [ "$CI_COMMIT_REF_SLUG" = "develop" ]; then
|
||||||
|
cp -ra $CACHE_DIR/* $CI_PROJECT_DIR/cache-artifact/
|
||||||
|
fi
|
||||||
|
|
||||||
|
# We're actually only interested in the develop branch creating the cache artifact, but there is no way to control this
|
||||||
|
# until https://gitlab.com/gitlab-org/gitlab-foss/issues/25478 gets implemented. Until then, we use an expiration time of
|
||||||
|
# 3 days and rely on daily builds to refresh the cache artifacts. We also keep non-develop artifacts at minimum size
|
||||||
|
artifacts:
|
||||||
|
name: cache-artifact
|
||||||
|
when: always
|
||||||
|
paths:
|
||||||
|
- $CI_PROJECT_DIR/cache-artifact
|
||||||
|
- $CI_PROJECT_DIR/testlogs
|
||||||
|
expire_in: 3 days
|
||||||
|
|
||||||
|
arm-linux:
|
||||||
|
<<: *build_template
|
||||||
|
|
||||||
|
win32:
|
||||||
|
<<: *build_template
|
||||||
|
|
||||||
|
win64:
|
||||||
|
<<: *build_template
|
||||||
|
|
||||||
|
linux32:
|
||||||
|
<<: *build_template
|
||||||
|
|
||||||
|
linux64:
|
||||||
|
<<: *build_template
|
||||||
|
|
||||||
|
linux64_nowallet:
|
||||||
|
<<: *build_template
|
||||||
|
|
||||||
|
linux64_release:
|
||||||
|
<<: *build_template
|
||||||
|
|
||||||
|
mac:
|
||||||
|
<<: *build_template
|
@ -125,6 +125,7 @@ install:
|
|||||||
- export HOST_SRC_DIR=$TRAVIS_BUILD_DIR
|
- export HOST_SRC_DIR=$TRAVIS_BUILD_DIR
|
||||||
- export HOST_CACHE_DIR=$HOME/cache
|
- export HOST_CACHE_DIR=$HOME/cache
|
||||||
- export TRAVIS_COMMIT_LOG=`git log --format=fuller -1`
|
- export TRAVIS_COMMIT_LOG=`git log --format=fuller -1`
|
||||||
|
- export PYTHON_DEBUG=1
|
||||||
- source ./ci/matrix.sh
|
- source ./ci/matrix.sh
|
||||||
- mkdir -p $HOST_CACHE_DIR/docker && mkdir -p $HOST_CACHE_DIR/ccache && mkdir -p $HOST_CACHE_DIR/depends && mkdir -p $HOST_CACHE_DIR/sdk-sources
|
- mkdir -p $HOST_CACHE_DIR/docker && mkdir -p $HOST_CACHE_DIR/ccache && mkdir -p $HOST_CACHE_DIR/depends && mkdir -p $HOST_CACHE_DIR/sdk-sources
|
||||||
# Keep this as it makes caching related debugging easier
|
# Keep this as it makes caching related debugging easier
|
||||||
|
@ -14,6 +14,7 @@ RUN apt-get update && apt-get install -y python3-pip
|
|||||||
|
|
||||||
# Python stuff
|
# Python stuff
|
||||||
RUN pip3 install pyzmq # really needed?
|
RUN pip3 install pyzmq # really needed?
|
||||||
|
RUN pip3 install jinja2
|
||||||
|
|
||||||
# dash_hash
|
# dash_hash
|
||||||
RUN git clone https://github.com/dashpay/dash_hash
|
RUN git clone https://github.com/dashpay/dash_hash
|
||||||
|
@ -27,7 +27,6 @@ export DOCKER_RUN_IN_BUILDER="docker run -t --rm -w $SRC_DIR $DOCKER_RUN_ARGS $B
|
|||||||
# Default values for targets
|
# Default values for targets
|
||||||
export GOAL="install"
|
export GOAL="install"
|
||||||
export SDK_URL=${SDK_URL:-https://bitcoincore.org/depends-sources/sdks}
|
export SDK_URL=${SDK_URL:-https://bitcoincore.org/depends-sources/sdks}
|
||||||
export PYTHON_DEBUG=1
|
|
||||||
export MAKEJOBS="-j4"
|
export MAKEJOBS="-j4"
|
||||||
|
|
||||||
export RUN_UNITTESTS=false
|
export RUN_UNITTESTS=false
|
||||||
|
@ -17,4 +17,30 @@ export LD_LIBRARY_PATH=$BUILD_DIR/depends/$HOST/lib
|
|||||||
|
|
||||||
cd build-ci/dashcore-$BUILD_TARGET
|
cd build-ci/dashcore-$BUILD_TARGET
|
||||||
|
|
||||||
./test/functional/test_runner.py --coverage --quiet $PASS_ARGS
|
set +e
|
||||||
|
./test/functional/test_runner.py --coverage --quiet --nocleanup --tmpdir=$(pwd)/testdatadirs $PASS_ARGS
|
||||||
|
RESULT=$?
|
||||||
|
set -e
|
||||||
|
|
||||||
|
echo "Collecting logs..."
|
||||||
|
BASEDIR=$(ls testdatadirs)
|
||||||
|
if [ "$BASEDIR" != "" ]; then
|
||||||
|
mkdir testlogs
|
||||||
|
for d in $(ls testdatadirs/$BASEDIR | grep -v '^cache$'); do
|
||||||
|
mkdir testlogs/$d
|
||||||
|
./test/functional/combine_logs.py -c ./testdatadirs/$BASEDIR/$d > ./testlogs/$d/combined.log
|
||||||
|
./test/functional/combine_logs.py --html ./testdatadirs/$BASEDIR/$d > ./testlogs/$d/combined.html
|
||||||
|
cd testdatadirs/$BASEDIR/$d
|
||||||
|
LOGFILES="$(find . -name 'debug.log' -or -name "test_framework.log")"
|
||||||
|
cd ../../..
|
||||||
|
for f in $LOGFILES; do
|
||||||
|
d2="testlogs/$d/$(dirname $f)"
|
||||||
|
mkdir -p $d2
|
||||||
|
cp testdatadirs/$BASEDIR/$d/$f $d2/
|
||||||
|
done
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
|
||||||
|
mv testlogs ../../
|
||||||
|
|
||||||
|
exit $RESULT
|
||||||
|
@ -106,7 +106,7 @@ def print_logs(log_events, color=False, html=False):
|
|||||||
except ImportError:
|
except ImportError:
|
||||||
print("jinja2 not found. Try `pip install jinja2`")
|
print("jinja2 not found. Try `pip install jinja2`")
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
print(jinja2.Environment(loader=jinja2.FileSystemLoader('./'))
|
print(jinja2.Environment(loader=jinja2.FileSystemLoader(os.path.dirname(os.path.abspath(__file__))))
|
||||||
.get_template('combined_log_template.html')
|
.get_template('combined_log_template.html')
|
||||||
.render(title="Combined Logs from testcase", log_events=[event._asdict() for event in log_events]))
|
.render(title="Combined Logs from testcase", log_events=[event._asdict() for event in log_events]))
|
||||||
|
|
||||||
|
@ -29,13 +29,14 @@ class LLMQChainLocksTest(DashTestFramework):
|
|||||||
sync_blocks(self.nodes, timeout=60*5)
|
sync_blocks(self.nodes, timeout=60*5)
|
||||||
|
|
||||||
self.nodes[0].spork("SPORK_17_QUORUM_DKG_ENABLED", 0)
|
self.nodes[0].spork("SPORK_17_QUORUM_DKG_ENABLED", 0)
|
||||||
self.nodes[0].spork("SPORK_19_CHAINLOCKS_ENABLED", 0)
|
|
||||||
self.wait_for_sporks_same()
|
self.wait_for_sporks_same()
|
||||||
|
|
||||||
self.log.info("Mining 4 quorums")
|
self.log.info("Mining 4 quorums")
|
||||||
for i in range(4):
|
for i in range(4):
|
||||||
self.mine_quorum()
|
self.mine_quorum()
|
||||||
|
|
||||||
|
self.nodes[0].spork("SPORK_19_CHAINLOCKS_ENABLED", 0)
|
||||||
|
|
||||||
self.log.info("Mine single block, wait for chainlock")
|
self.log.info("Mine single block, wait for chainlock")
|
||||||
self.nodes[0].generate(1)
|
self.nodes[0].generate(1)
|
||||||
self.wait_for_chainlocked_block_all_nodes(self.nodes[0].getbestblockhash())
|
self.wait_for_chainlocked_block_all_nodes(self.nodes[0].getbestblockhash())
|
||||||
|
@ -100,7 +100,7 @@ class WalletDumpTest(BitcoinTestFramework):
|
|||||||
#encrypt wallet, restart, unlock and dump
|
#encrypt wallet, restart, unlock and dump
|
||||||
self.nodes[0].node_encrypt_wallet('test')
|
self.nodes[0].node_encrypt_wallet('test')
|
||||||
self.start_node(0)
|
self.start_node(0)
|
||||||
self.nodes[0].walletpassphrase('test', 10)
|
self.nodes[0].walletpassphrase('test', 30)
|
||||||
# Should be a no-op:
|
# Should be a no-op:
|
||||||
self.nodes[0].keypoolrefill()
|
self.nodes[0].keypoolrefill()
|
||||||
self.nodes[0].dumpwallet(tmpdir + "/node0/wallet.encrypted.dump")
|
self.nodes[0].dumpwallet(tmpdir + "/node0/wallet.encrypted.dump")
|
||||||
|
Loading…
Reference in New Issue
Block a user