blob: 98526dcec22a77db366df6cdb34bc3bd9a5357fa (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
|
_templates:
common: &common
timeout_in_minutes: 30
retry:
automatic:
- exit_status: -1
limit: 10
- exit_status: "*"
limit: 2
benchmarks: &benchmarks
timeout_in_minutes: 120
retry:
automatic: false
soft_fail: true
if: build.branch == "master"
env:
# BENCHMARKS_OFFICIAL is set from hooks/pre-command, based
# on whether this is executing on the master branch.
BENCHMARKS_DATASET: buildkite
BENCHMARKS_PLATFORMS: "ptrace kvm"
BENCHMARKS_PROJECT: gvisor-benchmarks
BENCHMARKS_TABLE: benchmarks
BENCHMARKS_UPLOAD: true
steps:
# Run basic smoke tests before preceding to other tests.
- <<: *common
label: ":fire: Smoke tests"
command: make smoke-tests
- wait
# Check that the Go branch builds.
- <<: *common
label: ":golang: Go branch"
commands:
- tools/go_branch.sh
- git checkout go && git clean -xf .
- go build ./...
# Release workflow.
- <<: *common
label: ":ship: Release tests"
commands:
- make artifacts/x86_64
- make BAZEL_OPTIONS=--config=cross-aarch64 artifacts/aarch64
- make release
# Images tests.
- <<: *common
label: ":docker: Images (x86_64)"
command: make ARCH=x86_64 load-all-images
- <<: *common
label: ":docker: Images (aarch64)"
command: make ARCH=aarch64 load-all-images
# Basic unit tests.
- <<: *common
label: ":test_tube: Unit tests"
command: make unit-tests
- <<: *common
label: ":test_tube: runsc tests"
command: make runsc-tests
# All system call tests.
- <<: *common
label: ":toolbox: System call tests"
command: make syscall-tests
parallelism: 20
# Integration tests.
- <<: *common
label: ":parachute: FUSE tests"
command: make fuse-tests
- <<: *common
label: ":docker: Docker tests"
command: make docker-tests
- <<: *common
label: ":goggles: Overlay tests"
command: make overlay-tests
- <<: *common
label: ":safety_pin: Host network tests"
command: make hostnet-tests
- <<: *common
label: ":satellite: SWGSO tests"
command: make swgso-tests
- <<: *common
label: ":coffee: Do tests"
command: make do-tests
- <<: *common
label: ":person_in_lotus_position: KVM tests"
command: make kvm-tests
- <<: *common
label: ":docker: Containerd 1.3.9 tests"
command: make containerd-test-1.3.9
- <<: *common
label: ":docker: Containerd 1.4.3 tests"
command: make containerd-test-1.4.3
# Check the website builds.
- <<: *common
label: ":earth_americas: Website tests"
command: make website-build
# Networking tests.
- <<: *common
label: ":table_tennis_paddle_and_ball: IPTables tests"
command: make iptables-tests
- <<: *common
label: ":construction_worker: Packetdrill tests"
command: make packetdrill-tests
- <<: *common
label: ":hammer: Packetimpact tests"
command: make packetimpact-tests
# Runtime tests.
- <<: *common
label: ":php: PHP runtime tests"
command: make php7.3.6-runtime-tests_vfs2
parallelism: 10
- <<: *common
label: ":java: Java runtime tests"
command: make java11-runtime-tests_vfs2
parallelism: 40
- <<: *common
label: ":golang: Go runtime tests"
command: make go1.12-runtime-tests_vfs2
parallelism: 10
- <<: *common
label: ":node: NodeJS runtime tests"
command: make nodejs12.4.0-runtime-tests_vfs2
parallelism: 10
- <<: *common
label: ":python: Python runtime tests"
command: make python3.7.3-runtime-tests_vfs2
parallelism: 10
# Runtime tests (VFS1).
- <<: *common
label: ":php: PHP runtime tests (VFS1)"
command: make php7.3.6-runtime-tests
parallelism: 10
if: build.message =~ /VFS1/ || build.branch == "master"
- <<: *common
label: ":java: Java runtime tests (VFS1)"
command: make java11-runtime-tests
parallelism: 40
if: build.message =~ /VFS1/ || build.branch == "master"
- <<: *common
label: ":golang: Go runtime tests (VFS1)"
command: make go1.12-runtime-tests
parallelism: 10
if: build.message =~ /VFS1/ || build.branch == "master"
- <<: *common
label: ":node: NodeJS runtime tests (VFS1)"
command: make nodejs12.4.0-runtime-tests
parallelism: 10
if: build.message =~ /VFS1/ || build.branch == "master"
- <<: *common
label: ":python: Python runtime tests (VFS1)"
command: make python3.7.3-runtime-tests
parallelism: 10
if: build.message =~ /VFS1/ || build.branch == "master"
# ARM tests.
- <<: *common
label: ":mechanical_arm: ARM"
command: make arm-qemu-smoke-test
# Run basic benchmarks smoke tests (no upload).
- <<: *common
label: ":fire: Benchmarks smoke test"
command: make benchmark-platforms
# Use the opposite of the benchmarks filter.
if: build.branch != "master"
# Run all benchmarks.
- <<: *benchmarks
label: ":bazel: ABSL build benchmarks"
command: make benchmark-platforms BENCHMARKS_FILTER="ABSL/page_cache.clean" BENCHMARKS_SUITE=absl BENCHMARKS_TARGETS=test/benchmarks/fs:bazel_test
- <<: *benchmarks
label: ":go: runsc build benchmarks"
command: make benchmark-platforms BENCHMARKS_FILTER="Runsc/page_cache.clean/filesystem.bind" BENCHMARKS_SUITE=runsc BENCHMARKS_TARGETS=test/benchmarks/fs:bazel_test
- <<: *benchmarks
label: ":metal: FFMPEG benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=ffmpeg BENCHMARKS_TARGETS=test/benchmarks/media:ffmpeg_test
# For fio, running with --test.benchtime=Xs scales the written/read
# bytes to several GB. This is not a problem for root/bind/volume mounts,
# but for tmpfs mounts, the size can grow to more memory than the machine
# has availabe. Fix the runs to 1GB written/read for the benchmark.
- <<: *benchmarks
label: ":floppy_disk: FIO benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=fio BENCHMARKS_TARGETS=test/benchmarks/fs:fio_test BENCHMARKS_OPTIONS=--test.benchtime=1000x
- <<: *benchmarks
label: ":globe_with_meridians: HTTPD benchmarks"
command: make benchmark-platforms BENCHMARKS_FILTER="Continuous" BENCHMARKS_SUITE=httpd BENCHMARKS_TARGETS=test/benchmarks/network:httpd_test
- <<: *benchmarks
label: ":piedpiper: iperf benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=iperf BENCHMARKS_TARGETS=test/benchmarks/network:iperf_test
- <<: *benchmarks
label: ":nginx: nginx benchmarks"
command: make benchmark-platforms BENCHMARKS_FILTER="Continuous" BENCHMARKS_SUITE=nginx BENCHMARKS_TARGETS=test/benchmarks/network:nginx_test
- <<: *benchmarks
label: ":node: node benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=node BENCHMARKS_TARGETS=test/benchmarks/network:node_test
- <<: *benchmarks
label: ":redis: Redis benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=redis BENCHMARKS_TARGETS=test/benchmarks/database:redis_test BENCHMARKS_OPTIONS=-test.benchtime=15s
- <<: *benchmarks
label: ":ruby: Ruby benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=ruby BENCHMARKS_TARGETS=test/benchmarks/network:ruby_test
- <<: *benchmarks
label: ":weight_lifter: Size benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=size BENCHMARKS_TARGETS=test/benchmarks/base:size_test
- <<: *benchmarks
label: ":speedboat: Startup benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=startup BENCHMARKS_TARGETS=test/benchmarks/base:startup_test
- <<: *benchmarks
label: ":computer: sysbench benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=sysbench BENCHMARKS_TARGETS=test/benchmarks/base:sysbench_test
- <<: *benchmarks
label: ":tensorflow: TensorFlow benchmarks"
command: make benchmark-platforms BENCHMARKS_SUITE=tensorflow BENCHMARKS_TARGETS=test/benchmarks/ml:tensorflow_test
|