@@ -75,13 +75,13 @@ operator: configure-operator build-and-push-operator-image
7575
7676# build-push, (todo) restart database
7777database : aws_login
78- @ scripts/evergreen /run_python.sh pipeline.py --include database
78+ @ scripts/dev /run_python.sh pipeline.py --include database
7979
8080readiness_probe : aws_login
81- @ scripts/evergreen /run_python.sh pipeline.py --include readiness-probe
81+ @ scripts/dev /run_python.sh pipeline.py --include readiness-probe
8282
8383upgrade_hook : aws_login
84- @ scripts/evergreen /run_python.sh pipeline.py --include upgrade-hook
84+ @ scripts/dev /run_python.sh pipeline.py --include upgrade-hook
8585
8686# ensures cluster is up, cleans Kubernetes + OM, build-push-deploy operator,
8787# push-deploy database, create secrets, config map, resources etc
@@ -90,7 +90,7 @@ full: build-and-push-images
9090
9191# build-push appdb image
9292appdb : aws_login
93- @ scripts/evergreen /run_python.sh pipeline.py --include appdb
93+ @ scripts/dev /run_python.sh pipeline.py --include appdb
9494
9595# runs the e2e test: make e2e test=e2e_sharded_cluster_pv. The Operator is redeployed before the test, the namespace is cleaned.
9696# The e2e test image is built and pushed together with all main ones (operator, database, init containers)
@@ -112,7 +112,7 @@ mco-e2e: aws_login build-and-push-mco-test-image
112112
113113generate-env-file : # # generates a local-test.env for local testing
114114 mkdir -p .generated
115- { scripts/evergreen /run_python.sh mongodb-community-operator/scripts/dev/get_e2e_env_vars.py " .generated/config.json" | tee >( cut -d' ' -f2 > .generated/mco-test.env) ; } > .generated/mco-test.export.env
115+ { scripts/dev /run_python.sh mongodb-community-operator/scripts/dev/get_e2e_env_vars.py " .generated/config.json" | tee >( cut -d' ' -f2 > .generated/mco-test.env) ; } > .generated/mco-test.export.env
116116 . .generated/mco-test.export.env
117117
118118reset-helm-leftovers : # # sometimes you didn't cleanly uninstall a helm release, this cleans the existing helm artifacts
@@ -154,19 +154,19 @@ aws_cleanup:
154154 @ scripts/evergreen/prepare_aws.sh
155155
156156build-and-push-operator-image : aws_login
157- @ scripts/evergreen /run_python.sh pipeline.py --include operator-quick
157+ @ scripts/dev /run_python.sh pipeline.py --include operator-quick
158158
159159build-and-push-database-image : aws_login
160160 @ scripts/dev/build_push_database_image
161161
162162build-and-push-test-image : aws_login build-multi-cluster-binary
163163 @ if [[ -z " $( local) " ]]; then \
164- scripts/evergreen /run_python.sh pipeline.py --include test ; \
164+ scripts/dev /run_python.sh pipeline.py --include test ; \
165165 fi
166166
167167build-and-push-mco-test-image : aws_login
168168 @ if [[ -z " $( local) " ]]; then \
169- scripts/evergreen /run_python.sh pipeline.py --include mco-test; \
169+ scripts/dev /run_python.sh pipeline.py --include mco-test; \
170170 fi
171171
172172build-multi-cluster-binary :
@@ -181,27 +181,27 @@ build-and-push-images: build-and-push-operator-image appdb-init-image om-init-im
181181build-and-push-init-images : appdb-init-image om-init-image database-init-image
182182
183183database-init-image :
184- @ scripts/evergreen /run_python.sh pipeline.py --include init-database
184+ @ scripts/dev /run_python.sh pipeline.py --include init-database
185185
186186appdb-init-image :
187- @ scripts/evergreen /run_python.sh pipeline.py --include init-appdb
187+ @ scripts/dev /run_python.sh pipeline.py --include init-appdb
188188
189189# Not setting a parallel-factor will default to 0 which will lead to using all CPUs, that can cause docker to die.
190190# Here we are defaulting to 6, a higher value might work for you.
191191agent-image :
192- @ scripts/evergreen /run_python.sh pipeline.py --include agent --all-agents --parallel --parallel-factor 6
192+ @ scripts/dev /run_python.sh pipeline.py --include agent --all-agents --parallel --parallel-factor 6
193193
194194agent-image-slow :
195- @ scripts/evergreen /run_python.sh pipeline.py --include agent --parallel-factor 1
195+ @ scripts/dev /run_python.sh pipeline.py --include agent --parallel-factor 1
196196
197197operator-image :
198- @ scripts/evergreen /run_python.sh pipeline.py --include operator
198+ @ scripts/dev /run_python.sh pipeline.py --include operator
199199
200200om-init-image :
201- @ scripts/evergreen /run_python.sh pipeline.py --include init-ops-manager
201+ @ scripts/dev /run_python.sh pipeline.py --include init-ops-manager
202202
203203om-image :
204- @ scripts/evergreen /run_python.sh pipeline.py --include ops-manager
204+ @ scripts/dev /run_python.sh pipeline.py --include ops-manager
205205
206206configure-operator :
207207 @ scripts/dev/configure_operator.sh
@@ -284,16 +284,16 @@ golang-tests-race:
284284 USE_RACE=true scripts/evergreen/unit-tests.sh
285285
286286sbom-tests :
287- @ scripts/evergreen /run_python.sh -m pytest generate_ssdlc_report_test.py
287+ @ scripts/dev /run_python.sh -m pytest generate_ssdlc_report_test.py
288288
289289# e2e tests are also in python and we will need to ignore them as they are in the docker/mongodb-kubernetes-tests folder
290290# additionally, we have one lib which we want to test which is in the =docker/mongodb-kubernetes-tests folder.
291291python-tests :
292- @ scripts/evergreen /run_python.sh -m pytest docker/mongodb-kubernetes-tests/kubeobject
293- @ scripts/evergreen /run_python.sh -m pytest --ignore=docker/mongodb-kubernetes-tests
292+ @ scripts/dev /run_python.sh -m pytest docker/mongodb-kubernetes-tests/kubeobject
293+ @ scripts/dev /run_python.sh -m pytest --ignore=docker/mongodb-kubernetes-tests
294294
295295generate-ssdlc-report :
296- @ scripts/evergreen /run_python.sh generate_ssdlc_report.py
296+ @ scripts/dev /run_python.sh generate_ssdlc_report.py
297297
298298# test-race runs golang test with race enabled
299299test-race : generate fmt vet manifests golang-tests-race
0 commit comments