Skip to content

Commit b139449

Browse files
committed
test: update tests/plugin-codeflare/dashboard/inputs/1
Update this test input to have the latest schema
1 parent a41b2f7 commit b139449

File tree

11 files changed

+2507
-2234
lines changed

11 files changed

+2507
-2234
lines changed
Lines changed: 51 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -1,32 +1,66 @@
11
{
2+
"name": "default",
3+
"creationTime": 1657642748915,
4+
"lastModifiedTime": 1658146025488,
5+
"choices": {
6+
"madwizard/apriori/use-gpu": "don't use gpus",
7+
"madwizard/apriori/arch": "x64",
8+
"madwizard/apriori/platform": "darwin",
9+
"madwizard/apriori/mac-installer": "Homebrew",
10+
"madwizard/apriori/in-terminal": "HTML",
11+
"Training####Fine Tuning": "Fine Tuning",
12+
"GLUE": "GLUE",
13+
"AWS####IBM": "AWS",
14+
"S3 Bucket for Run Data.expand([ -n \"$MC_CONFIG_DIR\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls s3 | awk '{print substr($NF, 1, length($NF) - 1)}', S3 Buckets)####separator####📁 Create a new bucket": "browsey",
15+
"Run Locally####Run on a Kubernetes Cluster": "Run on a Kubernetes Cluster",
16+
"Choose the bucket that contains your model and glue data.madwizard/apriori/platform": "Darwin",
17+
"expand(kubectl config get-contexts -o name, Kubernetes contexts)": "default/api-codeflare-train-v11-codeflare-openshift-com:6443/kube:admin",
18+
"expand([ -z ${KUBE_CONTEXT} ] && exit 1 || kubectl --context ${KUBE_CONTEXT} get ns -o name | grep -Ev 'openshift|kube-' | sed 's#namespace/##', Kubernetes namespaces)####Create a namespace": "nvidia-gpu-operator",
19+
"Number of CPUs####Number of GPUs####Minimum Workers####Maximum Workers####Worker Memory####Head Memory": "{\"Number of CPUs\":\"1\",\"Number of GPUs\":\"1\",\"Minimum Workers\":\"1\",\"Maximum Workers\":\"1\",\"Worker Memory\":\"32Gi\",\"Head Memory\":\"32Gi\"}",
20+
"Choose the bucket that contains your model and glue data.expand([ -n \"$MC_CONFIG_DIR\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls s3 | awk '{print substr($NF, 1, length($NF) - 1)}', S3 Buckets)####separator####📁 Create a new bucket": "browsey",
21+
"Choose your Model File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "roberta-base",
22+
"Choose your Glue Data File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "glue_data",
23+
"BERT": "BERT",
24+
"Example: Using Ray Tasks to Parallelize a Function####Example: Using Ray Actors to Parallelize a Class####Example: Creating and Transforming Datasets####Example: Training Using PyTorch####Example: Hyperparameter Tuning####Example: Serving a scikit-learn gradient boosting classifier": "Example: Using Ray Tasks to Parallelize a Function",
25+
"Number of CPUs####Number of GPUs": "{\"Number of CPUs\":4,\"Number of GPUs\":3}",
26+
"expand(echo ${A-error} ; echo ${B-4} ; echo ${C-5})": "3",
27+
"XXXXXX.11111####222222": "11111",
28+
"YYYYYY.11111####222222": "222222",
29+
"My Cluster is Running Locally####My Cluster is Runing on Kubernetes": "My Cluster is Runing on Kubernetes",
30+
"expand([ -n \"$RAY_ADDRESS\" ] && ray job list --address $RAY_ADDRESS | tail +2 | awk '{print $1}' | sed \"s/[:{' ]//g\", Ray Runs)": "07a2647f-3656-4e3e-836c-95a2fa841af6",
31+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'keys | .[]', Ray Runs)": "d5a0d68f-a675-49ca-bff7-4ae762e6b146",
32+
"My Cluster is Running Locally####My Cluster is Running on Kubernetes": "My Cluster is Running on Kubernetes",
33+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2; $1=\"\";$2=\"\";print \"\\033;1m\" a, \"\\033[0;33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "\u001b;1ma88d4632-ab5c-4350-a770-d39a955c42c8 \u001b[0;33mRUNNING\u001b[0;2m -v --datapath /tmp/ --modelpath /tmp/ --logpath /tmp/ --tblogpath s3://browsey/codeflare/a88d4632-ab5c-4350-a770-d39a955c42c8/tensorboard/ --num_workers 1\u001b[0m",
34+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2; $1=\"\";$2=\"\";print a, \"\\033[33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "a88d4632-ab5c-4350-a770-d39a955c42c8 \u001b[33mRUNNING\u001b[0;2m -v --datapath /tmp/ --modelpath /tmp/ --logpath /tmp/ --tblogpath s3://browsey/codeflare/a88d4632-ab5c-4350-a770-d39a955c42c8/tensorboard/ --num_workers 1\u001b[0m",
35+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.start_time / 1000 | strflocaltime(\"%Y-%m-%dT%H:%M:%S\")) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2;c=$3; $1=\"\";$2=\"\";$3=\"\"; print a, \"\\033[0;36m\" c, \"\\033[0;1;33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "8443709b-74b3-4ae5-ae5a-aa4e1fbe2bb1 \u001b[0;36m2022-07-18T08:06:48 \u001b[0;1;33mPENDING\u001b[0;2m python main.py\u001b[0m",
36+
"Start a new Run####Connect Dashboard to an existing Run": "Connect Dashboard to an existing Run",
37+
"Start a new Run####Connect Dashboard to an existing Run####Shut down a Cloud Computer": "Shut down a Cloud Computer",
38+
"Start a new Run####Connect Dashboard to an existing Run####Boot up a Cloud Computer####Shut down a Cloud Computer": "Start a new Run",
39+
"SubTab1####SubTab2": "SubTab1",
40+
"Tab1####Tab2": "Tab2",
41+
"BERT####BYOT": "BYOT",
42+
"Location of your working directory": "{\"Location of your working directory\":\"/tmp/qiskit/ray_demo\"}",
43+
"Provide custom base image, if any": "{\"Provide custom base image, if any\":\"rayproject/ray:1.13.0-py37\"}"
44+
},
245
"madwizard/apriori/use-gpu": "don't use gpus",
346
"madwizard/apriori/arch": "x64",
447
"madwizard/apriori/platform": "darwin",
548
"madwizard/apriori/mac-installer": "Homebrew",
649
"madwizard/apriori/in-terminal": "HTML",
50+
"Run Locally####Run on a Kubernetes Cluster": "Run on a Kubernetes Cluster",
51+
"Example: Using Ray Tasks to Parallelize a Function####Example: Using Ray Actors to Parallelize a Class####Example: Creating and Transforming Datasets####Example: Training Using PyTorch####Example: Hyperparameter Tuning####Example: Serving a scikit-learn gradient boosting classifier": "Example: Using Ray Tasks to Parallelize a Function",
52+
"Start a new Run####Connect Dashboard to an existing Run####Boot up a Cloud Computer####Shut down a Cloud Computer": "Start a new Run",
753
"Training####Fine Tuning": "Fine Tuning",
8-
"GLUE": "GLUE",
54+
"BERT": "BERT",
955
"AWS####IBM": "AWS",
1056
"S3 Bucket for Run Data.expand([ -n \"$MC_CONFIG_DIR\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls s3 | awk '{print substr($NF, 1, length($NF) - 1)}', S3 Buckets)####separator####📁 Create a new bucket": "browsey",
11-
"Run Locally####Run on a Kubernetes Cluster": "Run on a Kubernetes Cluster",
12-
"Choose the bucket that contains your model and glue data.madwizard/apriori/platform": "Darwin",
1357
"expand(kubectl config get-contexts -o name, Kubernetes contexts)": "default/api-codeflare-train-v11-codeflare-openshift-com:6443/kube:admin",
1458
"expand([ -z ${KUBE_CONTEXT} ] && exit 1 || kubectl --context ${KUBE_CONTEXT} get ns -o name | grep -Ev 'openshift|kube-' | sed 's#namespace/##', Kubernetes namespaces)####Create a namespace": "nvidia-gpu-operator",
1559
"Number of CPUs####Number of GPUs####Minimum Workers####Maximum Workers####Worker Memory####Head Memory": "{\"Number of CPUs\":\"1\",\"Number of GPUs\":\"1\",\"Minimum Workers\":\"1\",\"Maximum Workers\":\"1\",\"Worker Memory\":\"32Gi\",\"Head Memory\":\"32Gi\"}",
60+
"My Cluster is Running Locally####My Cluster is Running on Kubernetes": "My Cluster is Running on Kubernetes",
61+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.start_time / 1000 | strflocaltime(\"%Y-%m-%dT%H:%M:%S\")) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2;c=$3; $1=\"\";$2=\"\";$3=\"\"; print a, \"\\033[0;36m\" c, \"\\033[0;1;33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "2c495c48-2202-4be6-a8f7-4cccbb217f34 \u001b[0;36m2022-07-12T13:27:59 \u001b[0;1;33mPENDING\u001b[0;2m -v -b browsey -m roberta-base -g glue_data -t WNLI -M -s 40 41 42 43\u001b[0m",
62+
"GLUE": "GLUE",
1663
"Choose the bucket that contains your model and glue data.expand([ -n \"$MC_CONFIG_DIR\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls s3 | awk '{print substr($NF, 1, length($NF) - 1)}', S3 Buckets)####separator####📁 Create a new bucket": "browsey",
1764
"Choose your Model File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "roberta-base",
18-
"Choose your Glue Data File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "glue_data",
19-
"BERT": "BERT",
20-
"Example: Using Ray Tasks to Parallelize a Function####Example: Using Ray Actors to Parallelize a Class####Example: Creating and Transforming Datasets####Example: Training Using PyTorch####Example: Hyperparameter Tuning####Example: Serving a scikit-learn gradient boosting classifier": "Example: Using Ray Tasks to Parallelize a Function",
21-
"Number of CPUs####Number of GPUs": "{\"Number of CPUs\":4,\"Number of GPUs\":3}",
22-
"expand(echo ${A-error} ; echo ${B-4} ; echo ${C-5})": "3",
23-
"XXXXXX.11111####222222": "11111",
24-
"YYYYYY.11111####222222": "222222",
25-
"My Cluster is Running Locally####My Cluster is Runing on Kubernetes": "My Cluster is Runing on Kubernetes",
26-
"expand([ -n \"$RAY_ADDRESS\" ] && ray job list --address $RAY_ADDRESS | tail +2 | awk '{print $1}' | sed \"s/[:{' ]//g\", Ray Runs)": "07a2647f-3656-4e3e-836c-95a2fa841af6",
27-
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'keys | .[]', Ray Runs)": "d5a0d68f-a675-49ca-bff7-4ae762e6b146",
28-
"My Cluster is Running Locally####My Cluster is Running on Kubernetes": "My Cluster is Running on Kubernetes",
29-
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2; $1=\"\";$2=\"\";print \"\\033;1m\" a, \"\\033[0;33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "\u001b;1ma88d4632-ab5c-4350-a770-d39a955c42c8 \u001b[0;33mRUNNING\u001b[0;2m -v --datapath /tmp/ --modelpath /tmp/ --logpath /tmp/ --tblogpath s3://browsey/codeflare/a88d4632-ab5c-4350-a770-d39a955c42c8/tensorboard/ --num_workers 1\u001b[0m",
30-
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2; $1=\"\";$2=\"\";print a, \"\\033[33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "a88d4632-ab5c-4350-a770-d39a955c42c8 \u001b[33mRUNNING\u001b[0;2m -v --datapath /tmp/ --modelpath /tmp/ --logpath /tmp/ --tblogpath s3://browsey/codeflare/a88d4632-ab5c-4350-a770-d39a955c42c8/tensorboard/ --num_workers 1\u001b[0m",
31-
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.start_time / 1000 | strflocaltime(\"%Y-%m-%dT%H:%M:%S\")) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2;c=$3; $1=\"\";$2=\"\";$3=\"\"; print a, \"\\033[0;36m\" c, \"\\033[0;1;33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "505b98b6-a258-4afd-bdc6-ddf84d3f2862 \u001b[0;36m2022-07-07T13:27:54 \u001b[0;1;33mRUNNING\u001b[0;2m -v --datapath /tmp/ --modelpath /tmp/ --logpath /tmp/ --tblogpath s3://browsey/codeflare/505b98b6-a258-4afd-bdc6-ddf84d3f2862/tensorboard/ --num_workers 1\u001b[0m"
65+
"Choose your Glue Data File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "glue_data"
3266
}
Lines changed: 20 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,20 @@
1+
LAST SEEN TYPE REASON OBJECT MESSAGE
2+
0s Normal ScalingReplicaSet deployment/ray-operator Scaled up replica set ray-operator-56f7b64d7b to 1
3+
0s Normal SuccessfulCreate replicaset/ray-operator-56f7b64d7b Created pod: ray-operator-56f7b64d7b-dkwcc
4+
0s Normal Scheduled pod/ray-operator-56f7b64d7b-dkwcc Successfully assigned nvidia-gpu-operator/ray-operator-56f7b64d7b-dkwcc to ip-10-0-135-114.ec2.internal
5+
0s Normal AddedInterface pod/ray-operator-56f7b64d7b-dkwcc Add eth0 [10.130.82.53/23] from openshift-sdn
6+
0s Normal Pulled pod/ray-operator-56f7b64d7b-dkwcc Container image "rayproject/ray:1.13.0-py37" already present on machine
7+
0s Normal Created pod/ray-operator-56f7b64d7b-dkwcc Created container ray
8+
0s Normal Started pod/ray-operator-56f7b64d7b-dkwcc Started container ray
9+
0s Normal Scheduled pod/mycluster-ray-head-type-2s9nz Successfully assigned nvidia-gpu-operator/mycluster-ray-head-type-2s9nz to ip-10-0-135-114.ec2.internal
10+
0s Normal AddedInterface pod/mycluster-ray-head-type-2s9nz Add eth0 [10.130.82.54/23] from openshift-sdn
11+
0s Normal Pulling pod/mycluster-ray-head-type-2s9nz Pulling image "rayproject/ray-ml:1.13.0-py37-gpu"
12+
0s Normal Pulled pod/mycluster-ray-head-type-2s9nz Successfully pulled image "rayproject/ray-ml:1.13.0-py37-gpu" in 5m31.366599351s
13+
0s Normal Created pod/mycluster-ray-head-type-2s9nz Created container ray-node
14+
0s Normal Started pod/mycluster-ray-head-type-2s9nz Started container ray-node
15+
0s Normal Scheduled pod/mycluster-ray-worker-type-p2s6h Successfully assigned nvidia-gpu-operator/mycluster-ray-worker-type-p2s6h to ip-10-0-130-229.ec2.internal
16+
0s Normal AddedInterface pod/mycluster-ray-worker-type-p2s6h Add eth0 [10.129.82.19/23] from openshift-sdn
17+
0s Normal Pulling pod/mycluster-ray-worker-type-p2s6h Pulling image "rayproject/ray-ml:1.13.0-py37-gpu"
18+
0s Normal Pulled pod/mycluster-ray-worker-type-p2s6h Successfully pulled image "rayproject/ray-ml:1.13.0-py37-gpu" in 5m25.620794291s
19+
0s Normal Created pod/mycluster-ray-worker-type-p2s6h Created container ray-node
20+
0s Normal Started pod/mycluster-ray-worker-type-p2s6h Started container ray-node

tests/plugin-codeflare/dashboard/inputs/1/job.json

Lines changed: 16 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -1,28 +1,28 @@
11
{
2-
"jobid": "bf3d5456-fba9-42b1-9b55-119366b409ee",
2+
"jobid": "1c219957-7da9-48ca-b3a8-f4f5e5b9d089",
33
"cmdline": {
4-
"appPart": "python3 tmp-10491-mEkonq1toz76 -v -b ${S3_BUCKET} -m ${S3_OBJECTMODEL} -g ${S3_OBJECTGLUEDATA} -t WNLI -M -s 40 41 42 43",
5-
"systemPart": "ray job submit --runtime-env=/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp-10491-BamAl1LRY98R --job-id ${JOB_ID} --no-wait"
4+
"appPart": "python3 tmp-66037-eEdZYdDtlEBh -v -b ${S3_BUCKET} -m ${S3_OBJECTMODEL} -g ${S3_OBJECTGLUEDATA} -t WNLI -M -s 40 41 42 43",
5+
"systemPart": "ray job submit --runtime-env=/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp-66037-Biargx8GGZ68 --job-id ${JOB_ID} --no-wait"
66
},
77
"runtimeEnv": {
88
"env_vars": {
9-
"JOB_ID": "bf3d5456-fba9-42b1-9b55-119366b409ee",
9+
"JOB_ID": "1c219957-7da9-48ca-b3a8-f4f5e5b9d089",
1010
"S3_PROVIDER": "aws",
1111
"S3_ENDPOINT": "https://s3.amazonaws.com",
1212
"S3_ACCESS_KEY_ID": "********",
1313
"AWS_ACCESS_KEY_ID": "********",
1414
"S3_SECRET_ACCESS_KEY": "********",
1515
"AWS_SECRET_ACCESS_KEY": "********",
16-
"MC_CONFIG_DIR": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp.8btLMMUx",
16+
"MC_CONFIG_DIR": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp.JHciytLd",
1717
"S3_BUCKETRAYLOGS": "browsey",
1818
"S3_FILEPATHRAYLOGS": "RAYLOGS",
19-
"LOGDIR_STAGE": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/logdir-stage.bj5hQCjp",
20-
"STREAMCONSUMER_LOGS": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/logdir-stage.bj5hQCjp/logs/",
21-
"STREAMCONSUMER_EVENTS": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/logdir-stage.bj5hQCjp/events/",
22-
"STREAMCONSUMER_RESOURCES": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/logdir-stage.bj5hQCjp/resources/",
23-
"S3_LOGDIR": "browsey/codeflare/bf3d5456-fba9-42b1-9b55-119366b409ee",
24-
"LOGDIR_URI": "s3://browsey/codeflare/bf3d5456-fba9-42b1-9b55-119366b409ee",
25-
"LOGDIR_MC": "s3/browsey/codeflare/bf3d5456-fba9-42b1-9b55-119366b409ee",
19+
"LOGDIR_STAGE": "/Users/nickm/Library/Application Support/madwizard-nodejs/default/jobs/1c219957-7da9-48ca-b3a8-f4f5e5b9d089",
20+
"STREAMCONSUMER_LOGS": "/Users/nickm/Library/Application Support/madwizard-nodejs/default/jobs/1c219957-7da9-48ca-b3a8-f4f5e5b9d089/logs/",
21+
"STREAMCONSUMER_EVENTS": "/Users/nickm/Library/Application Support/madwizard-nodejs/default/jobs/1c219957-7da9-48ca-b3a8-f4f5e5b9d089/events/",
22+
"STREAMCONSUMER_RESOURCES": "/Users/nickm/Library/Application Support/madwizard-nodejs/default/jobs/1c219957-7da9-48ca-b3a8-f4f5e5b9d089/resources/",
23+
"S3_LOGDIR": "browsey/codeflare/1c219957-7da9-48ca-b3a8-f4f5e5b9d089",
24+
"LOGDIR_URI": "s3://browsey/codeflare/1c219957-7da9-48ca-b3a8-f4f5e5b9d089",
25+
"LOGDIR_MC": "s3/browsey/codeflare/1c219957-7da9-48ca-b3a8-f4f5e5b9d089",
2626
"PATH": "/Users/nickm/.local/bin:/usr/local/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/bin:/Users/nickm/.krew/bin:/Library/Frameworks/Python.framework/Versions/2.7/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/Applications/VMware Fusion.app/Contents/Public:/Library/TeX/texbin:/usr/local/go/bin:/Library/Apple/usr/bin:/Users/nickm/bin",
2727
"KUBE_CONTEXT": "default/api-codeflare-train-v11-codeflare-openshift-com:6443/kube:admin",
2828
"KUBE_NS": "nvidia-gpu-operator",
@@ -37,10 +37,10 @@
3737
"HEAD_MEMORY": "32Gi",
3838
"RAY_OPERATOR_IMAGE": "rayproject/ray:1.13.0-py37",
3939
"RAY_IMAGE": "rayproject/ray-ml:1.13.0-py37-gpu",
40-
"HELM_CLONE_TEMPDIR": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp.LPMYQyK4",
40+
"HELM_CLONE_TEMPDIR": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp.seH7UeMl",
4141
"RAY_KUBE_CLUSTER_NAME": "mycluster",
42-
"RAY_KUBE_PORT": "8508",
43-
"RAY_ADDRESS": "http://127.0.0.1:8508",
42+
"RAY_KUBE_PORT": "8560",
43+
"RAY_ADDRESS": "http://127.0.0.1:8560",
4444
"S3_BUCKET": "browsey",
4545
"S3_FILEPATH": "browsey",
4646
"S3_OBJECTMODEL": "roberta-base",
@@ -50,7 +50,7 @@
5050
"WANDB_CONFIG_DIR": "/tmp",
5151
"WANDB_DISABLED": "true"
5252
},
53-
"working_dir": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp-10491-pvWSI0jqUa2s",
53+
"working_dir": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp-66037-nR3WsLSjtgIz",
5454
"pip": ["boto3", "ray[default]", "ray_lightning", "pytorch_lightning", "torchvision", "transformers==3.0.2"]
5555
},
5656
"language": "python",
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
bf3d5456-fba9-42b1-9b55-119366b409ee
1+
1c219957-7da9-48ca-b3a8-f4f5e5b9d089

0 commit comments

Comments
 (0)