Skip to content

Commit 5e5a417

Browse files
committed
test: update tests/plugin-codeflare/dashboard/inputs/2
Update this test input to have the latest schema
1 parent b139449 commit 5e5a417

File tree

12 files changed

+4337
-2650
lines changed

12 files changed

+4337
-2650
lines changed
Lines changed: 52 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,25 +1,66 @@
11
{
2+
"name": "default",
3+
"creationTime": 1657642748915,
4+
"lastModifiedTime": 1658152672874,
5+
"choices": {
6+
"madwizard/apriori/use-gpu": "don't use gpus",
7+
"madwizard/apriori/arch": "x64",
8+
"madwizard/apriori/platform": "darwin",
9+
"madwizard/apriori/mac-installer": "Homebrew",
10+
"madwizard/apriori/in-terminal": "HTML",
11+
"Training####Fine Tuning": "Fine Tuning",
12+
"GLUE": "GLUE",
13+
"AWS####IBM": "AWS",
14+
"S3 Bucket for Run Data.expand([ -n \"$MC_CONFIG_DIR\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls s3 | awk '{print substr($NF, 1, length($NF) - 1)}', S3 Buckets)####separator####📁 Create a new bucket": "browsey",
15+
"Run Locally####Run on a Kubernetes Cluster": "Run on a Kubernetes Cluster",
16+
"Choose the bucket that contains your model and glue data.madwizard/apriori/platform": "Darwin",
17+
"expand(kubectl config get-contexts -o name, Kubernetes contexts)": "default/api-codeflare-train-v11-codeflare-openshift-com:6443/kube:admin",
18+
"expand([ -z ${KUBE_CONTEXT} ] && exit 1 || kubectl --context ${KUBE_CONTEXT} get ns -o name | grep -Ev 'openshift|kube-' | sed 's#namespace/##', Kubernetes namespaces)####Create a namespace": "nvidia-gpu-operator",
19+
"Number of CPUs####Number of GPUs####Minimum Workers####Maximum Workers####Worker Memory####Head Memory": "{\"Number of CPUs\":\"1\",\"Number of GPUs\":\"1\",\"Minimum Workers\":\"4\",\"Maximum Workers\":\"4\",\"Worker Memory\":\"32Gi\",\"Head Memory\":\"32Gi\"}",
20+
"Choose the bucket that contains your model and glue data.expand([ -n \"$MC_CONFIG_DIR\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls s3 | awk '{print substr($NF, 1, length($NF) - 1)}', S3 Buckets)####separator####📁 Create a new bucket": "browsey",
21+
"Choose your Model File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "roberta-base",
22+
"Choose your Glue Data File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "glue_data",
23+
"BERT": "BERT",
24+
"Example: Using Ray Tasks to Parallelize a Function####Example: Using Ray Actors to Parallelize a Class####Example: Creating and Transforming Datasets####Example: Training Using PyTorch####Example: Hyperparameter Tuning####Example: Serving a scikit-learn gradient boosting classifier": "Example: Using Ray Tasks to Parallelize a Function",
25+
"Number of CPUs####Number of GPUs": "{\"Number of CPUs\":4,\"Number of GPUs\":3}",
26+
"expand(echo ${A-error} ; echo ${B-4} ; echo ${C-5})": "3",
27+
"XXXXXX.11111####222222": "11111",
28+
"YYYYYY.11111####222222": "222222",
29+
"My Cluster is Running Locally####My Cluster is Runing on Kubernetes": "My Cluster is Runing on Kubernetes",
30+
"expand([ -n \"$RAY_ADDRESS\" ] && ray job list --address $RAY_ADDRESS | tail +2 | awk '{print $1}' | sed \"s/[:{' ]//g\", Ray Runs)": "07a2647f-3656-4e3e-836c-95a2fa841af6",
31+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'keys | .[]', Ray Runs)": "d5a0d68f-a675-49ca-bff7-4ae762e6b146",
32+
"My Cluster is Running Locally####My Cluster is Running on Kubernetes": "My Cluster is Running on Kubernetes",
33+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2; $1=\"\";$2=\"\";print \"\\033;1m\" a, \"\\033[0;33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "\u001b;1ma88d4632-ab5c-4350-a770-d39a955c42c8 \u001b[0;33mRUNNING\u001b[0;2m -v --datapath /tmp/ --modelpath /tmp/ --logpath /tmp/ --tblogpath s3://browsey/codeflare/a88d4632-ab5c-4350-a770-d39a955c42c8/tensorboard/ --num_workers 1\u001b[0m",
34+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2; $1=\"\";$2=\"\";print a, \"\\033[33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "a88d4632-ab5c-4350-a770-d39a955c42c8 \u001b[33mRUNNING\u001b[0;2m -v --datapath /tmp/ --modelpath /tmp/ --logpath /tmp/ --tblogpath s3://browsey/codeflare/a88d4632-ab5c-4350-a770-d39a955c42c8/tensorboard/ --num_workers 1\u001b[0m",
35+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.start_time / 1000 | strflocaltime(\"%Y-%m-%dT%H:%M:%S\")) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2;c=$3; $1=\"\";$2=\"\";$3=\"\"; print a, \"\\033[0;36m\" c, \"\\033[0;1;33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "8443709b-74b3-4ae5-ae5a-aa4e1fbe2bb1 \u001b[0;36m2022-07-18T08:06:48 \u001b[0;1;33mPENDING\u001b[0;2m python main.py\u001b[0m",
36+
"Start a new Run####Connect Dashboard to an existing Run": "Connect Dashboard to an existing Run",
37+
"Start a new Run####Connect Dashboard to an existing Run####Shut down a Cloud Computer": "Shut down a Cloud Computer",
38+
"Start a new Run####Connect Dashboard to an existing Run####Boot up a Cloud Computer####Shut down a Cloud Computer": "Start a new Run",
39+
"SubTab1####SubTab2": "SubTab1",
40+
"Tab1####Tab2": "Tab2",
41+
"BERT####BYOT": "BYOT",
42+
"Location of your working directory": "{\"Location of your working directory\":\"/tmp/qiskit/ray_demo\"}",
43+
"Provide custom base image, if any": "{\"Provide custom base image, if any\":\"rayproject/ray:1.13.0-py37\"}"
44+
},
245
"madwizard/apriori/use-gpu": "don't use gpus",
346
"madwizard/apriori/arch": "x64",
447
"madwizard/apriori/platform": "darwin",
548
"madwizard/apriori/mac-installer": "Homebrew",
649
"madwizard/apriori/in-terminal": "HTML",
50+
"Run Locally####Run on a Kubernetes Cluster": "Run on a Kubernetes Cluster",
51+
"Example: Using Ray Tasks to Parallelize a Function####Example: Using Ray Actors to Parallelize a Class####Example: Creating and Transforming Datasets####Example: Training Using PyTorch####Example: Hyperparameter Tuning####Example: Serving a scikit-learn gradient boosting classifier": "Example: Using Ray Tasks to Parallelize a Function",
52+
"Start a new Run####Connect Dashboard to an existing Run####Boot up a Cloud Computer####Shut down a Cloud Computer": "Start a new Run",
753
"Training####Fine Tuning": "Fine Tuning",
8-
"GLUE": "GLUE",
54+
"BERT": "BERT",
955
"AWS####IBM": "AWS",
1056
"S3 Bucket for Run Data.expand([ -n \"$MC_CONFIG_DIR\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls s3 | awk '{print substr($NF, 1, length($NF) - 1)}', S3 Buckets)####separator####📁 Create a new bucket": "browsey",
11-
"Run Locally####Run on a Kubernetes Cluster": "Run on a Kubernetes Cluster",
12-
"Choose the bucket that contains your model and glue data.madwizard/apriori/platform": "Darwin",
1357
"expand(kubectl config get-contexts -o name, Kubernetes contexts)": "default/api-codeflare-train-v11-codeflare-openshift-com:6443/kube:admin",
1458
"expand([ -z ${KUBE_CONTEXT} ] && exit 1 || kubectl --context ${KUBE_CONTEXT} get ns -o name | grep -Ev 'openshift|kube-' | sed 's#namespace/##', Kubernetes namespaces)####Create a namespace": "nvidia-gpu-operator",
15-
"Number of CPUs####Number of GPUs####Minimum Workers####Maximum Workers####Worker Memory####Head Memory": "{\"Number of CPUs\":\"1\",\"Number of GPUs\":\"1\",\"Minimum Workers\":\"4\",\"Maximum Workers\":\"4\",\"Worker Memory\":\"32Gi\",\"Head Memory\":\"32Gi\"}",
59+
"Number of CPUs####Number of GPUs####Minimum Workers####Maximum Workers####Worker Memory####Head Memory": "{\"Number of CPUs\":\"1\",\"Number of GPUs\":\"1\",\"Minimum Workers\":\"1\",\"Maximum Workers\":\"1\",\"Worker Memory\":\"32Gi\",\"Head Memory\":\"32Gi\"}",
60+
"My Cluster is Running Locally####My Cluster is Running on Kubernetes": "My Cluster is Running on Kubernetes",
61+
"expand([ -n \"$RAY_ADDRESS\" ] && curl $RAY_ADDRESS/api/jobs/ | jq -r 'to_entries | sort_by(.value.start_time) | reverse | .[] | \"\\(.key) \\(.value.status) \\(.value.start_time / 1000 | strflocaltime(\"%Y-%m-%dT%H:%M:%S\")) \\(.value.entrypoint)\"' | sed -E 's/python3 ([^[:space:]])+ //g' | awk '{a=$1;b=$2;c=$3; $1=\"\";$2=\"\";$3=\"\"; print a, \"\\033[0;36m\" c, \"\\033[0;1;33m\" b \"\\033[0;2m\" $0 \"\\033[0m\"}', Ray Runs)": "2c495c48-2202-4be6-a8f7-4cccbb217f34 \u001b[0;36m2022-07-12T13:27:59 \u001b[0;1;33mPENDING\u001b[0;2m -v -b browsey -m roberta-base -g glue_data -t WNLI -M -s 40 41 42 43\u001b[0m",
62+
"GLUE": "GLUE",
1663
"Choose the bucket that contains your model and glue data.expand([ -n \"$MC_CONFIG_DIR\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls s3 | awk '{print substr($NF, 1, length($NF) - 1)}', S3 Buckets)####separator####📁 Create a new bucket": "browsey",
1764
"Choose your Model File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "roberta-base",
18-
"Choose your Glue Data File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "glue_data",
19-
"BERT": "BERT",
20-
"Example: Using Ray Tasks to Parallelize a Function####Example: Using Ray Actors to Parallelize a Class####Example: Creating and Transforming Datasets####Example: Training Using PyTorch####Example: Hyperparameter Tuning####Example: Serving a scikit-learn gradient boosting classifier": "Example: Using Ray Tasks to Parallelize a Function",
21-
"Number of CPUs####Number of GPUs": "{\"Number of CPUs\":4,\"Number of GPUs\":3}",
22-
"expand(echo ${A-error} ; echo ${B-4} ; echo ${C-5})": "3",
23-
"XXXXXX.11111####222222": "11111",
24-
"YYYYYY.11111####222222": "222222"
65+
"Choose your Glue Data File.expand([ -n \"$MC_CONFIG_DIR\" ] && [ -n \"$S3_FILEPATH\" ] && [ -n \"$S3_FILEPATH${S3_BUCKET_SUFFIX}\" ] && mc -q --config-dir ${MC_CONFIG_DIR} ls \"s3/$S3_FILEPATH${S3_BUCKET_SUFFIX}\" | awk '{print $NF}', S3 Objects)": "glue_data"
2566
}
Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,13 @@
1+
LAST SEEN TYPE REASON OBJECT MESSAGE
2+
3h1m Normal Pulled pod/nvidia-dcgm-exporter-7qs7v Successfully pulled image "nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37" in 1m16.137927284s
3+
3h1m Normal Created pod/nvidia-dcgm-exporter-7qs7v Created container toolkit-validation
4+
3h1m Normal Started pod/nvidia-dcgm-exporter-7qs7v Started container toolkit-validation
5+
3h1m Normal Pulled pod/nvidia-driver-daemonset-49.84.202201212103-0-zb76w Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:89f2c3b7c078a97682d4a375d04d6e45c73099679679dd2e2487ba3632be83d7" in 29.000491244s
6+
3h1m Normal Created pod/nvidia-driver-daemonset-49.84.202201212103-0-zb76w Created container openshift-driver-toolkit-ctr
7+
3h1m Normal Started pod/nvidia-driver-daemonset-49.84.202201212103-0-zb76w Started container openshift-driver-toolkit-ctr
8+
3h Normal Scheduled pod/mycluster-ray-head-type-jhrqg Successfully assigned nvidia-gpu-operator/mycluster-ray-head-type-jhrqg to ip-10-0-139-77.ec2.internal
9+
3h Normal AddedInterface pod/mycluster-ray-head-type-jhrqg Add eth0 [10.129.84.16/23] from openshift-sdn
10+
3h Normal Pulled pod/ray-operator-56f7b64d7b-wlzcb Successfully pulled image "rayproject/ray:1.13.0-py37" in 40.972650191s
11+
3h Normal Created pod/ray-operator-56f7b64d7b-wlzcb Created container ray
12+
3h Normal Started pod/ray-operator-56f7b64d7b-wlzcb Started container ray
13+
3h1m Normal Pulling pod/mycluster-ray-head-type-jhrqg Pulling image "rayproject/ray-ml:1.13.0-py37-gpu"

tests/plugin-codeflare/dashboard/inputs/2/job.json

Lines changed: 20 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -1,30 +1,34 @@
11
{
2-
"jobid": "f6c365c4-5fec-41b1-a811-3594c1c0f91d",
2+
"jobid": "a8998f9b-a4f4-4edd-a440-0340b46ff2bd",
33
"cmdline": {
4-
"appPart": "python3 tmp-65979-Ao7jcCUjajP4 -v -b ${S3_BUCKET} -m ${S3_OBJECTMODEL} -g ${S3_OBJECTGLUEDATA} -t WNLI -M -s 40 41 42 43",
5-
"systemPart": "ray job submit --runtime-env=/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp-65979-l1VQepk5OCSm --job-id ${JOB_ID} --no-wait"
4+
"appPart": "python3 tmp-80978-yaTBykEfMtbc -v -b ${S3_BUCKET} -m ${S3_OBJECTMODEL} -g ${S3_OBJECTGLUEDATA} -t WNLI -M -s 40 41 42 43",
5+
"systemPart": "ray job submit --runtime-env=/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp-80978-9S3zGVjVs2HG --job-id ${JOB_ID} --no-wait"
66
},
77
"runtimeEnv": {
88
"env_vars": {
9-
"JOB_ID": "f6c365c4-5fec-41b1-a811-3594c1c0f91d",
9+
"JOB_ID": "a8998f9b-a4f4-4edd-a440-0340b46ff2bd",
1010
"S3_PROVIDER": "aws",
1111
"S3_ENDPOINT": "https://s3.amazonaws.com",
1212
"S3_ACCESS_KEY_ID": "********",
1313
"AWS_ACCESS_KEY_ID": "********",
1414
"S3_SECRET_ACCESS_KEY": "********",
1515
"AWS_SECRET_ACCESS_KEY": "********",
16-
"MC_CONFIG_DIR": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp.kQkTy0d4",
16+
"MC_CONFIG_DIR": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp.eImBHFSt",
1717
"S3_BUCKETRAYLOGS": "browsey",
1818
"S3_FILEPATHRAYLOGS": "RAYLOGS",
19-
"S3_LOGDIR": "browsey/codeflare/f6c365c4-5fec-41b1-a811-3594c1c0f91d",
20-
"LOGDIR_STAGE": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/logdir-stage.TfJpqocD",
21-
"LOGDIR_URI": "s3://browsey/codeflare/f6c365c4-5fec-41b1-a811-3594c1c0f91d",
22-
"LOGDIR_MC": "s3/browsey/codeflare/f6c365c4-5fec-41b1-a811-3594c1c0f91d",
23-
"STREAMCONSUMER_LOGS": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/logdir-stage.TfJpqocD/logs/",
24-
"STREAMCONSUMER_EVENTS": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/logdir-stage.TfJpqocD/events/",
25-
"STREAMCONSUMER_RESOURCES": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/logdir-stage.TfJpqocD/resources/",
19+
"LOGDIR_STAGE": "/Users/nickm/Library/Application Support/madwizard-nodejs/default/jobs/a8998f9b-a4f4-4edd-a440-0340b46ff2bd",
20+
"STREAMCONSUMER_LOGS": "/Users/nickm/Library/Application Support/madwizard-nodejs/default/jobs/a8998f9b-a4f4-4edd-a440-0340b46ff2bd/logs/",
21+
"STREAMCONSUMER_EVENTS": "/Users/nickm/Library/Application Support/madwizard-nodejs/default/jobs/a8998f9b-a4f4-4edd-a440-0340b46ff2bd/events/",
22+
"STREAMCONSUMER_RESOURCES": "/Users/nickm/Library/Application Support/madwizard-nodejs/default/jobs/a8998f9b-a4f4-4edd-a440-0340b46ff2bd/resources/",
23+
"S3_LOGDIR": "browsey/codeflare/a8998f9b-a4f4-4edd-a440-0340b46ff2bd",
24+
"LOGDIR_URI": "s3://browsey/codeflare/a8998f9b-a4f4-4edd-a440-0340b46ff2bd",
25+
"LOGDIR_MC": "s3/browsey/codeflare/a8998f9b-a4f4-4edd-a440-0340b46ff2bd",
26+
"PATH": "/Users/nickm/.local/bin:/usr/local/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/bin:/Users/nickm/.krew/bin:/Library/Frameworks/Python.framework/Versions/2.7/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/Applications/VMware Fusion.app/Contents/Public:/Library/TeX/texbin:/usr/local/go/bin:/Library/Apple/usr/bin:/Users/nickm/bin",
2627
"KUBE_CONTEXT": "default/api-codeflare-train-v11-codeflare-openshift-com:6443/kube:admin",
2728
"KUBE_NS": "nvidia-gpu-operator",
29+
"KUBE_POD_LABEL_SELECTOR": "ray-user-node-type=rayWorkerType",
30+
"KUBE_POD_RAY_HEAD_LABEL_SELECTOR": "ray-user-node-type=rayHeadType",
31+
"KUBE_PODFULL_LABEL_SELECTOR": "ray-node-type",
2832
"NUM_CPUS": "1",
2933
"NUM_GPUS": "1",
3034
"MIN_WORKERS": "4",
@@ -33,12 +37,10 @@
3337
"HEAD_MEMORY": "32Gi",
3438
"RAY_OPERATOR_IMAGE": "rayproject/ray:1.13.0-py37",
3539
"RAY_IMAGE": "rayproject/ray-ml:1.13.0-py37-gpu",
36-
"HELM_CLONE_TEMPDIR": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp.f0NuCaF1",
40+
"HELM_CLONE_TEMPDIR": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp.3EMkfmLv",
3741
"RAY_KUBE_CLUSTER_NAME": "mycluster",
38-
"KUBE_POD_LABEL_SELECTOR": "ray-user-node-type=rayWorkerType",
39-
"KUBE_PODFULL_LABEL_SELECTOR": "ray-node-type",
40-
"RAY_KUBE_PORT": "8266",
41-
"RAY_ADDRESS": "http://127.0.0.1:8266",
42+
"RAY_KUBE_PORT": "9612",
43+
"RAY_ADDRESS": "http://127.0.0.1:9612",
4244
"S3_BUCKET": "browsey",
4345
"S3_FILEPATH": "browsey",
4446
"S3_OBJECTMODEL": "roberta-base",
@@ -48,7 +50,7 @@
4850
"WANDB_CONFIG_DIR": "/tmp",
4951
"WANDB_DISABLED": "true"
5052
},
51-
"working_dir": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp-65979-FpQKyyRyOv0T",
53+
"working_dir": "/var/folders/2k/7mgd1tq55gdbghf0xkl2t_l80000gp/T/tmp-80978-UYzNT1wjDJyI",
5254
"pip": ["boto3", "ray[default]", "ray_lightning", "pytorch_lightning", "torchvision", "transformers==3.0.2"]
5355
},
5456
"language": "python",
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
f6c365c4-5fec-41b1-a811-3594c1c0f91d
1+
a8998f9b-a4f4-4edd-a440-0340b46ff2bd

0 commit comments

Comments
 (0)