Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
* Add a Spark operator to Kubeflow along with integration tests. Start adding converted spark operator elements Can generate empty service account for Spark Create the service account for the spark-operator. Add clusterRole for Spark Add cluster role bindings for Spark Add deployment (todo cleanup name/image) Can now launch spark operator Put in a reasonable default for namespace (e.g default not null) and make the image used for spark-operator configurable Start working to add job type We can now launch and operator and launch a job, but the service accounts don't quite line up. TODO(holden) refactor the service accounts for the job to only be created in the job and move sparkJob inside of all.json as well then have an all / operator / job entry point in all.json maybe? Add two hacked up temporary test scripts for use during dev (TODO refactor later into proper workflow) Able to launch a job fixed coreLimit and added job arguments. Remaining TODOs are handling of nulls & svc account hack + test cleanup. Start trying to re-organize the operator/job Fix handling of optional jobArguments and mainClass and now it _works_ :) Auto format the ksonnet. Reviewer feedback: switch description of Spark operator to something meaningful, use sparkVersion param instead of hard coded v2.3.1-v1alpha1, and fix hardcoded namespace. Clarify jobName param, remove Fix this since it has been integrated into all.libsonnet as intended. CR feedback: change description typo and add opitonal param to spark operator for sparkVersion Start trying to add spark tests to test_deploy.py At @kunmingg suggestion Revert "Start trying to add spark tests to test_deploy.py" to focus on prow tests. This reverts commit 912a763. Start trying to add Spark to the e2e workflow for testing Looks like the prow tests call into the python tests normally so Revert "At @kunmingg suggestion Revert "Start trying to add spark tests to test_deploy.py" to focus on prow tests." This reverts commit 6c4c81f. autoformat jsonnet s/core/common/ and /var/log/syslog to README Race condition on first deployment Start adding SparkPI job to the workflow test. Generate spark operator during CI as well. Fix deploy kf indent Already covered by deploy. Install spark operator Revert "Install spark operator" This reverts commit cc559dd. Test against the PR not master. Fix string concat Take spark-deploy out of workflows since cover in kf presub anyways. Debug commit revert later. idk whats going on for real. hax Ok lets use where the sym link was coming from idk. Debug deploy kubeflow call... Pritn in maint oo. Specify a name. name Get all. More debugging also why do we eddit app.yaml; directly. don't gen common import for debug Just do spark-operator as verbose. spelling hmm namespace looked weird, lets run pytorch in verbose too so I can compare put verbose at the end Autoformat the json Add a deployment scope and give more things a namespace Format. Gen pytorch and spark ops as verbose idk wtf this is. Don't deploy the spark job in the releaser test no kfctl test either. Just use name We don't append any junk anymore format json Don't do spark in deploy_kubeflow anymore Spark job deployment with workflows Apply spark operator. Add a sleep hack Fix multi-line add a working dir for the ks app temp debug garbage specify working dir Working dir was not happy, just cd cause why not testdir not appDir change to tests.testDir Move operator deployment Make sure we are in the ks_app? Remove debugging and YOLO 90% less YOLO Add that comma Change deps well CD seems to work in the other command so uhhh who knows? Use runpath + pushd instead of kfctl generate Just generate for now Do both Generate k8s Install operator Break down setting up the spark operator into different steps We are in default rather than ghke Use the run script to do the dpeloy Change the namespace to stepsNamespace and add debug step cauise idk Append the params to generate cmd Remove params_str since we're doing list and a param of namespace s/extends/extend/ Move params to the right place Remove debug cluster step Remove local test since we now use the regular e2e argo triggered tests. Respond to the CR feedback Fix paramterization of spark executor config. Plumb through spark version to executor version label Remove unecessary whitespace change in otherwise unmodified file. * re-run autoformat * default doesn't seem to exists anymore * Debug the env list cause it changed * re-run autoformat again * Specify the env since env list shows default env is the only env present. * Remove debug env list since the operator now works * autofrmat and indent default * Address CR feedback: remove deploymentscope and just use clusterole, link to upstream base operator in doc, remove downstream job test since it's triggered in both minikube and kfctl tests and we don't want to test it in minikube right now * Take out the spark job from ther workflows in components test we just test the operator applies for now. * Remove namespace as a param and just use the env. * Fix end of line on namespace from ; to ,
- Loading branch information