-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
update change log #657
update change log #657
Conversation
/assign @yebrahim |
/lgtm |
/approve |
/approve |
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: IronPan The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
1 similar comment
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: IronPan The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: IronPan The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
* Create a blueprint reconciler to autodeploy and reconcile blueprints. * The reconciler decides whether we need to deploy a new blueprint and if it does it creates a Tekton PipelineRun to deploy Kubeflow. * Here are some differences in how we are deploying blueprints vs. kfctl deployments * We are using Tekton PipelineRuns as opposed to K8s jobs to do the deployment * We no longer use deployments.yaml to describe the group of deployments. Instead we just create a PipelineRun.yaml and that provides all the information the reconciler needs e.g. the branch to watch for changes. * Update the flask app to provide information about blueprints. * Include a link to the Tekton dashboard showing the PipelineRun that deployed Kubeflow. * Define a Pipeline to deploy Kubeflow so we don't have to inline the spec in the PipelienRun. * Remove Dockerfile.skaffold; we can use skaffold auto-sync in developer mode. Add a column in the webserver to redirect to the Tekton dashboard for the PipelineRun that deployed it. * GoogleCloudPlatform/kubeflow-distribution#5 Setup autodeploy for gcp blueprints.
* The reconciler is running amok because in kubeflow#657 we changed the code to skip over deployments with no manifest field because we couldn't get the zone. It turns out the lack of the manifest is an indication that the deployment failed. By not matching these deployments we end up retrying the auto-deployment. This is causing a cascading failure because we end up using up all our DM quota which in turn causes auto deployments to fail which in turn causes us to try again and eat up more deployments. * This PR fixes that by matching deployments even if the manifest field is missing. This should prevent us from repeatedly retrying. * Add exponential backoff to delete deployments. Because of the above error we are bumping to write quota limits per day which is impacting our ability to GC deployments.
* Added additional license info for python libraries * Added license file to images * Copied to correct image * Trying changes to dockerfileS * Removing go builds for now * Added third_party to docker builds
Signed-off-by: Michalina Kotwica <[email protected]>
This change is